r1-1776-distill-llama-70b

r1-1776-distill-llama-70b

perplexity-ai

70B parameter Llama-based model, distilled from R1 1776, designed to provide uncensored, unbiased responses while maintaining strong reasoning capabilities

PropertyValue
Base ModelLLaMA 70B
DeveloperPerplexity AI
Model TypeDistilled Language Model
Model URLHugging Face

What is r1-1776-distill-llama-70b?

R1-1776-Distill-LLaMA-70B is a sophisticated language model developed by Perplexity AI, representing a distilled version of the R1 1776 model. It's specifically designed to remove Chinese Communist Party censorship while maintaining high reasoning capabilities and performance metrics. The model has been carefully engineered to provide unbiased, accurate, and factual information across a wide range of topics.

Implementation Details

The model is built upon the LLaMA 70B architecture and underwent extensive post-training to eliminate censorship while preserving its core capabilities. Notable benchmark results include exceptional performance on MATH-500 (94.8%), MMLU (88.40%), and DROP (84.83%), demonstrating its strong reasoning abilities.

  • Comprehensive evaluation on 1000+ multilingual examples
  • Human and LLM-based assessment of censorship removal
  • Maintained mathematical and reasoning capabilities post-decensoring
  • Significant reduction in censorship score from 80.53 to 0.2

Core Capabilities

  • Uncensored information processing and generation
  • High-level mathematical reasoning (94.8% on MATH-500)
  • Strong performance on general knowledge (88.40% on MMLU)
  • Robust reading comprehension (84.83% on DROP)
  • Balanced performance on general purpose question answering (65.05% on GPQA)

Frequently Asked Questions

Q: What makes this model unique?

The model's primary distinction lies in its successful removal of censorship while maintaining high performance across various benchmarks. It achieves this without compromising its core reasoning capabilities, as evidenced by consistent scores across mathematical and analytical tasks.

Q: What are the recommended use cases?

The model is particularly suited for applications requiring unbiased information processing, complex reasoning tasks, mathematical problem-solving, and general knowledge applications. It's especially valuable in contexts where uncensored, accurate information is crucial.

Socials
PromptLayer
Company
All services online
Location IconPromptLayer is located in the heart of New York City
PromptLayer © 2026