R1 Distill Llama 70B: Pricing, Context Window & Benchmarks
70Bby DeepSeek
DeepSeek R1 Distill Llama 70B is a distilled large language model based on [Llama-3.3-70B-Instruct](/meta-llama/llama-3.3-70b-instruct), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). The model combines advanced distillation techniques to achieve high performance across multiple benchmarks, including: - AIME 2024 pass@1: 70.0 - MATH-500 pass@1: 94.5 - CodeForces Rating: 1633 The model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.
What you can do with R1 Distill Llama 70B
Everyday Q&A and clear explanations
Writing help (emails, posts, summaries)
Idea generation and brainstorming
Learning support with step-by-step guidance
Composite Indices
Intelligence, Coding, Math
Standard Benchmarks
Academic and industry benchmarks
Benchmark Highlights
| Metric | Value |
|---|---|
| Provider | DeepSeek |
| Context Window | 131,072 tokens |
| Input Price | $0.70/1M tokens |
| Output Price | $1.05/1M tokens |
| Release Date | Jan 20, 2025 |
| Modalities | text |
| Capabilities | N/A |
Compare R1 Distill Llama 70B to other models
See how it stacks up on price, quality, and overall performance.
Frequently asked questions
What is R1 Distill Llama 70B good for?
Use R1 Distill Llama 70B for everyday tasks like writing, summarizing, brainstorming, and getting clear explanations.
How much does R1 Distill Llama 70B cost?
Pricing is based on usage. Current rates are $0.70/1M tokens for input and $1.05/1M tokens for output.
Can I try R1 Distill Llama 70B for free?
Yes. You can start a chat instantly and test the model before deciding on a plan.
Does R1 Distill Llama 70B support images or audio?
R1 Distill Llama 70B focuses on text-based tasks.
Suggested comparisons
Similar models
Benchmarks and pricing are sourced from Artificial Analysis where available. OpenRouter specs are used as a fallback.
Compare Models
Select a model to compare with R1 Distill Llama 70B