DeepSeek: R1 Distill Llama 70B
DeepSeek R1 Distill Llama 70B is a distilled large language model based on [Llama-3.3-70B-Instruct](/meta-llama/llama-3.3-70b-instruct), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). The model combines advanced distillation techniques to achieve high performance across multiple benchmarks, including: - AIME 2024 pass@1: 70.0 - MATH-500 pass@1: 94.5 - CodeForces Rating: 1633 The model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.
Parameters
70B
Context Window
131,072
tokens
Input Price
$0.1
per 1M tokens
Output Price
$0.4
per 1M tokens
Capabilities
Model capabilities and supported modalities
Performance
Good reasoning with solid logical foundations
Strong mathematical capabilities, handles complex calculations well
Specialized in code generation with excellent programming capabilities
-
Modalities
text
text
LLM Price Calculator
Calculate the cost of using this model
Monthly Cost Estimator
Based on different usage levels