DeepSeek logo
DeepSeek
Open Weights

DeepSeek R1 Distill Llama 8B

Released Jan 2025

Intelligence
#352
Math
#152
Context128K
Parameters8B

DeepSeek R1 Distill Llama 8B is a reasoning-focused language model based on the Llama-3.1-8B-Instruct architecture. It was developed by DeepSeek through a distillation process that incorporates reasoning data from the larger DeepSeek-R1 model into smaller-scale architectures. This process allows the model to utilize reasoning chains, often referred to as Chain-of-Thought (CoT), to solve complex logic, mathematics, and programming tasks.

The distillation methodology involves supervised fine-tuning on approximately 800,000 reasoning samples generated by DeepSeek-R1. This training enables the 8B parameter model to achieve performance levels on reasoning benchmarks that are significantly higher than standard instruction-tuned models of similar size. By leveraging the Llama-3.1 foundation, the model maintains high computational efficiency and compatibility with existing software ecosystems while providing specialized capabilities in systematic problem-solving and multi-step inference.

Rankings & Comparison