DeepSeek logo
DeepSeek
Open Weights

DeepSeek V3.2 Speciale

Released Dec 2025

Intelligence
#122
Coding
#40
Math
#4
Context128K
Parameters685B

DeepSeek-V3.2-Speciale is a specialized, high-compute variant of the DeepSeek-V3.2 large language model family, released by DeepSeek in December 2025. It is a reasoning-centric model designed specifically for complex mathematical and algorithmic tasks, surpassing the standard V3.2 in its ability to generate extended thinking chains. The model utilizes a Mixture-of-Experts (MoE) architecture with a total of 685 billion parameters, activating approximately 37 billion parameters per token during inference.

Architecture and Performance

The model introduces DeepSeek Sparse Attention (DSA) and Multi-Head Latent Attention (MLA), mechanisms designed to reduce computational complexity and improve efficiency in long-context scenarios of up to 128,000 tokens. DeepSeek-V3.2-Speciale was developed using a reinforcement learning (RL) framework that relaxed length penalties, encouraging the model to engage in deeper, multi-step reasoning. This optimization led to gold-medal level performance in the 2025 International Mathematical Olympiad (IMO) and the International Olympiad in Informatics (IOI).

Capabilities and Limitations

While the Speciale variant offers reasoning proficiency comparable to frontier proprietary models, it is optimized for accuracy over token efficiency. It is characterized by higher latency and increased token usage compared to the standard DeepSeek-V3.2. Unlike its general-purpose counterpart, the Speciale variant is intended primarily for research and competitive reasoning; it does not support tool-calling functionality and is less optimized for general-purpose chat or writing scenarios.

Rankings & Comparison