DeepSeek logo
DeepSeek
Open Weights

DeepSeek V3.1 (Non-reasoning)

Released Aug 2025

DeepSeek V3.1 (Non-reasoning) is the standard conversational mode of the DeepSeek-V3.1 large language model, an iterative update to the DeepSeek-V3 architecture released on August 21, 2025. Developed by the Chinese AI laboratory DeepSeek, the model utilizes a Mixture-of-Experts (MoE) design with a total of 671 billion parameters, of which approximately 37 billion are activated per token during inference. This non-reasoning mode is optimized for high-efficiency text generation, software development, and general-purpose assistance, prioritizing direct output and low latency by bypassing the internal chain-of-thought processing used in reasoning-specialized variants. Technical highlights of this version include a 128,000-token context window, improved instruction-following, and superior performance in multi-step agentic tasks and tool calling compared to its predecessor. To ensure efficient resource utilization, the model incorporates Multi-head Latent Attention (MLA) and was trained using the FP8 precision format. It is distributed under the permissive MIT license, supporting both broad research and commercial applications.

Rankings & Comparison