Alibaba logo
Alibaba
Open Weights

Qwen3 0.6B (Reasoning)

Released Apr 2025

Intelligence
#455
Coding
#370
Math
#214
Context32K
Parameters0.6B

Qwen3 0.6B is a compact, dense language model developed by Alibaba Cloud, representing the entry-level tier of the third-generation Qwen family. Its primary distinction is the integration of a native Hybrid Reasoning capability, which allows the model to switch between a high-efficiency "Non-Thinking" mode for general tasks and a "Thinking" mode for multi-step logical deduction, mathematics, and code generation. In Thinking Mode, the model produces internal chain-of-thought traces, typically wrapped in <think> tags, before delivering a final response.

The model's architecture is built on a 28-layer Transformer decoder using Grouped Query Attention (GQA) to optimize inference speed and reduce memory footprint. It supports a context window of 32,768 tokens, making it optimized for deployment in resource-constrained environments or on-device applications that require sophisticated language understanding.

Trained on an extensive dataset of 36 trillion tokens across 119 languages, Qwen3 0.6B emphasizes cross-lingual mastery and instruction-following. It leverages techniques like strong-to-weak distillation from larger models in the Qwen3 suite, such as the 235B Mixture-of-Experts variant, to maintain competitive performance in reasoning benchmarks while remaining computationally efficient. The model is released under the Apache 2.0 license.

Rankings & Comparison