Alibaba logo
Alibaba
Open Weights

qwen3-32b

Released Apr 2025

Arena AI
#115
Parameters32.8B

Qwen3-32B is a dense large language model developed by the Qwen team at Alibaba Cloud, released in April 2025 as part of the Qwen3 series. It succeeds the Qwen2.5 family, offering significant improvements in reasoning, coding, and multilingual understanding. The model is released under the Apache 2.0 license, facilitating open research and commercial use.

Architecturally, the model utilizes a dense transformer framework with 32.8 billion parameters and 64 layers. It employs Grouped Query Attention (GQA) with 64 query heads and 8 key-value heads to optimize inference efficiency. Qwen3-32B supports a native context length of 32,768 tokens, which can be extended to 128,000 tokens using YaRN-based scaling techniques.

Key Features

A defining feature of the model is its hybrid operational architecture, allowing for seamless switching between a "thinking" mode and a "non-thinking" mode. In thinking mode, the model performs step-by-step chain-of-thought reasoning—often delimited by <think> tags—to solve complex mathematical and logical problems. Non-thinking mode is optimized for low-latency, general-purpose conversational tasks.

Qwen3-32B was trained on approximately 36 trillion tokens, covering 119 languages and dialects. It demonstrates high proficiency in tool-calling and agentic tasks, making it suitable for integration into complex automated workflows and RAG (Retrieval-Augmented Generation) systems.

Rankings & Comparison