Alibaba logo
Alibaba
Open Weights

Qwen3 235B A22B (Non-reasoning)

Released Apr 2025

Intelligence
#251
Coding
#226
Math
#202
Context33K
Parameters235B

Qwen3 235B A22B is a large-scale language model developed by Alibaba Cloud, utilizing a Mixture-of-Experts (MoE) architecture. It features a total of 235 billion parameters, with approximately 22 billion parameters activated for each token during inference. This specific Non-reasoning version (also known as the Instruct variant) is optimized for efficient, general-purpose dialogue and instruction following, prioritizing lower latency and direct responses over the extended chain-of-thought "thinking" blocks used by reasoning-enhanced editions of the Qwen3 series.

Technical Capabilities

The model was trained on an extensive dataset of 36 trillion tokens, providing support for over 119 languages and dialects. Its architecture includes 94 layers and utilizes Grouped-Query Attention (GQA) to optimize computational efficiency. It supports a native context window of 128,000 tokens, which can be extended to 1 million tokens using specialized configurations like YaRN or sparse attention mechanisms.

Compared to reasoning-focused models, the Non-reasoning variant is designed for high-throughput production environments. It excels in tasks such as creative writing, role-playing, and tool-calling (agentic interaction). In July 2025, Alibaba released an updated iteration (Instruct-2507) that further refined its instruction-following accuracy and expanded its long-tail knowledge across multiple languages, emphasizing its role as a high-performance generalist model.

Rankings & Comparison