Alibaba logo
Alibaba
Open Weights

Qwen3 30B A3B (Reasoning)

Released Apr 2025

Intelligence
#279
Coding
#270
Math
#90
Context33K
Parameters30.5B (3.3B active)

Qwen3 30B A3B (Reasoning) is a large-scale Mixture-of-Experts (MoE) language model developed by Alibaba's Qwen team. It is designed to provide high-performance reasoning while maintaining computational efficiency by activating only a fraction of its total parameters during inference. The model is specifically optimized for complex logical tasks, including mathematics, coding, and scientific problem-solving.

Architecture and Efficiency

The model features a total of 30.5 billion parameters, with approximately 3.3 billion parameters activated per token (denoted by the "A3B" suffix). It utilizes a sparse MoE structure with 128 experts, activating 8 for each task. This architecture allows the model to match the capabilities of larger dense models while significantly reducing inference latency and resource requirements.

Reasoning and Thinking Modes

A core innovation of the Qwen3 series is the implementation of a hybrid reasoning system. The model can transition between a Thinking Mode, which generates detailed chain-of-thought reasoning traces for complex logic, and a Non-Thinking Mode for efficient, general-purpose dialogue. The reasoning variant is further enhanced through specialized post-training and reinforcement learning to handle multi-step deduction and long-form logical processing.

Capabilities and Context

Qwen3 30B A3B supports a native context window of 32,768 tokens, which can be extended to 131,072 tokens or more using techniques like YaRN. It provides robust multilingual support across 119 languages and is engineered for agentic workflows, enabling precise tool integration and long-context understanding for complex document processing and interactive applications.

Rankings & Comparison