Alibaba logo
Alibaba
Open Weights

Qwen3.5 35B A3B (Reasoning)

Released Feb 2026

Intelligence
#71
Coding
#89
Arena AI
#74
Context262K
Parameters35B

Qwen3.5 35B A3B (Reasoning) is a mid-sized, multimodal large language model developed by Alibaba's Qwen team. Released as part of the Qwen3.5 "Medium" series, it serves as a high-efficiency alternative to larger dense models by employing a sparse Mixture-of-Experts (MoE) architecture. The model is designed for a variety of complex tasks, including agentic coding, visual reasoning, and long-context document analysis.

The model's architecture is a hybrid design that integrates Gated Delta Networks (linear attention mechanisms) with sparse MoE layers to optimize inference efficiency and throughput. Out of its total 35 billion parameters, only approximately 3 billion are activated per token during processing. This sparse activation allows for frontier-level performance while maintaining the computational requirements typically associated with much smaller models.

A key feature of Qwen3.5 35B A3B is its native vision-language foundation, which uses early-fusion training on multimodal tokens to process images and text simultaneously. It supports a dedicated "thinking" mode, enabling the model to perform internal chain-of-thought reasoning before outputting final answers. This capability significantly enhances its performance in logical reasoning and mathematical benchmarks.

The model supports a native context window of 262,144 tokens, which can be extended up to 1 million tokens for specialized long-form tasks. It provides extensive linguistic support for 201 languages and dialects and is released under the Apache 2.0 license, facilitating broad commercial and research adoption.

Rankings & Comparison