Alibaba logo
Alibaba

qwen3-max-preview

Released Sep 2025

Arena AI
#36
Parameters1T+

Qwen3-Max-Preview is a flagship large language model developed by Alibaba's Qwen team, released in September 2025. It represents the first model in the Qwen series to surpass the 1-trillion-parameter threshold, positioning it as a frontier-class model designed for high-end reasoning and complex technical tasks. Unlike several smaller models in the Qwen3 family, the Max-Preview version is primarily available as a closed-weight model through API access.

The model is built on a massive Mixture-of-Experts (MoE) architecture and was pretrained on 36 trillion tokens. This architectural approach is intended to provide stable training and higher Model FLOPs Utilization (MFU) compared to previous generations. It features a native context window of 256,000 tokens, which can be extended to 1 million tokens, and supports a context caching feature to improve efficiency for long-document processing and multi-turn conversations.

In terms of capabilities, Qwen3-Max-Preview is optimized for structured data extraction, coding assistance, and mathematical reasoning. Internal and public benchmarks place its performance in competition with leading models such as GPT-5 and Claude 4, particularly in technical domains and multilingual understanding across more than 100 languages. While it does not utilize an explicit "thinking" mode like some specialized reasoning models, it demonstrates strong performance on logic-heavy tasks and agentic tool-calling benchmarks.

Rankings & Comparison