Qwen3-Max-Thinking debuts with focus on hard math, code

Qwen3-Max-Thinking, Alibaba Cloud's new flagship reasoning model, is now in Qwen Chat and Model Studio, targeting tough math, code, and agent workflows.

· 1 min read
Qwen

Qwen has introduced Qwen3-Max-Thinking, a flagship reasoning model designed for tackling complex math, coding, and multi-step agent workflows. This model is now available for direct use in Qwen Chat and is intended to be integrated into applications via Alibaba Cloud’s Model Studio. Here, teams can utilize the model for deliberate “thinking” runs when accuracy is crucial, and switch back to faster responses for routine prompts.

In Model Studio, the Qwen3-Max line features a 262,144-token context window. The latest snapshot tag (qwen3-max-2026-01-23) is described as merging thinking and non-thinking capabilities into a single model. In thinking mode, it can interleave tool calls within the reasoning process, equipped with built-in web search, webpage content extraction, and a code interpreter. This setup is designed for questions that require evidence gathering, parsing, and calculation before an answer is produced.

Alibaba Cloud’s Qwen team has been expanding the Qwen3 family across chat and cloud APIs, targeting developers and enterprises seeking long-context reasoning and tool-using agents. Early testers have noted that the “thinking” variant excels when tasks necessitate tool use or deep verification. However, the day-to-day benefits can vary depending on the prompt mix and the amount of test-time compute allocated.

Source