Qwen-Max, based on Qwen2.5, provides the best inference performance among [Qwen models](/qwen), especially for complex multi-step tasks. It's a large-scale MoE model that has been pretrained on over 20 trillion tokens and further post-trained with curated Supervised Fine-Tuning (SFT) and Reinforcement Learning from Human Feedback (RLHF) methodologies. The parameter count is unknown.
Input
Output
Context
33K
Max Output
8K
Parameters
—
Input Modalities
Output Modalities
Input
$1.04
Output
$4.16
| Platform | Input | Output |
|---|---|---|
OpenRouter | $1.04 | $4.16 |
Data sourced from official provider APIs and documentation
Last updated: Mar 17, 2026
Explore models, compare pricing and benchmarks, and right-size your infrastructure — all in one place.