For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.
Input
Output
Context
1048K
Max Output
33K
Parameters
—
Input Modalities
Output Modalities
| Platform | Input | Output |
|---|---|---|
OpenRouter | $0.100 | $0.400 |
Data sourced from official provider APIs and documentation
Last updated: Mar 16, 2026
Explore models, compare pricing and benchmarks, and right-size your infrastructure — all in one place.