| Qwen3-235B-A22B |
Apr 2025 |
235 total / 22 active |
Flagship MoE model, multilingual, reasoning |
| Qwen3-30B-A3B |
Apr 2025 |
30 total / 3 active |
MoE, efficient reasoning |
| Qwen3-32B |
Apr 2025 |
32 |
Dense, general-purpose |
| Qwen3-14B |
Apr 2025 |
14 |
Dense, balanced performance |
| Qwen3-8B |
Apr 2025 |
8 |
Dense, efficient for mid-range tasks |
| Qwen3-4B |
Apr 2025 |
4 |
Dense, lightweight |
| Qwen3-1.7B |
Apr 2025 |
1.7 |
Dense, edge-capable |
| Qwen3-0.6B |
Apr 2025 |
0.6 |
Smallest, runs on low-end devices |
| ------------------------ |
-------------- |
------------------------- |
--------------------------------------------- |
| Qwen2.5-Coder-32B |
Nov 2024 |
32 |
Code generation, GPT-4o-level coding |
| Qwen2.5-Coder-14B |
Nov 2024 |
14 |
Code generation |
| Qwen2.5-Coder-7B |
Nov 2024 |
7 |
Code generation |
| Qwen2.5-Coder-3B |
Nov 2024 |
3 |
Code generation |
| Qwen2.5-Coder-0.5B |
Nov 2024 |
0.5 |
Code generation, edge-capable |
| Qwen2.5-72B |
2024 |
72 |
General-purpose |
| Qwen2.5-14B-Instruct |
2024 |
14 |
Instruction-tuned |
| Qwen2.5-7B-Instruct |
2024 |
7 |
Instruction-tuned |
| ------------------------ |
-------------- |
------------------------- |
--------------------------------------------- |
| QwQ-32B |
Nov 2024 |
32 |
Reasoning-focused, strong benchmarks |