AI 大模型排名 ArtificialAnalysis AI 大模型排行榜
信息查询
524
次浏览
100%
有帮助
·
1 人反馈
AI 大模型排名 Artificial Analysis AI 大模型排行榜,综合对超过 100 个 AI 模型(LLM)的性能进行了比较和排名,评估指标包括智能程度、价格以及常见AI基准测试的结果。
AI 大模型排行榜数据中心
| 排名 | 模型名称 | 综合指数 ▼ | 编程 | 价格 ($/1M) |
|---|---|---|---|---|
| 1 | Claude Opus 4.7 (Adaptive Reasoning, Max Effort) | 57.3 | 52.5 | $10 |
| 2 | Gemini 3.1 Pro Preview | 57.2 | 55.5 | $4.5 |
| 3 | GPT-5.4 (xhigh) | 56.8 | 57.3 | $5.625 |
| 4 | GPT-5.3 Codex (xhigh) | 53.6 | 53.1 | $4.813 |
| 5 | Claude Opus 4.6 (Adaptive Reasoning, Max Effort) | 53 | 48.1 | $10 |
| 6 | Muse Spark | 52.1 | 47.5 | $0 |
| 7 | Claude Sonnet 4.6 (Adaptive Reasoning, Max Effort) | 51.7 | 50.9 | $6 |
| 8 | GLM-5.1 (Reasoning) | 51.4 | 43.4 | $2.15 |
| 9 | GPT-5.2 (xhigh) | 51.3 | 48.7 | $4.813 |
| 10 | Qwen3.6 Plus | 50 | 42.9 | $1.125 |
| 11 | GLM-5 (Reasoning) | 49.8 | 44.2 | $1.55 |
| 12 | Claude Opus 4.5 (Reasoning) | 49.7 | 47.8 | $10 |
| 13 | MiniMax-M2.7 | 49.6 | 41.9 | $0.525 |
| 14 | Grok 4.20 0309 v2 (Reasoning) | 49.3 | 40.5 | $3 |
| 15 | MiMo-V2-Pro | 49.2 | 41.4 | $1.5 |
| 16 | GPT-5.2 Codex (xhigh) | 49 | 43 | $4.813 |
| 17 | GPT-5.4 mini (xhigh) | 48.9 | 51.5 | $1.688 |
| 18 | Grok 4.20 0309 (Reasoning) | 48.5 | 42.2 | $3 |
| 19 | Gemini 3 Pro Preview (high) | 48.4 | 46.5 | $4.5 |
| 20 | GPT-5.1 (high) | 47.7 | 44.7 | $3.438 |
| 21 | Kimi K2.5 (Reasoning) | 46.8 | 39.5 | $1.2 |
| 22 | GLM-5-Turbo | 46.8 | 36.8 | $0 |
| 23 | GPT-5.2 (medium) | 46.6 | 44.2 | $4.813 |
| 24 | Claude Opus 4.6 (Non-reasoning) | 46.5 | 47.6 | $10 |
| 25 | Gemini 3 Flash Preview (Reasoning) | 46.4 | 42.6 | $1.125 |
| 26 | Qwen3.5 397B A17B (Reasoning) | 45 | 41.3 | $1.35 |
| 27 | MiMo-V2-Omni-0327 | 44.9 | 36.9 | $0 |
| 28 | GPT-5 (high) | 44.6 | 36 | $3.438 |
| 29 | GPT-5 Codex (high) | 44.6 | 38.9 | $3.438 |
| 30 | Claude Sonnet 4.6 (Non-reasoning, High Effort) | 44.4 | 46.4 | $6 |
| 31 | GPT-5.4 nano (xhigh) | 44 | 43.9 | $0.463 |
| 32 | KAT Coder Pro V2 | 43.8 | 45.6 | $0.525 |
| 33 | GLM-5.1 (Non-reasoning) | 43.8 | 35.8 | $2.15 |
| 34 | Qwen3.6 35B A3B (Reasoning) | 43.5 | 35.1 | $0.844 |
| 35 | MiMo-V2-Omni | 43.4 | 35.5 | $0 |
| 36 | GPT-5.1 Codex (high) | 43.1 | 36.6 | $3.438 |
| 37 | Claude Opus 4.5 (Non-reasoning) | 43.1 | 42.9 | $10 |
| 38 | Claude 4.5 Sonnet (Reasoning) | 43 | 38.6 | $6 |
| 39 | GLM 5V Turbo (Reasoning) | 42.9 | 36.2 | $0 |
| 40 | Claude Sonnet 4.6 (Non-reasoning, Low Effort) | 42.6 | 43 | $6 |
| 41 | Qwen3.5 27B (Reasoning) | 42.1 | 34.9 | $0.825 |
| 42 | GLM-4.7 (Reasoning) | 42.1 | 36.3 | $1 |
| 43 | GPT-5 (medium) | 42 | 39 | $3.438 |
| 44 | Claude 4.1 Opus (Reasoning) | 42 | 36.5 | $30 |
| 45 | MiniMax-M2.5 | 41.9 | 37.4 | $0.525 |
| 46 | DeepSeek V3.2 (Reasoning) | 41.7 | 36.7 | $0.315 |
| 47 | Qwen3.5 122B A10B (Reasoning) | 41.6 | 34.7 | $1.1 |
| 48 | MiMo-V2-Flash (Feb 2026) | 41.5 | 33.5 | $0.15 |
| 49 | Grok 4 | 41.5 | 40.5 | $6 |
| 50 | Gemini 3 Pro Preview (low) | 41.3 | 39.4 | $4.5 |
| 51 | GPT-5 mini (high) | 41.2 | 35.3 | $0.688 |
| 52 | Kimi K2 Thinking | 40.9 | 34.8 | $1.075 |
| 53 | o3-pro | 40.7 | - | $35 |
| 54 | GLM-5 (Non-reasoning) | 40.6 | 39 | $1.55 |
| 55 | Qwen3.5 397B A17B (Non-reasoning) | 40.1 | 37.4 | $1.35 |
| 56 | Qwen3 Max Thinking | 39.9 | 30.5 | $2.4 |
| 57 | MiniMax-M2.1 | 39.4 | 32.8 | $0.525 |
| 58 | Gemma 4 31B (Reasoning) | 39.2 | 38.7 | $0 |
| 59 | GPT-5 (low) | 39.2 | 30.7 | $3.438 |
| 60 | MiMo-V2-Flash (Reasoning) | 39.2 | 31.8 | $0.15 |
| 61 | Claude 4 Opus (Reasoning) | 39 | 34 | $30 |
| 62 | GPT-5 mini (medium) | 38.9 | 32.9 | $0.688 |
| 63 | Claude 4 Sonnet (Reasoning) | 38.7 | 34.1 | $6 |
| 64 | Grok 4.1 Fast (Reasoning) | 38.6 | 30.9 | $0.275 |
| 65 | Qwen3.5 Omni Plus | 38.6 | 27.6 | $1.5 |
| 66 | GPT-5.1 Codex mini (high) | 38.6 | 36.4 | $0.688 |
| 67 | Step 3.5 Flash 2603 | 38.5 | 34.6 | $0 |
| 68 | o3 | 38.4 | 38.4 | $3.5 |
| 69 | GPT-5.4 nano (medium) | 38.1 | 35 | $0.463 |
| 70 | Step 3.5 Flash | 37.8 | 31.6 | $0.15 |
| 71 | GPT-5.4 mini (medium) | 37.7 | 37.5 | $1.688 |
| 72 | Kimi K2.5 (Non-reasoning) | 37.3 | 25.8 | $1.2 |
| 73 | Qwen3.5 27B (Non-reasoning) | 37.2 | 33.4 | $0.825 |
| 74 | Claude 4.5 Haiku (Reasoning) | 37.1 | 32.6 | $2 |
| 75 | Qwen3.5 35B A3B (Reasoning) | 37.1 | 30.3 | $0.688 |
| 76 | Claude 4.5 Sonnet (Non-reasoning) | 37.1 | 33.5 | $6 |
| 77 | MiniMax-M2 | 36.1 | 29.2 | $0.525 |
| 78 | NVIDIA Nemotron 3 Super 120B A12B (Reasoning) | 36 | 31.2 | $0.412 |
| 79 | KAT-Coder-Pro V1 | 36 | 18.3 | $0.525 |
| 80 | Claude 4.1 Opus (Non-reasoning) | 36 | - | $30 |
| 81 | Qwen3.5 122B A10B (Non-reasoning) | 35.9 | 31.6 | $1.1 |
| 82 | Nova 2.0 Pro Preview (medium) | 35.7 | 30.4 | $3.438 |
| 83 | GPT-5.4 (Non-reasoning) | 35.4 | 41 | $5.625 |
| 84 | Grok 4 Fast (Reasoning) | 35.1 | 27.4 | $0.275 |
| 85 | Gemini 3 Flash Preview (Non-reasoning) | 35 | 37.8 | $1.125 |
| 86 | Claude 3.7 Sonnet (Reasoning) | 34.7 | 27.6 | $6 |
| 87 | Gemini 2.5 Pro | 34.6 | 31.9 | $3.438 |
| 88 | Nova 2.0 Lite (high) | 34.5 | 23.4 | $0.85 |
| 89 | GLM-4.7 (Non-reasoning) | 34.2 | 32 | $1 |
| 90 | DeepSeek V3.1 Terminus (Reasoning) | 33.9 | 33.7 | $1.914 |
| 91 | GPT-5.2 (Non-reasoning) | 33.6 | 34.7 | $4.813 |
| 92 | Gemini 3.1 Flash-Lite Preview | 33.5 | 30.1 | $0.563 |
| 93 | Doubao Seed Code | 33.5 | 31.3 | $0 |
| 94 | gpt-oss-120B (high) | 33.3 | 28.6 | $0.263 |
| 95 | o4-mini (high) | 33.1 | 25.6 | $1.925 |
| 96 | Claude 4 Sonnet (Non-reasoning) | 33 | 30.6 | $6 |
| 97 | Claude 4 Opus (Non-reasoning) | 33 | - | $30 |
| 98 | DeepSeek V3.2 Exp (Reasoning) | 32.9 | 33.3 | $0.315 |
| 99 | Mercury 2 | 32.8 | 30.6 | $0.375 |
| 100 | GLM-4.6 (Reasoning) | 32.5 | 29.5 | $0.981 |
| 101 | Qwen3 Max Thinking (Preview) | 32.5 | 24.5 | $2.4 |
| 102 | Qwen3.5 9B (Reasoning) | 32.4 | 25.3 | $0.096 |
| 103 | Gemma 4 31B (Non-reasoning) | 32.3 | 33.9 | $0 |
| 104 | DeepSeek V3.2 (Non-reasoning) | 32.1 | 34.6 | $0.315 |
| 105 | Grok 3 mini Reasoning (high) | 32.1 | 25.2 | $0.35 |
| 106 | K-EXAONE (Reasoning) | 32.1 | 27 | $0 |
| 107 | Nova 2.0 Pro Preview (low) | 31.9 | 24.5 | $3.438 |
| 108 | Trinity Large Thinking | 31.9 | 27.2 | $0.395 |
| 109 | Qwen3 Max | 31.4 | 26.4 | $2.4 |
| 110 | Gemma 4 26B A4B (Reasoning) | 31.2 | 22.4 | $0.198 |
| 111 | Claude 4.5 Haiku (Non-reasoning) | 31.1 | 29.6 | $2 |
| 112 | Gemini 2.5 Flash Preview (Sep '25) (Reasoning) | 31.1 | 24.6 | $0 |
| 113 | Kimi K2 0905 | 30.9 | 25.9 | $1.075 |
| 114 | o1 | 30.8 | 20.5 | $26.25 |
| 115 | Claude 3.7 Sonnet (Non-reasoning) | 30.8 | 26.7 | $6 |
| 116 | Qwen3.5 35B A3B (Non-reasoning) | 30.7 | 16.8 | $0.688 |
| 117 | MiMo-V2-Flash (Non-reasoning) | 30.4 | 25.8 | $0.15 |
| 118 | Gemini 2.5 Pro Preview (Mar' 25) | 30.3 | 46.7 | $0 |
| 119 | GLM-4.6 (Non-reasoning) | 30.2 | 30.2 | $1 |
| 120 | GLM-4.7-Flash (Reasoning) | 30.1 | 25.9 | $0.152 |
| 121 | Nova 2.0 Lite (medium) | 29.7 | 23.9 | $0.85 |
| 122 | Grok 4.20 0309 (Non-reasoning) | 29.7 | 25.4 | $3 |
| 123 | Gemini 2.5 Pro Preview (May' 25) | 29.5 | - | $3.438 |
| 124 | Qwen3 235B A22B 2507 (Reasoning) | 29.5 | 23.2 | $2.625 |
| 125 | DeepSeek V3.2 Speciale | 29.4 | 37.9 | $0 |
| 126 | ERNIE 5.0 Thinking Preview | 29.1 | 29.2 | $0 |
| 127 | Grok 4.20 0309 v2 (Non-reasoning) | 29 | 22 | $3 |
| 128 | Grok Code Fast 1 | 28.7 | 23.7 | $0.525 |
| 129 | DeepSeek V3.1 Terminus (Non-reasoning) | 28.5 | 31.9 | $0.453 |
| 130 | Nemotron Cascade 2 30B A3B | 28.4 | 25.8 | $0 |
| 131 | DeepSeek V3.2 Exp (Non-reasoning) | 28.4 | 30 | $0.315 |
| 132 | Qwen3 Coder Next | 28.3 | 22.9 | $0.6 |
| 133 | Apriel-v1.5-15B-Thinker | 28.3 | 18.7 | $0 |
| 134 | DeepSeek V3.1 (Non-reasoning) | 28.1 | 28.4 | $0.834 |
| 135 | Nova 2.0 Omni (medium) | 28 | 15.1 | $0.85 |
| 136 | Mistral Small 4 (Reasoning) | 27.8 | 24.3 | $0.263 |
| 137 | DeepSeek V3.1 (Reasoning) | 27.7 | 29.7 | $0.865 |
| 138 | Apriel-v1.6-15B-Thinker | 27.6 | 22 | $0 |
| 139 | Qwen3 VL 235B A22B (Reasoning) | 27.6 | 20.9 | $2.625 |
| 140 | GPT-5.1 (Non-reasoning) | 27.4 | 27.3 | $3.438 |
| 141 | Qwen3.5 9B (Non-reasoning) | 27.3 | 21.4 | $0.08 |
| 142 | Gemma 4 26B A4B (Non-reasoning) | 27.1 | 29.1 | $0 |
| 143 | Magistral Medium 1.2 | 27.1 | 21.7 | $2.75 |
| 144 | DeepSeek R1 0528 (May '25) | 27.1 | 24 | $2.362 |
| 145 | Qwen3.5 4B (Reasoning) | 27.1 | 17.5 | $0.06 |
| 146 | Gemini 2.5 Flash (Reasoning) | 27 | 22.2 | $0.85 |
| 147 | GPT-5 nano (high) | 26.8 | 20.3 | $0.138 |
| 148 | Qwen3 Next 80B A3B (Reasoning) | 26.7 | 19.5 | $1.875 |
| 149 | GLM-4.5 (Reasoning) | 26.4 | 26.3 | $0.843 |
| 150 | GPT-4.1 | 26.3 | 21.8 | $3.5 |
| 151 | Kimi K2 | 26.3 | 22.1 | $1.039 |
| 152 | Qwen3 Max (Preview) | 26.1 | 25.5 | $2.4 |
| 153 | Solar Pro 3 | 25.9 | 13.3 | $0 |
| 154 | Qwen3.5 Omni Flash | 25.9 | 14 | $0.275 |
| 155 | GPT-5 nano (medium) | 25.9 | 22.9 | $0.138 |
| 156 | o3-mini | 25.9 | 17.9 | $1.925 |
| 157 | o1-pro | 25.8 | - | $262.5 |
| 158 | Gemini 2.5 Flash Preview (Sep '25) (Non-reasoning) | 25.7 | 22.1 | $0 |
| 159 | o3-mini (high) | 25.2 | 17.3 | $1.925 |
| 160 | Grok 3 | 25.2 | 19.8 | $6 |
| 161 | Seed-OSS-36B-Instruct | 25.2 | 16.7 | $0.3 |
| 162 | Qwen3 235B A22B 2507 Instruct | 25 | 22.1 | $1.225 |
| 163 | Qwen3 Coder 480B A35B Instruct | 24.8 | 24.6 | $3 |
| 164 | Qwen3 VL 32B (Reasoning) | 24.7 | 14.5 | $2.625 |
| 165 | Nova 2.0 Lite (low) | 24.6 | 13.6 | $0.85 |
| 166 | Sonar Reasoning Pro | 24.6 | - | $0 |
| 167 | gpt-oss-120B (low) | 24.5 | 15.5 | $0.263 |
| 168 | gpt-oss-20B (high) | 24.5 | 18.5 | $0.1 |
| 169 | GPT-5.4 nano (Non-Reasoning) | 24.4 | 27.9 | $0.463 |
| 170 | MiniMax M1 80k | 24.4 | 14.5 | $0.963 |
| 171 | NVIDIA Nemotron 3 Nano 30B A3B (Reasoning) | 24.3 | 19 | $0.105 |
| 172 | Gemini 2.5 Flash Preview (Reasoning) | 24.3 | - | $0 |
| 173 | K2 Think V2 | 24.1 | 15.5 | $0 |
| 174 | LongCat Flash Lite | 23.9 | 16.5 | $0 |
| 175 | GPT-5 (minimal) | 23.9 | 25.1 | $3.438 |
| 176 | HyperCLOVA X SEED Think (32B) | 23.7 | 17.5 | $0 |
| 177 | o1-preview | 23.7 | 34 | $28.875 |
| 178 | Grok 4.1 Fast (Non-reasoning) | 23.6 | 19.5 | $0.275 |
| 179 | K-EXAONE (Non-reasoning) | 23.4 | 13.5 | $0 |
| 180 | GLM-4.6V (Reasoning) | 23.4 | 19.7 | $0.45 |
| 181 | GPT-5.4 mini (Non-Reasoning) | 23.3 | 25.3 | $1.688 |
| 182 | Nova 2.0 Omni (low) | 23.2 | 13.9 | $0.85 |
| 183 | GLM-4.5-Air | 23.2 | 23.8 | $0.425 |
| 184 | Nova 2.0 Pro Preview (Non-reasoning) | 23.1 | 20.5 | $3.438 |
| 185 | Mi:dm K 2.5 Pro | 23.1 | 12.6 | $0 |
| 186 | Grok 4 Fast (Non-reasoning) | 23.1 | 19 | $0.275 |
| 187 | GPT-4.1 mini | 22.9 | 18.5 | $0.7 |
| 188 | Mistral Large 3 | 22.8 | 22.7 | $0.75 |
| 189 | Ring-1T | 22.8 | 16.8 | $0 |
| 190 | Qwen3.5 4B (Non-reasoning) | 22.6 | 13.7 | $0.06 |
| 191 | Qwen3 30B A3B 2507 (Reasoning) | 22.4 | 14.7 | $0.75 |
| 192 | DeepSeek V3 0324 | 22.3 | 22 | $1.25 |
| 193 | INTELLECT-3 | 22.2 | 19.1 | $0 |
| 194 | GLM-4.7-Flash (Non-reasoning) | 22.1 | 11 | $0.152 |
| 195 | Devstral 2 | 22 | 23.7 | $0 |
| 196 | GPT-5 (ChatGPT) | 21.8 | 21.2 | $3.438 |
| 197 | Solar Open 100B (Reasoning) | 21.7 | 10.5 | $0 |
| 198 | Gemini 2.5 Flash-Lite Preview (Sep '25) (Reasoning) | 21.6 | 18.1 | $0.175 |
| 199 | Grok 3 Reasoning Beta | 21.6 | - | $0 |
| 200 | Mistral Medium 3.1 | 21.3 | 18.3 | $0.8 |
| 201 | MiniMax M1 40k | 20.9 | 14.1 | $0 |
| 202 | gpt-oss-20B (low) | 20.8 | 14.4 | $0.1 |
| 203 | Qwen3 VL 235B A22B Instruct | 20.8 | 16.5 | $1.225 |
| 204 | GPT-5 mini (minimal) | 20.7 | 21.9 | $0.688 |
| 205 | K2-V2 (high) | 20.6 | 16.1 | $0 |
| 206 | Gemini 2.5 Flash (Non-reasoning) | 20.6 | 17.8 | $0.85 |
| 207 | o1-mini | 20.4 | - | $0 |
| 208 | Qwen3 Next 80B A3B Instruct | 20.1 | 15.3 | $0.875 |
| 209 | Tri-21B-think Preview | 20 | 7.4 | $0 |
| 210 | GPT-4.5 (Preview) | 20 | - | $0 |
| 211 | Qwen3 Coder 30B A3B Instruct | 20 | 19.4 | $0.9 |
| 212 | Qwen3 235B A22B (Reasoning) | 19.8 | 17.4 | $2.625 |
| 213 | QwQ 32B | 19.7 | - | $0.745 |
| 214 | Qwen3 VL 30B A3B (Reasoning) | 19.7 | 13.1 | $0.75 |
| 215 | Gemini 2.0 Flash Thinking Experimental (Jan '25) | 19.6 | 24.1 | $0 |
| 216 | Devstral Small 2 | 19.5 | 20.7 | $0 |
| 217 | Gemini 2.5 Flash-Lite Preview (Sep '25) (Non-reasoning) | 19.4 | 14.5 | $0.175 |
| 218 | Motif-2-12.7B-Reasoning | 19.1 | 11.9 | $0 |
| 219 | Nova Premier | 19 | 13.8 | $5 |
| 220 | Ling-1T | 19 | 18.8 | $0 |
| 221 | Gemma 4 E4B (Reasoning) | 18.8 | 13.7 | $0 |
| 222 | Magistral Medium 1 | 18.8 | 16 | $0 |
| 223 | Mistral Medium 3 | 18.8 | 13.6 | $0.8 |
| 224 | DeepSeek R1 (Jan '25) | 18.8 | 15.9 | $2.181 |
| 225 | Solar Pro 2 (Preview) (Reasoning) | 18.8 | - | $0 |
| 226 | Llama Nemotron Super 49B v1.5 (Reasoning) | 18.7 | 15.2 | $0.175 |
| 227 | K2-V2 (medium) | 18.7 | 14 | $0 |
| 228 | Claude 3.5 Haiku | 18.7 | 10.7 | $1.6 |
| 229 | Devstral Medium | 18.7 | 15.9 | $0.8 |
| 230 | Mistral Small 4 (Non-reasoning) | 18.6 | 16.4 | $0.263 |
| 231 | Hermes 4 - Llama-3.1 405B (Reasoning) | 18.6 | 16 | $1.5 |
| 232 | Tri-21B-Think | 18.6 | 6.3 | $0 |
| 233 | GPT-4o (Aug '24) | 18.6 | 16.6 | $4.375 |
| 234 | GPT-4o (March 2025, chatgpt-4o-latest) | 18.6 | - | $0 |
| 235 | Llama 3.3 Nemotron Super 49B v1 (Reasoning) | 18.5 | 9.4 | $0 |
| 236 | Gemini 2.0 Flash (Feb '25) | 18.5 | 13.6 | $0.263 |
| 237 | Llama 4 Maverick | 18.4 | 15.6 | $0.5 |
| 238 | Magistral Small 1.2 | 18.2 | 14.8 | $0.75 |
| 239 | Sarvam 105B (high) | 18.2 | 9.8 | $0 |
| 240 | Qwen3 4B 2507 (Reasoning) | 18.2 | 9.5 | $0 |
| 241 | Gemini 2.0 Pro Experimental (Feb '25) | 18.1 | 25.5 | $0 |
| 242 | Nova 2.0 Lite (Non-reasoning) | 18 | 12.5 | $0.85 |
| 243 | Claude 3 Opus | 18 | 19.5 | $30 |
| 244 | Devstral Small (May '25) | 18 | 12.2 | $0.075 |
| 245 | Sonar Reasoning | 17.9 | - | $0 |
| 246 | Gemini 2.5 Flash Preview (Non-reasoning) | 17.8 | - | $0 |
| 247 | Hermes 4 - Llama-3.1 405B (Non-reasoning) | 17.6 | 18.1 | $1.5 |
| 248 | Gemini 2.5 Flash-Lite (Reasoning) | 17.6 | 9.5 | $0.175 |
| 249 | Llama 3.1 Instruct 405B | 17.4 | 14.5 | $3.688 |
| 250 | GPT-4o (Nov '24) | 17.3 | 16.7 | $4.375 |
| 251 | DeepSeek R1 Distill Qwen 32B | 17.2 | - | $0.27 |
| 252 | Qwen3 VL 32B Instruct | 17.2 | 15.6 | $1.225 |
| 253 | GLM-4.6V (Non-reasoning) | 17.1 | 11.1 | $0.45 |
| 254 | Qwen3 235B A22B (Non-reasoning) | 17 | 14 | $1.225 |
| 255 | Gemini 2.0 Flash (experimental) | 16.8 | - | $0 |
| 256 | Magistral Small 1 | 16.8 | 11.1 | $0 |
| 257 | EXAONE 4.0 32B (Reasoning) | 16.7 | 14 | $0 |
| 258 | Qwen3 VL 8B (Reasoning) | 16.7 | 9.8 | $0.66 |
| 259 | Nova 2.0 Omni (Non-reasoning) | 16.6 | 13.8 | $0.85 |
| 260 | DeepSeek V3 (Dec '24) | 16.5 | 16.4 | $0.625 |
| 261 | Qwen3 32B (Reasoning) | 16.5 | 13.8 | $2.625 |
| 262 | DeepSeek R1 0528 Qwen3 8B | 16.4 | 7.8 | $0 |
| 263 | Qwen3.5 2B (Reasoning) | 16.3 | 3.5 | $0.04 |
| 264 | Qwen2.5 Max | 16.3 | - | $2.8 |
| 265 | Qwen3 14B (Reasoning) | 16.2 | 13.1 | $1.313 |
| 266 | Nanbeige4.1-3B | 16.1 | 8.9 | $0 |
| 267 | Qwen3 VL 30B A3B Instruct | 16.1 | 14.3 | $0.35 |
| 268 | Ministral 3 14B | 16 | 10.9 | $0.2 |
| 269 | DeepSeek R1 Distill Llama 70B | 16 | 11.4 | $0.875 |
| 270 | Hermes 4 - Llama-3.1 70B (Reasoning) | 16 | 14.4 | $0.198 |
| 271 | Gemini 1.5 Pro (Sep '24) | 16 | 23.6 | $0 |
| 272 | Solar Pro 2 (Preview) (Non-reasoning) | 16 | - | $0 |
| 273 | Claude 3.5 Sonnet (Oct '24) | 15.9 | 30.2 | $6 |
| 274 | Falcon-H1R-7B | 15.8 | 9.8 | $0 |
| 275 | DeepSeek R1 Distill Qwen 14B | 15.8 | - | $0 |
| 276 | Ling-flash-2.0 | 15.7 | 16.7 | $0.247 |
| 277 | Qwen3 Omni 30B A3B (Reasoning) | 15.6 | 12.7 | $0.43 |
| 278 | Qwen2.5 Instruct 72B | 15.6 | 11.9 | $0 |
| 279 | Sonar | 15.5 | - | $0 |
| 280 | Step3 VL 10B | 15.4 | 13.9 | $0 |
| 281 | Qwen3 30B A3B (Reasoning) | 15.3 | 11 | $0.75 |
| 282 | Gemma 4 E2B (Reasoning) | 15.2 | 9 | $0 |
| 283 | Devstral Small (Jul '25) | 15.2 | 12.1 | $0.15 |
| 284 | Sonar Pro | 15.2 | - | $0 |
| 285 | QwQ 32B-Preview | 15.2 | - | $0.135 |
| 286 | Mistral Large 2 (Nov '24) | 15.1 | 13.8 | $3 |
| 287 | Mistral Small 3.2 | 15.1 | 13.3 | $0.15 |
| 288 | GLM-4.5V (Reasoning) | 15.1 | 10.9 | $0.9 |
| 289 | Llama 3.1 Nemotron Ultra 253B v1 (Reasoning) | 15 | 13.1 | $0.9 |
| 290 | ERNIE 4.5 300B A47B | 15 | 14.5 | $0.485 |
| 291 | Qwen3 30B A3B 2507 Instruct | 15 | 14.2 | $0.35 |
| 292 | Solar Pro 2 (Reasoning) | 14.9 | 12.1 | $0 |
| 293 | NVIDIA Nemotron Nano 12B v2 VL (Reasoning) | 14.9 | 11.8 | $0.3 |
| 294 | Gemma 4 E4B (Non-reasoning) | 14.8 | 6.4 | $0 |
| 295 | Ministral 3 8B | 14.8 | 10 | $0.15 |
| 296 | NVIDIA Nemotron Nano 9B V2 (Reasoning) | 14.8 | 8.3 | $0.07 |
| 297 | NVIDIA Nemotron 3 Nano 4B | 14.7 | 10 | $0 |
| 298 | Qwen3.5 2B (Non-reasoning) | 14.7 | 4.9 | $0.04 |
| 299 | Gemini 2.0 Flash-Lite (Feb '25) | 14.7 | - | $0 |
| 300 | Llama Nemotron Super 49B v1.5 (Non-reasoning) | 14.6 | 10.5 | $0.175 |
| 301 | Llama 3.3 Instruct 70B | 14.5 | 10.7 | $0.675 |
| 302 | GPT-4o (May '24) | 14.5 | 24.2 | $7.5 |
| 303 | Gemini 2.0 Flash-Lite (Preview) | 14.5 | - | $0 |
| 304 | Mistral Small 3.1 | 14.5 | 13.9 | $0.15 |
| 305 | Qwen3 32B (Non-reasoning) | 14.5 | - | $1.225 |
| 306 | Llama 3.1 Nemotron Nano 4B v1.1 (Reasoning) | 14.4 | - | $0 |
| 307 | Kimi Linear 48B A3B Instruct | 14.4 | 14.2 | $0 |
| 308 | K2-V2 (low) | 14.4 | 10.5 | $0 |
| 309 | Llama 3.3 Nemotron Super 49B v1 (Non-reasoning) | 14.3 | 7.6 | $0 |
| 310 | Qwen3 VL 8B Instruct | 14.3 | 7.3 | $0.31 |
| 311 | Claude 3.5 Sonnet (June '24) | 14.2 | 26 | $6 |
| 312 | Qwen3 4B (Reasoning) | 14.2 | - | $0.398 |
| 313 | GPT-4o (ChatGPT) | 14.1 | - | $0 |
| 314 | Llama 3.1 Tulu3 405B | 14.1 | - | $0 |
| 315 | Ring-flash-2.0 | 14 | 10.6 | $0.247 |
| 316 | Pixtral Large | 14 | - | $3 |
| 317 | Olmo 3.1 32B Think | 13.9 | 9.8 | $0 |
| 318 | Grok 2 (Dec '24) | 13.9 | - | $0 |
| 319 | GPT-5 nano (minimal) | 13.8 | 14.2 | $0.138 |
| 320 | Gemini 1.5 Flash (Sep '24) | 13.8 | - | $0 |
| 321 | GPT-4 Turbo | 13.7 | 21.5 | $15 |
| 322 | Qwen3 VL 4B (Reasoning) | 13.7 | 6.7 | $0 |
| 323 | Solar Pro 2 (Non-reasoning) | 13.6 | 11.3 | $0 |
| 324 | Llama 4 Scout | 13.5 | 6.7 | $0.292 |
| 325 | Command A | 13.5 | 9.9 | $4.375 |
| 326 | Nova Pro | 13.5 | 11 | $1.4 |
| 327 | Llama 3.1 Nemotron Instruct 70B | 13.4 | 10.8 | $1.2 |
| 328 | Grok Beta | 13.3 | - | $0 |
| 329 | NVIDIA Nemotron 3 Nano 30B A3B (Non-reasoning) | 13.2 | 15.8 | $0.087 |
| 330 | NVIDIA Nemotron Nano 9B V2 (Non-reasoning) | 13.2 | 7.5 | $0.086 |
| 331 | Qwen2.5 Instruct 32B | 13.2 | - | $0 |
| 332 | Qwen3 8B (Reasoning) | 13.2 | 9 | $0.66 |
| 333 | GPT-4.1 nano | 13 | 11.2 | $0.175 |
| 334 | Mistral Large 2 (Jul '24) | 13 | - | $3 |
| 335 | Qwen2.5 Coder Instruct 32B | 12.9 | - | $0 |
| 336 | Qwen3 4B 2507 Instruct | 12.9 | 9.1 | $0 |
| 337 | GPT-4 | 12.8 | 13.1 | $37.5 |
| 338 | Qwen3 14B (Non-reasoning) | 12.8 | 12.4 | $0.613 |
| 339 | Gemini 2.5 Flash-Lite (Non-reasoning) | 12.7 | 7.4 | $0.175 |
| 340 | Mistral Small 3 | 12.7 | - | $0.15 |
| 341 | Nova Lite | 12.7 | 5.1 | $0.105 |
| 342 | GLM-4.5V (Non-reasoning) | 12.7 | 10.8 | $0.9 |
| 343 | Hermes 4 - Llama-3.1 70B (Non-reasoning) | 12.6 | 9.2 | $0.198 |
| 344 | GPT-4o mini | 12.6 | - | $0.263 |
| 345 | Llama 3.1 Instruct 70B | 12.5 | 10.9 | $0.56 |
| 346 | DeepSeek-V2.5 (Dec '24) | 12.5 | - | $0 |
| 347 | Qwen3 4B (Non-reasoning) | 12.5 | - | $0.188 |
| 348 | Qwen3 30B A3B (Non-reasoning) | 12.5 | 13.3 | $0.35 |
| 349 | Sarvam 30B (high) | 12.3 | 7.9 | $0 |
| 350 | Gemini 2.0 Flash Thinking Experimental (Dec '24) | 12.3 | - | $0 |
| 351 | Claude 3 Haiku | 12.3 | 6.7 | $0.5 |
| 352 | DeepSeek-V2.5 | 12.3 | - | $0 |
| 353 | Olmo 3.1 32B Instruct | 12.2 | 5.6 | $0.3 |
| 354 | Gemma 4 E2B (Non-reasoning) | 12.1 | 8.3 | $0 |
| 355 | Mistral Saba | 12.1 | - | $0 |
| 356 | DeepSeek R1 Distill Llama 8B | 12.1 | - | $0 |
| 357 | Olmo 3 32B Think | 12.1 | 10.5 | $0 |
| 358 | R1 1776 | 12 | - | $0 |
| 359 | Gemini 1.5 Pro (May '24) | 12 | 19.8 | $0 |
| 360 | Reka Flash (Sep '24) | 12 | - | $0.35 |
| 361 | Qwen2.5 Turbo | 12 | - | $0.087 |
| 362 | Llama 3.2 Instruct 90B (Vision) | 11.9 | - | $0.72 |
| 363 | Solar Mini | 11.9 | - | $0.15 |
| 364 | Llama 3.1 Instruct 8B | 11.8 | 4.9 | $0.1 |
| 365 | Grok-1 | 11.7 | - | $0 |
| 366 | EXAONE 4.0 32B (Non-reasoning) | 11.7 | 9.4 | $0 |
| 367 | Qwen2 Instruct 72B | 11.7 | - | $0 |
| 368 | Ministral 3 3B | 11.2 | 4.8 | $0.1 |
| 369 | Gemini 1.5 Flash-8B | 11.1 | - | $0 |
| 370 | DeepHermes 3 - Mistral 24B Preview (Non-reasoning) | 10.9 | - | $0 |
| 371 | Jamba 1.7 Large | 10.9 | 7.8 | $3.5 |
| 372 | Granite 4.0 H Small | 10.8 | 8.5 | $0.107 |
| 373 | Qwen3 Omni 30B A3B Instruct | 10.7 | 7.2 | $0.43 |
| 374 | Jamba 1.5 Large | 10.7 | - | $3.5 |
| 375 | DeepSeek-Coder-V2 | 10.6 | - | $0 |
| 376 | OLMo 2 32B | 10.6 | 2.7 | $0 |
| 377 | Hermes 3 - Llama-3.1 70B | 10.6 | - | $0.3 |
| 378 | Jamba 1.6 Large | 10.6 | - | $3.5 |
| 379 | Qwen3 8B (Non-reasoning) | 10.6 | 7.1 | $0.31 |
| 380 | LFM2 24B A2B | 10.5 | 3.6 | $0.052 |
| 381 | Qwen3.5 0.8B (Reasoning) | 10.5 | 0 | $0.02 |
| 382 | Gemini 1.5 Flash (May '24) | 10.5 | - | $0 |
| 383 | Phi-4 | 10.4 | 11.2 | $0.219 |
| 384 | Nova Micro | 10.3 | 4.1 | $0.061 |
| 385 | Gemma 3 27B Instruct | 10.3 | 9.6 | $0 |
| 386 | Claude 3 Sonnet | 10.3 | - | $6 |
| 387 | Mistral Small (Sep '24) | 10.2 | - | $0.3 |
| 388 | NVIDIA Nemotron Nano 12B v2 VL (Non-reasoning) | 10.1 | 5.9 | $0.3 |
| 389 | Gemma 3n E4B Instruct Preview (May '25) | 10.1 | - | $0 |
| 390 | Gemini 1.0 Ultra | 10.1 | 17.6 | $0 |
| 391 | Phi-3 Mini Instruct 3.8B | 10.1 | 3 | $0 |
| 392 | Phi-4 Multimodal Instruct | 10 | - | $0 |
| 393 | Qwen2.5 Coder Instruct 7B | 10 | - | $0 |
| 394 | Qwen3.5 0.8B (Non-reasoning) | 9.9 | 1 | $0.02 |
| 395 | Mistral Large (Feb '24) | 9.9 | - | $6 |
| 396 | Mixtral 8x22B Instruct | 9.8 | - | $0 |
| 397 | Llama 3.2 Instruct 3B | 9.7 | - | $0.085 |
| 398 | Llama 2 Chat 7B | 9.7 | - | $0.1 |
| 399 | Jamba Reasoning 3B | 9.6 | 2.5 | $0 |
| 400 | Qwen3 VL 4B Instruct | 9.6 | 4.5 | $0 |
| 401 | Reka Flash 3 | 9.5 | 8.9 | $0.35 |
| 402 | Qwen1.5 Chat 110B | 9.5 | - | $0 |
| 403 | Olmo 3 7B Think | 9.4 | 7.6 | $0 |
| 404 | Claude 2.1 | 9.3 | 14 | $0 |
| 405 | OLMo 2 7B | 9.3 | 1.2 | $0 |
| 406 | Molmo 7B-D | 9.2 | 1.2 | $0 |
| 407 | Ling-mini-2.0 | 9.2 | 5 | $0 |
| 408 | Claude 2.0 | 9.1 | 12.9 | $0 |
| 409 | DeepSeek R1 Distill Qwen 1.5B | 9.1 | - | $0 |
| 410 | DeepSeek-V2-Chat | 9.1 | - | $0 |
| 411 | GPT-3.5 Turbo | 9 | 10.7 | $0.75 |
| 412 | Mistral Small (Feb '24) | 9 | - | $1.5 |
| 413 | Mistral Medium | 9 | - | $4.088 |
| 414 | Llama 3 Instruct 70B | 8.9 | 6.8 | $0.871 |
| 415 | Gemma 3 12B Instruct | 8.8 | 6.3 | $0 |
| 416 | LFM 40B | 8.8 | - | $0 |
| 417 | Arctic Instruct | 8.8 | - | $0 |
| 418 | Qwen Chat 72B | 8.8 | - | $0 |
| 419 | Llama 3.2 Instruct 11B (Vision) | 8.7 | 4.3 | $0.245 |
| 420 | PALM-2 | 8.6 | 4.6 | $0 |
| 421 | Gemini 1.0 Pro | 8.5 | - | $0 |
| 422 | DeepSeek Coder V2 Lite Instruct | 8.5 | - | $0 |
| 423 | Phi-4 Mini Instruct | 8.4 | 3.6 | $0 |
| 424 | Llama 2 Chat 13B | 8.4 | - | $0 |
| 425 | Llama 2 Chat 70B | 8.4 | - | $0 |
| 426 | DeepSeek LLM 67B Chat (V1) | 8.4 | - | $0 |
| 427 | Sarvam M (Reasoning) | 8.4 | 7.5 | $0 |
| 428 | Exaone 4.0 1.2B (Reasoning) | 8.3 | 3.1 | $0 |
| 429 | OpenChat 3.5 (1210) | 8.3 | - | $0 |
| 430 | DBRX Instruct | 8.3 | - | $0 |
| 431 | Command-R+ (Apr '24) | 8.3 | - | $6 |
| 432 | Olmo 3 7B Instruct | 8.2 | 3.4 | $0.125 |
| 433 | LFM2.5-1.2B-Thinking | 8.1 | 1.4 | $0 |
| 434 | Exaone 4.0 1.2B (Non-reasoning) | 8.1 | 2.5 | $0 |
| 435 | Jamba 1.7 Mini | 8.1 | 3.1 | $0 |
| 436 | LFM2 2.6B | 8 | 1.4 | $0 |
| 437 | LFM2.5-1.2B-Instruct | 8 | 0.8 | $0 |
| 438 | Granite 4.0 H 1B | 8 | 2.7 | $0 |
| 439 | Jamba 1.5 Mini | 8 | - | $0.25 |
| 440 | Qwen3 1.7B (Reasoning) | 8 | 1.4 | $0.398 |
| 441 | Jamba 1.6 Mini | 7.9 | - | $0.25 |
| 442 | Gemma 3 270M | 7.7 | 0 | $0 |
| 443 | Granite 4.0 Micro | 7.7 | 5 | $0 |
| 444 | Apertus 70B Instruct | 7.7 | 1.9 | $1.345 |
| 445 | Mixtral 8x7B Instruct | 7.7 | - | $0.526 |
| 446 | DeepHermes 3 - Llama-3.1 8B Preview (Non-reasoning) | 7.6 | - | $0 |
| 447 | Llama 65B | 7.4 | - | $0 |
| 448 | Qwen Chat 14B | 7.4 | - | $0 |
| 449 | Claude Instant | 7.4 | 7.8 | $0 |
| 450 | Mistral 7B Instruct | 7.4 | - | $0.25 |
| 451 | Command-R (Mar '24) | 7.4 | - | $0.75 |
| 452 | Molmo2-8B | 7.3 | 4.4 | $0 |
| 453 | Granite 4.0 1B | 7.3 | 2.9 | $0 |
| 454 | LFM2 8B A1B | 7 | 2.3 | $0 |
| 455 | Granite 3.3 8B (Non-reasoning) | 7 | 3.4 | $0.085 |
| 456 | Qwen3 1.7B (Non-reasoning) | 6.8 | 2.3 | $0.188 |
| 457 | Qwen3 0.6B (Reasoning) | 6.5 | 0.9 | $0.398 |
| 458 | Llama 3 Instruct 8B | 6.4 | 4 | $0.07 |
| 459 | Gemma 3n E4B Instruct | 6.4 | 4.2 | $0.025 |
| 460 | Llama 3.2 Instruct 1B | 6.3 | 0.6 | $0.1 |
| 461 | Gemma 3 4B Instruct | 6.3 | 2.9 | $0 |
| 462 | LFM2 1.2B | 6.3 | 0.8 | $0 |
| 463 | LFM2.5-VL-1.6B | 6.2 | 1 | $0 |
| 464 | Granite 4.0 350M | 6.1 | 0.3 | $0 |
| 465 | Apertus 8B Instruct | 5.9 | 1.4 | $0.125 |
| 466 | Qwen3 0.6B (Non-reasoning) | 5.7 | 1.4 | $0.188 |
| 467 | Gemma 3 1B Instruct | 5.5 | 0.2 | $0 |
| 468 | Granite 4.0 H 350M | 5.4 | 0.6 | $0 |
| 469 | Gemma 3n E2B Instruct | 4.8 | 2.2 | $0 |
| 470 | Tiny Aya Global | 4.7 | 1.2 | $0 |
| 471 | GPT-5.4 Pro (xhigh) | - | - | $67.5 |
| 472 | Gemini 3 Deep Think | - | - | $0 |
| 473 | Cogito v2.1 (Reasoning) | - | 24.8 | $1.25 |
| 474 | Mi:dm K 2.5 Pro Preview | - | 11.9 | $0 |
| 475 | GPT-4o mini Realtime (Dec '24) | - | - | $0 |
| 476 | GPT-4o Realtime (Dec '24) | - | - | $0 |
| 477 | GPT-3.5 Turbo (0613) | - | - | $0 |
榜单解读建议
参考 AI 大模型排行榜 时,应综合考虑“综合指数”与“成本价格”。如果您是开发者,编程能力 (Coding) 是更核心的指标。
值品工具箱同步的 AI 大模型排行榜 数据每 24 小时更新,确保您获取到最新的模型性能对比。
指标说明
- ● 综合指数:评估通用理解与逻辑。
- ● 价格 $/1M:混合 3:1 输入输出比的平均成本。
- ● 编程能力:衡量代码生成的准确性。
AI 大模型排行榜 常见问题 (FAQ)
Q1: AI 大模型排行榜 的数据多久更新?
AI 大模型排行榜 数据每 24 小时自动抓取一次,确保最新模型加入列表。
Q2: 这个 AI 大模型排行榜 包含国产模型吗?
是的,只要国产模型通过了 Artificial Analysis 的全球测评,就会出现在 AI 大模型排行榜 中。
Q3: 综合指数在 AI 大模型排行榜 中代表什么?
它代表模型的全能表现。AI 大模型排行榜 通过加权算法给出这个综合评分。
Q4: 如何在 AI 大模型排行榜 中查找性价比最高的游戏?
在 AI 大模型排行榜 页面中,您可以点击“价格”标题进行排序,寻找低价高分的模型。
Q5: AI 大模型排行榜 的编程能力测试准吗?
AI 大模型排行榜 参考了 LiveCodeBench 等权威基准测试,具有极高的参考价值。
Q6: 为什么有的新模型没进入 AI 大模型排行榜?
模型进入 AI 大模型排行榜 需要经过一系列测试,通常在新模型发布后数日内会完成更新。
Q7: AI 大模型排行榜 中的价格计算标准是什么?
价格是基于百万 Token 的调用成本,由 AI 大模型排行榜 统一混合计算得出。
Q8: 手机上能查看 AI 大模型排行榜 吗?
当然可以。AI 大模型排行榜 进行了移动端响应式深度优化。
Q9: AI 大模型排行榜 这个工具免费吗?
是的,由值品工具箱免费提供 AI 大模型排行榜 信息查询服务。
Q10: 我该怎么利用 AI 大模型排行榜 做选型?
如果您需要智能客服,参考 AI 大模型排行榜 的综合指数;如果做翻译,参考编程外的语言指标。