Text
Vision
Image (LM Arena)
Image (Artificial Analysis)
Image (Fal)
Overall
Hard prompts
Hard prompts (english)
Longer query
Multiturn
English
Math
Instruction following
Creative writing
Coding
Chinese
French
German
Japanese
Korean
Russian
Spanish
Exclude <5 tok query
Exclude refusal
Style control
Rank
Model
Rating
95% CI
1
gemini-2.5-pro-exp-03-25
1379
+7/-7
2
chatgpt-4o-latest-20250326
1360
+6/-7
2
gpt-4.5-preview-2025-02-27
1356
+5/-6
4
grok-3-preview-02-24
1327
+5/-5
4
deepseek-v3-0324
open
1321
+8/-7
4
o1-2024-12-17
1320
+4/-4
4
chatgpt-4o-latest-20241120
1318
+4/-4
8
deepseek-r1
open
1316
+6/-6
8
claude-3-7-sonnet-20250219-thinking-32k
1314
+6/-7
8
gemini-2.0-flash-thinking-exp-01-21
1308
+4/-5
11
o1-preview
1303
+3/-4
11
claude-3-7-sonnet-20250219
1300
+5/-7
13
o3-mini-high
1287
+5/-6
13
qwen2.5-max
1286
+4/-6
13
claude-3-5-sonnet-20241022
1285
+3/-3
13
gemini-2.0-flash-001
1283
+5/-5
13
gemma-3-27b-it
open
1281
+6/-6
18
deepseek-v3
open
1277
+4/-5
18
o3-mini
1269
+5/-5
20
gemini-1.5-pro-002
1268
+3/-3
20
gemini-2.0-flash-lite-preview-02-05
1268
+5/-5
20
hunyuan-turbos-20250226
1265
+10/-12
20
command-a-03-2025
open
1263
+7/-8
20
gpt-4o-2024-05-13
1263
+2/-3
20
qwen-plus-0125
1262
+7/-7
26
claude-3-5-sonnet-20240620
1260
+2/-3
26
hunyuan-turbo-0110
1258
+9/-11
26
qwq-32b
open
1258
+6/-8
26
glm-4-plus-0111
1256
+8/-7
26
o1-mini
1255
+3/-4
26
llama-3.1-405b-instruct-bf16
open
1254
+4/-4
32
llama-3.1-405b-instruct-fp8
open
1253
+3/-3
32
llama-4-maverick-17b-128e-instruct
open
new
1252
+9/-10
32
step-2-16k-exp-202412
1251
+8/-8
32
gpt-4o-2024-08-06
1250
+3/-3
32
grok-2-2024-08-13
1249
+3/-4
37
yi-lightning
1246
+4/-4
37
llama-3.3-nemotron-49b-super-v1
open
1246
+12/-14
37
gpt-4-turbo-2024-04-09
1242
+2/-3
37
hunyuan-large-2025-02-10
1242
+9/-9
37
yi-lightning-lite
1239
+5/-5
42
claude-3-opus-20240229
1239
+2/-3
42
llama-3.3-70b-instruct
open
1235
+3/-4
42
glm-4-plus
1235
+3/-4
42
gpt-4-1106-preview
1235
+3/-3
42
gpt-4o-mini-2024-07-18
1234
+3/-3
42
claude-3-5-haiku-20241022
1233
+5/-4
48
mistral-large-2407
open
1232
+3/-3
48
qwen2.5-plus-1127
1231
+5/-6
48
qwen-max-0919
1231
+4/-5
48
gpt-4-0125-preview
1231
+3/-3
48
athene-v2-chat
open
1231
+5/-4
48
gemini-1.5-flash-002
1229
+4/-4
48
hunyuan-standard-2025-02-10
1229
+10/-7
55
grok-2-mini-2024-08-13
1224
+3/-4
55
athene-70b-0725
open
1223
+4/-6
55
qwen2.5-72b-instruct
open
1222
+4/-4
55
mistral-large-2411
open
1221
+4/-4
55
llama-3.1-nemotron-70b-instruct
open
1217
+5/-6
60
llama-3.1-70b-instruct
open
1214
+3/-3
60
llama-3.1-tulu-3-70b
open
1209
+10/-8
60
amazon-nova-pro-v1.0
1209
+6/-5
60
reka-core-20240904
1207
+8/-7
64
gemma-2-27b-it
open
1207
+3/-3
64
yi-large-preview
1207
+3/-3
64
jamba-1.5-large
open
1205
+5/-7
64
llama-3.1-nemotron-51b-instruct
open
1203
+11/-9
64
gemma-2-9b-it-simpo
open
1200
+6/-6
69
gpt-4-0314
1200
+3/-3
69
claude-3-sonnet-20240229
1197
+2/-2
69
command-r-plus-08-2024
open
1197
+5/-6
69
nemotron-4-340b-instruct
open
1196
+4/-5
69
llama-3-70b-instruct
open
1195
+2/-3
69
yi-large
1195
+5/-5
69
reka-flash-20240904
1193
+8/-7
76
mistral-small-24b-instruct-2501
open
1190
+5/-5
76
qwen2.5-coder-32b-instruct
open
1190
+6/-8
76
glm-4-0520
1188
+5/-6
76
gpt-4-0613
1185
+2/-3
76
c4ai-aya-expanse-32b
open
1185
+4/-4
76
command-r-plus
open
1183
+2/-3
76
amazon-nova-lite-v1.0
1182
+4/-5
83
qwen2-72b-instruct
open
1181
+3/-3
83
gemma-2-9b-it
open
1180
+3/-3
83
gemini-1.5-flash-8b-001
1179
+4/-4
83
claude-3-haiku-20240307
1178
+2/-3
83
phi-4
open
1176
+5/-5
83
command-r-08-2024
open
1176
+5/-7
89
qwen-max-0428
1172
+3/-4
90
amazon-nova-micro-v1.0
1164
+4/-6
90
glm-4-0116
1163
+7/-7
90
jamba-1.5-mini
open
1160
+7/-6
90
ministral-8b-2410
open
1160
+9/-7
90
claude-1
1159
+4/-4
90
mistral-large-2402
1158
+3/-3
90
hunyuan-standard-256k
1153
+11/-10
97
reka-flash-21b-20240226-online
1151
+4/-5
97
c4ai-aya-expanse-8b
open
1151
+5/-6
97
mixtral-8x22b-instruct-v0.1
open
1149
+2/-3
97
mistral-next
1149
+6/-6
97
command-r
open
1148
+2/-3
97
llama-3.1-tulu-3-8b
open
1147
+11/-10
97
claude-2.0
1146
+4/-5
104
llama-3-8b-instruct
open
1143
+2/-3
104
reka-flash-21b-20240226
1142
+4/-4
104
gpt-3.5-turbo-0314
1142
+9/-8
104
mistral-medium
1141
+3/-4
104
gpt-3.5-turbo-0125
1137
+3/-3
109
gpt-3.5-turbo-0613
1137
+3/-4
109
claude-2.1
1135
+4/-4
109
granite-3.1-8b-instruct
open
1135
+8/-12
109
yi-1.5-34b-chat
open
1133
+3/-4
109
llama-3.1-8b-instruct
open
1133
+3/-3
109
zephyr-orpo-141b-A35b-v0.1
open
1130
+10/-7
115
claude-instant-1
1127
+4/-5
116
phi-3-medium-4k-instruct
open
1118
+4/-4
116
gemma-2-2b-it
open
1118
+3/-4
116
internlm2_5-20b-chat
open
1116
+6/-5
116
mixtral-8x7b-instruct-v0.1
open
1114
+0/-0
116
dbrx-instruct-preview
open
1113
+4/-3
121
gpt-3.5-turbo-1106
1109
+4/-5
121
granite-3.0-8b-instruct
open
1109
+7/-7
121
wizardlm-70b
open
1105
+7/-6
121
granite-3.1-2b-instruct
open
1104
+10/-9
121
snowflake-arctic-instruct
open
1102
+3/-3
121
yi-34b-chat
open
1101
+4/-4
121
openchat-3.5-0106
open
1100
+5/-5
128
phi-3-small-8k-instruct
open
1098
+4/-5
128
openchat-3.5
open
1094
+5/-6
128
llama-3.2-3b-instruct
open
1093
+8/-5
128
starling-lm-7b-beta
open
1092
+5/-4
128
vicuna-33b
open
1091
+4/-4
128
openhermes-2.5-mistral-7b
open
1090
+8/-7
128
qwq-32b-preview
open
1088
+10/-10
135
starling-lm-7b-alpha
open
1083
+4/-5
135
granite-3.0-2b-instruct
open
1081
+8/-7
135
pplx-70b-online
1076
+6/-7
135
nous-hermes-2-mixtral-8x7b-dpo
open
1072
+10/-10
135
dolphin-2.2.1-mistral-7b
1069
+12/-12
140
phi-3-mini-4k-instruct-june-2024
open
1068
+5/-5
140
mistral-7b-instruct-v0.2
open
1067
+4/-4
140
solar-10.7b-instruct-v1.0
1066
+6/-8
140
wizardlm-13b
open
1063
+7/-6
140
mpt-30b-chat
open
1059
+11/-9
140
falcon-180b-chat
open
1059
+11/-15
140
vicuna-13b
open
1058
+5/-5
147
phi-3-mini-4k-instruct
open
1056
+4/-4
147
smollm2-1.7b-instruct
open
1051
+12/-14
147
zephyr-7b-beta
open
1049
+5/-7
147
phi-3-mini-128k-instruct
open
1049
+4/-4
147
codellama-34b-instruct
open
1047
+6/-6
147
zephyr-7b-alpha
open
1046
+12/-12
153
llama-3.2-1b-instruct
open
1044
+8/-6
153
palm-2
1042
+6/-6
153
pplx-7b-online
1041
+6/-7
153
guanaco-33b
open
1040
+9/-12
153
codellama-70b-instruct
open
1039
+15/-15
153
stripedhyena-nous-7b
open
1034
+6/-8
159
mistral-7b-instruct
open
1026
+6/-6
159
vicuna-7b
open
1021
+7/-6
161
olmo-7b-instruct
open
993
+6/-6
162
koala-13b
open
980
+7/-7
162
chatglm3-6b
open
974
+10/-8
162
gpt4all-13b-snoozy
open
973
+13/-15
162
mpt-7b-chat
open
967
+10/-11
162
alpaca-13b
967
+8/-7
167
RWKV-4-Raven-14B
open
949
+9/-8
167
chatglm2-6b
open
942
+10/-12
169
oasst-pythia-12b
open
931
+8/-9
170
fastchat-t5-3b
open
896
+9/-10
170
chatglm-6b
open
888
+9/-9
170
dolly-v2-12b
open
881
+9/-10
173
stablelm-tuned-alpha-7b
open
859
+13/-10
Settings
Open models only
Visualize scores
With moats
With charts
Price ranges
<$0.20
$0.20-$1
$1-$10
$10+
Filter models
Never
Drop deprecated
Drop old
Best / org
Remember: You need a 70 point difference for a 60% win rate
Share
Done