Models
stringlengths 5
34
| Data Source
stringclasses 4
values | Model Size(B)
stringlengths 1
5
| Overall
float64 0.11
0.93
| Biology
stringlengths 1
6
| Business
stringlengths 1
6
| Chemistry
stringlengths 1
6
| Computer Science
stringlengths 1
6
| Economics
stringlengths 1
6
| Engineering
stringlengths 1
6
| Health
stringlengths 1
6
| History
stringlengths 1
6
| Law
stringlengths 1
6
| Math
stringlengths 1
6
| Philosophy
stringlengths 1
6
| Physics
stringlengths 1
6
| Psychology
stringlengths 1
6
| Other
stringlengths 1
6
⌀ |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Yi-Lightning
|
TIGER-Lab
|
unk
| 0.6238 |
0.7964
|
0.6907
|
0.6193
|
0.6439
|
0.731
|
0.4221
|
0.6553
|
0.5748
|
0.3751
|
0.6913
|
0.5711
|
0.6251
|
0.7293
|
0.6677
|
SmolLM-135M
|
TIGER-Lab
|
0.135
| 0.1122 |
0.1227
|
0.1191
|
0.1051
|
0.1073
|
0.1173
|
0.0939
|
0.1247
|
0.0945
|
0.1153
|
0.1140
|
0.1563
|
0.0970
|
0.1090
|
0.1136
|
SmolLM-360M
|
TIGER-Lab
|
0.36
| 0.1095 |
0.1116
|
0.0989
|
0.1007
|
0.1098
|
0.1209
|
0.0877
|
0.1259
|
0.1129
|
0.1117
|
0.1044
|
0.1303
|
0.1001
|
0.1266
|
0.1169
|
Llama-3.2-1B
|
TIGER-Lab
|
1
| 0.1195 |
0.1325
|
0.1077
|
0.1095
|
0.1122
|
0.1256
|
0.1342
|
0.1247
|
0.0892
|
0.1317
|
0.1021
|
0.1323
|
0.1147
|
0.1253
|
0.1277
|
Llama-3.2-3B
|
TIGER-Lab
|
3
| 0.2217 |
0.3919
|
0.2041
|
0.1458
|
0.2049
|
0.3258
|
0.1538
|
0.2518
|
0.2782
|
0.1626
|
0.2021
|
0.2365
|
0.1493
|
0.3033
|
0.2543
|
SmolLM-1.7B
|
TIGER-Lab
|
1.7
| 0.1193 |
0.1046
|
0.1052
|
0.1069
|
0.1268
|
0.1197
|
0.1259
|
0.1467
|
0.1181
|
0.1371
|
0.1110
|
0.1423
|
0.1132
|
0.1115
|
0.1169
|
Granite-3.0-2B-Base
|
TIGER-Lab
|
2
| 0.2172 |
0.3445
|
0.1977
|
0.1564
|
0.2659
|
0.3033
|
0.1465
|
0.2298
|
0.1837
|
0.1653
|
0.2058
|
0.2385
|
0.1640
|
0.3271
|
0.2327
|
Granite-3.0-8B-Base
|
TIGER-Lab
|
8
| 0.3103 |
0.4728
|
0.3080
|
0.2217
|
0.3268
|
0.4135
|
0.2136
|
0.3863
|
0.3307
|
0.2316
|
0.2805
|
0.3467
|
0.2494
|
0.4336
|
0.3149
|
SmolLM2-1.7B
|
TIGER-Lab
|
1.7
| 0.1831 |
0.1925
|
0.1762
|
0.1422
|
0.1634
|
0.2536
|
0.1373
|
0.2054
|
0.1575
|
0.1444
|
0.1673
|
0.2144
|
0.1655
|
0.2845
|
0.2045
|
SmolLM2-360M
|
TIGER-Lab
|
0.36
| 0.1138 |
0.1269
|
0.1242
|
0.1131
|
0.1024
|
0.1209
|
0.1135
|
0.1112
|
0.1155
|
0.1299
|
0.1088
|
0.1202
|
0.0993
|
0.1015
|
0.1115
|
SmolLM2-135M
|
TIGER-Lab
|
0.135
| 0.1085 |
0.1185
|
0.1179
|
0.0945
|
0.0927
|
0.1209
|
0.0815
|
0.1186
|
0.1024
|
0.1199
|
0.1192
|
0.1182
|
0.1032
|
0.1140
|
0.0952
|
Aya-Expanse-8B
|
TIGER-Lab
|
8
| 0.3374 |
0.5579
|
0.3739
|
0.2217
|
0.3585
|
0.4336
|
0.2497
|
0.3741
|
0.3255
|
0.2307
|
0.3605
|
0.2966
|
0.2671
|
0.4599
|
0.3517
|
Aya-Expanse-32B
|
TIGER-Lab
|
32
| 0.4541 |
0.6165
|
0.4804
|
0.3481
|
0.5122
|
0.5829
|
0.2838
|
0.5428
|
0.4961
|
0.3415
|
0.4152
|
0.4609
|
0.3949
|
0.6103
|
0.5108
|
Athene-V2-Chat
|
TIGER-Lab
|
72
| 0.7021 |
0.8243
|
0.7351
|
0.7208
|
0.7341
|
0.8045
|
0.5470
|
0.6834
|
0.6273
|
0.4832
|
0.8061
|
0.6273
|
0.7336
|
0.7694
|
0.7056
|
Claude-3-5-Haiku-20241022
|
TIGER-Lab
|
unk
| 0.6212 |
0.8075
|
0.6857
|
0.5936
|
0.6463
|
0.7287
|
0.4231
|
0.6614
|
0.622
|
0.4668
|
0.6173
|
0.5852
|
0.5835
|
0.7544
|
0.6645
|
Gemma-2-27B-it
|
TIGER-Lab
|
27
| 0.5654 |
0.7796
|
0.6008
|
0.5371
|
0.5683
|
0.6979
|
0.3488
|
0.6186
|
0.5722
|
0.3951
|
0.5611
|
0.5230
|
0.5296
|
0.7155
|
0.6115
|
Mistral-Large-Instruct-2407
|
TIGER-Lab
|
123
| 0.6591 |
0.8271
|
0.6198
|
0.6307
|
0.7024
|
0.7571
|
0.4396
|
0.7152
|
0.6299
|
0.4823
|
0.7128
|
0.6453
|
0.6913
|
0.7707
|
0.6786
|
Mixtral-8x22B-Instruct-v0.1
|
TIGER-Lab
|
176
| 0.5633 |
0.7517
|
0.6033
|
0.4885
|
0.6415
|
0.6742
|
0.4324
|
0.6149
|
0.5643
|
0.3878
|
0.5655
|
0.5491
|
0.5089
|
0.7030
|
0.5996
|
Claude-3-Haiku-20240307
|
TIGER-Lab
|
unk
| 0.4229 |
0.7099
|
0.3663
|
0.2739
|
0.4049
|
0.5723
|
0.3209
|
0.5306
|
0.4357
|
0.3397
|
0.3397
|
0.4629
|
0.3226
|
0.6466
|
0.4556
|
Athene-V2-Chat (0-shot)
|
TIGER-Lab
|
72
| 0.7311 |
0.8689
|
0.7921
|
0.7473
|
0.7463
|
0.8081
|
0.5996
|
0.7164
|
0.6404
|
0.5095
|
0.8364
|
0.6293
|
0.7737
|
0.7857
|
0.7219
|
EXAONE-3.5-2.4B-Instruct
|
TIGER-Lab
|
2.4
| 0.391 |
0.6541
|
0.3942
|
0.3171
|
0.4415
|
0.5261
|
0.3168
|
0.3851
|
0.3727
|
0.2498
|
0.4323
|
0.3026
|
0.3472
|
0.5038
|
0.3387
|
EXAONE-3.5-7.8B-Instruct
|
TIGER-Lab
|
7.8
| 0.4624 |
0.7308
|
0.4740
|
0.3719
|
0.5537
|
0.6066
|
0.3767
|
0.4707
|
0.4514
|
0.2870
|
0.4996
|
0.4248
|
0.3818
|
0.5965
|
0.4426
|
EXAONE-3.5-32B-Instruct
|
TIGER-Lab
|
32
| 0.5891 |
0.7573
|
0.6502
|
0.5830
|
0.6634
|
0.6908
|
0.4654
|
0.6002
|
0.5328
|
0.4005
|
0.6691
|
0.5371
|
0.5350
|
0.6830
|
0.5617
|
QwQ-32B
|
TIGER-Lab
|
32
| 0.6907 |
0.8410
|
0.7668
|
0.7041
|
0.7366
|
0.7832
|
0.4871
|
0.7066
|
0.6325
|
0.4777
|
0.7876
|
0.6713
|
0.7028
|
0.7632
|
0.6537
|
Llama-3.3-70B-Instruct
|
TIGER-Lab
|
70
| 0.6592 |
0.8187
|
0.6857
|
0.6246
|
0.7073
|
0.7784
|
0.4665
|
0.7115
|
0.6614
|
0.4796
|
0.6891
|
0.6353
|
0.6428
|
0.7794
|
0.6818
|
Mistral-Large-Instruct-2411
|
TIGER-Lab
|
123
| 0.6794 |
0.8368
|
0.7186
|
0.6290
|
0.7683
|
0.7642
|
0.5119
|
0.7457
|
0.6640
|
0.5041
|
0.7091
|
0.6533
|
0.6605
|
0.7845
|
0.7067
|
Gemini-2.0-Flash-exp
|
TIGER-Lab
|
unk
| 0.7624 |
0.8836
|
0.7985
|
0.8004
|
0.799
|
0.8169
|
0.6155
|
0.7442
|
0.7008
|
0.5647
|
0.8638
|
0.6994
|
0.8127
|
0.7905
|
0.7476
|
Granite-3.1-1B-A400M-Base
|
TIGER-Lab
|
1
| 0.1234 |
0.1353
|
0.1153
|
0.1246
|
0.1415
|
0.1422
|
0.0980
|
0.1308
|
0.1234
|
0.1126
|
0.1310
|
0.1683
|
0.1124
|
0.1078
|
0.1212
|
Granite-3.1-1B-A400M-Instruct
|
TIGER-Lab
|
1
| 0.1327 |
0.1437
|
0.1267
|
0.1148
|
0.1610
|
0.1576
|
0.1125
|
0.1638
|
0.1129
|
0.1253
|
0.1303
|
0.1202
|
0.1209
|
0.1617
|
0.1288
|
Granite-3.1-2B-Base
|
TIGER-Lab
|
2
| 0.2389 |
0.3752
|
0.2560
|
0.1696
|
0.2439
|
0.3092
|
0.1930
|
0.2604
|
0.2178
|
0.1253
|
0.2487
|
0.2525
|
0.1986
|
0.3421
|
0.2565
|
Granite-3.1-2B-Instruct
|
TIGER-Lab
|
2
| 0.3197 |
0.5007
|
0.3308
|
0.2412
|
0.3707
|
0.4111
|
0.2528
|
0.3056
|
0.3045
|
0.2180
|
0.3442
|
0.2846
|
0.2648
|
0.4411
|
0.3258
|
Granite-3.1-3B-A800M-Base
|
TIGER-Lab
|
3
| 0.2039 |
0.2957
|
0.1762
|
0.1405
|
0.2268
|
0.2737
|
0.1527
|
0.2286
|
0.1995
|
0.1444
|
0.2198
|
0.2305
|
0.1640
|
0.3083
|
0.1926
|
Granite-3.1-3B-A800M-Instruct
|
TIGER-Lab
|
3
| 0.2542 |
0.3431
|
0.2725
|
0.1608
|
0.2829
|
0.3626
|
0.1744
|
0.2641
|
0.2415
|
0.1708
|
0.2754
|
0.2806
|
0.2017
|
0.3810
|
0.2706
|
Granite-3.1-8B-Base
|
TIGER-Lab
|
8
| 0.3308 |
0.4979
|
0.3181
|
0.2403
|
0.3390
|
0.4372
|
0.2425
|
0.3716
|
0.3412
|
0.2044
|
0.3249
|
0.3567
|
0.2748
|
0.4862
|
0.3636
|
Granite-3.1-8B-Instruct
|
TIGER-Lab
|
8
| 0.4103 |
0.5746
|
0.4563
|
0.3145
|
0.4244
|
0.5047
|
0.2910
|
0.4707
|
0.4121
|
0.2607
|
0.4189
|
0.4329
|
0.3472
|
0.5739
|
0.4405
|
Deepseek-V3
|
Self-Reported
|
671
| 0.7587 |
0.8689
|
0.8099
|
0.7951
|
0.7951
|
0.8175
|
0.6223
|
0.7372
|
0.6824
|
0.5477
|
0.8616
|
0.7154
|
0.7898
|
0.7882
|
0.7641
|
QwQ-32B-Preview
|
Self-Reported
|
32
| 0.7097 |
0.8490
|
0.7506
|
0.7429
|
0.4571
|
0.5469
|
0.6569
|
0.7784
|
0.7506
|
0.7605
|
0.6916
|
0.8494
|
0.6573
|
0.7732
| null |
SkyThought-T1
|
Self-Reported
|
32
| 0.692 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
MiniMax-Text-01
|
Self-Reported
|
456
| 0.757 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Llama-3.1-405B-Instruct
|
Self-Reported
|
405
| 0.733 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Llama-3.1-405B
|
Self-Reported
|
405
| 0.616 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Internlm3-8B-Instruct
|
Self-Reported
|
8
| 0.576 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Hunyuan-Large
|
Self-Reported
|
389
| 0.602 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Phi-4
|
Self-Reported
|
14
| 0.704 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
GPT-o1-mini
|
Self-Reported
|
unk
| 0.803 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
GPT-4o (2024-11-20)
|
Self-Reported
|
unk
| 0.779 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Claude-3.5-Sonnet (2024-10-22)
|
Self-Reported
|
unk
| 0.78 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
DeepSeek-R1
|
Self-Reported
|
671
| 0.84 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Doubao-1.5-Pro
|
Self-Reported
|
unk
| 0.801 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Qwen2.5-Max
|
Self-Reported
|
unk
| 0.761 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Mistral-Small-base
|
Self-Reported
|
24
| 0.544 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Mistral-Small-instruct
|
Self-Reported
|
24
| 0.663 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Gemini-2.0-Flash-Lite
|
Self-Reported
|
unk
| 0.716 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Gemini-2.0-Flash
|
Self-Reported
|
unk
| 0.776 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Gemini-2.0-Pro
|
Self-Reported
|
unk
| 0.791 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
azerogpt
|
Self-Reported
|
unk
| 0.6307 |
0.8215
|
0.6667
|
0.508
|
0.6683
|
0.7393
|
0.4828
|
0.665
|
0.6693
|
0.4587
|
0.6329
|
0.6212
|
0.5751
|
0.7506
|
0.6591
|
Grok3-mini-Beta
|
Self-Reported
|
unk
| 0.789 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Grok3-Beta
|
Self-Reported
|
unk
| 0.799 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Phi-4-mini
|
Self-Reported
|
5.6
| 0.528 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
LLaDA
|
Self-Reported
|
8
| 0.37 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
LLaDA
|
Self-Reported
|
8
| 0.37 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
HunyuanTurboS
|
Self-Reported
|
560
| 0.79 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Reka 3
|
Self-Reported
|
21
| 0.65 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Gemma-2-2B-it
|
Self-Reported
|
2
| 0.156 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Gemma-3-1B-it
|
Self-Reported
|
1
| 0.147 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Gemma-3-4B-it
|
Self-Reported
|
4
| 0.436 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Gemma-3-12B-it
|
Self-Reported
|
12
| 0.606 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Gemma-3-27B-it
|
Self-Reported
|
27
| 0.675 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Cohere-Aya-Vision
|
Self-Reported
|
32
| 0.472 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Mistral-3.1-Small
|
Sefl-Reported
|
24
| 0.668 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
GPT-o1
|
Self-Reported
|
unk
| 0.893 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
GPT-4.5
|
Self-Reported
|
unk
| 0.861 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Hunyuan-T1
|
Self-Reported
|
389
| 0.872 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Nemotron-H-56B-Base
|
Self-Reported
|
56
| 0.605 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Nemotron-H-8B-Base
|
Self-Reported
|
8
| 0.44 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Deepseek-V3-0324
|
Self-Reported
|
671
| 0.813 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Llama4-Maverick
|
Self-Reported
|
400
| 0.805 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Llama4-Scout
|
Self-Reported
|
109
| 0.743 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Llama4-Behemoth
|
Self-Reported
|
2000
| 0.828 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Gemini-2.5-Pro-Exp-03-25
|
TIGER-Lab
|
unk
| 0.8452 |
0.9299
|
0.893
|
0.8741
|
0.8556
|
0.8829
|
0.7358
|
0.817
|
0.7553
|
0.7244
|
0.8884
|
0.8327
|
0.8831
|
0.8726
|
0.8295
|
Grok-3-mini
|
Self-Reported
|
unk
| 0.83 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
GPT-4-mini (high)
|
Self-Reported
|
unk
| 0.83 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Claude-3.7-Sonnet-Thinking
|
Self-Reported
|
unk
| 0.84 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Qwen3-235B-A22B
|
Self-Reported
|
235
| 0.6818 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Mimo-7B-base
|
Self-Reported
|
7
| 0.419 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Mimo-7B-RL
|
Self-Reported
|
7
| 0.586 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
General-Reasoner-7B
|
Self-Reported
|
7
| 0.589 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
General-Reasoner-14B
|
Self-Reported
|
14
| 0.666 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Phi-4-reasoning
|
Self-Reported
|
14
| 0.743 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Phi-4-reasoning-plus
|
Self-Reported
|
14
| 0.76 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
GPT-o3-mini
|
Self-Reported
|
unk
| 0.794 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
ECHO_Ego_v2_14B
|
Self-Reported
|
14
| 0.7124 |
0.8396
|
0.7807
|
0.7624
|
0.739
|
0.7642
|
0.5641
|
0.6993
|
0.6063
|
0.4069
|
0.8764
|
0.6152
|
0.8014
|
0.7419
|
0.671
|
MindLink_Beta
|
Self-Reported
|
unk
| 0.932 |
0.9623
|
0.9404
|
0.9567
|
0.9415
|
0.9408
|
0.9309
|
0.8961
|
0.8583
|
0.8783
|
0.9793
|
0.8918
|
0.9623
|
0.9135
|
0.9123
|
Hunyuan-A13B
|
Self-Reported
|
80
| 0.673 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
ERNIE-4.5-21B-A3B-Base
|
Self-Reported
|
21
| 0.567 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
ERNIE-4.5-300B-A47B-Base
|
Self-Reported
|
300
| 0.695 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
ERNIE-4.5-300B-A47B
|
Self-Reported
|
300
| 0.784 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
MiniMax-M1
|
Self-Reported
|
456
| 0.811 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Seed-Thinking-v1.5
|
Self-Reported
|
200
| 0.87 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Kimi-K2-Instruct
|
Self-Reported
|
1000
| 0.81 |
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
-
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.