CultriX commited on
Commit
674fbc9
·
verified ·
1 Parent(s): d8ef618

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +608 -517
app.py CHANGED
@@ -282,525 +282,616 @@ def download_all_data():
282
  # This is your larger dataset, rank = 44..105
283
  benchmark_data = [
284
  {
285
- "rank": 44,
286
- "name": "sometimesanotion/Qwen2.5-14B-Vimarckoso-v3",
287
- "scores": {
288
- "average": 40.10,
289
- "IFEval": 72.57,
290
- "BBH": 48.58,
291
- "MATH": 34.44,
292
- "GPQA": 17.34,
293
- "MUSR": 19.39,
294
- "MMLU-PRO": 48.26
295
- },
296
- "hf_url": "https://huggingface.co/sometimesanotion/Qwen2.5-14B-Vimarckoso-v3",
297
- "known_config": {
298
- "models": [
299
- {"model": "CultriX/SeQwence-14Bv1"},
300
- {"model": "allknowingroger/Qwenslerp5-14B"}
301
- ],
302
- "merge_method": "slerp",
303
- "base_model": "CultriX/SeQwence-14Bv1",
304
- "dtype": "bfloat16",
305
- "parameters": {
306
- "t": [0, 0.5, 1, 0.5, 0]
307
- }
308
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
309
  },
310
  {
311
- "rank": 45,
312
- "name": "sthenno-com/miscii-14b-1225",
313
- "scores": {
314
- "average": 40.08,
315
- "IFEval": 78.78,
316
- "BBH": 50.91,
317
- "MATH": 31.57,
318
- "GPQA": 17.00,
319
- "MUSR": 14.77,
320
- "MMLU-PRO": 47.46
321
- },
322
- "hf_url": "https://huggingface.co/sthenno-com/miscii-14b-1225",
323
- "known_config": None
324
- },
325
- {
326
- "rank": 46,
327
- "name": "djuna/Q2.5-Veltha-14B-0.5",
328
- "scores": {
329
- "average": 39.96,
330
- "IFEval": 77.96,
331
- "BBH": 50.32,
332
- "MATH": 33.84,
333
- "GPQA": 15.77,
334
- "MUSR": 14.17,
335
- "MMLU-PRO": 47.72
336
- },
337
- "hf_url": "https://huggingface.co/djuna/Q2.5-Veltha-14B-0.5",
338
- "known_config": None
339
- },
340
- {
341
- "rank": 48,
342
- "name": "sometimesanotion/Qwen2.5-14B-Vimarckoso-v3-model_stock",
343
- "scores": {
344
- "average": 39.81,
345
- "IFEval": 71.62,
346
- "BBH": 48.76,
347
- "MATH": 33.99,
348
- "GPQA": 17.34,
349
- "MUSR": 19.23,
350
- "MMLU-PRO": 47.95
351
- },
352
- "hf_url": "https://huggingface.co/sometimesanotion/Qwen2.5-14B-Vimarckoso-v3-model_stock",
353
- "known_config": None
354
- },
355
- {
356
- "rank": 50,
357
- "name": "sometimesanotion/Qwen2.5-14B-Vimarckoso-v3-Prose01",
358
- "scores": {
359
- "average": 39.46,
360
- "IFEval": 68.72,
361
- "BBH": 47.71,
362
- "MATH": 35.05,
363
- "GPQA": 18.23,
364
- "MUSR": 19.56,
365
- "MMLU-PRO": 47.50
366
- },
367
- "hf_url": "https://huggingface.co/sometimesanotion/Qwen2.5-14B-Vimarckoso-v3-Prose01",
368
- "known_config": None
369
- },
370
- {
371
- "rank": 52,
372
- "name": "arcee-ai/Virtuoso-Small",
373
- "scores": {
374
- "average": 39.43,
375
- "IFEval": 79.35,
376
- "BBH": 50.40,
377
- "MATH": 34.29,
378
- "GPQA": 11.52,
379
- "MUSR": 14.44,
380
- "MMLU-PRO": 46.57
381
- },
382
- "hf_url": "https://huggingface.co/arcee-ai/Virtuoso-Small",
383
- "known_config": None
384
- },
385
- {
386
- "rank": 54,
387
- "name": "sometimesanotion/Qwentinuum-14B-v6",
388
- "scores": {
389
- "average": 39.23,
390
- "IFEval": 63.04,
391
- "BBH": 50.23,
392
- "MATH": 33.84,
393
- "GPQA": 18.23,
394
- "MUSR": 21.18,
395
- "MMLU-PRO": 48.89
396
- },
397
- "hf_url": "https://huggingface.co/sometimesanotion/Qwentinuum-14B-v6",
398
- "known_config": None
399
- },
400
- {
401
- "rank": 55,
402
- "name": "djuna/Q2.5-Veltha-14B",
403
- "scores": {
404
- "average": 39.21,
405
- "IFEval": 82.92,
406
- "BBH": 49.75,
407
- "MATH": 28.02,
408
- "GPQA": 14.54,
409
- "MUSR": 12.26,
410
- "MMLU-PRO": 47.76
411
- },
412
- "hf_url": "https://huggingface.co/djuna/Q2.5-Veltha-14B",
413
- "known_config": None
414
- },
415
- {
416
- "rank": 57,
417
- "name": "allknowingroger/QwenSlerp6-14B",
418
- "scores": {
419
- "average": 39.02,
420
- "IFEval": 68.67,
421
- "BBH": 47.59,
422
- "MATH": 34.14,
423
- "GPQA": 16.44,
424
- "MUSR": 18.32,
425
- "MMLU-PRO": 48.95
426
- },
427
- "hf_url": "https://huggingface.co/allknowingroger/QwenSlerp6-14B",
428
- "known_config": None
429
- },
430
- {
431
- "rank": 58,
432
- "name": "allknowingroger/QwenSlerp5-14B",
433
- "scores": {
434
- "average": 38.94,
435
- "IFEval": 71.19,
436
- "BBH": 47.39,
437
- "MATH": 33.16,
438
- "GPQA": 15.32,
439
- "MUSR": 17.81,
440
- "MMLU-PRO": 48.78
441
- },
442
- "hf_url": "https://huggingface.co/allknowingroger/QwenSlerp5-14B",
443
- "known_config": None
444
- },
445
- {
446
- "rank": 59,
447
- "name": "sometimesanotion/Qwentinuum-14B-v5",
448
- "scores": {
449
- "average": 38.87,
450
- "IFEval": 62.86,
451
- "BBH": 50.28,
452
- "MATH": 31.57,
453
- "GPQA": 18.34,
454
- "MUSR": 21.09,
455
- "MMLU-PRO": 49.09
456
- },
457
- "hf_url": "https://huggingface.co/sometimesanotion/Qwentinuum-14B-v5",
458
- "known_config": None
459
- },
460
- {
461
- "rank": 60,
462
- "name": "sometimesanotion/Qwenvergence-14B-v6-Prose",
463
- "scores": {
464
- "average": 38.82,
465
- "IFEval": 59.90,
466
- "BBH": 50.12,
467
- "MATH": 34.89,
468
- "GPQA": 18.46,
469
- "MUSR": 21.02,
470
- "MMLU-PRO": 48.56
471
- },
472
- "hf_url": "https://huggingface.co/sometimesanotion/Qwenvergence-14B-v6-Prose",
473
- "known_config": None
474
- },
475
- {
476
- "rank": 61,
477
- "name": "CultriX/Qwen2.5-14B-Brocav3",
478
- "scores": {
479
- "average": 38.76,
480
- "IFEval": 69.52,
481
- "BBH": 49.05,
482
- "MATH": 32.25,
483
- "GPQA": 14.54,
484
- "MUSR": 19.25,
485
- "MMLU-PRO": 47.97
486
- },
487
- "hf_url": "https://huggingface.co/CultriX/Qwen2.5-14B-Brocav3",
488
- "known_config": None
489
- },
490
- {
491
- "rank": 62,
492
- "name": "sometimesanotion/Qwentinuum-14B-v7",
493
- "scores": {
494
- "average": 38.76,
495
- "IFEval": 61.09,
496
- "BBH": 50.35,
497
- "MATH": 33.38,
498
- "GPQA": 18.79,
499
- "MUSR": 19.95,
500
- "MMLU-PRO": 49.00
501
- },
502
- "hf_url": "https://huggingface.co/sometimesanotion/Qwentinuum-14B-v7",
503
- "known_config": None
504
- },
505
- {
506
- "rank": 64,
507
- "name": "sometimesanotion/Qwentinuum-14B-v3",
508
- "scores": {
509
- "average": 38.74,
510
- "IFEval": 61.58,
511
- "BBH": 50.04,
512
- "MATH": 32.85,
513
- "GPQA": 18.34,
514
- "MUSR": 20.62,
515
- "MMLU-PRO": 49.03
516
- },
517
- "hf_url": "https://huggingface.co/sometimesanotion/Qwentinuum-14B-v3",
518
- "known_config": None
519
- },
520
- {
521
- "rank": 65,
522
- "name": "allura-org/TQ2.5-14B-Aletheia-v1",
523
- "scores": {
524
- "average": 38.74,
525
- "IFEval": 75.30,
526
- "BBH": 50.88,
527
- "MATH": 29.53,
528
- "GPQA": 14.99,
529
- "MUSR": 14.61,
530
- "MMLU-PRO": 47.12
531
- },
532
- "hf_url": "https://huggingface.co/allura-org/TQ2.5-14B-Aletheia-v1",
533
- "known_config": None
534
- },
535
- {
536
- "rank": 66,
537
- "name": "qingy2024/Fusion4-14B-Instruct",
538
- "scores": {
539
- "average": 38.73,
540
- "IFEval": 76.49,
541
- "BBH": 50.70,
542
- "MATH": 33.91,
543
- "GPQA": 10.74,
544
- "MUSR": 13.97,
545
- "MMLU-PRO": 46.60
546
- },
547
- "hf_url": "https://huggingface.co/qingy2024/Fusion4-14B-Instruct",
548
- "known_config": None
549
- },
550
- {
551
- "rank": 68,
552
- "name": "CultriX/Qwen2.5-14B-Brocav7",
553
- "scores": {
554
- "average": 38.52,
555
- "IFEval": 67.24,
556
- "BBH": 48.91,
557
- "MATH": 31.87,
558
- "GPQA": 15.66,
559
- "MUSR": 20.15,
560
- "MMLU-PRO": 47.31
561
- },
562
- "hf_url": "https://huggingface.co/CultriX/Qwen2.5-14B-Brocav7",
563
- "known_config": None
564
- },
565
- {
566
- "rank": 71,
567
- "name": "sometimesanotion/Qwentinuum-14B-v6-Prose",
568
- "scores": {
569
- "average": 38.46,
570
- "IFEval": 56.43,
571
- "BBH": 50.14,
572
- "MATH": 35.57,
573
- "GPQA": 18.46,
574
- "MUSR": 21.34,
575
- "MMLU-PRO": 48.80
576
- },
577
- "hf_url": "https://huggingface.co/sometimesanotion/Qwentinuum-14B-v6-Prose",
578
- "known_config": None
579
- },
580
- {
581
- "rank": 76,
582
- "name": "CultriX/Qwen2.5-14B-Brocav6",
583
- "scores": {
584
- "average": 38.32,
585
- "IFEval": 69.95,
586
- "BBH": 47.82,
587
- "MATH": 29.61,
588
- "GPQA": 15.66,
589
- "MUSR": 18.88,
590
- "MMLU-PRO": 47.99
591
- },
592
- "hf_url": "https://huggingface.co/CultriX/Qwen2.5-14B-Brocav6",
593
- "known_config": None
594
- },
595
- {
596
- "rank": 80,
597
- "name": "CultriX/SeQwence-14Bv1",
598
- "scores": {
599
- "average": 38.20,
600
- "IFEval": 66.78,
601
- "BBH": 47.19,
602
- "MATH": 33.53,
603
- "GPQA": 14.88,
604
- "MUSR": 18.80,
605
- "MMLU-PRO": 48.00
606
- },
607
- "hf_url": "https://huggingface.co/CultriX/SeQwence-14Bv1",
608
- "known_config": None
609
- },
610
- {
611
- "rank": 85,
612
- "name": "sometimesanotion/Qwentinuum-14B-v013",
613
- "scores": {
614
- "average": 37.96,
615
- "IFEval": 67.11,
616
- "BBH": 43.97,
617
- "MATH": 33.01,
618
- "GPQA": 14.32,
619
- "MUSR": 24.99,
620
- "MMLU-PRO": 44.34
621
- },
622
- "hf_url": "https://huggingface.co/sometimesanotion/Qwentinuum-14B-v013",
623
- "known_config": None
624
- },
625
- {
626
- "rank": 86,
627
- "name": "CultriX/Qwen2.5-14B-Wernickev3",
628
- "scores": {
629
- "average": 37.94,
630
- "IFEval": 70.48,
631
- "BBH": 44.58,
632
- "MATH": 32.78,
633
- "GPQA": 14.99,
634
- "MUSR": 18.69,
635
- "MMLU-PRO": 46.13
636
- },
637
- "hf_url": "https://huggingface.co/CultriX/Qwen2.5-14B-Wernickev3",
638
- "known_config": None
639
- },
640
- {
641
- "rank": 88,
642
- "name": "allknowingroger/QwenSlerp4-14B",
643
- "scores": {
644
- "average": 37.80,
645
- "IFEval": 63.28,
646
- "BBH": 49.38,
647
- "MATH": 30.97,
648
- "GPQA": 16.33,
649
- "MUSR": 17.59,
650
- "MMLU-PRO": 49.28
651
- },
652
- "hf_url": "https://huggingface.co/allknowingroger/QwenSlerp4-14B",
653
- "known_config": None
654
- },
655
- {
656
- "rank": 89,
657
- "name": "CultriX/Qwen2.5-14B-Broca",
658
- "scores": {
659
- "average": 37.72,
660
- "IFEval": 56.04,
661
- "BBH": 50.03,
662
- "MATH": 34.59,
663
- "GPQA": 18.23,
664
- "MUSR": 18.95,
665
- "MMLU-PRO": 48.49
666
- },
667
- "hf_url": "https://huggingface.co/CultriX/Qwen2.5-14B-Broca",
668
- "known_config": None
669
- },
670
- {
671
- "rank": 90,
672
- "name": "CultriX/Qwen2.5-14B-Emerged",
673
- "scores": {
674
- "average": 37.66,
675
- "IFEval": 70.00,
676
- "BBH": 45.93,
677
- "MATH": 30.74,
678
- "GPQA": 14.32,
679
- "MUSR": 18.47,
680
- "MMLU-PRO": 46.51
681
- },
682
- "hf_url": "https://huggingface.co/CultriX/Qwen2.5-14B-Emerged",
683
- "known_config": None
684
- },
685
- {
686
- "rank": 91,
687
- "name": "sometimesanotion/Qwentinuum-14B-v8",
688
- "scores": {
689
- "average": 37.65,
690
- "IFEval": 54.12,
691
- "BBH": 50.11,
692
- "MATH": 34.14,
693
- "GPQA": 17.79,
694
- "MUSR": 20.75,
695
- "MMLU-PRO": 49.02
696
- },
697
- "hf_url": "https://huggingface.co/sometimesanotion/Qwentinuum-14B-v8",
698
- "known_config": None
699
- },
700
- {
701
- "rank": 92,
702
- "name": "qingy2024/Fusion-14B-Instruct",
703
- "scores": {
704
- "average": 37.64,
705
- "IFEval": 72.60,
706
- "BBH": 48.58,
707
- "MATH": 30.97,
708
- "GPQA": 13.98,
709
- "MUSR": 14.81,
710
- "MMLU-PRO": 44.93
711
- },
712
- "hf_url": "https://huggingface.co/qingy2024/Fusion-14B-Instruct",
713
- "known_config": None
714
- },
715
- {
716
- "rank": 94,
717
- "name": "CultriX/Qwestion-14B",
718
- "scores": {
719
- "average": 37.63,
720
- "IFEval": 63.18,
721
- "BBH": 48.76,
722
- "MATH": 31.72,
723
- "GPQA": 15.77,
724
- "MUSR": 17.22,
725
- "MMLU-PRO": 49.14
726
- },
727
- "hf_url": "https://huggingface.co/CultriX/Qwestion-14B",
728
- "known_config": None
729
- },
730
- {
731
- "rank": 99,
732
- "name": "sometimesanotion/Qwenvergence-14B-v3-Prose",
733
- "scores": {
734
- "average": 37.37,
735
- "IFEval": 49.18,
736
- "BBH": 49.80,
737
- "MATH": 35.57,
738
- "GPQA": 19.35,
739
- "MUSR": 21.77,
740
- "MMLU-PRO": 48.55
741
- },
742
- "hf_url": "https://huggingface.co/sometimesanotion/Qwenvergence-14B-v3-Prose",
743
- "known_config": None
744
- },
745
- {
746
- "rank": 102,
747
- "name": "CultriX/SeQwence-14B-v5",
748
- "scores": {
749
- "average": 37.27,
750
- "IFEval": 59.20,
751
- "BBH": 50.00,
752
- "MATH": 31.04,
753
- "GPQA": 16.00,
754
- "MUSR": 18.33,
755
- "MMLU-PRO": 49.05
756
- },
757
- "hf_url": "https://huggingface.co/CultriX/SeQwence-14B-v5",
758
- "known_config": None
759
- },
760
- {
761
- "rank": 103,
762
- "name": "sometimesanotion/Qwen-14B-ProseStock-v4",
763
- "scores": {
764
- "average": 37.23,
765
- "IFEval": 49.42,
766
- "BBH": 49.54,
767
- "MATH": 35.50,
768
- "GPQA": 18.46,
769
- "MUSR": 21.70,
770
- "MMLU-PRO": 48.74
771
- },
772
- "hf_url": "https://huggingface.co/sometimesanotion/Qwen-14B-ProseStock-v4",
773
- "known_config": None
774
- },
775
- {
776
- "rank": 104,
777
- "name": "sometimesanotion/IF-reasoning-experiment-40",
778
- "scores": {
779
- "average": 37.21,
780
- "IFEval": 63.30,
781
- "BBH": 44.31,
782
- "MATH": 27.72,
783
- "GPQA": 17.34,
784
- "MUSR": 25.86,
785
- "MMLU-PRO": 44.72
786
- },
787
- "hf_url": "https://huggingface.co/sometimesanotion/IF-reasoning-experiment-40",
788
- "known_config": None
789
- },
790
- {
791
- "rank": 105,
792
- "name": "CultriX/SeQwence-14B-EvolMerge",
793
- "scores": {
794
- "average": 37.20,
795
- "IFEval": 53.82,
796
- "BBH": 50.78,
797
- "MATH": 31.80,
798
- "GPQA": 17.45,
799
- "MUSR": 20.26,
800
- "MMLU-PRO": 49.10
801
- },
802
- "hf_url": "https://huggingface.co/CultriX/SeQwence-14B-EvolMerge",
803
- "known_config": None
804
  }
805
  ]
806
 
 
282
  # This is your larger dataset, rank = 44..105
283
  benchmark_data = [
284
  {
285
+ rank: 1,
286
+ name: "sometimesanotion/Lamarck-14B-v0.7-rc4",
287
+ scores: {
288
+ average: 41.22,
289
+ IFEval: 72.11,
290
+ BBH: 49.85,
291
+ MATH: 36.86,
292
+ GPQA: 18.57,
293
+ MUSR: 21.07,
294
+ MMLU_PRO: 48.89,
295
+ Architecture: "Qwen2ForCausalLM",
296
+ Parameters: "14.766B",
297
+ Chat_Template: "No"
298
+ },
299
+ hf_url: "https://huggingface.co/sometimesanotion/Lamarck-14B-v0.7-rc4",
300
+ known_config: None
301
+ },
302
+ {
303
+ rank: 2,
304
+ name: "arcee-ai/Virtuoso-Small-v2",
305
+ scores: {
306
+ average: 41.08,
307
+ IFEval: 82.73,
308
+ BBH: 50.95,
309
+ MATH: 38.22,
310
+ GPQA: 13.76,
311
+ MUSR: 14.28,
312
+ MMLU_PRO: 46.53,
313
+ Architecture: "Qwen2ForCausalLM",
314
+ Parameters: "14.766B",
315
+ Chat_Template: "Yes"
316
+ },
317
+ hf_url: "https://huggingface.co/arcee-ai/Virtuoso-Small-v2",
318
+ known_config: None
319
+ },
320
+ {
321
+ rank: 3,
322
+ name: "sometimesanotion/Qwenvergence-14B-v12-Prose-DS",
323
+ scores: {
324
+ average: 41.08,
325
+ IFEval: 61.73,
326
+ BBH: 49.87,
327
+ MATH: 42.30,
328
+ GPQA: 19.24,
329
+ MUSR: 24.78,
330
+ MMLU_PRO: 48.54,
331
+ Architecture: "Qwen2ForCausalLM",
332
+ Parameters: "14.766B",
333
+ Chat_Template: "No"
334
+ },
335
+ hf_url: "https://huggingface.co/sometimesanotion/Qwenvergence-14B-v12-Prose-DS",
336
+ known_config: None
337
+ },
338
+ {
339
+ rank: 4,
340
+ name: "bunnycore/Phi-4-Model-Stock-v4",
341
+ scores: {
342
+ average: 41.03,
343
+ IFEval: 71.10,
344
+ BBH: 55.90,
345
+ MATH: 37.16,
346
+ GPQA: 15.88,
347
+ MUSR: 17.30,
348
+ MMLU_PRO: 48.82,
349
+ Architecture: "LlamaForCausalLM",
350
+ Parameters: "14.66B",
351
+ Chat_Template: "Yes"
352
+ },
353
+ hf_url: "https://huggingface.co/bunnycore/Phi-4-Model-Stock-v4",
354
+ known_config: None
355
+ },
356
+ {
357
+ rank: 5,
358
+ name: "sthenno/tempesthenno-nuslerp-0124",
359
+ scores: {
360
+ average: 40.97,
361
+ IFEval: 70.04,
362
+ BBH: 49.28,
363
+ MATH: 39.27,
364
+ GPQA: 18.68,
365
+ MUSR: 20.21,
366
+ MMLU_PRO: 48.36,
367
+ Architecture: "Qwen2ForCausalLM",
368
+ Parameters: "14.766B",
369
+ Chat_Template: "No"
370
+ },
371
+ hf_url: "https://huggingface.co/sthenno/tempesthenno-nuslerp-0124",
372
+ known_config: None
373
+ },
374
+ {
375
+ rank: 6,
376
+ name: "bunnycore/Phi-4-RR-Shoup",
377
+ scores: {
378
+ average: 40.95,
379
+ IFEval: 65.87,
380
+ BBH: 56.11,
381
+ MATH: 47.96,
382
+ GPQA: 11.63,
383
+ MUSR: 14.94,
384
+ MMLU_PRO: 49.21,
385
+ Architecture: "LlamaForCausalLM",
386
+ Parameters: "14.66B",
387
+ Chat_Template: "Yes"
388
+ },
389
+ hf_url: "https://huggingface.co/bunnycore/Phi-4-RR-Shoup",
390
+ known_config: None
391
+ },
392
+ {
393
+ rank: 7,
394
+ name: "sometimesanotion/Qwenvergence-14B-v10",
395
+ scores: {
396
+ average: 40.86,
397
+ IFEval: 67.57,
398
+ BBH: 46.75,
399
+ MATH: 44.18,
400
+ GPQA: 17.23,
401
+ MUSR: 22.33,
402
+ MMLU_PRO: 47.10,
403
+ Architecture: "Qwen2ForCausalLM",
404
+ Parameters: "14.766B",
405
+ Chat_Template: "No"
406
+ },
407
+ hf_url: "https://huggingface.co/sometimesanotion/Qwenvergence-14B-v10",
408
+ known_config: None
409
+ },
410
+ {
411
+ rank: 8,
412
+ name: "bunnycore/Phi-4-RStock-v0.1",
413
+ scores: {
414
+ average: 40.84,
415
+ IFEval: 70.03,
416
+ BBH: 55.98,
417
+ MATH: 38.07,
418
+ GPQA: 15.32,
419
+ MUSR: 16.73,
420
+ MMLU_PRO: 48.90,
421
+ Architecture: "LlamaForCausalLM",
422
+ Parameters: "14.66B",
423
+ Chat_Template: "Yes"
424
+ },
425
+ hf_url: "https://huggingface.co/bunnycore/Phi-4-RStock-v0.1",
426
+ known_config: None
427
+ },
428
+ {
429
+ rank: 9,
430
+ name: "jpacifico/Chocolatine-2-14B-Instruct-v2.0b3",
431
+ scores: {
432
+ average: 40.74,
433
+ IFEval: 73.23,
434
+ BBH: 49.57,
435
+ MATH: 36.93,
436
+ GPQA: 17.23,
437
+ MUSR: 19.30,
438
+ MMLU_PRO: 48.19,
439
+ Architecture: "Qwen2ForCausalLM",
440
+ Parameters: "14.766B",
441
+ Chat_Template: "No"
442
+ },
443
+ hf_url: "https://huggingface.co/jpacifico/Chocolatine-2-14B-Instruct-v2.0b3",
444
+ known_config: None
445
+ },
446
+ {
447
+ rank: 10,
448
+ name: "bunnycore/Phi-4-ReasoningRP",
449
+ scores: {
450
+ average: 40.73,
451
+ IFEval: 67.36,
452
+ BBH: 55.88,
453
+ MATH: 44.34,
454
+ GPQA: 12.53,
455
+ MUSR: 15.14,
456
+ MMLU_PRO: 49.12,
457
+ Architecture: "LlamaForCausalLM",
458
+ Parameters: "14.66B",
459
+ Chat_Template: "Yes"
460
+ },
461
+ hf_url: "https://huggingface.co/bunnycore/Phi-4-ReasoningRP",
462
+ known_config: None
463
+ },
464
+ {
465
+ rank: 11,
466
+ name: "sometimesanotion/Lamarck-14B-v0.7-rc1",
467
+ scores: {
468
+ average: 40.69,
469
+ IFEval: 73.05,
470
+ BBH: 49.51,
471
+ MATH: 35.80,
472
+ GPQA: 18.57,
473
+ MUSR: 18.13,
474
+ MMLU_PRO: 49.06,
475
+ Architecture: "Qwen2ForCausalLM",
476
+ Parameters: "14.766B",
477
+ Chat_Template: "No"
478
+ },
479
+ hf_url: "https://huggingface.co/sometimesanotion/Lamarck-14B-v0.7-rc1",
480
+ known_config: None
481
+ },
482
+ {
483
+ rank: 12,
484
+ name: "jpacifico/Chocolatine-2-14B-Instruct-v2.0b2",
485
+ scores: {
486
+ average: 40.62,
487
+ IFEval: 72.41,
488
+ BBH: 49.58,
489
+ MATH: 35.73,
490
+ GPQA: 17.79,
491
+ MUSR: 19.66,
492
+ MMLU_PRO: 48.54,
493
+ Architecture: "Qwen2ForCausalLM",
494
+ Parameters: "14.766B",
495
+ Chat_Template: "No"
496
+ },
497
+ hf_url: "https://huggingface.co/jpacifico/Chocolatine-2-14B-Instruct-v2.0b2",
498
+ known_config: None
499
+ },
500
+ {
501
+ rank: 13,
502
+ name: "sometimesanotion/Qwenvergence-14B-v11",
503
+ scores: {
504
+ average: 40.56,
505
+ IFEval: 71.92,
506
+ BBH: 47.55,
507
+ MATH: 40.71,
508
+ GPQA: 16.33,
509
+ MUSR: 18.76,
510
+ MMLU_PRO: 48.08,
511
+ Architecture: "Qwen2ForCausalLM",
512
+ Parameters: "14.766B",
513
+ Chat_Template: "No"
514
+ },
515
+ hf_url: "https://huggingface.co/sometimesanotion/Qwenvergence-14B-v11",
516
+ known_config: None
517
+ },
518
+ {
519
+ rank: 14,
520
+ name: "sthenno/tempesthenno-ppo-ckpt40",
521
+ scores: {
522
+ average: 40.55,
523
+ IFEval: 79.23,
524
+ BBH: 50.57,
525
+ MATH: 34.21,
526
+ GPQA: 17.00,
527
+ MUSR: 14.56,
528
+ MMLU_PRO: 47.69,
529
+ Architecture: "Qwen2ForCausalLM",
530
+ Parameters: "14.766B",
531
+ Chat_Template: "Yes"
532
+ },
533
+ hf_url: "https://huggingface.co/sthenno/tempesthenno-ppo-ckpt40",
534
+ known_config: None
535
+ },
536
+ {
537
+ rank: 15,
538
+ name: "tensopolis/virtuoso-small-v2-tensopolis-v1",
539
+ scores: {
540
+ average: 40.38,
541
+ IFEval: 80.94,
542
+ BBH: 50.46,
543
+ MATH: 35.88,
544
+ GPQA: 13.65,
545
+ MUSR: 15.82,
546
+ MMLU_PRO: 45.52,
547
+ Architecture: "Qwen2ForCausalLM",
548
+ Parameters: "14.766B",
549
+ Chat_Template: "Yes"
550
+ },
551
+ hf_url: "https://huggingface.co/tensopolis/virtuoso-small-v2-tensopolis-v1",
552
+ known_config: None
553
+ },
554
+ {
555
+ rank: 16,
556
+ name: "sometimesanotion/Lamarck-14B-v0.6",
557
+ scores: {
558
+ average: 40.37,
559
+ IFEval: 69.73,
560
+ BBH: 49.30,
561
+ MATH: 35.65,
562
+ GPQA: 18.57,
563
+ MUSR: 20.12,
564
+ MMLU_PRO: 48.89,
565
+ Architecture: "Qwen2ForCausalLM",
566
+ Parameters: "14.766B",
567
+ Chat_Template: "No"
568
+ },
569
+ hf_url: "https://huggingface.co/sometimesanotion/Lamarck-14B-v0.6",
570
+ known_config: None
571
+ },
572
+ {
573
+ rank: 17,
574
+ name: "sthenno-com/miscii-14b-0130",
575
+ scores: {
576
+ average: 40.29,
577
+ IFEval: 66.47,
578
+ BBH: 49.84,
579
+ MATH: 38.44,
580
+ GPQA: 17.56,
581
+ MUSR: 20.96,
582
+ MMLU_PRO: 48.48,
583
+ Architecture: "Qwen2ForCausalLM",
584
+ Parameters: "14.766B",
585
+ Chat_Template: "No"
586
+ },
587
+ hf_url: "https://huggingface.co/sthenno-com/miscii-14b-0130",
588
+ known_config: None
589
  },
590
  {
591
+ rank: 18,
592
+ name: "hotmailuser/QwenSlerp2-14B",
593
+ scores: {
594
+ average: 40.21,
595
+ IFEval: 70.37,
596
+ BBH: 49.68,
597
+ MATH: 35.73,
598
+ GPQA: 17.45,
599
+ MUSR: 19.35,
600
+ MMLU_PRO: 48.66,
601
+ Architecture: "Qwen2ForCausalLM",
602
+ Parameters: "14.766B",
603
+ Chat_Template: "No"
604
+ },
605
+ hf_url: "https://huggingface.co/hotmailuser/QwenSlerp2-14B",
606
+ known_config: None
607
+ },
608
+ {
609
+ rank: 19,
610
+ name: "Sakalti/ultiima-14B-v0.2",
611
+ scores: {
612
+ average: 40.18,
613
+ IFEval: 70.70,
614
+ BBH: 49.51,
615
+ MATH: 35.27,
616
+ GPQA: 17.67,
617
+ MUSR: 19.19,
618
+ MMLU_PRO: 48.75,
619
+ Architecture: "Qwen2ForCausalLM",
620
+ Parameters: "14.766B",
621
+ Chat_Template: "No"
622
+ },
623
+ hf_url: "https://huggingface.co/Sakalti/ultiima-14B-v0.2",
624
+ known_config: None
625
+ },
626
+ {
627
+ rank: 20,
628
+ name: "pankajmathur/orca_mini_phi-4",
629
+ scores: {
630
+ average: 40.16,
631
+ IFEval: 77.81,
632
+ BBH: 54.63,
633
+ MATH: 26.44,
634
+ GPQA: 16.55,
635
+ MUSR: 18.25,
636
+ MMLU_PRO: 47.28,
637
+ Architecture: "LlamaForCausalLM",
638
+ Parameters: "14.66B",
639
+ Chat_Template: "Yes"
640
+ },
641
+ hf_url: "https://huggingface.co/pankajmathur/orca_mini_phi-4",
642
+ known_config: None
643
+ },
644
+ {
645
+ rank: 21,
646
+ name: "pankajmathur/orca_mini_v9_2_14B",
647
+ scores: {
648
+ average: 40.16,
649
+ IFEval: 77.81,
650
+ BBH: 54.63,
651
+ MATH: 26.44,
652
+ GPQA: 16.55,
653
+ MUSR: 18.25,
654
+ MMLU_PRO: 47.28,
655
+ Architecture: "LlamaForCausalLM",
656
+ Parameters: "14.66B",
657
+ Chat_Template: "Yes"
658
+ },
659
+ hf_url: "https://huggingface.co/pankajmathur/orca_mini_v9_2_14B",
660
+ known_config: None
661
+ },
662
+ {
663
+ rank: 22,
664
+ name: "sometimesanotion/Qwen2.5-14B-Vimarckoso-v3",
665
+ scores: {
666
+ average: 40.10,
667
+ IFEval: 72.57,
668
+ BBH: 48.58,
669
+ MATH: 34.44,
670
+ GPQA: 17.34,
671
+ MUSR: 19.39,
672
+ MMLU_PRO: 48.26,
673
+ Architecture: "Qwen2ForCausalLM",
674
+ Parameters: "14B",
675
+ Chat_Template: "No"
676
+ },
677
+ hf_url: "https://huggingface.co/sometimesanotion/Qwen2.5-14B-Vimarckoso-v3",
678
+ known_config: None
679
+ },
680
+ {
681
+ rank: 23,
682
+ name: "sthenno-com/miscii-14b-1225",
683
+ scores: {
684
+ average: 40.08,
685
+ IFEval: 78.78,
686
+ BBH: 50.91,
687
+ MATH: 31.57,
688
+ GPQA: 17.00,
689
+ MUSR: 14.77,
690
+ MMLU_PRO: 47.46,
691
+ Architecture: "Qwen2ForCausalLM",
692
+ Parameters: "14.766B",
693
+ Chat_Template: "Yes"
694
+ },
695
+ hf_url: "https://huggingface.co/sthenno-com/miscii-14b-1225",
696
+ known_config: None
697
+ },
698
+ {
699
+ rank: 24,
700
+ name: "bunnycore/Phi-4-Model-Stock",
701
+ scores: {
702
+ average: 40.06,
703
+ IFEval: 68.79,
704
+ BBH: 55.32,
705
+ MATH: 38.60,
706
+ GPQA: 13.98,
707
+ MUSR: 15.12,
708
+ MMLU_PRO: 48.54,
709
+ Architecture: "LlamaForCausalLM",
710
+ Parameters: "14.66B",
711
+ Chat_Template: "Yes"
712
+ },
713
+ hf_url: "https://huggingface.co/bunnycore/Phi-4-Model-Stock",
714
+ known_config: None
715
+ },
716
+ {
717
+ rank: 25,
718
+ name: "djuna/Q2.5-Veltha-14B-0.5",
719
+ scores: {
720
+ average: 39.96,
721
+ IFEval: 77.96,
722
+ BBH: 50.32,
723
+ MATH: 33.84,
724
+ GPQA: 15.77,
725
+ MUSR: 14.17,
726
+ MMLU_PRO: 47.72,
727
+ Architecture: "Qwen2ForCausalLM",
728
+ Parameters: "14.766B",
729
+ Chat_Template: "Yes"
730
+ },
731
+ hf_url: "https://huggingface.co/djuna/Q2.5-Veltha-14B-0.5",
732
+ known_config: None
733
+ },
734
+ {
735
+ rank: 26,
736
+ name: "ehristoforu/fp4-14b-v1-fix",
737
+ scores: {
738
+ average: 39.96,
739
+ IFEval: 67.42,
740
+ BBH: 54.33,
741
+ MATH: 39.58,
742
+ GPQA: 13.87,
743
+ MUSR: 16.18,
744
+ MMLU_PRO: 48.37,
745
+ Architecture: "LlamaForCausalLM",
746
+ Parameters: "14.66B",
747
+ Chat_Template: "Yes"
748
+ },
749
+ hf_url: "https://huggingface.co/ehristoforu/fp4-14b-v1-fix",
750
+ known_config: None
751
+ },
752
+ {
753
+ rank: 27,
754
+ name: "sthenno/tempesthenno-nuslerp-001",
755
+ scores: {
756
+ average: 39.94,
757
+ IFEval: 79.26,
758
+ BBH: 51.04,
759
+ MATH: 31.72,
760
+ GPQA: 16.44,
761
+ MUSR: 13.88,
762
+ MMLU_PRO: 47.30,
763
+ Architecture: "Qwen2ForCausalLM",
764
+ Parameters: "14.766B",
765
+ Chat_Template: "Yes"
766
+ },
767
+ hf_url: "https://huggingface.co/sthenno/tempesthenno-nuslerp-001",
768
+ known_config: None
769
+ },
770
+ {
771
+ rank: 28,
772
+ name: "bunnycore/Phi-4-Stock-Ex",
773
+ scores: {
774
+ average: 39.93,
775
+ IFEval: 65.75,
776
+ BBH: 55.20,
777
+ MATH: 39.12,
778
+ GPQA: 13.42,
779
+ MUSR: 17.46,
780
+ MMLU_PRO: 48.61,
781
+ Architecture: "LlamaForCausalLM",
782
+ Parameters: "14.66B",
783
+ Chat_Template: "Yes"
784
+ },
785
+ hf_url: "https://huggingface.co/bunnycore/Phi-4-Stock-Ex",
786
+ known_config: None
787
+ },
788
+ {
789
+ rank: 29,
790
+ name: "hotmailuser/QwenSlerp-14B",
791
+ scores: {
792
+ average: 39.87,
793
+ IFEval: 70.25,
794
+ BBH: 49.42,
795
+ MATH: 35.50,
796
+ GPQA: 18.34,
797
+ MUSR: 16.83,
798
+ MMLU_PRO: 48.89,
799
+ Architecture: "Qwen2ForCausalLM",
800
+ Parameters: "14.766B",
801
+ Chat_Template: "No"
802
+ },
803
+ hf_url: "https://huggingface.co/hotmailuser/QwenSlerp-14B",
804
+ known_config: None
805
+ },
806
+ {
807
+ rank: 30,
808
+ name: "sometimesanotion/Qwen2.5-14B-Vimarckoso-v3-model_stock",
809
+ scores: {
810
+ average: 39.81,
811
+ IFEval: 71.62,
812
+ BBH: 48.76,
813
+ MATH: 33.99,
814
+ GPQA: 17.34,
815
+ MUSR: 19.23,
816
+ MMLU_PRO: 47.95,
817
+ Architecture: "Qwen2ForCausalLM",
818
+ Parameters: "14B",
819
+ Chat_Template: "No"
820
+ },
821
+ hf_url: "https://huggingface.co/sometimesanotion/Qwen2.5-14B-Vimarckoso-v3-model_stock",
822
+ known_config: None
823
+ },
824
+ {
825
+ rank: 31,
826
+ name: "tensopolis/virtuoso-small-tensopolis-v1",
827
+ scores: {
828
+ average: 39.69,
829
+ IFEval: 79.50,
830
+ BBH: 50.70,
831
+ MATH: 36.03,
832
+ GPQA: 10.85,
833
+ MUSR: 14.70,
834
+ MMLU_PRO: 46.36,
835
+ Architecture: "Qwen2ForCausalLM",
836
+ Parameters: "14.77B",
837
+ Chat_Template: "Yes"
838
+ },
839
+ hf_url: "https://huggingface.co/tensopolis/virtuoso-small-tensopolis-v1",
840
+ known_config: None
841
+ },
842
+ {
843
+ rank: 32,
844
+ name: "sometimesanotion/Lamarck-14B-v0.6-model_stock",
845
+ scores: {
846
+ average: 39.58,
847
+ IFEval: 67.90,
848
+ BBH: 46.49,
849
+ MATH: 35.88,
850
+ GPQA: 17.90,
851
+ MUSR: 22.68,
852
+ MMLU_PRO: 46.64,
853
+ Architecture: "Qwen2ForCausalLM",
854
+ Parameters: "14B",
855
+ Chat_Template: "No"
856
+ },
857
+ hf_url: "https://huggingface.co/sometimesanotion/Lamarck-14B-v0.6-model_stock",
858
+ known_config: None
859
+ },
860
+ {
861
+ rank: 33,
862
+ name: "tensopolis/virtuoso-small-tensopolis-v2",
863
+ scores: {
864
+ average: 39.53,
865
+ IFEval: 80.20,
866
+ BBH: 50.23,
867
+ MATH: 35.27,
868
+ GPQA: 10.51,
869
+ MUSR: 14.84,
870
+ MMLU_PRO: 46.15,
871
+ Architecture: "Qwen2ForCausalLM",
872
+ Parameters: "14.77B",
873
+ Chat_Template: "Yes"
874
+ },
875
+ hf_url: "https://huggingface.co/tensopolis/virtuoso-small-tensopolis-v2",
876
+ known_config: None
877
+ },
878
+ {
879
+ rank: 34,
880
+ name: "Sakalti/ultiima-14B-v0.3",
881
+ scores: {
882
+ average: 39.53,
883
+ IFEval: 70.40,
884
+ BBH: 48.45,
885
+ MATH: 34.52,
886
+ GPQA: 16.89,
887
+ MUSR: 18.73,
888
+ MMLU_PRO: 48.18,
889
+ Architecture: "Qwen2ForCausalLM",
890
+ Parameters: "14.766B",
891
+ Chat_Template: "No"
892
+ },
893
+ hf_url: "https://huggingface.co/Sakalti/ultiima-14B-v0.3",
894
+ known_config: None
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
895
  }
896
  ]
897