ai-forever commited on
Commit
7d9acaa
·
verified ·
1 Parent(s): add7f1a

Add/update results for Llama 4 Scout 17B (version 1.33.1, guid ef91e41e46fd4351b30c5cf2e2e6f9fc)

Browse files
Files changed (1) hide show
  1. results.json +74 -0
results.json CHANGED
@@ -450,6 +450,80 @@
450
  "n_questions": 158,
451
  "submit_timestamp": ""
452
  }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
453
  }
454
  }
455
  },
 
450
  "n_questions": 158,
451
  "submit_timestamp": ""
452
  }
453
+ },
454
+ "ef91e41e46fd4351b30c5cf2e2e6f9fc": {
455
+ "model_name": "Llama 4 Scout 17B",
456
+ "timestamp": "2025-05-22T14:47:39",
457
+ "config": {
458
+ "embedding_model": "sentence-transformers/distiluse-base-multilingual-cased",
459
+ "retriever_type": "mmr",
460
+ "retrieval_config": {
461
+ "top_k": 4,
462
+ "chunk_size": 1000,
463
+ "chunk_overlap": 100
464
+ }
465
+ },
466
+ "metrics": {
467
+ "mh": {
468
+ "retrieval": {
469
+ "hit_rate": 0.4791666666666667,
470
+ "mrr": 0.328125,
471
+ "precision": 0.13541666666666666
472
+ },
473
+ "generation": {
474
+ "rouge1": 0.04071969696969697,
475
+ "rougeL": 0.04071969696969697
476
+ }
477
+ },
478
+ "cond": {
479
+ "retrieval": {
480
+ "hit_rate": 0.5686274509803921,
481
+ "mrr": 0.43137254901960786,
482
+ "precision": 0.18627450980392157
483
+ },
484
+ "generation": {
485
+ "rouge1": 0.012118736383442266,
486
+ "rougeL": 0.012118736383442266
487
+ }
488
+ },
489
+ "simple": {
490
+ "retrieval": {
491
+ "hit_rate": 0.5555555555555556,
492
+ "mrr": 0.39351851851851843,
493
+ "precision": 0.1527777777777778
494
+ },
495
+ "generation": {
496
+ "rouge1": 0.04487118921642731,
497
+ "rougeL": 0.04487118921642731
498
+ }
499
+ },
500
+ "set": {
501
+ "retrieval": {
502
+ "hit_rate": 0.6,
503
+ "mrr": 0.25,
504
+ "precision": 0.2
505
+ },
506
+ "generation": {
507
+ "rouge1": 0.0,
508
+ "rougeL": 0.0
509
+ }
510
+ },
511
+ "overall": {
512
+ "retrieval": {
513
+ "hit_rate": 0.5379746835443038,
514
+ "mrr": 0.38132911392405067,
515
+ "precision": 0.15981012658227847
516
+ },
517
+ "generation": {
518
+ "rouge1": 0.03161800777081066,
519
+ "rougeL": 0.03161800777081066
520
+ }
521
+ }
522
+ },
523
+ "metadata": {
524
+ "n_questions": 158,
525
+ "submit_timestamp": ""
526
+ }
527
  }
528
  }
529
  },