Spaces:
Running
Running
Add/update results for Llama 4 Scout 17B (version 1.33.1, guid ef91e41e46fd4351b30c5cf2e2e6f9fc)
Browse files- results.json +74 -0
results.json
CHANGED
@@ -450,6 +450,80 @@
|
|
450 |
"n_questions": 158,
|
451 |
"submit_timestamp": ""
|
452 |
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
453 |
}
|
454 |
}
|
455 |
},
|
|
|
450 |
"n_questions": 158,
|
451 |
"submit_timestamp": ""
|
452 |
}
|
453 |
+
},
|
454 |
+
"ef91e41e46fd4351b30c5cf2e2e6f9fc": {
|
455 |
+
"model_name": "Llama 4 Scout 17B",
|
456 |
+
"timestamp": "2025-05-22T14:47:39",
|
457 |
+
"config": {
|
458 |
+
"embedding_model": "sentence-transformers/distiluse-base-multilingual-cased",
|
459 |
+
"retriever_type": "mmr",
|
460 |
+
"retrieval_config": {
|
461 |
+
"top_k": 4,
|
462 |
+
"chunk_size": 1000,
|
463 |
+
"chunk_overlap": 100
|
464 |
+
}
|
465 |
+
},
|
466 |
+
"metrics": {
|
467 |
+
"mh": {
|
468 |
+
"retrieval": {
|
469 |
+
"hit_rate": 0.4791666666666667,
|
470 |
+
"mrr": 0.328125,
|
471 |
+
"precision": 0.13541666666666666
|
472 |
+
},
|
473 |
+
"generation": {
|
474 |
+
"rouge1": 0.04071969696969697,
|
475 |
+
"rougeL": 0.04071969696969697
|
476 |
+
}
|
477 |
+
},
|
478 |
+
"cond": {
|
479 |
+
"retrieval": {
|
480 |
+
"hit_rate": 0.5686274509803921,
|
481 |
+
"mrr": 0.43137254901960786,
|
482 |
+
"precision": 0.18627450980392157
|
483 |
+
},
|
484 |
+
"generation": {
|
485 |
+
"rouge1": 0.012118736383442266,
|
486 |
+
"rougeL": 0.012118736383442266
|
487 |
+
}
|
488 |
+
},
|
489 |
+
"simple": {
|
490 |
+
"retrieval": {
|
491 |
+
"hit_rate": 0.5555555555555556,
|
492 |
+
"mrr": 0.39351851851851843,
|
493 |
+
"precision": 0.1527777777777778
|
494 |
+
},
|
495 |
+
"generation": {
|
496 |
+
"rouge1": 0.04487118921642731,
|
497 |
+
"rougeL": 0.04487118921642731
|
498 |
+
}
|
499 |
+
},
|
500 |
+
"set": {
|
501 |
+
"retrieval": {
|
502 |
+
"hit_rate": 0.6,
|
503 |
+
"mrr": 0.25,
|
504 |
+
"precision": 0.2
|
505 |
+
},
|
506 |
+
"generation": {
|
507 |
+
"rouge1": 0.0,
|
508 |
+
"rougeL": 0.0
|
509 |
+
}
|
510 |
+
},
|
511 |
+
"overall": {
|
512 |
+
"retrieval": {
|
513 |
+
"hit_rate": 0.5379746835443038,
|
514 |
+
"mrr": 0.38132911392405067,
|
515 |
+
"precision": 0.15981012658227847
|
516 |
+
},
|
517 |
+
"generation": {
|
518 |
+
"rouge1": 0.03161800777081066,
|
519 |
+
"rougeL": 0.03161800777081066
|
520 |
+
}
|
521 |
+
}
|
522 |
+
},
|
523 |
+
"metadata": {
|
524 |
+
"n_questions": 158,
|
525 |
+
"submit_timestamp": ""
|
526 |
+
}
|
527 |
}
|
528 |
}
|
529 |
},
|