Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- lm-evaluation/tests/testdata/arc_challenge-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/arithmetic_4ds-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/arithmetic_5da-v0-res.json +1 -0
- lm-evaluation/tests/testdata/blimp_complex_NP_island-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/blimp_coordinate_structure_constraint_object_extraction-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/blimp_determiner_noun_agreement_2-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/blimp_distractor_agreement_relative_clause-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/blimp_ellipsis_n_bar_1-v0-res.json +1 -0
- lm-evaluation/tests/testdata/blimp_inchoative-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/blimp_passive_2-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/blimp_principle_A_c_command-v0-res.json +1 -0
- lm-evaluation/tests/testdata/blimp_principle_A_domain_2-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/blimp_principle_A_domain_3-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/blimp_regular_plural_subject_verb_agreement_2-v0-res.json +1 -0
- lm-evaluation/tests/testdata/blimp_wh_questions_object_gap-v0-res.json +1 -0
- lm-evaluation/tests/testdata/coqa-v0-res.json +1 -0
- lm-evaluation/tests/testdata/crows_pairs_english_autre-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/crows_pairs_english_race_color-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/crows_pairs_english_race_color-v0-res.json +1 -0
- lm-evaluation/tests/testdata/crows_pairs_french_autre-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/crows_pairs_french_nationality-v0-res.json +1 -0
- lm-evaluation/tests/testdata/crows_pairs_french_race_color-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/crows_pairs_french_socioeconomic-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/ethics_cm-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/ethics_cm-v0-res.json +1 -0
- lm-evaluation/tests/testdata/ethics_utilitarianism_original-v0-res.json +1 -0
- lm-evaluation/tests/testdata/gguf_test_44e268d15decc4d2d0f99e57e1476269826cd3b54262f7a0981f75ddd45b25d0.pkl +3 -0
- lm-evaluation/tests/testdata/hellaswag-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/hendrycksTest-college_computer_science-v0-res.json +1 -0
- lm-evaluation/tests/testdata/hendrycksTest-college_mathematics-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/hendrycksTest-high_school_biology-v0-res.json +1 -0
- lm-evaluation/tests/testdata/hendrycksTest-high_school_macroeconomics-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/hendrycksTest-high_school_physics-v0-res.json +1 -0
- lm-evaluation/tests/testdata/hendrycksTest-high_school_psychology-v0-res.json +1 -0
- lm-evaluation/tests/testdata/hendrycksTest-high_school_statistics-v0-res.json +1 -0
- lm-evaluation/tests/testdata/hendrycksTest-jurisprudence-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/hendrycksTest-moral_disputes-v0-res.json +1 -0
- lm-evaluation/tests/testdata/hendrycksTest-professional_law-v0-res.json +1 -0
- lm-evaluation/tests/testdata/hendrycksTest-public_relations-v0-res.json +1 -0
- lm-evaluation/tests/testdata/lambada_openai-v2.0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/lambada_openai_mt_de-v0-res.json +1 -0
- lm-evaluation/tests/testdata/lambada_openai_mt_es-v0-res.json +1 -0
- lm-evaluation/tests/testdata/lambada_openai_mt_fr-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/lambada_openai_mt_it-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/logiqa-v0-loglikelihood +1 -0
- lm-evaluation/tests/testdata/math_counting_and_prob-v1-res.json +1 -0
- lm-evaluation/tests/testdata/math_num_theory-v0-res.json +1 -0
- lm-evaluation/tests/testdata/math_prealgebra-v0-greedy_until +1 -0
- lm-evaluation/tests/testdata/math_prealgebra-v1-greedy_until +1 -0
- lm-evaluation/tests/testdata/math_prealgebra-v1-res.json +1 -0
lm-evaluation/tests/testdata/arc_challenge-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
41c34c96cca8ace661911d0033d630c554b283f5a3953bcdc50720ae6b00a9c1
|
lm-evaluation/tests/testdata/arithmetic_4ds-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
d915830b8621e66331383bb2ae4c60acebf008e2f94741092ef4c33ea5441037
|
lm-evaluation/tests/testdata/arithmetic_5da-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"arithmetic_5da": {"acc": 0.0, "acc_stderr": 0.0}}, "versions": {"arithmetic_5da": 0}}
|
lm-evaluation/tests/testdata/blimp_complex_NP_island-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
f46cfcc7e43050a235fd2a6b989cabbfbcce76786df74db9f0d4a9cd1caa1628
|
lm-evaluation/tests/testdata/blimp_coordinate_structure_constraint_object_extraction-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
23ddafdff7b1ebe331b146e23b2c21aa109fe57aa1ce8ca201a0d239fcbdd166
|
lm-evaluation/tests/testdata/blimp_determiner_noun_agreement_2-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
123e2acd00fbba60aba1fbae607c79a062e512c9e79c7d8dfafff63e30111d76
|
lm-evaluation/tests/testdata/blimp_distractor_agreement_relative_clause-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
bf78e2b53c0f3531303c668c96bd3897a0a35e960da37439e63724ecba4e371a
|
lm-evaluation/tests/testdata/blimp_ellipsis_n_bar_1-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"blimp_ellipsis_n_bar_1": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_ellipsis_n_bar_1": 0}}
|
lm-evaluation/tests/testdata/blimp_inchoative-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
3ff73629fb4473986a0e8ae2fcb7c40e88292189ab0d8755d20836c5aa5a2f99
|
lm-evaluation/tests/testdata/blimp_passive_2-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
755bdfe2c89737c43001ff1dc83d68ad33e444aaf0669af66aaf82dcd09f2eca
|
lm-evaluation/tests/testdata/blimp_principle_A_c_command-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"blimp_principle_A_c_command": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_principle_A_c_command": 0}}
|
lm-evaluation/tests/testdata/blimp_principle_A_domain_2-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
eb5ddf0a97982373ab1a4e58267cfcdebdecdb86c376dfd5ebf46737c9d3ee12
|
lm-evaluation/tests/testdata/blimp_principle_A_domain_3-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
38454befedcf1f3f6ef27d3bef9ccfdfb3e94a7ab32d86a63493a920d2d50093
|
lm-evaluation/tests/testdata/blimp_regular_plural_subject_verb_agreement_2-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"blimp_regular_plural_subject_verb_agreement_2": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_regular_plural_subject_verb_agreement_2": 0}}
|
lm-evaluation/tests/testdata/blimp_wh_questions_object_gap-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"blimp_wh_questions_object_gap": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_wh_questions_object_gap": 0}}
|
lm-evaluation/tests/testdata/coqa-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"coqa": {"em": 0.0, "em_stderr": 0.0, "f1": 0.0, "f1_stderr": 0.0}}, "versions": {"coqa": 0}}
|
lm-evaluation/tests/testdata/crows_pairs_english_autre-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
a197ccc8538231404a8e43f5ed0fbbfb2c317b4da337f6e7aa9642131aeb426a
|
lm-evaluation/tests/testdata/crows_pairs_english_race_color-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
0a750596d77cd96502dc414ff699a399b1b91c2078adeec1d3dd982b3d591089
|
lm-evaluation/tests/testdata/crows_pairs_english_race_color-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"crows_pairs_english_race_color": {"likelihood_difference": 0.3322827903840805, "likelihood_difference_stderr": 0.01019838186372816, "pct_stereotype": 0.4822834645669291, "pct_stereotype_stderr": 0.022191835500120254}}, "versions": {"crows_pairs_english_race_color": 0}}
|
lm-evaluation/tests/testdata/crows_pairs_french_autre-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
f145ad5086da0bf8c76f0730258529fa243efe32b7ab792d3c4716284b4b5495
|
lm-evaluation/tests/testdata/crows_pairs_french_nationality-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"crows_pairs_french_nationality": {"likelihood_difference": 0.33534193269044926, "likelihood_difference_stderr": 0.01429836309463257, "pct_stereotype": 0.4743083003952569, "pct_stereotype_stderr": 0.031455431847992904}}, "versions": {"crows_pairs_french_nationality": 0}}
|
lm-evaluation/tests/testdata/crows_pairs_french_race_color-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
6f9119026abff33c5c882d6172e092e806a8b21bd86864022978b1961839350f
|
lm-evaluation/tests/testdata/crows_pairs_french_socioeconomic-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
8ba0a525c65f795c99f6416e70c998e75e4b6cc43bf9a4bd7ccacd3c3591e9cb
|
lm-evaluation/tests/testdata/ethics_cm-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
92d136ebb2bd86cd036e61699ad9a1417dbb48651f0a3afa5045cf57cef5a3f6
|
lm-evaluation/tests/testdata/ethics_cm-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"ethics_cm": {"acc": 0.49987129987129986, "acc_stderr": 0.008022881531793336}}, "versions": {"ethics_cm": 0}}
|
lm-evaluation/tests/testdata/ethics_utilitarianism_original-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"ethics_utilitarianism_original": {"acc": 0.5214226289517471, "acc_stderr": 0.007204999520618661}}, "versions": {"ethics_utilitarianism_original": 0}}
|
lm-evaluation/tests/testdata/gguf_test_44e268d15decc4d2d0f99e57e1476269826cd3b54262f7a0981f75ddd45b25d0.pkl
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d672564bf8af93738531c26d02efda9ed844d1fdcdd619a06e37fce7ef6d8485
|
3 |
+
size 153
|
lm-evaluation/tests/testdata/hellaswag-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
abb808c97d6529eda6c11067837a132c62d25cba0394d720f80cca6df9f7196e
|
lm-evaluation/tests/testdata/hendrycksTest-college_computer_science-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"hendrycksTest-college_computer_science": {"acc": 0.22, "acc_norm": 0.24, "acc_norm_stderr": 0.04292346959909282, "acc_stderr": 0.041633319989322695}}, "versions": {"hendrycksTest-college_computer_science": 0}}
|
lm-evaluation/tests/testdata/hendrycksTest-college_mathematics-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
e9fe80752686527281f834d2397875b4580581434b94799f9de6aaa450bd73ff
|
lm-evaluation/tests/testdata/hendrycksTest-high_school_biology-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"hendrycksTest-high_school_biology": {"acc": 0.23870967741935484, "acc_norm": 0.2709677419354839, "acc_norm_stderr": 0.025284416114900152, "acc_stderr": 0.024251071262208834}}, "versions": {"hendrycksTest-high_school_biology": 0}}
|
lm-evaluation/tests/testdata/hendrycksTest-high_school_macroeconomics-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
ce4faae2fb6628caa48f6fc74cbc848880db49e6ff51079392778a2322bcefef
|
lm-evaluation/tests/testdata/hendrycksTest-high_school_physics-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"hendrycksTest-high_school_physics": {"acc": 0.2582781456953642, "acc_norm": 0.271523178807947, "acc_norm_stderr": 0.03631329803969653, "acc_stderr": 0.035737053147634576}}, "versions": {"hendrycksTest-high_school_physics": 0}}
|
lm-evaluation/tests/testdata/hendrycksTest-high_school_psychology-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"hendrycksTest-high_school_psychology": {"acc": 0.24587155963302754, "acc_norm": 0.23302752293577983, "acc_norm_stderr": 0.018125669180861493, "acc_stderr": 0.018461940968708436}}, "versions": {"hendrycksTest-high_school_psychology": 0}}
|
lm-evaluation/tests/testdata/hendrycksTest-high_school_statistics-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"hendrycksTest-high_school_statistics": {"acc": 0.2962962962962963, "acc_norm": 0.3055555555555556, "acc_norm_stderr": 0.03141554629402544, "acc_stderr": 0.03114144782353604}}, "versions": {"hendrycksTest-high_school_statistics": 0}}
|
lm-evaluation/tests/testdata/hendrycksTest-jurisprudence-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
cac440189f1ec778e82f4975d88b74689553ecc5116aaa7f76587a50c1a610e0
|
lm-evaluation/tests/testdata/hendrycksTest-moral_disputes-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"hendrycksTest-moral_disputes": {"acc": 0.24855491329479767, "acc_norm": 0.27167630057803466, "acc_norm_stderr": 0.023948512905468365, "acc_stderr": 0.023267528432100174}}, "versions": {"hendrycksTest-moral_disputes": 0}}
|
lm-evaluation/tests/testdata/hendrycksTest-professional_law-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"hendrycksTest-professional_law": {"acc": 0.2561929595827901, "acc_norm": 0.2470664928292047, "acc_norm_stderr": 0.011015752255279352, "acc_stderr": 0.011149173153110582}}, "versions": {"hendrycksTest-professional_law": 0}}
|
lm-evaluation/tests/testdata/hendrycksTest-public_relations-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"hendrycksTest-public_relations": {"acc": 0.3090909090909091, "acc_norm": 0.2636363636363636, "acc_norm_stderr": 0.04220224692971987, "acc_stderr": 0.044262946482000985}}, "versions": {"hendrycksTest-public_relations": 0}}
|
lm-evaluation/tests/testdata/lambada_openai-v2.0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
9ca5643bbaafed2f027eab5b68cc438e9e268f6df9a678e956e61726a985cf0b
|
lm-evaluation/tests/testdata/lambada_openai_mt_de-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"lambada_openai_mt_de": {"acc": 0.0, "acc_stderr": 0.0, "ppl": 1.6479047769869253, "ppl_stderr": 0.006497321146240192}}, "versions": {"lambada_openai_mt_de": 0}}
|
lm-evaluation/tests/testdata/lambada_openai_mt_es-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"lambada_openai_mt_es": {"acc": 0.0, "acc_stderr": 0.0, "ppl": 1.6479047769869253, "ppl_stderr": 0.006497321146240192}}, "versions": {"lambada_openai_mt_es": 0}}
|
lm-evaluation/tests/testdata/lambada_openai_mt_fr-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
5d16f4a0c51dc6d7b6df2ebeba2bbfa51e700b843779b559b3d90183d7b02a11
|
lm-evaluation/tests/testdata/lambada_openai_mt_it-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
fd87c6c5cf4e0499c5f9f80e5bd7ee6a4f3d2991902a0cc3ec9e6eaf22d6760a
|
lm-evaluation/tests/testdata/logiqa-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
12495c50454ba5e1ce0753bd18c09aaca516bebd27648d815e37b15229dbf198
|
lm-evaluation/tests/testdata/math_counting_and_prob-v1-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"math_counting_and_prob": {"acc": 0.0, "acc_stderr": 0.0}}, "versions": {"math_counting_and_prob": 1}}
|
lm-evaluation/tests/testdata/math_num_theory-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"math_num_theory": {"acc": 0.0, "acc_stderr": 0.0}}, "versions": {"math_num_theory": 0}}
|
lm-evaluation/tests/testdata/math_prealgebra-v0-greedy_until
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
752cdf343d7152e476b0273065024f6ea0e0f47ea385c6bdf9067736cb39724a
|
lm-evaluation/tests/testdata/math_prealgebra-v1-greedy_until
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
752cdf343d7152e476b0273065024f6ea0e0f47ea385c6bdf9067736cb39724a
|
lm-evaluation/tests/testdata/math_prealgebra-v1-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"math_prealgebra": {"acc": 0.0, "acc_stderr": 0.0}}, "versions": {"math_prealgebra": 1}}
|