Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- llama13b_500M/23-04-2024-13:05:50/checkpoints_zero_stage_2/global_step2/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
- llama13b_500M/23-04-2024-13:05:50/checkpoints_zero_stage_2/global_step2/layer_17-model_01-model_states.pt +3 -0
- llama13b_500M/23-04-2024-13:05:50/checkpoints_zero_stage_2/global_step2/layer_21-model_01-model_states.pt +3 -0
- lm-evaluation-harness/tests/testdata/anli_r1-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/arc_easy-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/arithmetic_2ds-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/arithmetic_4da-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/arithmetic_4da-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/arithmetic_4ds-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/blimp_anaphor_number_agreement-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/blimp_animate_subject_passive-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/blimp_determiner_noun_agreement_1-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/blimp_determiner_noun_agreement_2-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/blimp_irregular_plural_subject_verb_agreement_1-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/blimp_left_branch_island_echo_question-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/blimp_only_npi_licensor_present-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/blimp_principle_A_domain_3-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/blimp_regular_plural_subject_verb_agreement_2-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/blimp_wh_island-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/blimp_wh_vs_that_with_gap_long_distance-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/cb-v1-res.json +1 -0
- lm-evaluation-harness/tests/testdata/cola-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/copa-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/crows_pairs_english_age-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/crows_pairs_english_age-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/crows_pairs_english_autre-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/crows_pairs_english_disability-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/crows_pairs_french-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/crows_pairs_french_gender-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/crows_pairs_french_sexual_orientation-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/drop-v1-res.json +1 -0
- lm-evaluation-harness/tests/testdata/ethics_utilitarianism-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/hendrycksTest-college_medicine-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/hendrycksTest-high_school_us_history-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/hendrycksTest-high_school_world_history-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/hendrycksTest-logical_fallacies-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/hendrycksTest-marketing-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/hendrycksTest-moral_disputes-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/hendrycksTest-professional_medicine-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/hendrycksTest-security_studies-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/hendrycksTest-sociology-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/lambada_cloze-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/lambada_cloze-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/lambada_mt_en-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/lambada_mt_it-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/lambada_openai_cloze-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/lambada_openai_mt_fr-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/math_counting_and_prob-v1-greedy_until +1 -0
- lm-evaluation-harness/tests/testdata/mnli_mismatched-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/multirc-v0-res.json +1 -0
llama13b_500M/23-04-2024-13:05:50/checkpoints_zero_stage_2/global_step2/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8a0c1bc556363498d873e0b01790199fcb44c82ba812eb230b30d511e4209919
|
3 |
+
size 4193899376
|
llama13b_500M/23-04-2024-13:05:50/checkpoints_zero_stage_2/global_step2/layer_17-model_01-model_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f3153d27d1206319754c857616c86378cedd60abacb635639be3bfb7a74aa244
|
3 |
+
size 317218260
|
llama13b_500M/23-04-2024-13:05:50/checkpoints_zero_stage_2/global_step2/layer_21-model_01-model_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f5622453486cfcb162e7b2aa0308b8a80e310d5e596c2b514120d915e37eb69f
|
3 |
+
size 258213263
|
lm-evaluation-harness/tests/testdata/anli_r1-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"anli_r1": {"acc": 0.334, "acc_stderr": 0.014922019523732967}}, "versions": {"anli_r1": 0}}
|
lm-evaluation-harness/tests/testdata/arc_easy-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"arc_easy": {"acc": 0.2474747474747475, "acc_norm": 0.24074074074074073, "acc_norm_stderr": 0.008772796145221907, "acc_stderr": 0.008855114414834707}}, "versions": {"arc_easy": 0}}
|
lm-evaluation-harness/tests/testdata/arithmetic_2ds-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
66f7ff3b40251ee38fadcbee658e309a200224356fc3efa07d0a490a2c24bfa3
|
lm-evaluation-harness/tests/testdata/arithmetic_4da-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
d3557beb8b9e5704122c2fc6362b11fbe2c3f2f3cb72aed4462b208767c40e01
|
lm-evaluation-harness/tests/testdata/arithmetic_4da-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"arithmetic_4da": {"acc": 0.0, "acc_stderr": 0.0}}, "versions": {"arithmetic_4da": 0}}
|
lm-evaluation-harness/tests/testdata/arithmetic_4ds-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"arithmetic_4ds": {"acc": 0.0, "acc_stderr": 0.0}}, "versions": {"arithmetic_4ds": 0}}
|
lm-evaluation-harness/tests/testdata/blimp_anaphor_number_agreement-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"blimp_anaphor_number_agreement": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_anaphor_number_agreement": 0}}
|
lm-evaluation-harness/tests/testdata/blimp_animate_subject_passive-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"blimp_animate_subject_passive": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_animate_subject_passive": 0}}
|
lm-evaluation-harness/tests/testdata/blimp_determiner_noun_agreement_1-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
2df8cc7f17089f7e8c7d974dcb324c809d30ef059a5be22aed6b69f44230809f
|
lm-evaluation-harness/tests/testdata/blimp_determiner_noun_agreement_2-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"blimp_determiner_noun_agreement_2": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_determiner_noun_agreement_2": 0}}
|
lm-evaluation-harness/tests/testdata/blimp_irregular_plural_subject_verb_agreement_1-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
7084358b1b7dd7fb5ead1a58f4b499d6f7610eca897bfac25a986d0f9a91aa5d
|
lm-evaluation-harness/tests/testdata/blimp_left_branch_island_echo_question-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"blimp_left_branch_island_echo_question": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_left_branch_island_echo_question": 0}}
|
lm-evaluation-harness/tests/testdata/blimp_only_npi_licensor_present-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
d2d0711611b5b218c6fa8c7278494749252b7868c396451919b761303556bd66
|
lm-evaluation-harness/tests/testdata/blimp_principle_A_domain_3-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
38454befedcf1f3f6ef27d3bef9ccfdfb3e94a7ab32d86a63493a920d2d50093
|
lm-evaluation-harness/tests/testdata/blimp_regular_plural_subject_verb_agreement_2-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"blimp_regular_plural_subject_verb_agreement_2": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_regular_plural_subject_verb_agreement_2": 0}}
|
lm-evaluation-harness/tests/testdata/blimp_wh_island-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
91a9e4b60b0f3572a7fdbd7648d0e69f36e5eb34db715315b0082558d7ed8b65
|
lm-evaluation-harness/tests/testdata/blimp_wh_vs_that_with_gap_long_distance-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
eed67491bdf493a1dad8f1d9766bc7bd0e79946365b833c0f7eb81ac998e3dca
|
lm-evaluation-harness/tests/testdata/cb-v1-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"cb": {"acc": 0.3392857142857143, "acc_stderr": 0.06384226561930825, "f1": 0.2819143819143819}}, "versions": {"cb": 1}}
|
lm-evaluation-harness/tests/testdata/cola-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
e8635578ed8ee70b707a666d35e468b9321db24470f80c92080651e2bfa01751
|
lm-evaluation-harness/tests/testdata/copa-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
66276b9045b5300cba4b81340db06f674f031fa0b8883714ad0d03be464cd799
|
lm-evaluation-harness/tests/testdata/crows_pairs_english_age-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
de74d2ac7f926f2f486c045d84aae8f71711102f9d77b31f758fd148810d13d3
|
lm-evaluation-harness/tests/testdata/crows_pairs_english_age-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"crows_pairs_english_age": {"likelihood_difference": 0.3160680928470684, "likelihood_difference_stderr": 0.02397758321605678, "pct_stereotype": 0.43956043956043955, "pct_stereotype_stderr": 0.05231815698566189}}, "versions": {"crows_pairs_english_age": 0}}
|
lm-evaluation-harness/tests/testdata/crows_pairs_english_autre-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
a197ccc8538231404a8e43f5ed0fbbfb2c317b4da337f6e7aa9642131aeb426a
|
lm-evaluation-harness/tests/testdata/crows_pairs_english_disability-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
90c1bcfdeec0ff51d891ee8cf00ae2a5ec61bab6739faea9865809b8ffed2cdb
|
lm-evaluation-harness/tests/testdata/crows_pairs_french-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"crows_pairs_french": {"likelihood_difference": 0.3367363060632734, "likelihood_difference_stderr": 0.005827747024053628, "pct_stereotype": 0.5062611806797853, "pct_stereotype_stderr": 0.012212341600228745}}, "versions": {"crows_pairs_french": 0}}
|
lm-evaluation-harness/tests/testdata/crows_pairs_french_gender-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"crows_pairs_french_gender": {"likelihood_difference": 0.3364019171359413, "likelihood_difference_stderr": 0.012815700745990895, "pct_stereotype": 0.4766355140186916, "pct_stereotype_stderr": 0.027920316348204986}}, "versions": {"crows_pairs_french_gender": 0}}
|
lm-evaluation-harness/tests/testdata/crows_pairs_french_sexual_orientation-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
2ce823fdb93d325aa8fb40db5d335b093b4b69792763532d940a752440ee3a76
|
lm-evaluation-harness/tests/testdata/drop-v1-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"drop": {"em": 0.0, "em_stderr": 0.0, "f1": 0.0, "f1_stderr": 0.0}}, "versions": {"drop": 1}}
|
lm-evaluation-harness/tests/testdata/ethics_utilitarianism-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"ethics_utilitarianism": {"acc": 0.49771214642262895, "acc_stderr": 0.007211546310787838}}, "versions": {"ethics_utilitarianism": 0}}
|
lm-evaluation-harness/tests/testdata/hendrycksTest-college_medicine-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
dd6e0a9be1407890e9f8cd4434fb6aa4752ab3d2473837fd465ad99f60ad685e
|
lm-evaluation-harness/tests/testdata/hendrycksTest-high_school_us_history-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
8c65c1a28330dd001d395ac11f1bb80c3b33f5935f503e74067aef6e9e1d9d9b
|
lm-evaluation-harness/tests/testdata/hendrycksTest-high_school_world_history-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"hendrycksTest-high_school_world_history": {"acc": 0.23628691983122363, "acc_norm": 0.24472573839662448, "acc_norm_stderr": 0.02798569938703642, "acc_stderr": 0.027652153144159263}}, "versions": {"hendrycksTest-high_school_world_history": 0}}
|
lm-evaluation-harness/tests/testdata/hendrycksTest-logical_fallacies-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
2e9449dd803f9e2334dc562d9f04031fd013ed36b883b44ab500533a5dbbface
|
lm-evaluation-harness/tests/testdata/hendrycksTest-marketing-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"hendrycksTest-marketing": {"acc": 0.2863247863247863, "acc_norm": 0.2905982905982906, "acc_norm_stderr": 0.029745048572674043, "acc_stderr": 0.029614323690456648}}, "versions": {"hendrycksTest-marketing": 0}}
|
lm-evaluation-harness/tests/testdata/hendrycksTest-moral_disputes-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"hendrycksTest-moral_disputes": {"acc": 0.24855491329479767, "acc_norm": 0.27167630057803466, "acc_norm_stderr": 0.023948512905468365, "acc_stderr": 0.023267528432100174}}, "versions": {"hendrycksTest-moral_disputes": 0}}
|
lm-evaluation-harness/tests/testdata/hendrycksTest-professional_medicine-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
7a30599858398169cde61430c18efdd7fb4dcd09c34aa9baba70f0f8cf17a9f1
|
lm-evaluation-harness/tests/testdata/hendrycksTest-security_studies-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
92dfffe2acf3278256486d3e1cf1edb5a739ad0a54c0f9c67695f7a411ed5f76
|
lm-evaluation-harness/tests/testdata/hendrycksTest-sociology-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"hendrycksTest-sociology": {"acc": 0.23383084577114427, "acc_norm": 0.24875621890547264, "acc_norm_stderr": 0.030567675938916707, "acc_stderr": 0.02992941540834838}}, "versions": {"hendrycksTest-sociology": 0}}
|
lm-evaluation-harness/tests/testdata/lambada_cloze-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
7655e748b63ae7e9911411d2d2a2577221d6c861ca4448509992541294d689f3
|
lm-evaluation-harness/tests/testdata/lambada_cloze-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"lambada_cloze": {"acc": 0.0, "acc_stderr": 0.0, "ppl": 1.6479047769869253, "ppl_stderr": 0.006497321146240192}}, "versions": {"lambada_cloze": 0}}
|
lm-evaluation-harness/tests/testdata/lambada_mt_en-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"lambada_mt_en": {"acc": 0.0, "acc_stderr": 0.0, "ppl": 1.6479047769869253, "ppl_stderr": 0.006497321146240192}}, "versions": {"lambada_mt_en": 0}}
|
lm-evaluation-harness/tests/testdata/lambada_mt_it-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
fd87c6c5cf4e0499c5f9f80e5bd7ee6a4f3d2991902a0cc3ec9e6eaf22d6760a
|
lm-evaluation-harness/tests/testdata/lambada_openai_cloze-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
7655e748b63ae7e9911411d2d2a2577221d6c861ca4448509992541294d689f3
|
lm-evaluation-harness/tests/testdata/lambada_openai_mt_fr-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
5d16f4a0c51dc6d7b6df2ebeba2bbfa51e700b843779b559b3d90183d7b02a11
|
lm-evaluation-harness/tests/testdata/math_counting_and_prob-v1-greedy_until
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
2aa9ae43ee9dbb2457525247d7b65358632c5eaa9cbfc40cf95a4f17f5d942ad
|
lm-evaluation-harness/tests/testdata/mnli_mismatched-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"mnli_mismatched": {"acc": 0.3360455655004068, "acc_stderr": 0.004763973908606819}}, "versions": {"mnli_mismatched": 0}}
|
lm-evaluation-harness/tests/testdata/multirc-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"multirc": {"acc": 0.07450157397691501, "acc_stderr": 0.008510441526175931}}, "versions": {"multirc": 0}}
|