applied-ai-018 commited on
Commit
cf0b273
·
verified ·
1 Parent(s): c9133a7

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. llama13b_500M/23-04-2024-13:05:50/checkpoints_zero_stage_2/global_step2/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
  2. llama13b_500M/23-04-2024-13:05:50/checkpoints_zero_stage_2/global_step2/layer_17-model_01-model_states.pt +3 -0
  3. llama13b_500M/23-04-2024-13:05:50/checkpoints_zero_stage_2/global_step2/layer_21-model_01-model_states.pt +3 -0
  4. lm-evaluation-harness/tests/testdata/anli_r1-v0-res.json +1 -0
  5. lm-evaluation-harness/tests/testdata/arc_easy-v0-res.json +1 -0
  6. lm-evaluation-harness/tests/testdata/arithmetic_2ds-v0-loglikelihood +1 -0
  7. lm-evaluation-harness/tests/testdata/arithmetic_4da-v0-loglikelihood +1 -0
  8. lm-evaluation-harness/tests/testdata/arithmetic_4da-v0-res.json +1 -0
  9. lm-evaluation-harness/tests/testdata/arithmetic_4ds-v0-res.json +1 -0
  10. lm-evaluation-harness/tests/testdata/blimp_anaphor_number_agreement-v0-res.json +1 -0
  11. lm-evaluation-harness/tests/testdata/blimp_animate_subject_passive-v0-res.json +1 -0
  12. lm-evaluation-harness/tests/testdata/blimp_determiner_noun_agreement_1-v0-loglikelihood +1 -0
  13. lm-evaluation-harness/tests/testdata/blimp_determiner_noun_agreement_2-v0-res.json +1 -0
  14. lm-evaluation-harness/tests/testdata/blimp_irregular_plural_subject_verb_agreement_1-v0-loglikelihood +1 -0
  15. lm-evaluation-harness/tests/testdata/blimp_left_branch_island_echo_question-v0-res.json +1 -0
  16. lm-evaluation-harness/tests/testdata/blimp_only_npi_licensor_present-v0-loglikelihood +1 -0
  17. lm-evaluation-harness/tests/testdata/blimp_principle_A_domain_3-v0-loglikelihood +1 -0
  18. lm-evaluation-harness/tests/testdata/blimp_regular_plural_subject_verb_agreement_2-v0-res.json +1 -0
  19. lm-evaluation-harness/tests/testdata/blimp_wh_island-v0-loglikelihood +1 -0
  20. lm-evaluation-harness/tests/testdata/blimp_wh_vs_that_with_gap_long_distance-v0-loglikelihood +1 -0
  21. lm-evaluation-harness/tests/testdata/cb-v1-res.json +1 -0
  22. lm-evaluation-harness/tests/testdata/cola-v0-loglikelihood +1 -0
  23. lm-evaluation-harness/tests/testdata/copa-v0-loglikelihood +1 -0
  24. lm-evaluation-harness/tests/testdata/crows_pairs_english_age-v0-loglikelihood +1 -0
  25. lm-evaluation-harness/tests/testdata/crows_pairs_english_age-v0-res.json +1 -0
  26. lm-evaluation-harness/tests/testdata/crows_pairs_english_autre-v0-loglikelihood +1 -0
  27. lm-evaluation-harness/tests/testdata/crows_pairs_english_disability-v0-loglikelihood +1 -0
  28. lm-evaluation-harness/tests/testdata/crows_pairs_french-v0-res.json +1 -0
  29. lm-evaluation-harness/tests/testdata/crows_pairs_french_gender-v0-res.json +1 -0
  30. lm-evaluation-harness/tests/testdata/crows_pairs_french_sexual_orientation-v0-loglikelihood +1 -0
  31. lm-evaluation-harness/tests/testdata/drop-v1-res.json +1 -0
  32. lm-evaluation-harness/tests/testdata/ethics_utilitarianism-v0-res.json +1 -0
  33. lm-evaluation-harness/tests/testdata/hendrycksTest-college_medicine-v0-loglikelihood +1 -0
  34. lm-evaluation-harness/tests/testdata/hendrycksTest-high_school_us_history-v0-loglikelihood +1 -0
  35. lm-evaluation-harness/tests/testdata/hendrycksTest-high_school_world_history-v0-res.json +1 -0
  36. lm-evaluation-harness/tests/testdata/hendrycksTest-logical_fallacies-v0-loglikelihood +1 -0
  37. lm-evaluation-harness/tests/testdata/hendrycksTest-marketing-v0-res.json +1 -0
  38. lm-evaluation-harness/tests/testdata/hendrycksTest-moral_disputes-v0-res.json +1 -0
  39. lm-evaluation-harness/tests/testdata/hendrycksTest-professional_medicine-v0-loglikelihood +1 -0
  40. lm-evaluation-harness/tests/testdata/hendrycksTest-security_studies-v0-loglikelihood +1 -0
  41. lm-evaluation-harness/tests/testdata/hendrycksTest-sociology-v0-res.json +1 -0
  42. lm-evaluation-harness/tests/testdata/lambada_cloze-v0-loglikelihood +1 -0
  43. lm-evaluation-harness/tests/testdata/lambada_cloze-v0-res.json +1 -0
  44. lm-evaluation-harness/tests/testdata/lambada_mt_en-v0-res.json +1 -0
  45. lm-evaluation-harness/tests/testdata/lambada_mt_it-v0-loglikelihood +1 -0
  46. lm-evaluation-harness/tests/testdata/lambada_openai_cloze-v0-loglikelihood +1 -0
  47. lm-evaluation-harness/tests/testdata/lambada_openai_mt_fr-v0-loglikelihood +1 -0
  48. lm-evaluation-harness/tests/testdata/math_counting_and_prob-v1-greedy_until +1 -0
  49. lm-evaluation-harness/tests/testdata/mnli_mismatched-v0-res.json +1 -0
  50. lm-evaluation-harness/tests/testdata/multirc-v0-res.json +1 -0
llama13b_500M/23-04-2024-13:05:50/checkpoints_zero_stage_2/global_step2/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a0c1bc556363498d873e0b01790199fcb44c82ba812eb230b30d511e4209919
3
+ size 4193899376
llama13b_500M/23-04-2024-13:05:50/checkpoints_zero_stage_2/global_step2/layer_17-model_01-model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3153d27d1206319754c857616c86378cedd60abacb635639be3bfb7a74aa244
3
+ size 317218260
llama13b_500M/23-04-2024-13:05:50/checkpoints_zero_stage_2/global_step2/layer_21-model_01-model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5622453486cfcb162e7b2aa0308b8a80e310d5e596c2b514120d915e37eb69f
3
+ size 258213263
lm-evaluation-harness/tests/testdata/anli_r1-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"anli_r1": {"acc": 0.334, "acc_stderr": 0.014922019523732967}}, "versions": {"anli_r1": 0}}
lm-evaluation-harness/tests/testdata/arc_easy-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"arc_easy": {"acc": 0.2474747474747475, "acc_norm": 0.24074074074074073, "acc_norm_stderr": 0.008772796145221907, "acc_stderr": 0.008855114414834707}}, "versions": {"arc_easy": 0}}
lm-evaluation-harness/tests/testdata/arithmetic_2ds-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ 66f7ff3b40251ee38fadcbee658e309a200224356fc3efa07d0a490a2c24bfa3
lm-evaluation-harness/tests/testdata/arithmetic_4da-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ d3557beb8b9e5704122c2fc6362b11fbe2c3f2f3cb72aed4462b208767c40e01
lm-evaluation-harness/tests/testdata/arithmetic_4da-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"arithmetic_4da": {"acc": 0.0, "acc_stderr": 0.0}}, "versions": {"arithmetic_4da": 0}}
lm-evaluation-harness/tests/testdata/arithmetic_4ds-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"arithmetic_4ds": {"acc": 0.0, "acc_stderr": 0.0}}, "versions": {"arithmetic_4ds": 0}}
lm-evaluation-harness/tests/testdata/blimp_anaphor_number_agreement-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"blimp_anaphor_number_agreement": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_anaphor_number_agreement": 0}}
lm-evaluation-harness/tests/testdata/blimp_animate_subject_passive-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"blimp_animate_subject_passive": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_animate_subject_passive": 0}}
lm-evaluation-harness/tests/testdata/blimp_determiner_noun_agreement_1-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ 2df8cc7f17089f7e8c7d974dcb324c809d30ef059a5be22aed6b69f44230809f
lm-evaluation-harness/tests/testdata/blimp_determiner_noun_agreement_2-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"blimp_determiner_noun_agreement_2": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_determiner_noun_agreement_2": 0}}
lm-evaluation-harness/tests/testdata/blimp_irregular_plural_subject_verb_agreement_1-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ 7084358b1b7dd7fb5ead1a58f4b499d6f7610eca897bfac25a986d0f9a91aa5d
lm-evaluation-harness/tests/testdata/blimp_left_branch_island_echo_question-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"blimp_left_branch_island_echo_question": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_left_branch_island_echo_question": 0}}
lm-evaluation-harness/tests/testdata/blimp_only_npi_licensor_present-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ d2d0711611b5b218c6fa8c7278494749252b7868c396451919b761303556bd66
lm-evaluation-harness/tests/testdata/blimp_principle_A_domain_3-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ 38454befedcf1f3f6ef27d3bef9ccfdfb3e94a7ab32d86a63493a920d2d50093
lm-evaluation-harness/tests/testdata/blimp_regular_plural_subject_verb_agreement_2-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"blimp_regular_plural_subject_verb_agreement_2": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_regular_plural_subject_verb_agreement_2": 0}}
lm-evaluation-harness/tests/testdata/blimp_wh_island-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ 91a9e4b60b0f3572a7fdbd7648d0e69f36e5eb34db715315b0082558d7ed8b65
lm-evaluation-harness/tests/testdata/blimp_wh_vs_that_with_gap_long_distance-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ eed67491bdf493a1dad8f1d9766bc7bd0e79946365b833c0f7eb81ac998e3dca
lm-evaluation-harness/tests/testdata/cb-v1-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"cb": {"acc": 0.3392857142857143, "acc_stderr": 0.06384226561930825, "f1": 0.2819143819143819}}, "versions": {"cb": 1}}
lm-evaluation-harness/tests/testdata/cola-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ e8635578ed8ee70b707a666d35e468b9321db24470f80c92080651e2bfa01751
lm-evaluation-harness/tests/testdata/copa-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ 66276b9045b5300cba4b81340db06f674f031fa0b8883714ad0d03be464cd799
lm-evaluation-harness/tests/testdata/crows_pairs_english_age-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ de74d2ac7f926f2f486c045d84aae8f71711102f9d77b31f758fd148810d13d3
lm-evaluation-harness/tests/testdata/crows_pairs_english_age-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"crows_pairs_english_age": {"likelihood_difference": 0.3160680928470684, "likelihood_difference_stderr": 0.02397758321605678, "pct_stereotype": 0.43956043956043955, "pct_stereotype_stderr": 0.05231815698566189}}, "versions": {"crows_pairs_english_age": 0}}
lm-evaluation-harness/tests/testdata/crows_pairs_english_autre-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ a197ccc8538231404a8e43f5ed0fbbfb2c317b4da337f6e7aa9642131aeb426a
lm-evaluation-harness/tests/testdata/crows_pairs_english_disability-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ 90c1bcfdeec0ff51d891ee8cf00ae2a5ec61bab6739faea9865809b8ffed2cdb
lm-evaluation-harness/tests/testdata/crows_pairs_french-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"crows_pairs_french": {"likelihood_difference": 0.3367363060632734, "likelihood_difference_stderr": 0.005827747024053628, "pct_stereotype": 0.5062611806797853, "pct_stereotype_stderr": 0.012212341600228745}}, "versions": {"crows_pairs_french": 0}}
lm-evaluation-harness/tests/testdata/crows_pairs_french_gender-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"crows_pairs_french_gender": {"likelihood_difference": 0.3364019171359413, "likelihood_difference_stderr": 0.012815700745990895, "pct_stereotype": 0.4766355140186916, "pct_stereotype_stderr": 0.027920316348204986}}, "versions": {"crows_pairs_french_gender": 0}}
lm-evaluation-harness/tests/testdata/crows_pairs_french_sexual_orientation-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ 2ce823fdb93d325aa8fb40db5d335b093b4b69792763532d940a752440ee3a76
lm-evaluation-harness/tests/testdata/drop-v1-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"drop": {"em": 0.0, "em_stderr": 0.0, "f1": 0.0, "f1_stderr": 0.0}}, "versions": {"drop": 1}}
lm-evaluation-harness/tests/testdata/ethics_utilitarianism-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"ethics_utilitarianism": {"acc": 0.49771214642262895, "acc_stderr": 0.007211546310787838}}, "versions": {"ethics_utilitarianism": 0}}
lm-evaluation-harness/tests/testdata/hendrycksTest-college_medicine-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ dd6e0a9be1407890e9f8cd4434fb6aa4752ab3d2473837fd465ad99f60ad685e
lm-evaluation-harness/tests/testdata/hendrycksTest-high_school_us_history-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ 8c65c1a28330dd001d395ac11f1bb80c3b33f5935f503e74067aef6e9e1d9d9b
lm-evaluation-harness/tests/testdata/hendrycksTest-high_school_world_history-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"hendrycksTest-high_school_world_history": {"acc": 0.23628691983122363, "acc_norm": 0.24472573839662448, "acc_norm_stderr": 0.02798569938703642, "acc_stderr": 0.027652153144159263}}, "versions": {"hendrycksTest-high_school_world_history": 0}}
lm-evaluation-harness/tests/testdata/hendrycksTest-logical_fallacies-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ 2e9449dd803f9e2334dc562d9f04031fd013ed36b883b44ab500533a5dbbface
lm-evaluation-harness/tests/testdata/hendrycksTest-marketing-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"hendrycksTest-marketing": {"acc": 0.2863247863247863, "acc_norm": 0.2905982905982906, "acc_norm_stderr": 0.029745048572674043, "acc_stderr": 0.029614323690456648}}, "versions": {"hendrycksTest-marketing": 0}}
lm-evaluation-harness/tests/testdata/hendrycksTest-moral_disputes-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"hendrycksTest-moral_disputes": {"acc": 0.24855491329479767, "acc_norm": 0.27167630057803466, "acc_norm_stderr": 0.023948512905468365, "acc_stderr": 0.023267528432100174}}, "versions": {"hendrycksTest-moral_disputes": 0}}
lm-evaluation-harness/tests/testdata/hendrycksTest-professional_medicine-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ 7a30599858398169cde61430c18efdd7fb4dcd09c34aa9baba70f0f8cf17a9f1
lm-evaluation-harness/tests/testdata/hendrycksTest-security_studies-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ 92dfffe2acf3278256486d3e1cf1edb5a739ad0a54c0f9c67695f7a411ed5f76
lm-evaluation-harness/tests/testdata/hendrycksTest-sociology-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"hendrycksTest-sociology": {"acc": 0.23383084577114427, "acc_norm": 0.24875621890547264, "acc_norm_stderr": 0.030567675938916707, "acc_stderr": 0.02992941540834838}}, "versions": {"hendrycksTest-sociology": 0}}
lm-evaluation-harness/tests/testdata/lambada_cloze-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ 7655e748b63ae7e9911411d2d2a2577221d6c861ca4448509992541294d689f3
lm-evaluation-harness/tests/testdata/lambada_cloze-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"lambada_cloze": {"acc": 0.0, "acc_stderr": 0.0, "ppl": 1.6479047769869253, "ppl_stderr": 0.006497321146240192}}, "versions": {"lambada_cloze": 0}}
lm-evaluation-harness/tests/testdata/lambada_mt_en-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"lambada_mt_en": {"acc": 0.0, "acc_stderr": 0.0, "ppl": 1.6479047769869253, "ppl_stderr": 0.006497321146240192}}, "versions": {"lambada_mt_en": 0}}
lm-evaluation-harness/tests/testdata/lambada_mt_it-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ fd87c6c5cf4e0499c5f9f80e5bd7ee6a4f3d2991902a0cc3ec9e6eaf22d6760a
lm-evaluation-harness/tests/testdata/lambada_openai_cloze-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ 7655e748b63ae7e9911411d2d2a2577221d6c861ca4448509992541294d689f3
lm-evaluation-harness/tests/testdata/lambada_openai_mt_fr-v0-loglikelihood ADDED
@@ -0,0 +1 @@
 
 
1
+ 5d16f4a0c51dc6d7b6df2ebeba2bbfa51e700b843779b559b3d90183d7b02a11
lm-evaluation-harness/tests/testdata/math_counting_and_prob-v1-greedy_until ADDED
@@ -0,0 +1 @@
 
 
1
+ 2aa9ae43ee9dbb2457525247d7b65358632c5eaa9cbfc40cf95a4f17f5d942ad
lm-evaluation-harness/tests/testdata/mnli_mismatched-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"mnli_mismatched": {"acc": 0.3360455655004068, "acc_stderr": 0.004763973908606819}}, "versions": {"mnli_mismatched": 0}}
lm-evaluation-harness/tests/testdata/multirc-v0-res.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"results": {"multirc": {"acc": 0.07450157397691501, "acc_stderr": 0.008510441526175931}}, "versions": {"multirc": 0}}