Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +4 -0
- ckpts/universal/global_step40/zero/17.mlp.dense_h_to_4h_swiglu.weight/exp_avg.pt +3 -0
- ckpts/universal/global_step40/zero/17.mlp.dense_h_to_4h_swiglu.weight/exp_avg_sq.pt +3 -0
- ckpts/universal/global_step40/zero/17.mlp.dense_h_to_4h_swiglu.weight/fp32.pt +3 -0
- ckpts/universal/global_step40/zero/21.mlp.dense_4h_to_h.weight/exp_avg_sq.pt +3 -0
- ckpts/universal/global_step40/zero/3.post_attention_layernorm.weight/exp_avg.pt +3 -0
- ckpts/universal/global_step40/zero/3.post_attention_layernorm.weight/fp32.pt +3 -0
- lm-evaluation-harness/tests/testdata/arithmetic_1dc-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/blimp_irregular_past_participle_verbs-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/blimp_left_branch_island_simple_question-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/blimp_only_npi_scope-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/blimp_passive_1-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/blimp_wh_vs_that_no_gap_long_distance-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/crows_pairs_english_gender-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/crows_pairs_french_race_color-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/crows_pairs_french_race_color-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/crows_pairs_french_sexual_orientation-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/ethics_virtue-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/hendrycksTest-college_mathematics-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/hendrycksTest-electrical_engineering-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/hendrycksTest-high_school_biology-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/hendrycksTest-logical_fallacies-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/hendrycksTest-moral_scenarios-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/lambada_cloze-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/lambada_openai-v2.0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/lambada_openai_mt_it-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/lambada_standard-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/math_algebra-v0-greedy_until +1 -0
- lm-evaluation-harness/tests/testdata/math_algebra-v1-greedy_until +1 -0
- lm-evaluation-harness/tests/testdata/math_intermediate_algebra-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/pile_arxiv-v1-loglikelihood_rolling +1 -0
- lm-evaluation-harness/tests/testdata/pile_arxiv-v1-res.json +1 -0
- lm-evaluation-harness/tests/testdata/pile_bookcorpus2-v1-res.json +1 -0
- lm-evaluation-harness/tests/testdata/pile_europarl-v1-loglikelihood_rolling +1 -0
- lm-evaluation-harness/tests/testdata/pile_hackernews-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/pile_philpapers-v1-loglikelihood_rolling +1 -0
- lm-evaluation-harness/tests/testdata/pile_pile-cc-v1-loglikelihood_rolling +1 -0
- lm-evaluation-harness/tests/testdata/pile_youtubesubtitles-v1-loglikelihood_rolling +1 -0
- lm-evaluation-harness/tests/testdata/qa4mre_2011-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/sst-v0-loglikelihood +1 -0
- lm-evaluation-harness/tests/testdata/wmt16-en-ro-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/wmt20-cs-en-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/wmt20-en-de-v0-greedy_until +1 -0
- lm-evaluation-harness/tests/testdata/wmt20-en-iu-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/wmt20-en-pl-v0-greedy_until +1 -0
- lm-evaluation-harness/tests/testdata/wmt20-en-ps-v0-greedy_until +1 -0
- lm-evaluation-harness/tests/testdata/wmt20-en-zh-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/wmt20-en-zh-v1-res.json +1 -0
- lm-evaluation-harness/tests/testdata/wmt20-fr-de-v0-res.json +1 -0
- lm-evaluation-harness/tests/testdata/wmt20-pl-en-v0-res.json +1 -0
.gitattributes
CHANGED
@@ -72,3 +72,7 @@ venv/lib/python3.10/site-packages/pydantic_core/_pydantic_core.cpython-310-x86_6
|
|
72 |
venv/lib/python3.10/site-packages/zstandard/_cffi.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
|
73 |
venv/lib/python3.10/site-packages/yaml/_yaml.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
|
74 |
venv/lib/python3.10/site-packages/safetensors/_safetensors_rust.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
72 |
venv/lib/python3.10/site-packages/zstandard/_cffi.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
|
73 |
venv/lib/python3.10/site-packages/yaml/_yaml.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
|
74 |
venv/lib/python3.10/site-packages/safetensors/_safetensors_rust.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
|
75 |
+
venv/lib/python3.10/site-packages/lxml/objectify.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
|
76 |
+
venv/lib/python3.10/site-packages/pyarrow/_dataset.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
|
77 |
+
venv/lib/python3.10/site-packages/lxml/etree.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
|
78 |
+
venv/lib/python3.10/site-packages/pyarrow/libparquet.so.1600 filter=lfs diff=lfs merge=lfs -text
|
ckpts/universal/global_step40/zero/17.mlp.dense_h_to_4h_swiglu.weight/exp_avg.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0e03a35ada3cb32db3a50c91fa987c020e7cc3cd40d29299cd451c2d07654b09
|
3 |
+
size 33555612
|
ckpts/universal/global_step40/zero/17.mlp.dense_h_to_4h_swiglu.weight/exp_avg_sq.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e69f123a14465a13101b5e6c579c7333c5b44c1505ada2b19b8408c84c9d1bcc
|
3 |
+
size 33555627
|
ckpts/universal/global_step40/zero/17.mlp.dense_h_to_4h_swiglu.weight/fp32.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:459f609bbc9bd2e507784de5dea7fde30e6227ba8c994af991478d78c9e4332d
|
3 |
+
size 33555533
|
ckpts/universal/global_step40/zero/21.mlp.dense_4h_to_h.weight/exp_avg_sq.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a20bc4a005fdf87fa7206b9f6fc7a22ac2726348813efeeb490f96e3b6a8deb6
|
3 |
+
size 33555627
|
ckpts/universal/global_step40/zero/3.post_attention_layernorm.weight/exp_avg.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6b6e3741eee58ba1e686f474f5b75272d6c0f142250242f97c037b984977eecf
|
3 |
+
size 9372
|
ckpts/universal/global_step40/zero/3.post_attention_layernorm.weight/fp32.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:55baea1be4d87c71e5b29923772c5262815e5a7e3a187b3b741af9af0b9264b1
|
3 |
+
size 9293
|
lm-evaluation-harness/tests/testdata/arithmetic_1dc-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
04c3a63a6b3c579bd3775d92b3076ba9130041d5ce7cf9244d3f86e95c804387
|
lm-evaluation-harness/tests/testdata/blimp_irregular_past_participle_verbs-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"blimp_irregular_past_participle_verbs": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_irregular_past_participle_verbs": 0}}
|
lm-evaluation-harness/tests/testdata/blimp_left_branch_island_simple_question-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"blimp_left_branch_island_simple_question": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_left_branch_island_simple_question": 0}}
|
lm-evaluation-harness/tests/testdata/blimp_only_npi_scope-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"blimp_only_npi_scope": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_only_npi_scope": 0}}
|
lm-evaluation-harness/tests/testdata/blimp_passive_1-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
fa4addddd8e380031b8e0871776cabcb707c0f21dcaf5d8b3defec66cce55043
|
lm-evaluation-harness/tests/testdata/blimp_wh_vs_that_no_gap_long_distance-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"blimp_wh_vs_that_no_gap_long_distance": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_wh_vs_that_no_gap_long_distance": 0}}
|
lm-evaluation-harness/tests/testdata/crows_pairs_english_gender-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"crows_pairs_english_gender": {"likelihood_difference": 0.3361377482385407, "likelihood_difference_stderr": 0.012853081126751691, "pct_stereotype": 0.478125, "pct_stereotype_stderr": 0.027967820983765136}}, "versions": {"crows_pairs_english_gender": 0}}
|
lm-evaluation-harness/tests/testdata/crows_pairs_french_race_color-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
6f9119026abff33c5c882d6172e092e806a8b21bd86864022978b1961839350f
|
lm-evaluation-harness/tests/testdata/crows_pairs_french_race_color-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"crows_pairs_french_race_color": {"likelihood_difference": 0.33233909422443764, "likelihood_difference_stderr": 0.010623405969915857, "pct_stereotype": 0.4782608695652174, "pct_stereotype_stderr": 0.023315932363473738}}, "versions": {"crows_pairs_french_race_color": 0}}
|
lm-evaluation-harness/tests/testdata/crows_pairs_french_sexual_orientation-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
2ce823fdb93d325aa8fb40db5d335b093b4b69792763532d940a752440ee3a76
|
lm-evaluation-harness/tests/testdata/ethics_virtue-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"ethics_virtue": {"acc": 0.5035175879396985, "acc_stderr": 0.0070893491553555765, "em": 0.036180904522613064}}, "versions": {"ethics_virtue": 0}}
|
lm-evaluation-harness/tests/testdata/hendrycksTest-college_mathematics-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
e9fe80752686527281f834d2397875b4580581434b94799f9de6aaa450bd73ff
|
lm-evaluation-harness/tests/testdata/hendrycksTest-electrical_engineering-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
b9b5d8b8bb02696302ec6bc2a99bf987a5504d3bae0e529d2c8f263538c97518
|
lm-evaluation-harness/tests/testdata/hendrycksTest-high_school_biology-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
d4dc051f37a49dc75c218741e87bc826fd44f31ee1309b55e0f33bd191c1bc78
|
lm-evaluation-harness/tests/testdata/hendrycksTest-logical_fallacies-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"hendrycksTest-logical_fallacies": {"acc": 0.20245398773006135, "acc_norm": 0.2147239263803681, "acc_norm_stderr": 0.03226219377286774, "acc_stderr": 0.03157065078911902}}, "versions": {"hendrycksTest-logical_fallacies": 0}}
|
lm-evaluation-harness/tests/testdata/hendrycksTest-moral_scenarios-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
a8e1882e77728b53c8b86312254d08320d8363fb606d746a8dd145b812f62cf5
|
lm-evaluation-harness/tests/testdata/lambada_cloze-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"lambada_cloze": {"acc": 0.0, "acc_stderr": 0.0, "ppl": 1.6479047769869253, "ppl_stderr": 0.006497321146240192}}, "versions": {"lambada_cloze": 0}}
|
lm-evaluation-harness/tests/testdata/lambada_openai-v2.0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
9ca5643bbaafed2f027eab5b68cc438e9e268f6df9a678e956e61726a985cf0b
|
lm-evaluation-harness/tests/testdata/lambada_openai_mt_it-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
fd87c6c5cf4e0499c5f9f80e5bd7ee6a4f3d2991902a0cc3ec9e6eaf22d6760a
|
lm-evaluation-harness/tests/testdata/lambada_standard-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
8958d9f8d8145046b692fadd8a9cc9c8bad5617c10774280cf7c24c21d2be160
|
lm-evaluation-harness/tests/testdata/math_algebra-v0-greedy_until
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
f19182ce697a2c095d9e5b56ee6659dc38c93994b69ca75d7c3d3f5fd87572b4
|
lm-evaluation-harness/tests/testdata/math_algebra-v1-greedy_until
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
f19182ce697a2c095d9e5b56ee6659dc38c93994b69ca75d7c3d3f5fd87572b4
|
lm-evaluation-harness/tests/testdata/math_intermediate_algebra-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"math_intermediate_algebra": {"acc": 0.0, "acc_stderr": 0.0}}, "versions": {"math_intermediate_algebra": 0}}
|
lm-evaluation-harness/tests/testdata/pile_arxiv-v1-loglikelihood_rolling
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
814f9954e44368559602c00f7e85fa3971acdfd0315f508ec7df6318a79c55ec
|
lm-evaluation-harness/tests/testdata/pile_arxiv-v1-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"pile_arxiv": {"bits_per_byte": 1.55095665856779e-05, "byte_perplexity": 1.0000107504701365, "word_perplexity": 1.0000819333090385}}, "versions": {"pile_arxiv": 1}}
|
lm-evaluation-harness/tests/testdata/pile_bookcorpus2-v1-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"pile_bookcorpus2": {"bits_per_byte": 1.6780040419457868e-06, "byte_perplexity": 1.000001163104447, "word_perplexity": 1.0000066499426599}}, "versions": {"pile_bookcorpus2": 1}}
|
lm-evaluation-harness/tests/testdata/pile_europarl-v1-loglikelihood_rolling
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
e67d3dbccd47d308bfc5b0e66b76d0dfc5e386ebfa94e056562c2281c395543f
|
lm-evaluation-harness/tests/testdata/pile_hackernews-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"pile_hackernews": {"bits_per_byte": 0.00010170276359193358, "byte_perplexity": 1.0001017079354932, "word_perplexity": 1.0006273924348839}}, "versions": {"pile_hackernews": 0}}
|
lm-evaluation-harness/tests/testdata/pile_philpapers-v1-loglikelihood_rolling
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
339ba5d8c044c4a3ff9b9a8eaa24da1d6c01b72972074eb671a7da049eeb7047
|
lm-evaluation-harness/tests/testdata/pile_pile-cc-v1-loglikelihood_rolling
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
731fdef4a43949b179ba0c540148ebc2fa41583dd583ef580dd812076c66a451
|
lm-evaluation-harness/tests/testdata/pile_youtubesubtitles-v1-loglikelihood_rolling
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
68263c52adc0086011e2220b619983935cabb1cc1f5f9f8ee1a74ab2a7457967
|
lm-evaluation-harness/tests/testdata/qa4mre_2011-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"qa4mre_2011": {"acc": 0.225, "acc_norm": 0.23333333333333334, "acc_norm_stderr": 0.03877199986918664, "acc_stderr": 0.0382797091741014}}, "versions": {"qa4mre_2011": 0}}
|
lm-evaluation-harness/tests/testdata/sst-v0-loglikelihood
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
d2ebe3a63517d1d481aa1513bebe124c57a0904554a1e95f566979cfe67b1a7f
|
lm-evaluation-harness/tests/testdata/wmt16-en-ro-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"wmt16-en-ro": {"bleu": 0.0, "bleu_stderr": 0.0, "chrf": 0.012004814364156886, "chrf_stderr": 6.424423961332661e-05, "ter": 1.0, "ter_stderr": 0.0}}, "versions": {"wmt16-en-ro": 0}}
|
lm-evaluation-harness/tests/testdata/wmt20-cs-en-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"wmt20-cs-en": {"bleu": 0.0, "bleu_stderr": 0.0, "chrf": 0.006212086270964023, "chrf_stderr": 0.0001119165191795531, "ter": 1.0, "ter_stderr": 0.0}}, "versions": {"wmt20-cs-en": 0}}
|
lm-evaluation-harness/tests/testdata/wmt20-en-de-v0-greedy_until
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
b6e9c305766ea23ce1027309f83c6d4c2ce8948d70b63a7858586ca34050d7fb
|
lm-evaluation-harness/tests/testdata/wmt20-en-iu-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"wmt20-en-iu": {"bleu": 0.0, "bleu_stderr": 0.0, "chrf": 0.00011803644548940443, "chrf_stderr": 2.175287038623409e-05, "ter": 1.0, "ter_stderr": 0.0}}, "versions": {"wmt20-en-iu": 0}}
|
lm-evaluation-harness/tests/testdata/wmt20-en-pl-v0-greedy_until
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
952f02575d4936d93c4d2808d86c4bf5f1f3a0901212acee6cbc1f9cbd30d39e
|
lm-evaluation-harness/tests/testdata/wmt20-en-ps-v0-greedy_until
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
8411c2cb73114cbd0c6e0f17eab2625d486cc3a601105deb0ea1338a401df689
|
lm-evaluation-harness/tests/testdata/wmt20-en-zh-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"wmt20-en-zh": {"bleu": 0.0, "bleu_stderr": 0.0, "chrf": 0.00014170297316825535, "chrf_stderr": 6.590669847391838e-05, "ter": 1.0, "ter_stderr": 0.0}}, "versions": {"wmt20-en-zh": 0}}
|
lm-evaluation-harness/tests/testdata/wmt20-en-zh-v1-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"wmt20-en-zh": {"bleu": 0.0, "bleu_stderr": 0.0, "chrf": 0.00014170297316825535, "chrf_stderr": 6.590669847391838e-05, "ter": 1.0, "ter_stderr": 0.0}}, "versions": {"wmt20-en-zh": 1}}
|
lm-evaluation-harness/tests/testdata/wmt20-fr-de-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"wmt20-fr-de": {"bleu": 0.0, "bleu_stderr": 0.0, "chrf": 0.01143193767396364, "chrf_stderr": 0.00012555271954563658, "ter": 1.0, "ter_stderr": 0.0}}, "versions": {"wmt20-fr-de": 0}}
|
lm-evaluation-harness/tests/testdata/wmt20-pl-en-v0-res.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"results": {"wmt20-pl-en": {"bleu": 0.0, "bleu_stderr": 0.0, "chrf": 0.01353367757716276, "chrf_stderr": 0.00018386199249976465, "ter": 1.0, "ter_stderr": 0.0}}, "versions": {"wmt20-pl-en": 0}}
|