diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/analogical_similarity.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/analogical_similarity.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5cc6550a6075a991bce4826c95188e0c7b3d2a94 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/analogical_similarity.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: analogical_similarity_zero_shot +include: ../generate_until_template_yaml +task: bigbench_analogical_similarity_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/analytic_entailment.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/analytic_entailment.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4ae5cfe90f02a8154c49c23ff2aad2cbb40cbbc1 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/analytic_entailment.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: analytic_entailment_zero_shot +include: ../generate_until_template_yaml +task: bigbench_analytic_entailment_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/ascii_word_recognition.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/ascii_word_recognition.yaml new file mode 100644 index 0000000000000000000000000000000000000000..60eaa0be986950cc508431170accc8a9ae644c36 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/ascii_word_recognition.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: ascii_word_recognition_zero_shot +include: ../generate_until_template_yaml +task: bigbench_ascii_word_recognition_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/auto_categorization.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/auto_categorization.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d90a0e7cc31f1c7a04f7b509a26513d6bdb22c00 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/auto_categorization.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: auto_categorization_zero_shot +include: ../generate_until_template_yaml +task: bigbench_auto_categorization_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cause_and_effect.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cause_and_effect.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c34bfdc26ecc1dc3f2f8e023e13eefc85d3fad71 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cause_and_effect.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: cause_and_effect_zero_shot +include: ../generate_until_template_yaml +task: bigbench_cause_and_effect_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/checkmate_in_one.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/checkmate_in_one.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e0736f96ba0ca4bb0cd042ef325132b81a06f3d5 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/checkmate_in_one.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: checkmate_in_one_zero_shot +include: ../generate_until_template_yaml +task: bigbench_checkmate_in_one_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/chess_state_tracking.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/chess_state_tracking.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8b3dde85706c6b50ca3c597443efb6686037fe8b --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/chess_state_tracking.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: chess_state_tracking_zero_shot +include: ../generate_until_template_yaml +task: bigbench_chess_state_tracking_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/chinese_remainder_theorem.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/chinese_remainder_theorem.yaml new file mode 100644 index 0000000000000000000000000000000000000000..872e809b8637380fd3eafa0bb4a5a57e7ce6335c --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/chinese_remainder_theorem.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: chinese_remainder_theorem_zero_shot +include: ../generate_until_template_yaml +task: bigbench_chinese_remainder_theorem_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/codenames.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/codenames.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e71510b4ba4215c91aca96d4a2c2d7fb676498e6 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/codenames.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: codenames_zero_shot +include: ../generate_until_template_yaml +task: bigbench_codenames_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/color.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/color.yaml new file mode 100644 index 0000000000000000000000000000000000000000..18793a9977a0d84bf32470e1f5ba0493549e31fd --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/color.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: color_zero_shot +include: ../generate_until_template_yaml +task: bigbench_color_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/common_morpheme.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/common_morpheme.yaml new file mode 100644 index 0000000000000000000000000000000000000000..09a8b9f407385400214d48478a6e2cf9b24a70cc --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/common_morpheme.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: common_morpheme_zero_shot +include: ../generate_until_template_yaml +task: bigbench_common_morpheme_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conceptual_combinations.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conceptual_combinations.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b36c1d5c2a2ac9a6d6a0b633c2777135122610b0 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conceptual_combinations.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: conceptual_combinations_zero_shot +include: ../generate_until_template_yaml +task: bigbench_conceptual_combinations_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conlang_translation.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conlang_translation.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ec9cccc8c72e887e047a5871c496d68498f7f576 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conlang_translation.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: conlang_translation_zero_shot +include: ../generate_until_template_yaml +task: bigbench_conlang_translation_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/contextual_parametric_knowledge_conflicts.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/contextual_parametric_knowledge_conflicts.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e4da8946fd98ef021df67902ba5dc4857f34a227 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/contextual_parametric_knowledge_conflicts.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: contextual_parametric_knowledge_conflicts_zero_shot +include: ../generate_until_template_yaml +task: bigbench_contextual_parametric_knowledge_conflicts_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/crash_blossom.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/crash_blossom.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3b551e5d8aa4e8963fbcb6f6476c76c0db64b609 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/crash_blossom.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: crash_blossom_zero_shot +include: ../generate_until_template_yaml +task: bigbench_crash_blossom_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cryptonite.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cryptonite.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3393c36805d6b29cd3d59481b11c8b8dd45e2910 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cryptonite.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: cryptonite_zero_shot +include: ../generate_until_template_yaml +task: bigbench_cryptonite_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cs_algorithms.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cs_algorithms.yaml new file mode 100644 index 0000000000000000000000000000000000000000..938fc4aff312eabeda39e95f46eaa787f9526ef2 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cs_algorithms.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: cs_algorithms_zero_shot +include: ../generate_until_template_yaml +task: bigbench_cs_algorithms_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/date_understanding.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/date_understanding.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0fdca6abd643776f45e4bd7163fd0fbe01f6087f --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/date_understanding.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: date_understanding_zero_shot +include: ../generate_until_template_yaml +task: bigbench_date_understanding_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/disfl_qa.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/disfl_qa.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4c6b9567bef7165ab725f1286ea33b2c62c0fc48 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/disfl_qa.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: disfl_qa_zero_shot +include: ../generate_until_template_yaml +task: bigbench_disfl_qa_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/elementary_math_qa.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/elementary_math_qa.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9fe807bc645a88d7f2e87da1d094a2ec1bb51805 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/elementary_math_qa.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: elementary_math_qa_zero_shot +include: ../generate_until_template_yaml +task: bigbench_elementary_math_qa_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/english_proverbs.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/english_proverbs.yaml new file mode 100644 index 0000000000000000000000000000000000000000..cdd014d9c64b37666cc54c9b7097941fcb2a54a2 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/english_proverbs.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: english_proverbs_zero_shot +include: ../generate_until_template_yaml +task: bigbench_english_proverbs_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/english_russian_proverbs.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/english_russian_proverbs.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4e6da1e0ce03973656fdceb8854cf2b6adbeeedf --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/english_russian_proverbs.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: english_russian_proverbs_zero_shot +include: ../generate_until_template_yaml +task: bigbench_english_russian_proverbs_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/entailed_polarity.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/entailed_polarity.yaml new file mode 100644 index 0000000000000000000000000000000000000000..cb2ecba07ebf5bd97f7482e1adb535e064f8a146 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/entailed_polarity.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: entailed_polarity_zero_shot +include: ../generate_until_template_yaml +task: bigbench_entailed_polarity_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/epistemic_reasoning.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/epistemic_reasoning.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f080bcf3988c2dcbcee08bae53025f6ce18ece13 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/epistemic_reasoning.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: epistemic_reasoning_zero_shot +include: ../generate_until_template_yaml +task: bigbench_epistemic_reasoning_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/fantasy_reasoning.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/fantasy_reasoning.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b94f4c05b924d9ca001addc50ba76a03fc3a32f7 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/fantasy_reasoning.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: fantasy_reasoning_zero_shot +include: ../generate_until_template_yaml +task: bigbench_fantasy_reasoning_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/few_shot_nlg.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/few_shot_nlg.yaml new file mode 100644 index 0000000000000000000000000000000000000000..718837f1c086b955d97d5ab0661dc350d482ae20 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/few_shot_nlg.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: few_shot_nlg_zero_shot +include: ../generate_until_template_yaml +task: bigbench_few_shot_nlg_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/figure_of_speech_detection.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/figure_of_speech_detection.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ffbb5f60f4fac85de3847fb7c26c0b5b98403409 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/figure_of_speech_detection.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: figure_of_speech_detection_zero_shot +include: ../generate_until_template_yaml +task: bigbench_figure_of_speech_detection_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/gender_inclusive_sentences_german.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/gender_inclusive_sentences_german.yaml new file mode 100644 index 0000000000000000000000000000000000000000..12dd01b8b299a1fd703c8853653eea979543b0a5 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/gender_inclusive_sentences_german.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: gender_inclusive_sentences_german_zero_shot +include: ../generate_until_template_yaml +task: bigbench_gender_inclusive_sentences_german_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/general_knowledge.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/general_knowledge.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1c0a2ea65470661e5e8822ac7b46e89d01bdebca --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/general_knowledge.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: general_knowledge_zero_shot +include: ../generate_until_template_yaml +task: bigbench_general_knowledge_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/gre_reading_comprehension.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/gre_reading_comprehension.yaml new file mode 100644 index 0000000000000000000000000000000000000000..449b09c47ed4638e2773772b0ce27264cd694be0 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/gre_reading_comprehension.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: gre_reading_comprehension_zero_shot +include: ../generate_until_template_yaml +task: bigbench_gre_reading_comprehension_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hhh_alignment.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hhh_alignment.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c5c437a4ad0322775013c80ff48cd1d875eb2cff --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hhh_alignment.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: hhh_alignment_zero_shot +include: ../generate_until_template_yaml +task: bigbench_hhh_alignment_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hindu_knowledge.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hindu_knowledge.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7fef48a443c5256290c90650834832ebf2008000 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hindu_knowledge.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: hindu_knowledge_zero_shot +include: ../generate_until_template_yaml +task: bigbench_hindu_knowledge_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hinglish_toxicity.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hinglish_toxicity.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7ad63dda3e7cd433a29e34282ceaec71f188fa76 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hinglish_toxicity.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: hinglish_toxicity_zero_shot +include: ../generate_until_template_yaml +task: bigbench_hinglish_toxicity_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/human_organs_senses.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/human_organs_senses.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2334fd6dc7d0a02751be1672d5f21eed837cb07b --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/human_organs_senses.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: human_organs_senses_zero_shot +include: ../generate_until_template_yaml +task: bigbench_human_organs_senses_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/identify_math_theorems.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/identify_math_theorems.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4d0028e03dcb1af695e98fdb619c7d6d101e290c --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/identify_math_theorems.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: identify_math_theorems_zero_shot +include: ../generate_until_template_yaml +task: bigbench_identify_math_theorems_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/international_phonetic_alphabet_nli.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/international_phonetic_alphabet_nli.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1497c7802888d83da4c99cb1c0845e15da887584 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/international_phonetic_alphabet_nli.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: international_phonetic_alphabet_nli_zero_shot +include: ../generate_until_template_yaml +task: bigbench_international_phonetic_alphabet_nli_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kanji_ascii.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kanji_ascii.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f9a8a5b86f69a9966116c203a114d2d0ca5428e7 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kanji_ascii.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: kanji_ascii_zero_shot +include: ../generate_until_template_yaml +task: bigbench_kanji_ascii_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kannada.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kannada.yaml new file mode 100644 index 0000000000000000000000000000000000000000..047e7049b4a5adb0f4a16d31f0018ece6be0e72e --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kannada.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: kannada_zero_shot +include: ../generate_until_template_yaml +task: bigbench_kannada_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/key_value_maps.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/key_value_maps.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3ea697d1f7664866050ecbd0615ea3e957a13602 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/key_value_maps.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: key_value_maps_zero_shot +include: ../generate_until_template_yaml +task: bigbench_key_value_maps_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/language_games.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/language_games.yaml new file mode 100644 index 0000000000000000000000000000000000000000..560223007d7670499ec5064dddf200c0a252fc89 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/language_games.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: language_games_zero_shot +include: ../generate_until_template_yaml +task: bigbench_language_games_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/linguistic_mappings.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/linguistic_mappings.yaml new file mode 100644 index 0000000000000000000000000000000000000000..cc351ce11290861bdf9d9ce71fb46ee832282265 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/linguistic_mappings.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: linguistic_mappings_zero_shot +include: ../generate_until_template_yaml +task: bigbench_linguistic_mappings_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/list_functions.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/list_functions.yaml new file mode 100644 index 0000000000000000000000000000000000000000..658630ac7a0ba0e0dfbc7c86e08a518866e6746c --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/list_functions.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: list_functions_zero_shot +include: ../generate_until_template_yaml +task: bigbench_list_functions_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_args.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_args.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e85c142962ef552e5727de69763c01c912ac5716 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_args.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: logical_args_zero_shot +include: ../generate_until_template_yaml +task: bigbench_logical_args_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_fallacy_detection.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_fallacy_detection.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a74d11ea422980037b47c95d8f7aad02f7a76f5e --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_fallacy_detection.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: logical_fallacy_detection_zero_shot +include: ../generate_until_template_yaml +task: bigbench_logical_fallacy_detection_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_sequence.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_sequence.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b55c057b1e67b4d5af232a9f9710dbbd56f10899 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_sequence.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: logical_sequence_zero_shot +include: ../generate_until_template_yaml +task: bigbench_logical_sequence_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/matrixshapes.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/matrixshapes.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1a162eae1b4226ba93f7dce1f0d8c46800512f9e --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/matrixshapes.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: matrixshapes_zero_shot +include: ../generate_until_template_yaml +task: bigbench_matrixshapes_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/minute_mysteries_qa.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/minute_mysteries_qa.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d453fd941b840482073260cb55a095d4534baeeb --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/minute_mysteries_qa.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: minute_mysteries_qa_zero_shot +include: ../generate_until_template_yaml +task: bigbench_minute_mysteries_qa_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/mnist_ascii.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/mnist_ascii.yaml new file mode 100644 index 0000000000000000000000000000000000000000..db7ce738e76e2de4b5af98a034e517f48ed493e7 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/mnist_ascii.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: mnist_ascii_zero_shot +include: ../generate_until_template_yaml +task: bigbench_mnist_ascii_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/multiemo.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/multiemo.yaml new file mode 100644 index 0000000000000000000000000000000000000000..465ccd0ce4f15270edcc4a4e2585764ee59d4e71 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/multiemo.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: multiemo_zero_shot +include: ../generate_until_template_yaml +task: bigbench_multiemo_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/natural_instructions.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/natural_instructions.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9b77c895577fa3894b4f6646702c7e237436864b --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/natural_instructions.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: natural_instructions_zero_shot +include: ../generate_until_template_yaml +task: bigbench_natural_instructions_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/navigate.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/navigate.yaml new file mode 100644 index 0000000000000000000000000000000000000000..549ed37058fb3c2a9db7eb9d0d6e6ba4c2868983 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/navigate.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: navigate_zero_shot +include: ../generate_until_template_yaml +task: bigbench_navigate_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/nonsense_words_grammar.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/nonsense_words_grammar.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0ed30902f6ec63439564b5e021807eb4ae672967 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/nonsense_words_grammar.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: nonsense_words_grammar_zero_shot +include: ../generate_until_template_yaml +task: bigbench_nonsense_words_grammar_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/novel_concepts.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/novel_concepts.yaml new file mode 100644 index 0000000000000000000000000000000000000000..12f388f8ef8164c30c0843d0a0cda59bc108d66d --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/novel_concepts.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: novel_concepts_zero_shot +include: ../generate_until_template_yaml +task: bigbench_novel_concepts_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/odd_one_out.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/odd_one_out.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a58d7b5fb25068bb2149f4112355106f91fe263a --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/odd_one_out.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: odd_one_out_zero_shot +include: ../generate_until_template_yaml +task: bigbench_odd_one_out_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/operators.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/operators.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d6aaa8b61799f665645249c19d833593576709c6 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/operators.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: operators_zero_shot +include: ../generate_until_template_yaml +task: bigbench_operators_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/phrase_relatedness.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/phrase_relatedness.yaml new file mode 100644 index 0000000000000000000000000000000000000000..037da053e4e5ed3869f75976ebca9ae81d394314 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/phrase_relatedness.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: phrase_relatedness_zero_shot +include: ../generate_until_template_yaml +task: bigbench_phrase_relatedness_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physics.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..39bc786bae05862d66b4f358313feee70ee8d14a --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physics.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: physics_zero_shot +include: ../generate_until_template_yaml +task: bigbench_physics_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/play_dialog_same_or_different.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/play_dialog_same_or_different.yaml new file mode 100644 index 0000000000000000000000000000000000000000..57b65cfd3b09ccbf473a6788f28777d05b71112b --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/play_dialog_same_or_different.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: play_dialog_same_or_different_zero_shot +include: ../generate_until_template_yaml +task: bigbench_play_dialog_same_or_different_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/presuppositions_as_nli.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/presuppositions_as_nli.yaml new file mode 100644 index 0000000000000000000000000000000000000000..70da2d747022062c552856c3594c5033b1401562 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/presuppositions_as_nli.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: presuppositions_as_nli_zero_shot +include: ../generate_until_template_yaml +task: bigbench_presuppositions_as_nli_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/question_selection.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/question_selection.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8e2321a8db770ea9e20761f5b7b117cbdeb7b583 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/question_selection.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: question_selection_zero_shot +include: ../generate_until_template_yaml +task: bigbench_question_selection_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/reasoning_about_colored_objects.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/reasoning_about_colored_objects.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0b371d6e37baabaadb7a7e7424a12cd9dd7b81b9 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/reasoning_about_colored_objects.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: reasoning_about_colored_objects_zero_shot +include: ../generate_until_template_yaml +task: bigbench_reasoning_about_colored_objects_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/repeat_copy_logic.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/repeat_copy_logic.yaml new file mode 100644 index 0000000000000000000000000000000000000000..bd8cd4d8563d4be2b92e18fcd48adc13d6c06f9e --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/repeat_copy_logic.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: repeat_copy_logic_zero_shot +include: ../generate_until_template_yaml +task: bigbench_repeat_copy_logic_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/scientific_press_release.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/scientific_press_release.yaml new file mode 100644 index 0000000000000000000000000000000000000000..122f66e7da0ec45e780fbb727809452c6ef64036 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/scientific_press_release.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: scientific_press_release_zero_shot +include: ../generate_until_template_yaml +task: bigbench_scientific_press_release_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/semantic_parsing_spider.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/semantic_parsing_spider.yaml new file mode 100644 index 0000000000000000000000000000000000000000..39307d92fc3d5f78037102153cfd4e9cc0bb4b48 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/semantic_parsing_spider.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: semantic_parsing_spider_zero_shot +include: ../generate_until_template_yaml +task: bigbench_semantic_parsing_spider_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/simp_turing_concept.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/simp_turing_concept.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6eb9cd87e782bdb6aa857d2550c515a2db9382fe --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/simp_turing_concept.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: simp_turing_concept_zero_shot +include: ../generate_until_template_yaml +task: bigbench_simp_turing_concept_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/simple_arithmetic_json.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/simple_arithmetic_json.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3ff5a1b1a8f51346978d03fd34cb6ad780f85f0b --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/simple_arithmetic_json.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: simple_arithmetic_json_zero_shot +include: ../generate_until_template_yaml +task: bigbench_simple_arithmetic_json_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/simple_arithmetic_multiple_targets_json.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/simple_arithmetic_multiple_targets_json.yaml new file mode 100644 index 0000000000000000000000000000000000000000..393ec8843a009267ea2515fe21105b50fed672e2 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/simple_arithmetic_multiple_targets_json.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: simple_arithmetic_multiple_targets_json_zero_shot +include: ../generate_until_template_yaml +task: bigbench_simple_arithmetic_multiple_targets_json_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/social_iqa.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/social_iqa.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4ba7721de1664e92a1f2de1359c44a5a1bf2e23c --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/social_iqa.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: social_iqa_zero_shot +include: ../generate_until_template_yaml +task: bigbench_social_iqa_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/strange_stories.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/strange_stories.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f5405d92e2eea8199985004288270fc1c50bce96 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/strange_stories.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: strange_stories_zero_shot +include: ../generate_until_template_yaml +task: bigbench_strange_stories_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/sufficient_information.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/sufficient_information.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0705a250288610ebd7162a6a730dd1fef58973c3 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/sufficient_information.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: sufficient_information_zero_shot +include: ../generate_until_template_yaml +task: bigbench_sufficient_information_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/timedial.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/timedial.yaml new file mode 100644 index 0000000000000000000000000000000000000000..854d8642b93197453e8e2d5242c8c1aeb30b519f --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/timedial.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: timedial_zero_shot +include: ../generate_until_template_yaml +task: bigbench_timedial_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/understanding_fables.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/understanding_fables.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9972f4034148bd4f8f4b59b122a89a416f3d5c2f --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/understanding_fables.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: understanding_fables_zero_shot +include: ../generate_until_template_yaml +task: bigbench_understanding_fables_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/what_is_the_tao.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/what_is_the_tao.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3a1487ab41c445cda992e30235947c6e8e9f01db --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/what_is_the_tao.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: what_is_the_tao_zero_shot +include: ../generate_until_template_yaml +task: bigbench_what_is_the_tao_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/which_wiki_edit.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/which_wiki_edit.yaml new file mode 100644 index 0000000000000000000000000000000000000000..bc05c377785c652d603e275b6e9df7608eeef5fc --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/which_wiki_edit.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: which_wiki_edit_zero_shot +include: ../generate_until_template_yaml +task: bigbench_which_wiki_edit_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/word_sorting.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/word_sorting.yaml new file mode 100644 index 0000000000000000000000000000000000000000..16be6060b7700a43fb4f1084fd753e72d370b20e --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/word_sorting.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: word_sorting_zero_shot +include: ../generate_until_template_yaml +task: bigbench_word_sorting_generate_until