diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/abstract_narrative_understanding.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/abstract_narrative_understanding.yaml new file mode 100644 index 0000000000000000000000000000000000000000..dce5238b65beb5e1eb7d579f72abac0e91079984 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/abstract_narrative_understanding.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: abstract_narrative_understanding_zero_shot +include: ../generate_until_template_yaml +task: bigbench_abstract_narrative_understanding_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/anachronisms.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/anachronisms.yaml new file mode 100644 index 0000000000000000000000000000000000000000..831361984ab186fb29835595db2853469ee0f7e6 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/anachronisms.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: anachronisms_zero_shot +include: ../generate_until_template_yaml +task: bigbench_anachronisms_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/arithmetic.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/arithmetic.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d6ae791f5f3b7057f4d7927a986ec57bc27cb7cb --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/arithmetic.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: arithmetic_zero_shot +include: ../generate_until_template_yaml +task: bigbench_arithmetic_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/authorship_verification.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/authorship_verification.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3d7510dfc80d4e52db0cc020f5f2abcdf9952795 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/authorship_verification.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: authorship_verification_zero_shot +include: ../generate_until_template_yaml +task: bigbench_authorship_verification_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/bridging_anaphora_resolution_barqa.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/bridging_anaphora_resolution_barqa.yaml new file mode 100644 index 0000000000000000000000000000000000000000..28e7309f9f0e3ef74e662bdf0cd372c165400ee9 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/bridging_anaphora_resolution_barqa.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: bridging_anaphora_resolution_barqa_zero_shot +include: ../generate_until_template_yaml +task: bigbench_bridging_anaphora_resolution_barqa_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/causal_judgment.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/causal_judgment.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1e1656800ad5d19d72508aaa35e68af0b55da624 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/causal_judgment.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: causal_judgment_zero_shot +include: ../generate_until_template_yaml +task: bigbench_causal_judgment_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/code_line_description.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/code_line_description.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4bd83353a5fcebc5abcded346ab4d38f26bbd7ee --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/code_line_description.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: code_line_description_zero_shot +include: ../generate_until_template_yaml +task: bigbench_code_line_description_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cryobiology_spanish.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cryobiology_spanish.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5fc59ee24bb455dff7cb77cfdb73ad11b7f1f572 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cryobiology_spanish.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: cryobiology_spanish_zero_shot +include: ../generate_until_template_yaml +task: bigbench_cryobiology_spanish_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/dark_humor_detection.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/dark_humor_detection.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f13ec2a4a0fc2dd244aefb53cb7e409fdb2bdad1 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/dark_humor_detection.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: dark_humor_detection_zero_shot +include: ../generate_until_template_yaml +task: bigbench_dark_humor_detection_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/disambiguation_qa.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/disambiguation_qa.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b671d715e1fe69c06c20385bc07b493ecc4d4d6f --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/disambiguation_qa.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: disambiguation_qa_zero_shot +include: ../generate_until_template_yaml +task: bigbench_disambiguation_qa_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/dyck_languages.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/dyck_languages.yaml new file mode 100644 index 0000000000000000000000000000000000000000..814a95de6b16fb6ceb57cb9991bdec00bdffabb7 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/dyck_languages.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: dyck_languages_zero_shot +include: ../generate_until_template_yaml +task: bigbench_dyck_languages_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/emojis_emotion_prediction.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/emojis_emotion_prediction.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3eafb81943aec74feb620500ba8281f62249873b --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/emojis_emotion_prediction.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: emojis_emotion_prediction_zero_shot +include: ../generate_until_template_yaml +task: bigbench_emojis_emotion_prediction_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/evaluating_information_essentiality.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/evaluating_information_essentiality.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b640b9430ad8a11758152c63ad0c77497fd16d50 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/evaluating_information_essentiality.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: evaluating_information_essentiality_zero_shot +include: ../generate_until_template_yaml +task: bigbench_evaluating_information_essentiality_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/geometric_shapes.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/geometric_shapes.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d586c3cb372b95a43243c59e6e7abc04f61f6513 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/geometric_shapes.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: geometric_shapes_zero_shot +include: ../generate_until_template_yaml +task: bigbench_geometric_shapes_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/goal_step_wikihow.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/goal_step_wikihow.yaml new file mode 100644 index 0000000000000000000000000000000000000000..22748246128e774650563a8652a94d57b0e5a338 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/goal_step_wikihow.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: goal_step_wikihow_zero_shot +include: ../generate_until_template_yaml +task: bigbench_goal_step_wikihow_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hindi_question_answering.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hindi_question_answering.yaml new file mode 100644 index 0000000000000000000000000000000000000000..463450b0cb275e2ea6391eb5bed44782ad3265da --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hindi_question_answering.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: hindi_question_answering_zero_shot +include: ../generate_until_template_yaml +task: bigbench_hindi_question_answering_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hyperbaton.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hyperbaton.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1e428c2a5304d43efc1b00ff53e8d3de493c115b --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hyperbaton.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: hyperbaton_zero_shot +include: ../generate_until_template_yaml +task: bigbench_hyperbaton_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/identify_odd_metaphor.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/identify_odd_metaphor.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b4e1f9aa86cd9e29ad5109673b767dc33bde1e00 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/identify_odd_metaphor.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: identify_odd_metaphor_zero_shot +include: ../generate_until_template_yaml +task: bigbench_identify_odd_metaphor_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/implicatures.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/implicatures.yaml new file mode 100644 index 0000000000000000000000000000000000000000..cf19c32aad8960cc8427d7269927fd67ae732f14 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/implicatures.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: implicatures_zero_shot +include: ../generate_until_template_yaml +task: bigbench_implicatures_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/implicit_relations.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/implicit_relations.yaml new file mode 100644 index 0000000000000000000000000000000000000000..361f0435ef63a75bd2413100c434841f206ee5f8 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/implicit_relations.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: implicit_relations_zero_shot +include: ../generate_until_template_yaml +task: bigbench_implicit_relations_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/international_phonetic_alphabet_transliterate.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/international_phonetic_alphabet_transliterate.yaml new file mode 100644 index 0000000000000000000000000000000000000000..71ad3b9d4a7f980529e64ce4ebba38a4db026f05 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/international_phonetic_alphabet_transliterate.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: international_phonetic_alphabet_transliterate_zero_shot +include: ../generate_until_template_yaml +task: bigbench_international_phonetic_alphabet_transliterate_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/irony_identification.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/irony_identification.yaml new file mode 100644 index 0000000000000000000000000000000000000000..556c5a62a7e31b56732dd158efca9111fa2b8f60 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/irony_identification.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: irony_identification_zero_shot +include: ../generate_until_template_yaml +task: bigbench_irony_identification_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/known_unknowns.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/known_unknowns.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b1a8bb0640198dd3a1e288e80804a9dadeb2c806 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/known_unknowns.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: known_unknowns_zero_shot +include: ../generate_until_template_yaml +task: bigbench_known_unknowns_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/language_identification.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/language_identification.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9cb7b27408b9a82c308ebac33b89e799df0763a0 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/language_identification.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: language_identification_zero_shot +include: ../generate_until_template_yaml +task: bigbench_language_identification_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logic_grid_puzzle.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logic_grid_puzzle.yaml new file mode 100644 index 0000000000000000000000000000000000000000..aa8f2c2fefbed31d42e61db5261810b49e7ff35e --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logic_grid_puzzle.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: logic_grid_puzzle_zero_shot +include: ../generate_until_template_yaml +task: bigbench_logic_grid_puzzle_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/metaphor_boolean.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/metaphor_boolean.yaml new file mode 100644 index 0000000000000000000000000000000000000000..28922b3f1b498e073db5835c94bf3ee03fa07ebd --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/metaphor_boolean.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: metaphor_boolean_zero_shot +include: ../generate_until_template_yaml +task: bigbench_metaphor_boolean_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/metaphor_understanding.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/metaphor_understanding.yaml new file mode 100644 index 0000000000000000000000000000000000000000..029a4c0a073ccaefc8975ae37937319b27f1e7ee --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/metaphor_understanding.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: metaphor_understanding_zero_shot +include: ../generate_until_template_yaml +task: bigbench_metaphor_understanding_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/misconceptions_russian.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/misconceptions_russian.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a5e5e102ae68e5c472cfb368652064f4f67259fe --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/misconceptions_russian.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: misconceptions_russian_zero_shot +include: ../generate_until_template_yaml +task: bigbench_misconceptions_russian_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/moral_permissibility.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/moral_permissibility.yaml new file mode 100644 index 0000000000000000000000000000000000000000..277bf69feff29559672655e47ce037df3c42c454 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/moral_permissibility.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: moral_permissibility_zero_shot +include: ../generate_until_template_yaml +task: bigbench_moral_permissibility_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/movie_dialog_same_or_different.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/movie_dialog_same_or_different.yaml new file mode 100644 index 0000000000000000000000000000000000000000..27cc6228f092b33652b0adcc5597fe71365128b0 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/movie_dialog_same_or_different.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: movie_dialog_same_or_different_zero_shot +include: ../generate_until_template_yaml +task: bigbench_movie_dialog_same_or_different_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/paragraph_segmentation.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/paragraph_segmentation.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5f982c5db5ccb458e9815708a26493f309ea436a --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/paragraph_segmentation.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: paragraph_segmentation_zero_shot +include: ../generate_until_template_yaml +task: bigbench_paragraph_segmentation_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/parsinlu_qa.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/parsinlu_qa.yaml new file mode 100644 index 0000000000000000000000000000000000000000..552f8c6068fde183ab744a1e322c41c8744070e0 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/parsinlu_qa.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: parsinlu_qa_zero_shot +include: ../generate_until_template_yaml +task: bigbench_parsinlu_qa_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/parsinlu_reading_comprehension.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/parsinlu_reading_comprehension.yaml new file mode 100644 index 0000000000000000000000000000000000000000..358184e11ced80305697c7e5f18317af2161bab9 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/parsinlu_reading_comprehension.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: parsinlu_reading_comprehension_zero_shot +include: ../generate_until_template_yaml +task: bigbench_parsinlu_reading_comprehension_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/penguins_in_a_table.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/penguins_in_a_table.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6dc70030d9ef8ea5671a780bd88a186122a0fb47 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/penguins_in_a_table.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: penguins_in_a_table_zero_shot +include: ../generate_until_template_yaml +task: bigbench_penguins_in_a_table_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/periodic_elements.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/periodic_elements.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c5c96cec606f6ba3e749c970b20f71d9ed200799 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/periodic_elements.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: periodic_elements_zero_shot +include: ../generate_until_template_yaml +task: bigbench_periodic_elements_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physical_intuition.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physical_intuition.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ecef1581c907281e920a08651434a15313f0dc39 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physical_intuition.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: physical_intuition_zero_shot +include: ../generate_until_template_yaml +task: bigbench_physical_intuition_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physics_questions.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physics_questions.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3fcfd47776ba5be480ed396fb98534e3cc7316aa --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physics_questions.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: physics_questions_zero_shot +include: ../generate_until_template_yaml +task: bigbench_physics_questions_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/polish_sequence_labeling.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/polish_sequence_labeling.yaml new file mode 100644 index 0000000000000000000000000000000000000000..23775493c1ffe4e82c833515eef998f767b401db --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/polish_sequence_labeling.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: polish_sequence_labeling_zero_shot +include: ../generate_until_template_yaml +task: bigbench_polish_sequence_labeling_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/qa_wikidata.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/qa_wikidata.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9fb5b23036510e8256774fb0d32964a590ff9dfe --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/qa_wikidata.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: qa_wikidata_zero_shot +include: ../generate_until_template_yaml +task: bigbench_qa_wikidata_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/real_or_fake_text.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/real_or_fake_text.yaml new file mode 100644 index 0000000000000000000000000000000000000000..948bfb0c478b96a8e1285819748f905acfc004b1 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/real_or_fake_text.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: real_or_fake_text_zero_shot +include: ../generate_until_template_yaml +task: bigbench_real_or_fake_text_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/rephrase.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/rephrase.yaml new file mode 100644 index 0000000000000000000000000000000000000000..16a337dbc2a8568cc36245f34b7eccaf28ed2548 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/rephrase.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: rephrase_zero_shot +include: ../generate_until_template_yaml +task: bigbench_rephrase_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/riddle_sense.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/riddle_sense.yaml new file mode 100644 index 0000000000000000000000000000000000000000..745cdb3244845caa9914fae7073b29f64f9773bb --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/riddle_sense.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: riddle_sense_zero_shot +include: ../generate_until_template_yaml +task: bigbench_riddle_sense_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/salient_translation_error_detection.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/salient_translation_error_detection.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4968e441daa4b119bcaf0e5ae5f33d2acfd5a4a6 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/salient_translation_error_detection.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: salient_translation_error_detection_zero_shot +include: ../generate_until_template_yaml +task: bigbench_salient_translation_error_detection_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/sentence_ambiguity.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/sentence_ambiguity.yaml new file mode 100644 index 0000000000000000000000000000000000000000..263b453fac68a15afa2b8d4ac14328fe6e096124 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/sentence_ambiguity.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: sentence_ambiguity_zero_shot +include: ../generate_until_template_yaml +task: bigbench_sentence_ambiguity_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/similarities_abstraction.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/similarities_abstraction.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c33b1c8b1f0be9a26c8c5bc165195828a692d6d5 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/similarities_abstraction.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: similarities_abstraction_zero_shot +include: ../generate_until_template_yaml +task: bigbench_similarities_abstraction_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/simple_arithmetic_json_subtasks.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/simple_arithmetic_json_subtasks.yaml new file mode 100644 index 0000000000000000000000000000000000000000..57052288e7fed1fabbe9a2c572b10c99f9a1fdcd --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/simple_arithmetic_json_subtasks.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: simple_arithmetic_json_subtasks_zero_shot +include: ../generate_until_template_yaml +task: bigbench_simple_arithmetic_json_subtasks_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/simple_text_editing.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/simple_text_editing.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d3310fa2126ea3c2601e4e4e16cdf22df06e8c4f --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/simple_text_editing.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: simple_text_editing_zero_shot +include: ../generate_until_template_yaml +task: bigbench_simple_text_editing_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/social_support.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/social_support.yaml new file mode 100644 index 0000000000000000000000000000000000000000..dc00bb83755f75220a068b9c97047ec02e1eafed --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/social_support.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: social_support_zero_shot +include: ../generate_until_template_yaml +task: bigbench_social_support_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/strategyqa.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/strategyqa.yaml new file mode 100644 index 0000000000000000000000000000000000000000..47c4b25c971fbbf78c5d62ee79de7c0699af2ba9 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/strategyqa.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: strategyqa_zero_shot +include: ../generate_until_template_yaml +task: bigbench_strategyqa_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/suicide_risk.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/suicide_risk.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e276c4a051d1507991e00499f344c72fe42a4147 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/suicide_risk.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: suicide_risk_zero_shot +include: ../generate_until_template_yaml +task: bigbench_suicide_risk_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/swedish_to_german_proverbs.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/swedish_to_german_proverbs.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5a13d6f7fe014a2ab9a55fdb86cff68f8cb3401d --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/swedish_to_german_proverbs.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: swedish_to_german_proverbs_zero_shot +include: ../generate_until_template_yaml +task: bigbench_swedish_to_german_proverbs_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/symbol_interpretation.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/symbol_interpretation.yaml new file mode 100644 index 0000000000000000000000000000000000000000..cca33bf67e954f18336d8becfb39d75c0e37df56 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/symbol_interpretation.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: symbol_interpretation_zero_shot +include: ../generate_until_template_yaml +task: bigbench_symbol_interpretation_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/temporal_sequences.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/temporal_sequences.yaml new file mode 100644 index 0000000000000000000000000000000000000000..414dc51b137fb55037b5b9bc109bba116ee72d34 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/temporal_sequences.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: temporal_sequences_zero_shot +include: ../generate_until_template_yaml +task: bigbench_temporal_sequences_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/topical_chat.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/topical_chat.yaml new file mode 100644 index 0000000000000000000000000000000000000000..47a301cf705d5abc403ddfa122b989bef2e82099 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/topical_chat.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: topical_chat_zero_shot +include: ../generate_until_template_yaml +task: bigbench_topical_chat_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/tracking_shuffled_objects.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/tracking_shuffled_objects.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9c02866c8f07d5d8d9fdfd0459bbd01f327d19b3 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/tracking_shuffled_objects.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: tracking_shuffled_objects_zero_shot +include: ../generate_until_template_yaml +task: bigbench_tracking_shuffled_objects_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/vitaminc_fact_verification.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/vitaminc_fact_verification.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6f2ad8d3fd46a37ffc4fad10c1d927324054e043 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/vitaminc_fact_verification.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: vitaminc_fact_verification_zero_shot +include: ../generate_until_template_yaml +task: bigbench_vitaminc_fact_verification_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/word_unscrambling.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/word_unscrambling.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5632a79c639f23b9635a810176a5ea10343c506f --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/word_unscrambling.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: word_unscrambling_zero_shot +include: ../generate_until_template_yaml +task: bigbench_word_unscrambling_generate_until diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/arithmetic.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/arithmetic.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9b19b92fde9418c5df171864249e414862e673d3 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/arithmetic.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: arithmetic_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_arithmetic_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/chinese_remainder_theorem.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/chinese_remainder_theorem.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c24d5761fd9efcab1e0b4d69e9b9e1474c1f1aaa --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/chinese_remainder_theorem.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: chinese_remainder_theorem_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_chinese_remainder_theorem_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/empirical_judgments.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/empirical_judgments.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c848740b2c4ad2f73fd3c54ba1ec5a48ea0e1d72 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/empirical_judgments.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: empirical_judgments_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_empirical_judgments_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/figure_of_speech_detection.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/figure_of_speech_detection.yaml new file mode 100644 index 0000000000000000000000000000000000000000..68a83956eb58a15aaf9ae8fa705f2883cbf3a9a6 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/figure_of_speech_detection.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: figure_of_speech_detection_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_figure_of_speech_detection_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/formal_fallacies_syllogisms_negation.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/formal_fallacies_syllogisms_negation.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7ff37fd7b390f252adc11541a7c37e313d2a378b --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/formal_fallacies_syllogisms_negation.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: formal_fallacies_syllogisms_negation_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_formal_fallacies_syllogisms_negation_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hyperbaton.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hyperbaton.yaml new file mode 100644 index 0000000000000000000000000000000000000000..34b377101839b652ea0143e8c9ba318c8c9542fd --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hyperbaton.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: hyperbaton_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_hyperbaton_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/international_phonetic_alphabet_nli.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/international_phonetic_alphabet_nli.yaml new file mode 100644 index 0000000000000000000000000000000000000000..89d7742d5eb11f19f80e59a2afaf1401e74e2169 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/international_phonetic_alphabet_nli.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: international_phonetic_alphabet_nli_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_international_phonetic_alphabet_nli_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/international_phonetic_alphabet_transliterate.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/international_phonetic_alphabet_transliterate.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c8e866e2cc525d0b716a6747a7272dbed252fd8c --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/international_phonetic_alphabet_transliterate.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: international_phonetic_alphabet_transliterate_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_international_phonetic_alphabet_transliterate_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/operators.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/operators.yaml new file mode 100644 index 0000000000000000000000000000000000000000..951db6f99efd524ca65a7c0349f27e2f9ad45b84 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/operators.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: operators_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_operators_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/penguins_in_a_table.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/penguins_in_a_table.yaml new file mode 100644 index 0000000000000000000000000000000000000000..de024e2e7f7caa176955e65cf83989991306e5fc --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/penguins_in_a_table.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: penguins_in_a_table_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_penguins_in_a_table_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/presuppositions_as_nli.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/presuppositions_as_nli.yaml new file mode 100644 index 0000000000000000000000000000000000000000..71a56aa805c70cfb32013c7895ffe5412475c446 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/presuppositions_as_nli.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: presuppositions_as_nli_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_presuppositions_as_nli_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/reasoning_about_colored_objects.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/reasoning_about_colored_objects.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3ab6d5e06243e5701634cd0f23b3366349a7c9fd --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/reasoning_about_colored_objects.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: reasoning_about_colored_objects_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_reasoning_about_colored_objects_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/ruin_names.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/ruin_names.yaml new file mode 100644 index 0000000000000000000000000000000000000000..32c38ba3786f863626e4ae23fe12c016ade5b8af --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/ruin_names.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: ruin_names_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_ruin_names_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/scientific_press_release.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/scientific_press_release.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f23190e7acc67fd26699f5b80652e3ec674b49c8 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/scientific_press_release.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: scientific_press_release_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_scientific_press_release_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/sufficient_information.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/sufficient_information.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f53d677caa8fd15389d25592db83dfafb56c768e --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/sufficient_information.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: sufficient_information_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_sufficient_information_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/swahili_english_proverbs.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/swahili_english_proverbs.yaml new file mode 100644 index 0000000000000000000000000000000000000000..40103274e9781f1d2736bf1f4aabdaa08dc9aa21 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/swahili_english_proverbs.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: swahili_english_proverbs_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_swahili_english_proverbs_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/symbol_interpretation.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/symbol_interpretation.yaml new file mode 100644 index 0000000000000000000000000000000000000000..98e3d5b3694354c11446c8efc8583694c33d462d --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/symbol_interpretation.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: symbol_interpretation_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_symbol_interpretation_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/vitaminc_fact_verification.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/vitaminc_fact_verification.yaml new file mode 100644 index 0000000000000000000000000000000000000000..84305bf33bbdcb441efd5efc3d8da8b1ac1bc167 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/vitaminc_fact_verification.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: vitaminc_fact_verification_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_vitaminc_fact_verification_multiple_choice