diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/abstract_narrative_understanding.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/abstract_narrative_understanding.yaml new file mode 100644 index 0000000000000000000000000000000000000000..34cefc2543a16a02883ad493eb9d44634c186ea6 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/abstract_narrative_understanding.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: abstract_narrative_understanding_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_abstract_narrative_understanding_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/auto_categorization.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/auto_categorization.yaml new file mode 100644 index 0000000000000000000000000000000000000000..16e62e69ba2f183473fea70c68513a87534797e6 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/auto_categorization.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: auto_categorization_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_auto_categorization_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/auto_debugging.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/auto_debugging.yaml new file mode 100644 index 0000000000000000000000000000000000000000..72db1d8ee2a6cd04694e4bf6f48937e13bb7a692 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/auto_debugging.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: auto_debugging_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_auto_debugging_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/bridging_anaphora_resolution_barqa.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/bridging_anaphora_resolution_barqa.yaml new file mode 100644 index 0000000000000000000000000000000000000000..73448ad929c76ab0e3b59cdd244f0cb429f9b92b --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/bridging_anaphora_resolution_barqa.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: bridging_anaphora_resolution_barqa_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_bridging_anaphora_resolution_barqa_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/causal_judgment.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/causal_judgment.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1d09f2d463394f3300e533a3013f021153195a09 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/causal_judgment.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: causal_judgment_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_causal_judgment_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/checkmate_in_one.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/checkmate_in_one.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0a9883d0eb304b2cf1425e3b04a657e4fb7b0903 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/checkmate_in_one.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: checkmate_in_one_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_checkmate_in_one_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cifar10_classification.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cifar10_classification.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f5918e604dd83fe4c748fe0e509374d32ed065ca --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cifar10_classification.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: cifar10_classification_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_cifar10_classification_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/color.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/color.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7350013f1bc15f7ea30fbfac6c2ea1e2bdc31beb --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/color.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: color_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_color_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/crass_ai.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/crass_ai.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ac7c1820d4c2ca98c0055c4ed5c4593eb682eb25 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/crass_ai.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: crass_ai_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_crass_ai_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cryobiology_spanish.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cryobiology_spanish.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c187505d302f723db6d4c7be0d6c464cce79047c --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cryobiology_spanish.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: cryobiology_spanish_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_cryobiology_spanish_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cryptonite.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cryptonite.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c5e0519f0fd1e7b0e7210f2b4fc84caaa45c2843 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cryptonite.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: cryptonite_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_cryptonite_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/dark_humor_detection.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/dark_humor_detection.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3a77ea44766557982463fffeccf0f18c8627c66b --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/dark_humor_detection.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: dark_humor_detection_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_dark_humor_detection_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/date_understanding.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/date_understanding.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2851f0bbbbd5265dc7b2cab72a8c8ffb8d85f22f --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/date_understanding.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: date_understanding_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_date_understanding_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/discourse_marker_prediction.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/discourse_marker_prediction.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5a18733fb7b5698c2649a57ad883cd3e1436130d --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/discourse_marker_prediction.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: discourse_marker_prediction_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_discourse_marker_prediction_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/dyck_languages.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/dyck_languages.yaml new file mode 100644 index 0000000000000000000000000000000000000000..48d6f32e4504687fd22d6715d935eb404d279a4d --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/dyck_languages.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: dyck_languages_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_dyck_languages_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/english_russian_proverbs.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/english_russian_proverbs.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ed26147aec06003944012d1e8fb1f6d49363fd2d --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/english_russian_proverbs.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: english_russian_proverbs_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_english_russian_proverbs_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/evaluating_information_essentiality.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/evaluating_information_essentiality.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b85acd95aedcfc2e197984c7bf9901b28c975f14 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/evaluating_information_essentiality.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: evaluating_information_essentiality_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_evaluating_information_essentiality_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/fact_checker.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/fact_checker.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4fbed8039d6a6a5442ea23afb20cec36754a88e9 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/fact_checker.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: fact_checker_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_fact_checker_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/fantasy_reasoning.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/fantasy_reasoning.yaml new file mode 100644 index 0000000000000000000000000000000000000000..68a55e473930188eec62273d46823c23677d0b51 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/fantasy_reasoning.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: fantasy_reasoning_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_fantasy_reasoning_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/few_shot_nlg.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/few_shot_nlg.yaml new file mode 100644 index 0000000000000000000000000000000000000000..39fcd9cf49e2e42a54e9cb3894607c6f74f20482 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/few_shot_nlg.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: few_shot_nlg_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_few_shot_nlg_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/gem.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/gem.yaml new file mode 100644 index 0000000000000000000000000000000000000000..bf81e88006f0ef68f26af3fdad9fc2aa48fd92c0 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/gem.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: gem_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_gem_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/gender_inclusive_sentences_german.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/gender_inclusive_sentences_german.yaml new file mode 100644 index 0000000000000000000000000000000000000000..39eee21af5dfc4d90850e1bfd6e034c784b6cd3d --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/gender_inclusive_sentences_german.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: gender_inclusive_sentences_german_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_gender_inclusive_sentences_german_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/general_knowledge.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/general_knowledge.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8083b8698ece8e253404a17fed5c13b46aaad9b3 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/general_knowledge.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: general_knowledge_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_general_knowledge_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/geometric_shapes.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/geometric_shapes.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7b80acbf1d32fa7f44b092e0fa105a2f180da252 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/geometric_shapes.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: geometric_shapes_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_geometric_shapes_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/goal_step_wikihow.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/goal_step_wikihow.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6413fb033763cdf18cf351afecae5442c680755f --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/goal_step_wikihow.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: goal_step_wikihow_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_goal_step_wikihow_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/gre_reading_comprehension.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/gre_reading_comprehension.yaml new file mode 100644 index 0000000000000000000000000000000000000000..53523c33219911c0a2c82b4b1481fd1f0b4f1f53 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/gre_reading_comprehension.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: gre_reading_comprehension_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_gre_reading_comprehension_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hhh_alignment.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hhh_alignment.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c5e4f24aa7f2a3ac109468474b903f62f891b437 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hhh_alignment.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: hhh_alignment_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_hhh_alignment_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hindi_question_answering.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hindi_question_answering.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ed1ed278627da24ee1307ad5eee87892ee25797f --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hindi_question_answering.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: hindi_question_answering_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_hindi_question_answering_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hindu_knowledge.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hindu_knowledge.yaml new file mode 100644 index 0000000000000000000000000000000000000000..321f7513755a6b9555e6ef712e723725df698b43 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hindu_knowledge.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: hindu_knowledge_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_hindu_knowledge_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hinglish_toxicity.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hinglish_toxicity.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5dac090fd4cfa1d4bcd739a4ec93998305ad19d1 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hinglish_toxicity.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: hinglish_toxicity_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_hinglish_toxicity_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/identify_odd_metaphor.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/identify_odd_metaphor.yaml new file mode 100644 index 0000000000000000000000000000000000000000..93c4c24487c60205c3176c3c960d3fba770fe7f2 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/identify_odd_metaphor.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: identify_odd_metaphor_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_identify_odd_metaphor_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/implicatures.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/implicatures.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9a26fd55cef16b9367bf6ee836ab6432de42b776 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/implicatures.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: implicatures_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_implicatures_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/intersect_geometry.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/intersect_geometry.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6014a175f1c3278ece40a6d1f77d2a944a9f1601 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/intersect_geometry.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: intersect_geometry_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_intersect_geometry_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/kannada.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/kannada.yaml new file mode 100644 index 0000000000000000000000000000000000000000..910cec477c3d4f0201dec825585619937125e7ee --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/kannada.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: kannada_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_kannada_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/known_unknowns.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/known_unknowns.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1c5f629386f4598bf3c8e67507f1a6adb077bd1f --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/known_unknowns.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: known_unknowns_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_known_unknowns_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/language_games.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/language_games.yaml new file mode 100644 index 0000000000000000000000000000000000000000..07e2711b457b0276f06b5489d406169bdf63149f --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/language_games.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: language_games_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_language_games_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/linguistic_mappings.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/linguistic_mappings.yaml new file mode 100644 index 0000000000000000000000000000000000000000..50800d9deb31a5107c865318ba82e5bdb0fd21a8 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/linguistic_mappings.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: linguistic_mappings_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_linguistic_mappings_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/linguistics_puzzles.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/linguistics_puzzles.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e269cd04e915e745b05714aa35b384d0a4305ab3 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/linguistics_puzzles.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: linguistics_puzzles_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_linguistics_puzzles_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/logic_grid_puzzle.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/logic_grid_puzzle.yaml new file mode 100644 index 0000000000000000000000000000000000000000..da6a018fa8dd7d44e6b0a26bb481a65061bb8988 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/logic_grid_puzzle.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: logic_grid_puzzle_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_logic_grid_puzzle_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/logical_fallacy_detection.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/logical_fallacy_detection.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1c6411afc8f5e50c6b7e88dfe55db027f1da0b49 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/logical_fallacy_detection.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: logical_fallacy_detection_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_logical_fallacy_detection_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/matrixshapes.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/matrixshapes.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9facf63967bd8a281e2053203d90a6622c7d82bb --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/matrixshapes.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: matrixshapes_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_matrixshapes_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/metaphor_understanding.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/metaphor_understanding.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6661a54f7f74ac9cbf97d739ee8af018e8b97106 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/metaphor_understanding.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: metaphor_understanding_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_metaphor_understanding_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/minute_mysteries_qa.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/minute_mysteries_qa.yaml new file mode 100644 index 0000000000000000000000000000000000000000..67109c8cbb941013dd106d486f57c7caa0a2cff4 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/minute_mysteries_qa.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: minute_mysteries_qa_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_minute_mysteries_qa_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/misconceptions.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/misconceptions.yaml new file mode 100644 index 0000000000000000000000000000000000000000..63d0fcda69e3109695b0a250b00b214f822e1568 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/misconceptions.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: misconceptions_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_misconceptions_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/misconceptions_russian.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/misconceptions_russian.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f9c5db38f81186c27a588c70b8856c8881be6310 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/misconceptions_russian.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: misconceptions_russian_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_misconceptions_russian_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/modified_arithmetic.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/modified_arithmetic.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c8a2373588920c97d03c0dafd829a73174433161 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/modified_arithmetic.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: modified_arithmetic_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_modified_arithmetic_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/moral_permissibility.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/moral_permissibility.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3829555221c2cd7ab0359d4f4074c4c00da57adc --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/moral_permissibility.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: moral_permissibility_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_moral_permissibility_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/movie_dialog_same_or_different.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/movie_dialog_same_or_different.yaml new file mode 100644 index 0000000000000000000000000000000000000000..89b93d9d807f8fdc73a4447aaec6b37d0779e69a --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/movie_dialog_same_or_different.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: movie_dialog_same_or_different_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_movie_dialog_same_or_different_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/mult_data_wrangling.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/mult_data_wrangling.yaml new file mode 100644 index 0000000000000000000000000000000000000000..17b67bcc6d6950a8296d0dddb50ed3b9d383e231 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/mult_data_wrangling.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: mult_data_wrangling_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_mult_data_wrangling_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/multiemo.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/multiemo.yaml new file mode 100644 index 0000000000000000000000000000000000000000..10ff48ea585e99e2bc45b5632e117788b9ad5be5 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/multiemo.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: multiemo_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_multiemo_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/navigate.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/navigate.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e69f27904bcf22b2d64c1b22040f860a9855fd06 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/navigate.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: navigate_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_navigate_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/nonsense_words_grammar.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/nonsense_words_grammar.yaml new file mode 100644 index 0000000000000000000000000000000000000000..52d25bcacd61f9cf13f68f08c95bbaa6d5da7c21 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/nonsense_words_grammar.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: nonsense_words_grammar_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_nonsense_words_grammar_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/novel_concepts.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/novel_concepts.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3fc74aa9ce7e6bddb32868680a723e76fa3575df --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/novel_concepts.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: novel_concepts_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_novel_concepts_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/paragraph_segmentation.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/paragraph_segmentation.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2cfc8283e831283b55d7bab4c3e801fae2232fc5 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/paragraph_segmentation.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: paragraph_segmentation_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_paragraph_segmentation_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/parsinlu_qa.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/parsinlu_qa.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7a9b61fb16af05b6dd34025fe6fbac184839cb61 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/parsinlu_qa.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: parsinlu_qa_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_parsinlu_qa_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/parsinlu_reading_comprehension.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/parsinlu_reading_comprehension.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5fa0eccce97772dbd70dd7c5a77ae1a3fe466c35 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/parsinlu_reading_comprehension.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: parsinlu_reading_comprehension_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_parsinlu_reading_comprehension_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/periodic_elements.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/periodic_elements.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b7a644f9d7448f6ce49ed3836c3c1ed06e1f33a5 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/periodic_elements.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: periodic_elements_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_periodic_elements_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/persian_idioms.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/persian_idioms.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6fa92ed3a8ec59135f3ffa8c245efbef327bb1b1 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/persian_idioms.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: persian_idioms_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_persian_idioms_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/physical_intuition.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/physical_intuition.yaml new file mode 100644 index 0000000000000000000000000000000000000000..089376dd8ee05da574c844f37ad9c1c7a23cd162 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/physical_intuition.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: physical_intuition_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_physical_intuition_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/physics.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/physics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..bc06f79dff07e43a966c3c776b5d421b2f826f90 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/physics.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: physics_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_physics_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/play_dialog_same_or_different.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/play_dialog_same_or_different.yaml new file mode 100644 index 0000000000000000000000000000000000000000..85aac7f4b66045c7e0d1ea68710c4bdb06c1a1fb --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/play_dialog_same_or_different.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: play_dialog_same_or_different_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_play_dialog_same_or_different_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/question_selection.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/question_selection.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3b3dd0d70e0f3cc25bd8be5fbc4b95c9c016c782 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/question_selection.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: question_selection_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_question_selection_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/real_or_fake_text.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/real_or_fake_text.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8138791fffc5ff4f8a20d28113459c54d44b4385 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/real_or_fake_text.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: real_or_fake_text_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_real_or_fake_text_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/repeat_copy_logic.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/repeat_copy_logic.yaml new file mode 100644 index 0000000000000000000000000000000000000000..666aa49b060cd26a95d919cc4e920006e38ee6ca --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/repeat_copy_logic.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: repeat_copy_logic_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_repeat_copy_logic_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/rephrase.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/rephrase.yaml new file mode 100644 index 0000000000000000000000000000000000000000..49e3cb4b8dfc5eba4e84c94b4776c4e1f682ddd4 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/rephrase.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: rephrase_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_rephrase_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/semantic_parsing_in_context_sparc.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/semantic_parsing_in_context_sparc.yaml new file mode 100644 index 0000000000000000000000000000000000000000..00574b2f53d940acb881e3b80bbb736366199843 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/semantic_parsing_in_context_sparc.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: semantic_parsing_in_context_sparc_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_semantic_parsing_in_context_sparc_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/similarities_abstraction.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/similarities_abstraction.yaml new file mode 100644 index 0000000000000000000000000000000000000000..82b86d1b47b857cb7a5b1d8b7789ecadddce8bed --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/similarities_abstraction.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: similarities_abstraction_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_similarities_abstraction_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/simp_turing_concept.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/simp_turing_concept.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7b1849d5e56a1e6dfbfa0939e975af6c86708182 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/simp_turing_concept.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: simp_turing_concept_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_simp_turing_concept_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/simple_ethical_questions.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/simple_ethical_questions.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0983381ba2031eaa858d615cb13b1c3825b6d464 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/simple_ethical_questions.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: simple_ethical_questions_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_simple_ethical_questions_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/social_support.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/social_support.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1b3bd5936ed64c7d0183484c7eef577be9300c54 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/social_support.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: social_support_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_social_support_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/strange_stories.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/strange_stories.yaml new file mode 100644 index 0000000000000000000000000000000000000000..30877750e5ba4c9a0de36019ec03effdb2ac1791 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/strange_stories.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: strange_stories_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_strange_stories_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/tense.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/tense.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6a2676f087cfddcc47856d674df6d6dc673e6808 --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/tense.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: tense_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_tense_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/timedial.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/timedial.yaml new file mode 100644 index 0000000000000000000000000000000000000000..350d4e786c1a46ddc057edeffde8695772b5f24a --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/timedial.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: timedial_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_timedial_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/topical_chat.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/topical_chat.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b9a03639a22c4a0e0ace177c4f433bc71b66284b --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/topical_chat.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: topical_chat_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_topical_chat_multiple_choice diff --git a/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/what_is_the_tao.yaml b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/what_is_the_tao.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7879d1661eed9dd083ab3acabc737f756a29735e --- /dev/null +++ b/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/what_is_the_tao.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: what_is_the_tao_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_what_is_the_tao_multiple_choice