diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/abstract_narrative_understanding.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/abstract_narrative_understanding.yaml new file mode 100644 index 0000000000000000000000000000000000000000..34cefc2543a16a02883ad493eb9d44634c186ea6 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/abstract_narrative_understanding.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: abstract_narrative_understanding_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_abstract_narrative_understanding_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/analogical_similarity.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/analogical_similarity.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6e20092e9dce7594545786eb54ed587813158ba4 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/analogical_similarity.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: analogical_similarity_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_analogical_similarity_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/auto_categorization.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/auto_categorization.yaml new file mode 100644 index 0000000000000000000000000000000000000000..16e62e69ba2f183473fea70c68513a87534797e6 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/auto_categorization.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: auto_categorization_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_auto_categorization_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/bbq_lite_json.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/bbq_lite_json.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3c4be304435bac358b7ddb732f60605ab029ed82 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/bbq_lite_json.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: bbq_lite_json_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_bbq_lite_json_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/bridging_anaphora_resolution_barqa.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/bridging_anaphora_resolution_barqa.yaml new file mode 100644 index 0000000000000000000000000000000000000000..73448ad929c76ab0e3b59cdd244f0cb429f9b92b --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/bridging_anaphora_resolution_barqa.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: bridging_anaphora_resolution_barqa_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_bridging_anaphora_resolution_barqa_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/causal_judgment.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/causal_judgment.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1d09f2d463394f3300e533a3013f021153195a09 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/causal_judgment.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: causal_judgment_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_causal_judgment_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/cause_and_effect.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/cause_and_effect.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c39ec2780916b80be6e6841c4d8fa8babf916d15 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/cause_and_effect.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: cause_and_effect_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_cause_and_effect_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/checkmate_in_one.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/checkmate_in_one.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0a9883d0eb304b2cf1425e3b04a657e4fb7b0903 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/checkmate_in_one.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: checkmate_in_one_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_checkmate_in_one_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/chess_state_tracking.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/chess_state_tracking.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ea29979786c98ae463812876dfc5d7027d14e3dd --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/chess_state_tracking.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: chess_state_tracking_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_chess_state_tracking_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/cifar10_classification.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/cifar10_classification.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f5918e604dd83fe4c748fe0e509374d32ed065ca --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/cifar10_classification.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: cifar10_classification_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_cifar10_classification_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/code_line_description.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/code_line_description.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9360f759ce40760244ee2bf98206b92a72bd9b67 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/code_line_description.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: code_line_description_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_code_line_description_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/codenames.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/codenames.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5655ea1f5ac17b0c8772a5f4a0443e38e94f1bfa --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/codenames.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: codenames_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_codenames_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/common_morpheme.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/common_morpheme.yaml new file mode 100644 index 0000000000000000000000000000000000000000..bf8f3aca16eacc0bb33cf2240576b9dfe6361bed --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/common_morpheme.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: common_morpheme_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_common_morpheme_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/crash_blossom.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/crash_blossom.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4aca69ad45a59b9773c040afb2e0c76327f19e87 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/crash_blossom.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: crash_blossom_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_crash_blossom_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/crass_ai.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/crass_ai.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ac7c1820d4c2ca98c0055c4ed5c4593eb682eb25 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/crass_ai.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: crass_ai_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_crass_ai_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/cryptonite.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/cryptonite.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c5e0519f0fd1e7b0e7210f2b4fc84caaa45c2843 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/cryptonite.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: cryptonite_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_cryptonite_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/dark_humor_detection.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/dark_humor_detection.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3a77ea44766557982463fffeccf0f18c8627c66b --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/dark_humor_detection.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: dark_humor_detection_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_dark_humor_detection_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/emoji_movie.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/emoji_movie.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0604d97d834d15a4db5adc57b4d1240cabacbb33 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/emoji_movie.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: emoji_movie_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_emoji_movie_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/emojis_emotion_prediction.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/emojis_emotion_prediction.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ff648d9c8f4aa1d4a16500ebf591e5b38abd0b6a --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/emojis_emotion_prediction.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: emojis_emotion_prediction_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_emojis_emotion_prediction_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/empirical_judgments.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/empirical_judgments.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c848740b2c4ad2f73fd3c54ba1ec5a48ea0e1d72 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/empirical_judgments.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: empirical_judgments_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_empirical_judgments_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/english_proverbs.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/english_proverbs.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8adc12e96ee2726b36027b0ed604faf502c43f3c --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/english_proverbs.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: english_proverbs_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_english_proverbs_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/english_russian_proverbs.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/english_russian_proverbs.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ed26147aec06003944012d1e8fb1f6d49363fd2d --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/english_russian_proverbs.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: english_russian_proverbs_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_english_russian_proverbs_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/epistemic_reasoning.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/epistemic_reasoning.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2c35581af46c580527af70b12aebe60aa808181d --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/epistemic_reasoning.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: epistemic_reasoning_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_epistemic_reasoning_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/formal_fallacies_syllogisms_negation.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/formal_fallacies_syllogisms_negation.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7ff37fd7b390f252adc11541a7c37e313d2a378b --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/formal_fallacies_syllogisms_negation.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: formal_fallacies_syllogisms_negation_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_formal_fallacies_syllogisms_negation_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/gem.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/gem.yaml new file mode 100644 index 0000000000000000000000000000000000000000..bf81e88006f0ef68f26af3fdad9fc2aa48fd92c0 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/gem.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: gem_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_gem_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/general_knowledge.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/general_knowledge.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8083b8698ece8e253404a17fed5c13b46aaad9b3 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/general_knowledge.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: general_knowledge_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_general_knowledge_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/hhh_alignment.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/hhh_alignment.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c5e4f24aa7f2a3ac109468474b903f62f891b437 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/hhh_alignment.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: hhh_alignment_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_hhh_alignment_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/hindu_knowledge.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/hindu_knowledge.yaml new file mode 100644 index 0000000000000000000000000000000000000000..321f7513755a6b9555e6ef712e723725df698b43 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/hindu_knowledge.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: hindu_knowledge_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_hindu_knowledge_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/hinglish_toxicity.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/hinglish_toxicity.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5dac090fd4cfa1d4bcd739a4ec93998305ad19d1 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/hinglish_toxicity.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: hinglish_toxicity_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_hinglish_toxicity_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/identify_math_theorems.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/identify_math_theorems.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f716129d6d37cc4dac56b81dc083fec1c799f085 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/identify_math_theorems.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: identify_math_theorems_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_identify_math_theorems_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/identify_odd_metaphor.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/identify_odd_metaphor.yaml new file mode 100644 index 0000000000000000000000000000000000000000..93c4c24487c60205c3176c3c960d3fba770fe7f2 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/identify_odd_metaphor.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: identify_odd_metaphor_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_identify_odd_metaphor_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/implicit_relations.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/implicit_relations.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9bb0844203dd229e0f68eea232d2d2c14a1ae733 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/implicit_relations.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: implicit_relations_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_implicit_relations_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/intent_recognition.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/intent_recognition.yaml new file mode 100644 index 0000000000000000000000000000000000000000..720ac92ae42d54aa469b8c999c518cfc068c2d78 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/intent_recognition.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: intent_recognition_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_intent_recognition_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/international_phonetic_alphabet_transliterate.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/international_phonetic_alphabet_transliterate.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c8e866e2cc525d0b716a6747a7272dbed252fd8c --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/international_phonetic_alphabet_transliterate.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: international_phonetic_alphabet_transliterate_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_international_phonetic_alphabet_transliterate_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/irony_identification.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/irony_identification.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a19ff99e55b6c61967b850dc0e356d0d474dc8fb --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/irony_identification.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: irony_identification_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_irony_identification_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/kanji_ascii.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/kanji_ascii.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a90a82860909a072cbc0b9ff431439d5b9488b94 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/kanji_ascii.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: kanji_ascii_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_kanji_ascii_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/kannada.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/kannada.yaml new file mode 100644 index 0000000000000000000000000000000000000000..910cec477c3d4f0201dec825585619937125e7ee --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/kannada.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: kannada_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_kannada_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/language_games.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/language_games.yaml new file mode 100644 index 0000000000000000000000000000000000000000..07e2711b457b0276f06b5489d406169bdf63149f --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/language_games.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: language_games_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_language_games_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/linguistics_puzzles.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/linguistics_puzzles.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e269cd04e915e745b05714aa35b384d0a4305ab3 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/linguistics_puzzles.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: linguistics_puzzles_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_linguistics_puzzles_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/logical_deduction.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/logical_deduction.yaml new file mode 100644 index 0000000000000000000000000000000000000000..592d2afa8b7b78bf25daf47b43188dec090d7406 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/logical_deduction.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: logical_deduction_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_logical_deduction_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/logical_fallacy_detection.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/logical_fallacy_detection.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1c6411afc8f5e50c6b7e88dfe55db027f1da0b49 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/logical_fallacy_detection.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: logical_fallacy_detection_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_logical_fallacy_detection_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/logical_sequence.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/logical_sequence.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6567189438e8418ae49ee05225ad0451bcfaf0a7 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/logical_sequence.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: logical_sequence_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_logical_sequence_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/movie_dialog_same_or_different.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/movie_dialog_same_or_different.yaml new file mode 100644 index 0000000000000000000000000000000000000000..89b93d9d807f8fdc73a4447aaec6b37d0779e69a --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/movie_dialog_same_or_different.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: movie_dialog_same_or_different_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_movie_dialog_same_or_different_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/movie_recommendation.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/movie_recommendation.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7055028ee950f7a56449e88854b7c2971b825d50 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/movie_recommendation.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: movie_recommendation_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_movie_recommendation_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/natural_instructions.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/natural_instructions.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4874dd155bbffe0b3e749583d8d989dd548ea537 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/natural_instructions.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: natural_instructions_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_natural_instructions_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/nonsense_words_grammar.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/nonsense_words_grammar.yaml new file mode 100644 index 0000000000000000000000000000000000000000..52d25bcacd61f9cf13f68f08c95bbaa6d5da7c21 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/nonsense_words_grammar.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: nonsense_words_grammar_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_nonsense_words_grammar_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/novel_concepts.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/novel_concepts.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3fc74aa9ce7e6bddb32868680a723e76fa3575df --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/novel_concepts.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: novel_concepts_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_novel_concepts_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/operators.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/operators.yaml new file mode 100644 index 0000000000000000000000000000000000000000..951db6f99efd524ca65a7c0349f27e2f9ad45b84 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/operators.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: operators_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_operators_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/paragraph_segmentation.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/paragraph_segmentation.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2cfc8283e831283b55d7bab4c3e801fae2232fc5 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/paragraph_segmentation.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: paragraph_segmentation_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_paragraph_segmentation_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/penguins_in_a_table.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/penguins_in_a_table.yaml new file mode 100644 index 0000000000000000000000000000000000000000..de024e2e7f7caa176955e65cf83989991306e5fc --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/penguins_in_a_table.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: penguins_in_a_table_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_penguins_in_a_table_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/periodic_elements.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/periodic_elements.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b7a644f9d7448f6ce49ed3836c3c1ed06e1f33a5 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/periodic_elements.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: periodic_elements_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_periodic_elements_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/polish_sequence_labeling.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/polish_sequence_labeling.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d61345feb5c10ef9ded6ca39bf1f8b9505a40f2b --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/polish_sequence_labeling.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: polish_sequence_labeling_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_polish_sequence_labeling_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/presuppositions_as_nli.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/presuppositions_as_nli.yaml new file mode 100644 index 0000000000000000000000000000000000000000..71a56aa805c70cfb32013c7895ffe5412475c446 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/presuppositions_as_nli.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: presuppositions_as_nli_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_presuppositions_as_nli_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/question_selection.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/question_selection.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3b3dd0d70e0f3cc25bd8be5fbc4b95c9c016c782 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/question_selection.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: question_selection_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_question_selection_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/reasoning_about_colored_objects.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/reasoning_about_colored_objects.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3ab6d5e06243e5701634cd0f23b3366349a7c9fd --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/reasoning_about_colored_objects.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: reasoning_about_colored_objects_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_reasoning_about_colored_objects_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/salient_translation_error_detection.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/salient_translation_error_detection.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d930e7419a4175762e8cacf1f5297cc4424dd0d6 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/salient_translation_error_detection.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: salient_translation_error_detection_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_salient_translation_error_detection_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/similarities_abstraction.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/similarities_abstraction.yaml new file mode 100644 index 0000000000000000000000000000000000000000..82b86d1b47b857cb7a5b1d8b7789ecadddce8bed --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/similarities_abstraction.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: similarities_abstraction_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_similarities_abstraction_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simp_turing_concept.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simp_turing_concept.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7b1849d5e56a1e6dfbfa0939e975af6c86708182 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simp_turing_concept.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: simp_turing_concept_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_simp_turing_concept_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_arithmetic_json.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_arithmetic_json.yaml new file mode 100644 index 0000000000000000000000000000000000000000..cd1b61b9b0fe70c08c83ebf6c791c5d1437cdd08 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_arithmetic_json.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: simple_arithmetic_json_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_simple_arithmetic_json_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_arithmetic_json_multiple_choice.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_arithmetic_json_multiple_choice.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4e63fce945afc5a55f565e0f0e6c4cd5ce4012dc --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_arithmetic_json_multiple_choice.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: simple_arithmetic_json_multiple_choice_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_simple_arithmetic_json_multiple_choice_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_arithmetic_json_subtasks.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_arithmetic_json_subtasks.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8688512bda759ae89b34230ffefa5ff477e69eb7 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_arithmetic_json_subtasks.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: simple_arithmetic_json_subtasks_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_simple_arithmetic_json_subtasks_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_arithmetic_multiple_targets_json.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_arithmetic_multiple_targets_json.yaml new file mode 100644 index 0000000000000000000000000000000000000000..685ec17c1ad672bc07df05bb140c1400043bc2d6 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_arithmetic_multiple_targets_json.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: simple_arithmetic_multiple_targets_json_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_simple_arithmetic_multiple_targets_json_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_ethical_questions.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_ethical_questions.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0983381ba2031eaa858d615cb13b1c3825b6d464 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_ethical_questions.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: simple_ethical_questions_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_simple_ethical_questions_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_text_editing.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_text_editing.yaml new file mode 100644 index 0000000000000000000000000000000000000000..13b67888cd767063dc1c0ceeceeb92c256b54a7d --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/simple_text_editing.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: simple_text_editing_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_simple_text_editing_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/social_support.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/social_support.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1b3bd5936ed64c7d0183484c7eef577be9300c54 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/social_support.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: social_support_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_social_support_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/strange_stories.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/strange_stories.yaml new file mode 100644 index 0000000000000000000000000000000000000000..30877750e5ba4c9a0de36019ec03effdb2ac1791 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/strange_stories.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: strange_stories_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_strange_stories_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/sufficient_information.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/sufficient_information.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f53d677caa8fd15389d25592db83dfafb56c768e --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/sufficient_information.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: sufficient_information_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_sufficient_information_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/swahili_english_proverbs.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/swahili_english_proverbs.yaml new file mode 100644 index 0000000000000000000000000000000000000000..40103274e9781f1d2736bf1f4aabdaa08dc9aa21 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/swahili_english_proverbs.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: swahili_english_proverbs_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_swahili_english_proverbs_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/temporal_sequences.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/temporal_sequences.yaml new file mode 100644 index 0000000000000000000000000000000000000000..abd8834b0f30b30768ab9aec4524c3c4142d0530 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/temporal_sequences.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: temporal_sequences_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_temporal_sequences_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/tense.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/tense.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6a2676f087cfddcc47856d674df6d6dc673e6808 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/tense.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: tense_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_tense_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/tracking_shuffled_objects.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/tracking_shuffled_objects.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f9aa366b7abcc27a0efcbc825068c0fdfdd4c929 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/tracking_shuffled_objects.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: tracking_shuffled_objects_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_tracking_shuffled_objects_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/unit_conversion.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/unit_conversion.yaml new file mode 100644 index 0000000000000000000000000000000000000000..21a67c437b76dd6275e1e0f8d40f77379e82648c --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/unit_conversion.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: unit_conversion_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_unit_conversion_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/unit_interpretation.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/unit_interpretation.yaml new file mode 100644 index 0000000000000000000000000000000000000000..68614cfddfcbb41a317d8277842e5f2e268dabe1 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/unit_interpretation.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: unit_interpretation_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_unit_interpretation_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/vitaminc_fact_verification.yaml b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/vitaminc_fact_verification.yaml new file mode 100644 index 0000000000000000000000000000000000000000..84305bf33bbdcb441efd5efc3d8da8b1ac1bc167 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/multiple_choice/vitaminc_fact_verification.yaml @@ -0,0 +1,4 @@ +# Generated by utils.py +dataset_name: vitaminc_fact_verification_zero_shot +include: ../multiple_choice_template_yaml +task: bigbench_vitaminc_fact_verification_multiple_choice diff --git a/lm-evaluation/lm_eval/tasks/bigbench/push_bigbench_dataset.py b/lm-evaluation/lm_eval/tasks/bigbench/push_bigbench_dataset.py new file mode 100644 index 0000000000000000000000000000000000000000..44577fa5d41fae16cbbc1936d036a8aa0bf148f2 --- /dev/null +++ b/lm-evaluation/lm_eval/tasks/bigbench/push_bigbench_dataset.py @@ -0,0 +1,31 @@ +""" +A utility script that pushes all Bigbench subtasks from their form in the `bigbench` HF dataset +into `{org name}/bigbench`. + +Prior to running, log into HF Hub for the target HF hub org via `huggingface-cli login`. + +Requires the installation of +`pip install "bigbench @ https://storage.googleapis.com/public_research_data/bigbench/bigbench-0.0.1.tar.gz"` +and is included so that the bigbench dependency can be avoided. +""" +import bigbench.api.util as bb_utils +import datasets +from tqdm import tqdm + + +all_task_names = bb_utils.get_all_json_task_names() + +num_shots = [0] + +for shots in num_shots: + for task_name in tqdm(all_task_names): + try: + print(f"Loading '{task_name}' with num_shots={shots}...") + task_ds = datasets.load_dataset("bigbench", name=task_name, num_shots=shots) + + print(f"Pushing '{task_name}' with num_shots={shots}...") + task_ds.push_to_hub("hails/bigbench", task_name + "_zero_shot") + + del task_ds + except Exception as e: + raise e