diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/_default_template_yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/_default_template_yaml new file mode 100644 index 0000000000000000000000000000000000000000..37e8bb16490fce1ca376d6f13027adab3e6061f5 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/_default_template_yaml @@ -0,0 +1,15 @@ +dataset_path: hails/mmlu_no_train # a copy of `cais/mmlu` with no auxiliary_train split +test_split: test +fewshot_split: dev +fewshot_config: + sampler: first_n +output_type: multiple_choice +doc_to_text: "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:" +doc_to_choice: ["A", "B", "C", "D"] +doc_to_target: answer +metric_list: + - metric: acc + aggregation: mean + higher_is_better: true +metadata: + version: 0.0 diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/_mmlu.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/_mmlu.yaml new file mode 100644 index 0000000000000000000000000000000000000000..584de02993acedc37b6d508967334f9a55675f6b --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/_mmlu.yaml @@ -0,0 +1,6 @@ +group: mmlu +task: + - mmlu_stem + - mmlu_other + - mmlu_social_sciences + - mmlu_humanities diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_abstract_algebra.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_abstract_algebra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..90f3cc50b3f8e54b584237edfa50fcdc96d0e625 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_abstract_algebra.yaml @@ -0,0 +1,8 @@ +"dataset_name": "abstract_algebra" +"description": "The following are multiple choice questions (with answers) about abstract\ + \ algebra.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_abstract_algebra" +"task_alias": "abstract_algebra" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_anatomy.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_anatomy.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0e9e09b27fa40e222d6284c58826a089391e1bb4 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_anatomy.yaml @@ -0,0 +1,8 @@ +"dataset_name": "anatomy" +"description": "The following are multiple choice questions (with answers) about anatomy.\n\ + \n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_anatomy" +"task_alias": "anatomy" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_astronomy.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_astronomy.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e3bdfc9521f3b98bf684a1c022f69e3c14c9d406 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_astronomy.yaml @@ -0,0 +1,8 @@ +"dataset_name": "astronomy" +"description": "The following are multiple choice questions (with answers) about astronomy.\n\ + \n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_astronomy" +"task_alias": "astronomy" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_business_ethics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_business_ethics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ea0d1fe2a68cb893759f3c550775ff3554f4fc13 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_business_ethics.yaml @@ -0,0 +1,8 @@ +"dataset_name": "business_ethics" +"description": "The following are multiple choice questions (with answers) about business\ + \ ethics.\n\n" +"group": "mmlu_other" +"group_alias": "other" +"include": "_default_template_yaml" +"task": "mmlu_business_ethics" +"task_alias": "business_ethics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_clinical_knowledge.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_clinical_knowledge.yaml new file mode 100644 index 0000000000000000000000000000000000000000..20bab1470e8b58b9e56652edf8128a4e7959bd33 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_clinical_knowledge.yaml @@ -0,0 +1,8 @@ +"dataset_name": "clinical_knowledge" +"description": "The following are multiple choice questions (with answers) about clinical\ + \ knowledge.\n\n" +"group": "mmlu_other" +"group_alias": "other" +"include": "_default_template_yaml" +"task": "mmlu_clinical_knowledge" +"task_alias": "clinical_knowledge" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_biology.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_biology.yaml new file mode 100644 index 0000000000000000000000000000000000000000..afb4d9c6c0d209644806ca8376a25cf28d6fa2ed --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_biology.yaml @@ -0,0 +1,8 @@ +"dataset_name": "college_biology" +"description": "The following are multiple choice questions (with answers) about college\ + \ biology.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_college_biology" +"task_alias": "college_biology" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_chemistry.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_chemistry.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a7de3532359a418dc0d1ba8f9096700e2a888907 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_chemistry.yaml @@ -0,0 +1,8 @@ +"dataset_name": "college_chemistry" +"description": "The following are multiple choice questions (with answers) about college\ + \ chemistry.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_college_chemistry" +"task_alias": "college_chemistry" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_computer_science.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_computer_science.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9786cc6ebd0d14936bb0138d322161b01a4574dd --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_computer_science.yaml @@ -0,0 +1,8 @@ +"dataset_name": "college_computer_science" +"description": "The following are multiple choice questions (with answers) about college\ + \ computer science.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_college_computer_science" +"task_alias": "college_computer_science" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_mathematics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_mathematics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e7699f8bfcedffdcc83984f96c6d337f41135b54 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_mathematics.yaml @@ -0,0 +1,8 @@ +"dataset_name": "college_mathematics" +"description": "The following are multiple choice questions (with answers) about college\ + \ mathematics.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_college_mathematics" +"task_alias": "college_mathematics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_medicine.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_medicine.yaml new file mode 100644 index 0000000000000000000000000000000000000000..df9e890136328d4caf3ca3c4715919358dcf92dc --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_medicine.yaml @@ -0,0 +1,8 @@ +"dataset_name": "college_medicine" +"description": "The following are multiple choice questions (with answers) about college\ + \ medicine.\n\n" +"group": "mmlu_other" +"group_alias": "other" +"include": "_default_template_yaml" +"task": "mmlu_college_medicine" +"task_alias": "college_medicine" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_physics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_physics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3c5e7462e3fefe122e9df4bb5ff9458f8ae9dfc8 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_college_physics.yaml @@ -0,0 +1,8 @@ +"dataset_name": "college_physics" +"description": "The following are multiple choice questions (with answers) about college\ + \ physics.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_college_physics" +"task_alias": "college_physics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_computer_security.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_computer_security.yaml new file mode 100644 index 0000000000000000000000000000000000000000..df9c4a51b6ca2715b921f2a33ef4755ebfacbc46 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_computer_security.yaml @@ -0,0 +1,8 @@ +"dataset_name": "computer_security" +"description": "The following are multiple choice questions (with answers) about computer\ + \ security.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_computer_security" +"task_alias": "computer_security" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_conceptual_physics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_conceptual_physics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8ab59ed10f1cf02f47eacf7930a9efcbfc230fcf --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_conceptual_physics.yaml @@ -0,0 +1,8 @@ +"dataset_name": "conceptual_physics" +"description": "The following are multiple choice questions (with answers) about conceptual\ + \ physics.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_conceptual_physics" +"task_alias": "conceptual_physics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_econometrics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_econometrics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a974fc8462d649c27ac09192d3cf84d9b162f82d --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_econometrics.yaml @@ -0,0 +1,8 @@ +"dataset_name": "econometrics" +"description": "The following are multiple choice questions (with answers) about econometrics.\n\ + \n" +"group": "mmlu_social_sciences" +"group_alias": "social_sciences" +"include": "_default_template_yaml" +"task": "mmlu_econometrics" +"task_alias": "econometrics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_electrical_engineering.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_electrical_engineering.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9c45cc6109d2b53322b2303efcb1f57e99d8d60c --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_electrical_engineering.yaml @@ -0,0 +1,8 @@ +"dataset_name": "electrical_engineering" +"description": "The following are multiple choice questions (with answers) about electrical\ + \ engineering.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_electrical_engineering" +"task_alias": "electrical_engineering" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_elementary_mathematics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_elementary_mathematics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2154ab65454c9234b57d76782315c6559b0e7e0f --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_elementary_mathematics.yaml @@ -0,0 +1,8 @@ +"dataset_name": "elementary_mathematics" +"description": "The following are multiple choice questions (with answers) about elementary\ + \ mathematics.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_elementary_mathematics" +"task_alias": "elementary_mathematics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_formal_logic.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_formal_logic.yaml new file mode 100644 index 0000000000000000000000000000000000000000..689d3d1f87503953be03cd54973d240b64f846f4 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_formal_logic.yaml @@ -0,0 +1,8 @@ +"dataset_name": "formal_logic" +"description": "The following are multiple choice questions (with answers) about formal\ + \ logic.\n\n" +"group": "mmlu_humanities" +"group_alias": "humanities" +"include": "_default_template_yaml" +"task": "mmlu_formal_logic" +"task_alias": "formal_logic" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_global_facts.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_global_facts.yaml new file mode 100644 index 0000000000000000000000000000000000000000..60b5c12921ee2ad1e2200392ecc3ffaef44f4105 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_global_facts.yaml @@ -0,0 +1,8 @@ +"dataset_name": "global_facts" +"description": "The following are multiple choice questions (with answers) about global\ + \ facts.\n\n" +"group": "mmlu_other" +"group_alias": "other" +"include": "_default_template_yaml" +"task": "mmlu_global_facts" +"task_alias": "global_facts" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_biology.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_biology.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c7e055dcc8dcfffb2572a8f4ab1770479e84daeb --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_biology.yaml @@ -0,0 +1,8 @@ +"dataset_name": "high_school_biology" +"description": "The following are multiple choice questions (with answers) about high\ + \ school biology.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_high_school_biology" +"task_alias": "high_school_biology" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_chemistry.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_chemistry.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8e9421c1a6f7271ddfc87ad099a914f81bc031e9 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_chemistry.yaml @@ -0,0 +1,8 @@ +"dataset_name": "high_school_chemistry" +"description": "The following are multiple choice questions (with answers) about high\ + \ school chemistry.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_high_school_chemistry" +"task_alias": "high_school_chemistry" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_computer_science.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_computer_science.yaml new file mode 100644 index 0000000000000000000000000000000000000000..87ec15cc3ad2bcd2d9d1d802389cb9e3d26e5abf --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_computer_science.yaml @@ -0,0 +1,8 @@ +"dataset_name": "high_school_computer_science" +"description": "The following are multiple choice questions (with answers) about high\ + \ school computer science.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_high_school_computer_science" +"task_alias": "high_school_computer_science" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_european_history.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_european_history.yaml new file mode 100644 index 0000000000000000000000000000000000000000..be0d696a962c38a9ffb5404ea75cb2d402f10ec0 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_european_history.yaml @@ -0,0 +1,8 @@ +"dataset_name": "high_school_european_history" +"description": "The following are multiple choice questions (with answers) about high\ + \ school european history.\n\n" +"group": "mmlu_humanities" +"group_alias": "humanities" +"include": "_default_template_yaml" +"task": "mmlu_high_school_european_history" +"task_alias": "high_school_european_history" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_geography.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_geography.yaml new file mode 100644 index 0000000000000000000000000000000000000000..57c5261a8be0e498111d834b8787341104f24775 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_geography.yaml @@ -0,0 +1,8 @@ +"dataset_name": "high_school_geography" +"description": "The following are multiple choice questions (with answers) about high\ + \ school geography.\n\n" +"group": "mmlu_social_sciences" +"group_alias": "social_sciences" +"include": "_default_template_yaml" +"task": "mmlu_high_school_geography" +"task_alias": "high_school_geography" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_government_and_politics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_government_and_politics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2e92f152081e102b07feb394038aa0b30ca827a2 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_government_and_politics.yaml @@ -0,0 +1,8 @@ +"dataset_name": "high_school_government_and_politics" +"description": "The following are multiple choice questions (with answers) about high\ + \ school government and politics.\n\n" +"group": "mmlu_social_sciences" +"group_alias": "social_sciences" +"include": "_default_template_yaml" +"task": "mmlu_high_school_government_and_politics" +"task_alias": "high_school_government_and_politics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_macroeconomics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_macroeconomics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..988d132aad3309d086e8f8549bd6710f2bd2b817 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_macroeconomics.yaml @@ -0,0 +1,8 @@ +"dataset_name": "high_school_macroeconomics" +"description": "The following are multiple choice questions (with answers) about high\ + \ school macroeconomics.\n\n" +"group": "mmlu_social_sciences" +"group_alias": "social_sciences" +"include": "_default_template_yaml" +"task": "mmlu_high_school_macroeconomics" +"task_alias": "high_school_macroeconomics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_mathematics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_mathematics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f7c07a608592709c08d4e2130e9f99abd5058880 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_mathematics.yaml @@ -0,0 +1,8 @@ +"dataset_name": "high_school_mathematics" +"description": "The following are multiple choice questions (with answers) about high\ + \ school mathematics.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_high_school_mathematics" +"task_alias": "high_school_mathematics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_microeconomics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_microeconomics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5339a0235cd8e27e51c7292a7cd2201c46cbc2f1 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_microeconomics.yaml @@ -0,0 +1,8 @@ +"dataset_name": "high_school_microeconomics" +"description": "The following are multiple choice questions (with answers) about high\ + \ school microeconomics.\n\n" +"group": "mmlu_social_sciences" +"group_alias": "social_sciences" +"include": "_default_template_yaml" +"task": "mmlu_high_school_microeconomics" +"task_alias": "high_school_microeconomics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_physics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_physics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0fae0405aa093a0fb2eaea8fa8a48d0e4e220c9a --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_physics.yaml @@ -0,0 +1,8 @@ +"dataset_name": "high_school_physics" +"description": "The following are multiple choice questions (with answers) about high\ + \ school physics.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_high_school_physics" +"task_alias": "high_school_physics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_psychology.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_psychology.yaml new file mode 100644 index 0000000000000000000000000000000000000000..31ecb18eb91397d2bad070d11d141f3375c66eaf --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_psychology.yaml @@ -0,0 +1,8 @@ +"dataset_name": "high_school_psychology" +"description": "The following are multiple choice questions (with answers) about high\ + \ school psychology.\n\n" +"group": "mmlu_social_sciences" +"group_alias": "social_sciences" +"include": "_default_template_yaml" +"task": "mmlu_high_school_psychology" +"task_alias": "high_school_psychology" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_statistics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_statistics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..54d70880eff3e5aa6ab8f8b233ee45b9f30ca25c --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_statistics.yaml @@ -0,0 +1,8 @@ +"dataset_name": "high_school_statistics" +"description": "The following are multiple choice questions (with answers) about high\ + \ school statistics.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_high_school_statistics" +"task_alias": "high_school_statistics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_us_history.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_us_history.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e4432fe4459fcaa3c1ccf311c094ceb9122d0637 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_us_history.yaml @@ -0,0 +1,8 @@ +"dataset_name": "high_school_us_history" +"description": "The following are multiple choice questions (with answers) about high\ + \ school us history.\n\n" +"group": "mmlu_humanities" +"group_alias": "humanities" +"include": "_default_template_yaml" +"task": "mmlu_high_school_us_history" +"task_alias": "high_school_us_history" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_world_history.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_world_history.yaml new file mode 100644 index 0000000000000000000000000000000000000000..08773a20089a5551710cb82447d6ce23d65c367d --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_high_school_world_history.yaml @@ -0,0 +1,8 @@ +"dataset_name": "high_school_world_history" +"description": "The following are multiple choice questions (with answers) about high\ + \ school world history.\n\n" +"group": "mmlu_humanities" +"group_alias": "humanities" +"include": "_default_template_yaml" +"task": "mmlu_high_school_world_history" +"task_alias": "high_school_world_history" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_human_aging.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_human_aging.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c9e1feb154293d0a75b25837f365ec892756a465 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_human_aging.yaml @@ -0,0 +1,8 @@ +"dataset_name": "human_aging" +"description": "The following are multiple choice questions (with answers) about human\ + \ aging.\n\n" +"group": "mmlu_other" +"group_alias": "other" +"include": "_default_template_yaml" +"task": "mmlu_human_aging" +"task_alias": "human_aging" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_human_sexuality.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_human_sexuality.yaml new file mode 100644 index 0000000000000000000000000000000000000000..715859a1e8ccd52ac406b9f9dcaf4dc80dbaa42f --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_human_sexuality.yaml @@ -0,0 +1,8 @@ +"dataset_name": "human_sexuality" +"description": "The following are multiple choice questions (with answers) about human\ + \ sexuality.\n\n" +"group": "mmlu_social_sciences" +"group_alias": "social_sciences" +"include": "_default_template_yaml" +"task": "mmlu_human_sexuality" +"task_alias": "human_sexuality" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_international_law.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_international_law.yaml new file mode 100644 index 0000000000000000000000000000000000000000..68765225bbee355a00a2d0d3e9a8931624fc45dd --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_international_law.yaml @@ -0,0 +1,8 @@ +"dataset_name": "international_law" +"description": "The following are multiple choice questions (with answers) about international\ + \ law.\n\n" +"group": "mmlu_humanities" +"group_alias": "humanities" +"include": "_default_template_yaml" +"task": "mmlu_international_law" +"task_alias": "international_law" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_jurisprudence.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_jurisprudence.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e16de5c40b2a29cfc0cadeaad5ed11df9b7503e8 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_jurisprudence.yaml @@ -0,0 +1,8 @@ +"dataset_name": "jurisprudence" +"description": "The following are multiple choice questions (with answers) about jurisprudence.\n\ + \n" +"group": "mmlu_humanities" +"group_alias": "humanities" +"include": "_default_template_yaml" +"task": "mmlu_jurisprudence" +"task_alias": "jurisprudence" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_logical_fallacies.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_logical_fallacies.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8b12057bc78ffc195b4d016bfdc1c931f08e6267 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_logical_fallacies.yaml @@ -0,0 +1,8 @@ +"dataset_name": "logical_fallacies" +"description": "The following are multiple choice questions (with answers) about logical\ + \ fallacies.\n\n" +"group": "mmlu_humanities" +"group_alias": "humanities" +"include": "_default_template_yaml" +"task": "mmlu_logical_fallacies" +"task_alias": "logical_fallacies" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_machine_learning.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_machine_learning.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2387d680b2d57d2d409ecc1bf744f512dd4fcef3 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_machine_learning.yaml @@ -0,0 +1,8 @@ +"dataset_name": "machine_learning" +"description": "The following are multiple choice questions (with answers) about machine\ + \ learning.\n\n" +"group": "mmlu_stem" +"group_alias": "stem" +"include": "_default_template_yaml" +"task": "mmlu_machine_learning" +"task_alias": "machine_learning" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_management.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_management.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d0cdc812d7e94eb51a9f1d3148b3d51253c1c92a --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_management.yaml @@ -0,0 +1,8 @@ +"dataset_name": "management" +"description": "The following are multiple choice questions (with answers) about management.\n\ + \n" +"group": "mmlu_other" +"group_alias": "other" +"include": "_default_template_yaml" +"task": "mmlu_management" +"task_alias": "management" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_marketing.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_marketing.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a614db2977421ad88dcf4cd963647534f0d23b88 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_marketing.yaml @@ -0,0 +1,8 @@ +"dataset_name": "marketing" +"description": "The following are multiple choice questions (with answers) about marketing.\n\ + \n" +"group": "mmlu_other" +"group_alias": "other" +"include": "_default_template_yaml" +"task": "mmlu_marketing" +"task_alias": "marketing" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_medical_genetics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_medical_genetics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5d7ce7083bc62f5859988fb76f84c4dc102175e8 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_medical_genetics.yaml @@ -0,0 +1,8 @@ +"dataset_name": "medical_genetics" +"description": "The following are multiple choice questions (with answers) about medical\ + \ genetics.\n\n" +"group": "mmlu_other" +"group_alias": "other" +"include": "_default_template_yaml" +"task": "mmlu_medical_genetics" +"task_alias": "medical_genetics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_miscellaneous.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_miscellaneous.yaml new file mode 100644 index 0000000000000000000000000000000000000000..77e819cf831fb69381d75a1c039ef9ffaee3028f --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_miscellaneous.yaml @@ -0,0 +1,8 @@ +"dataset_name": "miscellaneous" +"description": "The following are multiple choice questions (with answers) about miscellaneous.\n\ + \n" +"group": "mmlu_other" +"group_alias": "other" +"include": "_default_template_yaml" +"task": "mmlu_miscellaneous" +"task_alias": "miscellaneous" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_moral_disputes.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_moral_disputes.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2df1a1ddb6302f7e563852fdd56cdb0d1668dad1 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_moral_disputes.yaml @@ -0,0 +1,8 @@ +"dataset_name": "moral_disputes" +"description": "The following are multiple choice questions (with answers) about moral\ + \ disputes.\n\n" +"group": "mmlu_humanities" +"group_alias": "humanities" +"include": "_default_template_yaml" +"task": "mmlu_moral_disputes" +"task_alias": "moral_disputes" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_moral_scenarios.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_moral_scenarios.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6da63cb24847aeaf580fc54f701326a187836517 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_moral_scenarios.yaml @@ -0,0 +1,8 @@ +"dataset_name": "moral_scenarios" +"description": "The following are multiple choice questions (with answers) about moral\ + \ scenarios.\n\n" +"group": "mmlu_humanities" +"group_alias": "humanities" +"include": "_default_template_yaml" +"task": "mmlu_moral_scenarios" +"task_alias": "moral_scenarios" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_nutrition.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_nutrition.yaml new file mode 100644 index 0000000000000000000000000000000000000000..df70fbb2de2f40a5cb10a9093d4a0243e45c6ffc --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_nutrition.yaml @@ -0,0 +1,8 @@ +"dataset_name": "nutrition" +"description": "The following are multiple choice questions (with answers) about nutrition.\n\ + \n" +"group": "mmlu_other" +"group_alias": "other" +"include": "_default_template_yaml" +"task": "mmlu_nutrition" +"task_alias": "nutrition" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_philosophy.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_philosophy.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9dba09c10ebe140068f8054ee4d391f00682745d --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_philosophy.yaml @@ -0,0 +1,8 @@ +"dataset_name": "philosophy" +"description": "The following are multiple choice questions (with answers) about philosophy.\n\ + \n" +"group": "mmlu_humanities" +"group_alias": "humanities" +"include": "_default_template_yaml" +"task": "mmlu_philosophy" +"task_alias": "philosophy" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_prehistory.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_prehistory.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d787898c62a174ddfd3895da3b164651f30d266b --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_prehistory.yaml @@ -0,0 +1,8 @@ +"dataset_name": "prehistory" +"description": "The following are multiple choice questions (with answers) about prehistory.\n\ + \n" +"group": "mmlu_humanities" +"group_alias": "humanities" +"include": "_default_template_yaml" +"task": "mmlu_prehistory" +"task_alias": "prehistory" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_professional_accounting.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_professional_accounting.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3443c336f2bc8dcf33e03bffa438a0e60a98ad3e --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_professional_accounting.yaml @@ -0,0 +1,8 @@ +"dataset_name": "professional_accounting" +"description": "The following are multiple choice questions (with answers) about professional\ + \ accounting.\n\n" +"group": "mmlu_other" +"group_alias": "other" +"include": "_default_template_yaml" +"task": "mmlu_professional_accounting" +"task_alias": "professional_accounting" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_professional_law.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_professional_law.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f3a02631e28703dc8d3a6091175c49b9eb5676eb --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_professional_law.yaml @@ -0,0 +1,8 @@ +"dataset_name": "professional_law" +"description": "The following are multiple choice questions (with answers) about professional\ + \ law.\n\n" +"group": "mmlu_humanities" +"group_alias": "humanities" +"include": "_default_template_yaml" +"task": "mmlu_professional_law" +"task_alias": "professional_law" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_professional_medicine.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_professional_medicine.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e8c49b5eb084d9cb2133c333f573ccbe3b8d9824 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_professional_medicine.yaml @@ -0,0 +1,8 @@ +"dataset_name": "professional_medicine" +"description": "The following are multiple choice questions (with answers) about professional\ + \ medicine.\n\n" +"group": "mmlu_other" +"group_alias": "other" +"include": "_default_template_yaml" +"task": "mmlu_professional_medicine" +"task_alias": "professional_medicine" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_professional_psychology.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_professional_psychology.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ec48a06fe11d98facbf5c4d24e03595ceb2fb2bd --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_professional_psychology.yaml @@ -0,0 +1,8 @@ +"dataset_name": "professional_psychology" +"description": "The following are multiple choice questions (with answers) about professional\ + \ psychology.\n\n" +"group": "mmlu_social_sciences" +"group_alias": "social_sciences" +"include": "_default_template_yaml" +"task": "mmlu_professional_psychology" +"task_alias": "professional_psychology" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_public_relations.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_public_relations.yaml new file mode 100644 index 0000000000000000000000000000000000000000..db36fb498f6886549b420eaee34fb68dddfb126c --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_public_relations.yaml @@ -0,0 +1,8 @@ +"dataset_name": "public_relations" +"description": "The following are multiple choice questions (with answers) about public\ + \ relations.\n\n" +"group": "mmlu_social_sciences" +"group_alias": "social_sciences" +"include": "_default_template_yaml" +"task": "mmlu_public_relations" +"task_alias": "public_relations" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_security_studies.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_security_studies.yaml new file mode 100644 index 0000000000000000000000000000000000000000..072dfd704815409d0a21576fa9620b1b5cb1a15b --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_security_studies.yaml @@ -0,0 +1,8 @@ +"dataset_name": "security_studies" +"description": "The following are multiple choice questions (with answers) about security\ + \ studies.\n\n" +"group": "mmlu_social_sciences" +"group_alias": "social_sciences" +"include": "_default_template_yaml" +"task": "mmlu_security_studies" +"task_alias": "security_studies" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_sociology.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_sociology.yaml new file mode 100644 index 0000000000000000000000000000000000000000..efcbd27b8d653f2df89f52880d47fe5a461e1246 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_sociology.yaml @@ -0,0 +1,8 @@ +"dataset_name": "sociology" +"description": "The following are multiple choice questions (with answers) about sociology.\n\ + \n" +"group": "mmlu_social_sciences" +"group_alias": "social_sciences" +"include": "_default_template_yaml" +"task": "mmlu_sociology" +"task_alias": "sociology" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_us_foreign_policy.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_us_foreign_policy.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d80ee94adf4fd7b74dc368696924065862b9ce9e --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_us_foreign_policy.yaml @@ -0,0 +1,8 @@ +"dataset_name": "us_foreign_policy" +"description": "The following are multiple choice questions (with answers) about us\ + \ foreign policy.\n\n" +"group": "mmlu_social_sciences" +"group_alias": "social_sciences" +"include": "_default_template_yaml" +"task": "mmlu_us_foreign_policy" +"task_alias": "us_foreign_policy" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_virology.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_virology.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d935f92aa80e30f773960e6127466a4d7f768448 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_virology.yaml @@ -0,0 +1,8 @@ +"dataset_name": "virology" +"description": "The following are multiple choice questions (with answers) about virology.\n\ + \n" +"group": "mmlu_other" +"group_alias": "other" +"include": "_default_template_yaml" +"task": "mmlu_virology" +"task_alias": "virology" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_world_religions.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_world_religions.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8681354f7057d2d3ff2d68a98b3a5e6f23c960cc --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/default/mmlu_world_religions.yaml @@ -0,0 +1,8 @@ +"dataset_name": "world_religions" +"description": "The following are multiple choice questions (with answers) about world\ + \ religions.\n\n" +"group": "mmlu_humanities" +"group_alias": "humanities" +"include": "_default_template_yaml" +"task": "mmlu_world_religions" +"task_alias": "world_religions" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_fewshot/mmlu_college_chemistry.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_fewshot/mmlu_college_chemistry.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b924eb867e3f4353022757da7b7027a468d686a5 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_fewshot/mmlu_college_chemistry.yaml @@ -0,0 +1,38 @@ +"dataset_name": "college_chemistry" +"description": "The following are multiple choice questions (with answers) about college\ + \ chemistry.\n\nQ: 3 Cl−(aq) + 4 CrO_4^2−(aq) + 23 H+(aq) → 3 HClO2(aq) + 4 Cr3+(aq)\ + \ + 10 H2O(l). In the reaction shown above, Cl−(aq) behaves as\n(A) an acid (B)\ + \ a base (C) a catalyst (D) a reducing agent\nA: Let's think step by step. A molecule\ + \ that behaves as a base accepts an H+ ion (or proton) from another molecule, whereas\ + \ a molecule that behaves as an acid donates an H+ ion (or proton) to another molecule.\ + \ Neither of these is the case for Cl in this reaction, which rules out (A) and\ + \ (B). A catalyst is a substance that only accelerates a reaction without itself\ + \ undergoing chemical change, which is not the case here. This rules out (C). Instead,\ + \ the $Cl^{-} molecules carry a negative charge, which they donate in the reaction\ + \ to form 3 HClO2. This is the behavior of a reducing agent, or (D). The answer\ + \ is (D).\n\nQ: Which of the following statements about the lanthanide elements\ + \ is NOT true?\n(A) The most common oxidation state for the lanthanide elements\ + \ is +3. (B) Lanthanide complexes often have high coordination numbers (> 6). (C)\ + \ All of the lanthanide elements react with aqueous acid to liberate hydrogen. (D)\ + \ The atomic radii of the lanthanide elements increase across the period from La\ + \ to Lu.\nA: Let's think step by step. The atomic radii of the lanthanide elements\ + \ in fact decrease across the period from La to Lu. Options (A), (B), and (C) are\ + \ all true. This means that only (D) is NOT true. The answer is (D).\n\nQ: Which\ + \ of the following lists the hydrides of group-14 elements in order of thermal stability,\ + \ from lowest to highest?\n(A) PbH4 < SnH4 < GeH4 < SiH4 < CH4 (B) PbH4 < SnH4 <\ + \ CH4 < GeH4 < SiH4 (C) CH4 < SiH4 < GeH4 < SnH4 < PbH4 (D) CH4 < PbH4 < GeH4 <\ + \ SnH4 < SiH4\nA: Let's think step by step. The thermal stability of group-14 hydrides\ + \ decreases as we move from the top of group 14 to the bottom. The order of elements\ + \ in the group from top to bottom is C, Si, Ge, Sn, Pb. Therefore in order of increasing\ + \ thermal stability we have PbH4, SnH4, GeH4, SiH4, and CH4, or answer (A). The\ + \ answer is (A).\n\nQ: Predict the number of lines in the EPR spectrum of a solution\ + \ of 13C-labelled methyl radical (13CH3•), assuming the lines do not overlap.\n\ + (A) 4 (B) 3 (C) 6 (D) 24 (E) 8\nA: Let's think step by step. The electron paramagnetic\ + \ resonance spectrum will be split by two forms of interactions. The first is the\ + \ hyperfine interaction with the 13C (nuclear spin $I = \nrac{1}{2}$) which will\ + \ split the spectrum into 2 lines. This will be further split into 4 lines by the\ + \ interaction with three equivalent 1H nuclei. The total number of lines is therefore\ + \ $2 \\cdot 4 = 8$. The answer is (E).\n\n" +"group": "mmlu_flan_cot_fewshot_stem" +"include": "_mmlu_flan_cot_fewshot_template_yaml" +"task": "mmlu_flan_cot_fewshot_college_chemistry" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_fewshot/mmlu_computer_security.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_fewshot/mmlu_computer_security.yaml new file mode 100644 index 0000000000000000000000000000000000000000..943ab5ac8ecf5ba53701763cb1bb18e621600fa4 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_fewshot/mmlu_computer_security.yaml @@ -0,0 +1,36 @@ +"dataset_name": "computer_security" +"description": "The following are multiple choice questions (with answers) about computer\ + \ security.\n\nQ: SHA-1 has a message digest of\n(A) 160 bits (B) 512 bits (C) 628\ + \ bits (D) 820 bits\nA: Let's think step by step. Since SHA-1 is a hash function\ + \ which takes an input and produces a 160-bit (20-byte) hash value, its message\ + \ digest is 160 bits. The answer is (A).\n\nQ: _____________ can modify data on\ + \ your system – so that your system doesn’t run correctly or you can no longer access\ + \ specific data, or it may even ask for ransom in order to give your access.\n(A)\ + \ IM – Trojans (B) Backdoor Trojans (C) Trojan-Downloader (D) Ransom Trojan\nA:\ + \ Let's think step by step. The system is asking for trojans, which are for ransom,\ + \ which means ransom trojan. The answer is (D).\n\nQ: What is ethical hacking?\n\ + (A) \"Hacking\" ethics so they justify unintended selfish behavior (B) Hacking systems\ + \ (e.g., during penetration testing) to expose vulnerabilities so they can be fixed,\ + \ rather than exploited (C) Hacking into systems run by those whose ethics you disagree\ + \ with (D) A slang term for rapid software development, e.g., as part of hackathons\n\ + A: Let's think step by step. Ethical hacking is a process of detecting vulnerabilities\ + \ in an application, system, or organization's infrastructure that an attacker can\ + \ use to exploit an individual or organization. They use this process to prevent\ + \ cyberattacks and security breaches by lawfully hacking into the systems and looking\ + \ for weak points. The answer is (B).\n\nQ: The ____________ is anything which your\ + \ search engine cannot search.\n(A) Haunted web (B) World Wide Web (C) Surface web\ + \ (D) Deep Web\nA: Let's think step by step. The search engine searches on the Surface\ + \ Web, which is the portion of the world wide web which is visible so (B,C) are\ + \ wrong. The Haunted Web doesn’t correspond to an internet concept. The Deep Web\ + \ is the part of the World Wide Web which is not indexed. The answer is (D).\n\n\ + Q: Exploitation of the Heartbleed bug permits\n(A) overwriting cryptographic keys\ + \ in memory (B) a kind of code injection (C) a read outside bounds of a buffer (D)\ + \ a format string attack\nA: Let's think step by step. The Heartbleed Bug is a serious\ + \ vulnerability in the popular OpenSSL cryptographic software library. Heartbleed\ + \ resulted from improper input validation (due to a missing bounds check) in the\ + \ implementation of the TLS heartbeat extension. The vulnerability was classified\ + \ as a buffer over-read, a situation where more data can be read than should be\ + \ allowed. The answer is (C).\n\n" +"group": "mmlu_flan_cot_fewshot_stem" +"include": "_mmlu_flan_cot_fewshot_template_yaml" +"task": "mmlu_flan_cot_fewshot_computer_security" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_fewshot/mmlu_security_studies.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_fewshot/mmlu_security_studies.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f9a441fa3b4fe298381f152063ef79f2b4f98f27 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_fewshot/mmlu_security_studies.yaml @@ -0,0 +1,86 @@ +"dataset_name": "security_studies" +"description": "The following are multiple choice questions (with answers) about security\ + \ studies.\n\nQ: What are the frameworks of analysis within which terrorism has\ + \ been considered (as of 2020)?\n(A) Competition between larger nations has resulted\ + \ in some countries actively supporting terrorist groups to undermine the strength\ + \ of rival states. Terrorist networks are extended patronage clubs maintained and\ + \ paid for by their donor states and are conceptualised as being like state actors,\ + \ to be dealt with using military force. (B) Globalization has enabled the internationalization\ + \ of terrorist activities by opening up their operational space, although coordination\ + \ is still managed from a geographical base. This suggests that terrorist groups\ + \ are nationally structured which means that terrorism cannot be considered in terms\ + \ of a war to be defeated militarily without having serious implications on the\ + \ indigenous population. (C) Terrorism can be viewed as a problem to be resolved\ + \ by military means (war on terrorism), by normal police techniques (terrorism as\ + \ crime), or as a medical problem with underlying causes and symptoms (terrorism\ + \ as disease). (D) Terrorism is viewed as a criminal problem. The criminalization\ + \ of terrorism has two important implications. Firstly, it suggests that terrorism\ + \ can be eradicated - terrorists can be caught and brought to trial by normal judicial\ + \ proceedings thereby removing the threat from society - and secondly, it suggests\ + \ that preventative crime techniques are applicable to prevent its development.\n\ + A: Let's think step by step. We refer to Wikipedia articles on security studies\ + \ for help. (A) is wrong because it is not competition between larger nations that\ + \ causes terrorism. \n(B) is wrong because globalization is not the cause of terrorism.\n\ + (C) is correct because the US undertook the war on terrorism. \n(D) is wrong because\ + \ preventative crime techniques will likely not end terrorism. The answer is (C).\n\ + \nQ: Which of the following is the best lens through which to investigate the role\ + \ of child soldiers?\n(A) Child soldiers are victims of combat that need re-education\ + \ and rehabilitation. (B) Children and their mothers are not active subjects in\ + \ warfare and are best considered as subjects in the private sphere. (C) Children\ + \ are most often innocent bystanders in war and are best used as signifiers of peace.\ + \ (D) Children have political subjecthood that is missed when they are considered\ + \ as passive victims of warfare.\nA: Let's think step by step. We refer to Wikipedia\ + \ articles on security studies for help. Child soliders as a political topic can\ + \ be missed when they are considered passive victims of warfare. The answer is (D).\n\ + \nQ: How can we best describe the relationship between the state-centric approach\ + \ and the concept of human security?\n(A) There are such wide divisions within the\ + \ human security framework regarding the nature of threats and referent objects\ + \ that no widely applicable comparisons between state-centric approaches and human\ + \ security can be drawn. (B) By adopting the framework of human security, the limitations\ + \ of the realist state-centric approach become evident. Whilst human security defines\ + \ the referent object as the person or population, state-centric approaches prioritise\ + \ the security of the state, de-prioritizing the pursuit of human security. (C)\ + \ The state-centric approach to security is a faction of human security, usually\ + \ defined within the broad school of human security. By being state-centric this\ + \ approach prioritises the individual as the referent object in security studies.\ + \ (D) Both the state-centric and human-centric approaches to security are mutually\ + \ exclusive and offer a sufficient analytic framework with which to understand the\ + \ international security system. It is therefore the role of security analysts to\ + \ determine which of these substantial concepts is correct, and which should be\ + \ discarded.\nA: Let's think step by step. We refer to Wikipedia articles on security\ + \ studies for help. Human security focuses on a person or population whereas state-centric\ + \ approaches focus on the state while deprioritizing human security. The answer\ + \ is (B).\n\nQ: In order to become securitized, a threat must be presented in which\ + \ of these ways?\n(A) As an existential threat that requires immediate and extraordinary\ + \ action, posing a threat to the survival of the state or to societal security.\ + \ (B) As requiring immediate and extraordinary action by the state, threatening\ + \ the survival of a referent object and therefore warranting the use of measures\ + \ not normally employed in the political realm. (C) As an urgent threat to the survival\ + \ of the referent object, so serious that it legitimises the employment of extraordinary\ + \ action in response. (D) As an urgent threat to the survival of the audience that\ + \ requires extraordinary or emergency measures.\nA: Let's think step by step. We\ + \ refer to Wikipedia articles on security studies for help. To be securitized, a\ + \ threat must be an urgent threat to the survival of the referent object. The answer\ + \ is (C).\n\nQ: What distinguishes coercive diplomacy from military force?\n(A)\ + \ Compellence is another term for coercive diplomacy, but covering a narrower set\ + \ of criteria; compellence covers those threats aimed at initiating adversary action.\ + \ A threat to coerce a state to give up part of its territory would count as coercive\ + \ diplomacy, as long as that threat proactively initiates action before reactive\ + \ diplomacy is taken. (B) Coercive diplomacy constitutes the threats of limited\ + \ force to induce adversary's incentive to comply with the coercer's demands. It\ + \ is an influence strategy that is intended to obtain compliance: the use of force\ + \ to defeat an opponent first does not count. It leaves an element of choice with\ + \ the target to comply, or to continue. (C) Military force, or the threat of military\ + \ force, utilises fear to achieve strategic objectives. Coercive diplomacy is differentiated\ + \ from this approach, because it does not use fear as a tool for coercing an adversary.\ + \ (D) Coercive diplomacy is employed to use force but to limit its effects on the\ + \ international community. Coercive diplomacy is an aggressive strategy that is\ + \ intended to obtain compliance through defeat. It does not leave an element of\ + \ choice with the target, the target either being forced to comply or engage in\ + \ conflict. It seeks to control by imposing compliance by removing any opportunity\ + \ for negotiation or concession.\nA: Let's think step by step. We refer to Wikipedia\ + \ articles on security studies for help. Coercive diplomacy uses the threat of force\ + \ to induce the opponent to comply with demands. The answer is (B).\n\n" +"group": "mmlu_flan_cot_fewshot_social_sciences" +"include": "_mmlu_flan_cot_fewshot_template_yaml" +"task": "mmlu_flan_cot_fewshot_security_studies" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_fewshot/mmlu_sociology.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_fewshot/mmlu_sociology.yaml new file mode 100644 index 0000000000000000000000000000000000000000..53f69bb05b4b89fcf4bb192d6c2510a86bfe23cd --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_fewshot/mmlu_sociology.yaml @@ -0,0 +1,43 @@ +"dataset_name": "sociology" +"description": "The following are multiple choice questions (with answers) about sociology.\n\ + \nQ: Which of the following is not a problem associated with official statistics\ + \ on strike action?\n(A) most strikes go unnoticed by employers and the mass media\ + \ (B) not all industrial disputes will be reported by the employer (C) the definition\ + \ of strikes excludes those that involve fewer than ten workers or last less than\ + \ one day (D) it is hard to compare strikes that were measured in different ways\n\ + A: Let's think step by step. We refer to Wikipedia articles on sociology for help.\ + \ Official statistics on strike action can be problematic because not all industrial\ + \ disputes will be reported by employers, the definition of strikes excludes those\ + \ that involves fewer than ten workers or last less than one day, and it is hard\ + \ to compare strikes that were measured in different ways. Thus, (A) is not a problem\ + \ associated with official statistics on strike action. The answer is (A).\n\nQ:\ + \ What does Berger (1963) describe as a metaphor for social reality?\n(A) a fairground\ + \ ride (B) a circus (C) a puppet theatre (D) a ballet\nA: Let's think step by step.\ + \ We refer to Wikipedia articles on sociology for help. Berger describes social\ + \ reality using the metaphor of a puppet theatre. The answer is (C).\n\nQ: The term\ + \ 'hegemony' refers to:\n(A) the tendency for the working class not to realize their\ + \ own interests (B) a dominant ideology that legitimates economic, political and\ + \ cultural power (C) a form of dual consciousness based on ideology and everyday\ + \ experiences (D) a mode of payment given for outstanding topiary\nA: Let's think\ + \ step by step. We refer to Wikipedia articles on sociology for help. Hegemony refers\ + \ to a dominant ideology that legitimates economic, policital, and cultural power.\ + \ The answer is (B).\n\nQ: The shift from 'civil religion' to 'common religion'\ + \ means that:\n(A) the increasing bureaucracy of the state has made religion only\ + \ a marginal part of our lives (B) despite the weakening of traditional authority,\ + \ our everyday lives and 'common sense' remain shaped by religious beliefs and values\ + \ (C) religious participation in collective worship may have declined, but people\ + \ still practise their faiths in private (D) people are much more likely to discuss\ + \ their religious beliefs in public, informal settings\nA: Let's think step by step.\ + \ We refer to Wikipedia articles on sociology for help. The shift from civil religion\ + \ to common religion means that despite the weakening of traditional authority,\ + \ our everyday lives and common sense remain shaped by religious beliefs and values.\ + \ The answer is (B).\n\nQ: Which of the following did the post-war welfare state\ + \ of 1948 not aim to provide:\n(A) free health care and education for all (B) a\ + \ minimum wage (C) full employment (D) universal welfare\nA: Let's think step by\ + \ step. We refer to Wikipedia articles on sociology for help. The post-war welfare\ + \ state of 1948 aimed to provide free healthcare and education, full employment,\ + \ and universal welfare. But it did not aim to provide a minimum wage. The answer\ + \ is (B).\n\n" +"group": "mmlu_flan_cot_fewshot_social_sciences" +"include": "_mmlu_flan_cot_fewshot_template_yaml" +"task": "mmlu_flan_cot_fewshot_sociology" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/_mmlu_flan_cot_zeroshot_template_yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/_mmlu_flan_cot_zeroshot_template_yaml new file mode 100644 index 0000000000000000000000000000000000000000..83a857b08c1d6e1f4e6b0d55993c10c2fd0b5bae --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/_mmlu_flan_cot_zeroshot_template_yaml @@ -0,0 +1,36 @@ +dataset_path: hails/mmlu_no_train # a copy of `cais/mmlu` with no auxiliary_train split +validation_split: validation +fewshot_split: dev +output_type: generate_until +doc_to_text: "Q: {{question.strip()}}\n(A) {{choices[0]}} (B) {{choices[1]}} (C) {{choices[2]}} (D) {{choices[3]}}\nA: Let's think step by step." +doc_to_target: "{{['(A)', '(B)', '(C)', '(D)'][answer]}}" +filter_list: + - name: "strict-match" + filter: + - function: "regex" + regex_pattern: "((?<=The answer is )(.*)(?=.)|(?<=the answer is )(.*)(?=.)|(?<=The answer: )(.*)(?=.)|(?<=The final answer: )(.*)(?=.))" + - function: "take_first" + - name: "flexible-extract" + filter: + - function: !function utils.MultiChoiceRegexFilter + group_select: -1 + ignore_case: true + ignore_punctuation: true + regex_pattern: "(\\([A-Z]\\))" + - function: "take_first" +generation_kwargs: + until: + - "" + - "Q:" + - "<|im_end|>" + do_sample: false + temperature: 0.0 +num_fewshot: 0 +metric_list: + - metric: exact_match + aggregation: mean + higher_is_better: true + ignore_case: true + ignore_punctuation: true +metadata: + version: 1.0 diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_college_computer_science.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_college_computer_science.yaml new file mode 100644 index 0000000000000000000000000000000000000000..185f2a6614dad536c1a16d1b1017599a18427a67 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_college_computer_science.yaml @@ -0,0 +1,6 @@ +"dataset_name": "college_computer_science" +"description": "The following are multiple choice questions (with answers) about college\ + \ computer science.\n\n" +"group": "mmlu_flan_cot_zeroshot_stem" +"include": "_mmlu_flan_cot_zeroshot_template_yaml" +"task": "mmlu_flan_cot_zeroshot_college_computer_science" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_college_physics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_college_physics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..319c72142454df340260c1a6dd3f59f244a57eab --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_college_physics.yaml @@ -0,0 +1,6 @@ +"dataset_name": "college_physics" +"description": "The following are multiple choice questions (with answers) about college\ + \ physics.\n\n" +"group": "mmlu_flan_cot_zeroshot_stem" +"include": "_mmlu_flan_cot_zeroshot_template_yaml" +"task": "mmlu_flan_cot_zeroshot_college_physics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_conceptual_physics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_conceptual_physics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2e1e43dbad9432de41c580779108843761280313 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_conceptual_physics.yaml @@ -0,0 +1,6 @@ +"dataset_name": "conceptual_physics" +"description": "The following are multiple choice questions (with answers) about conceptual\ + \ physics.\n\n" +"group": "mmlu_flan_cot_zeroshot_stem" +"include": "_mmlu_flan_cot_zeroshot_template_yaml" +"task": "mmlu_flan_cot_zeroshot_conceptual_physics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_elementary_mathematics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_elementary_mathematics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..065c92d200e4df5a303cc1f4269ada803524cd89 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_elementary_mathematics.yaml @@ -0,0 +1,6 @@ +"dataset_name": "elementary_mathematics" +"description": "The following are multiple choice questions (with answers) about elementary\ + \ mathematics.\n\n" +"group": "mmlu_flan_cot_zeroshot_stem" +"include": "_mmlu_flan_cot_zeroshot_template_yaml" +"task": "mmlu_flan_cot_zeroshot_elementary_mathematics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_high_school_world_history.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_high_school_world_history.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c89dd0faa47730d507b7337abbf38e00879389b5 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_high_school_world_history.yaml @@ -0,0 +1,6 @@ +"dataset_name": "high_school_world_history" +"description": "The following are multiple choice questions (with answers) about high\ + \ school world history.\n\n" +"group": "mmlu_flan_cot_zeroshot_humanities" +"include": "_mmlu_flan_cot_zeroshot_template_yaml" +"task": "mmlu_flan_cot_zeroshot_high_school_world_history" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_human_sexuality.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_human_sexuality.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ed2116dd24e9804e2ae4def9fd0cd9518618cf3a --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_human_sexuality.yaml @@ -0,0 +1,6 @@ +"dataset_name": "human_sexuality" +"description": "The following are multiple choice questions (with answers) about human\ + \ sexuality.\n\n" +"group": "mmlu_flan_cot_zeroshot_social_sciences" +"include": "_mmlu_flan_cot_zeroshot_template_yaml" +"task": "mmlu_flan_cot_zeroshot_human_sexuality" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_marketing.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_marketing.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0827f78df766331ccdf6235a4b4120a089cdd9c9 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_marketing.yaml @@ -0,0 +1,6 @@ +"dataset_name": "marketing" +"description": "The following are multiple choice questions (with answers) about marketing.\n\ + \n" +"group": "mmlu_flan_cot_zeroshot_other" +"include": "_mmlu_flan_cot_zeroshot_template_yaml" +"task": "mmlu_flan_cot_zeroshot_marketing" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_medical_genetics.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_medical_genetics.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1706ee5b4ac151a36e7a75b638d75e8194696f8d --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_medical_genetics.yaml @@ -0,0 +1,6 @@ +"dataset_name": "medical_genetics" +"description": "The following are multiple choice questions (with answers) about medical\ + \ genetics.\n\n" +"group": "mmlu_flan_cot_zeroshot_other" +"include": "_mmlu_flan_cot_zeroshot_template_yaml" +"task": "mmlu_flan_cot_zeroshot_medical_genetics" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_professional_medicine.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_professional_medicine.yaml new file mode 100644 index 0000000000000000000000000000000000000000..47cf9573d08e1f911af7fbfbd984358fecd275d9 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_professional_medicine.yaml @@ -0,0 +1,6 @@ +"dataset_name": "professional_medicine" +"description": "The following are multiple choice questions (with answers) about professional\ + \ medicine.\n\n" +"group": "mmlu_flan_cot_zeroshot_other" +"include": "_mmlu_flan_cot_zeroshot_template_yaml" +"task": "mmlu_flan_cot_zeroshot_professional_medicine" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_professional_psychology.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_professional_psychology.yaml new file mode 100644 index 0000000000000000000000000000000000000000..cc055d5bacacf77b1dd5f70b68bbaa81d1aad2ff --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_professional_psychology.yaml @@ -0,0 +1,6 @@ +"dataset_name": "professional_psychology" +"description": "The following are multiple choice questions (with answers) about professional\ + \ psychology.\n\n" +"group": "mmlu_flan_cot_zeroshot_social_sciences" +"include": "_mmlu_flan_cot_zeroshot_template_yaml" +"task": "mmlu_flan_cot_zeroshot_professional_psychology" diff --git a/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_public_relations.yaml b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_public_relations.yaml new file mode 100644 index 0000000000000000000000000000000000000000..14d02c3a3c015e78cb780c646461fb7ac70a5ce4 --- /dev/null +++ b/lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_cot_zeroshot/mmlu_public_relations.yaml @@ -0,0 +1,6 @@ +"dataset_name": "public_relations" +"description": "The following are multiple choice questions (with answers) about public\ + \ relations.\n\n" +"group": "mmlu_flan_cot_zeroshot_social_sciences" +"include": "_mmlu_flan_cot_zeroshot_template_yaml" +"task": "mmlu_flan_cot_zeroshot_public_relations"