applied-ai-018 commited on
Commit
8b56070
·
verified ·
1 Parent(s): 234b9f4

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/_mmlu_flan_generative_template_yaml +30 -0
  2. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_biology.yaml +6 -0
  3. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_chemistry.yaml +6 -0
  4. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_computer_science.yaml +6 -0
  5. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_physics.yaml +6 -0
  6. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_econometrics.yaml +6 -0
  7. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_formal_logic.yaml +6 -0
  8. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_biology.yaml +6 -0
  9. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_chemistry.yaml +6 -0
  10. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_macroeconomics.yaml +6 -0
  11. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_mathematics.yaml +6 -0
  12. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_microeconomics.yaml +6 -0
  13. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_world_history.yaml +6 -0
  14. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_human_aging.yaml +6 -0
  15. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_logical_fallacies.yaml +6 -0
  16. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_marketing.yaml +6 -0
  17. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_medical_genetics.yaml +6 -0
  18. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_moral_disputes.yaml +6 -0
  19. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_nutrition.yaml +6 -0
  20. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_prehistory.yaml +6 -0
  21. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_professional_medicine.yaml +6 -0
  22. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_us_foreign_policy.yaml +6 -0
  23. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/_mmlu.yaml +6 -0
  24. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/_mmlu_flan_loglikelihood_template_yaml +16 -0
  25. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_abstract_algebra.yaml +6 -0
  26. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_anatomy.yaml +6 -0
  27. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_astronomy.yaml +6 -0
  28. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_business_ethics.yaml +6 -0
  29. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_clinical_knowledge.yaml +6 -0
  30. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_biology.yaml +6 -0
  31. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_computer_science.yaml +6 -0
  32. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_mathematics.yaml +6 -0
  33. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_medicine.yaml +6 -0
  34. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_computer_security.yaml +6 -0
  35. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_conceptual_physics.yaml +6 -0
  36. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_econometrics.yaml +6 -0
  37. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_electrical_engineering.yaml +6 -0
  38. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_elementary_mathematics.yaml +6 -0
  39. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_formal_logic.yaml +6 -0
  40. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_global_facts.yaml +6 -0
  41. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_biology.yaml +6 -0
  42. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_chemistry.yaml +6 -0
  43. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_computer_science.yaml +6 -0
  44. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_european_history.yaml +6 -0
  45. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_geography.yaml +6 -0
  46. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_government_and_politics.yaml +6 -0
  47. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_microeconomics.yaml +6 -0
  48. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_physics.yaml +6 -0
  49. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_statistics.yaml +6 -0
  50. lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_us_history.yaml +6 -0
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/_mmlu_flan_generative_template_yaml ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ group: mmlu_flan_n_shot_generative
2
+ dataset_path: hails/mmlu_no_train # a copy of `cais/mmlu` with no auxiliary_train split
3
+ test_split: test
4
+ fewshot_split: dev
5
+ output_type: generate_until
6
+ doc_to_text: "Q: {{question.strip()}}\n(A) {{choices[0]}} (B) {{choices[1]}} (C) {{choices[2]}} (D) {{choices[3]}}\nA: "
7
+ doc_to_target: "{{['(A)', '(B)', '(C)', '(D)'][answer]}}"
8
+ filter_list:
9
+ - name: "strict-match"
10
+ filter:
11
+ - function: "take_first"
12
+ - name: "flexible-extract"
13
+ filter:
14
+ - function: !function utils.MultiChoiceRegexFilter
15
+ group_select: 0
16
+ regex_pattern: "(\\([A-Z]\\))"
17
+ ignore_case: true
18
+ ignore_punctuation: true
19
+ - function: "take_first"
20
+ generation_kwargs:
21
+ until:
22
+ - "</s>"
23
+ - "Q:"
24
+ - "<|im_end|>"
25
+ metric_list:
26
+ - metric: exact_match
27
+ aggregation: mean
28
+ higher_is_better: true
29
+ metadata:
30
+ version: 1.0
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_biology.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "college_biology"
2
+ "description": "The following are multiple choice questions (with answers) about college\
3
+ \ biology.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_stem"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_college_biology"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_chemistry.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "college_chemistry"
2
+ "description": "The following are multiple choice questions (with answers) about college\
3
+ \ chemistry.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_stem"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_college_chemistry"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_computer_science.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "college_computer_science"
2
+ "description": "The following are multiple choice questions (with answers) about college\
3
+ \ computer science.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_stem"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_college_computer_science"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_physics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "college_physics"
2
+ "description": "The following are multiple choice questions (with answers) about college\
3
+ \ physics.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_stem"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_college_physics"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_econometrics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "econometrics"
2
+ "description": "The following are multiple choice questions (with answers) about econometrics.\n\
3
+ \n"
4
+ "group": "mmlu_flan_n_shot_generative_social_sciences"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_econometrics"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_formal_logic.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "formal_logic"
2
+ "description": "The following are multiple choice questions (with answers) about formal\
3
+ \ logic.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_humanities"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_formal_logic"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_biology.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_biology"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school biology.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_stem"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_high_school_biology"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_chemistry.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_chemistry"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school chemistry.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_stem"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_high_school_chemistry"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_macroeconomics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_macroeconomics"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school macroeconomics.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_social_sciences"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_high_school_macroeconomics"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_mathematics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_mathematics"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school mathematics.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_stem"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_high_school_mathematics"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_microeconomics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_microeconomics"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school microeconomics.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_social_sciences"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_high_school_microeconomics"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_world_history.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_world_history"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school world history.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_humanities"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_high_school_world_history"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_human_aging.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "human_aging"
2
+ "description": "The following are multiple choice questions (with answers) about human\
3
+ \ aging.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_other"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_human_aging"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_logical_fallacies.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "logical_fallacies"
2
+ "description": "The following are multiple choice questions (with answers) about logical\
3
+ \ fallacies.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_humanities"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_logical_fallacies"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_marketing.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "marketing"
2
+ "description": "The following are multiple choice questions (with answers) about marketing.\n\
3
+ \n"
4
+ "group": "mmlu_flan_n_shot_generative_other"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_marketing"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_medical_genetics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "medical_genetics"
2
+ "description": "The following are multiple choice questions (with answers) about medical\
3
+ \ genetics.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_other"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_medical_genetics"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_moral_disputes.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "moral_disputes"
2
+ "description": "The following are multiple choice questions (with answers) about moral\
3
+ \ disputes.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_humanities"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_moral_disputes"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_nutrition.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "nutrition"
2
+ "description": "The following are multiple choice questions (with answers) about nutrition.\n\
3
+ \n"
4
+ "group": "mmlu_flan_n_shot_generative_other"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_nutrition"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_prehistory.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "prehistory"
2
+ "description": "The following are multiple choice questions (with answers) about prehistory.\n\
3
+ \n"
4
+ "group": "mmlu_flan_n_shot_generative_humanities"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_prehistory"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_professional_medicine.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "professional_medicine"
2
+ "description": "The following are multiple choice questions (with answers) about professional\
3
+ \ medicine.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_other"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_professional_medicine"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_us_foreign_policy.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "us_foreign_policy"
2
+ "description": "The following are multiple choice questions (with answers) about us\
3
+ \ foreign policy.\n\n"
4
+ "group": "mmlu_flan_n_shot_generative_social_sciences"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_us_foreign_policy"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/_mmlu.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ group: mmlu_flan_n_shot_loglikelihood
2
+ task:
3
+ - mmlu_flan_n_shot_loglikelihood_stem
4
+ - mmlu_flan_n_shot_loglikelihood_other
5
+ - mmlu_flan_n_shot_loglikelihood_social_sciences
6
+ - mmlu_flan_n_shot_loglikelihood_humanities
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/_mmlu_flan_loglikelihood_template_yaml ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ dataset_path: hails/mmlu_no_train # a copy of `cais/mmlu` with no auxiliary_train split
2
+ test_split: test
3
+ fewshot_split: dev
4
+ output_type: multiple_choice
5
+ doc_to_text: "Q: {{question.strip()}}\n(A) {{choices[0]}} (B) {{choices[1]}} (C) {{choices[2]}} (D) {{choices[3]}}\nA: "
6
+ doc_to_choice: ["(A)", "(B)", "(C)", "(D)"]
7
+ doc_to_target: answer
8
+ metric_list:
9
+ - metric: acc
10
+ aggregation: mean
11
+ higher_is_better: true
12
+ - metric: acc_norm
13
+ aggregation: mean
14
+ higher_is_better: true
15
+ metadata:
16
+ version: 0.0
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_abstract_algebra.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "abstract_algebra"
2
+ "description": "The following are multiple choice questions (with answers) about abstract\
3
+ \ algebra.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_abstract_algebra"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_anatomy.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "anatomy"
2
+ "description": "The following are multiple choice questions (with answers) about anatomy.\n\
3
+ \n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_anatomy"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_astronomy.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "astronomy"
2
+ "description": "The following are multiple choice questions (with answers) about astronomy.\n\
3
+ \n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_astronomy"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_business_ethics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "business_ethics"
2
+ "description": "The following are multiple choice questions (with answers) about business\
3
+ \ ethics.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_other"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_business_ethics"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_clinical_knowledge.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "clinical_knowledge"
2
+ "description": "The following are multiple choice questions (with answers) about clinical\
3
+ \ knowledge.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_other"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_clinical_knowledge"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_biology.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "college_biology"
2
+ "description": "The following are multiple choice questions (with answers) about college\
3
+ \ biology.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_college_biology"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_computer_science.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "college_computer_science"
2
+ "description": "The following are multiple choice questions (with answers) about college\
3
+ \ computer science.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_college_computer_science"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_mathematics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "college_mathematics"
2
+ "description": "The following are multiple choice questions (with answers) about college\
3
+ \ mathematics.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_college_mathematics"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_medicine.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "college_medicine"
2
+ "description": "The following are multiple choice questions (with answers) about college\
3
+ \ medicine.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_other"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_college_medicine"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_computer_security.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "computer_security"
2
+ "description": "The following are multiple choice questions (with answers) about computer\
3
+ \ security.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_computer_security"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_conceptual_physics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "conceptual_physics"
2
+ "description": "The following are multiple choice questions (with answers) about conceptual\
3
+ \ physics.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_conceptual_physics"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_econometrics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "econometrics"
2
+ "description": "The following are multiple choice questions (with answers) about econometrics.\n\
3
+ \n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_social_sciences"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_econometrics"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_electrical_engineering.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "electrical_engineering"
2
+ "description": "The following are multiple choice questions (with answers) about electrical\
3
+ \ engineering.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_electrical_engineering"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_elementary_mathematics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "elementary_mathematics"
2
+ "description": "The following are multiple choice questions (with answers) about elementary\
3
+ \ mathematics.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_elementary_mathematics"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_formal_logic.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "formal_logic"
2
+ "description": "The following are multiple choice questions (with answers) about formal\
3
+ \ logic.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_humanities"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_formal_logic"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_global_facts.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "global_facts"
2
+ "description": "The following are multiple choice questions (with answers) about global\
3
+ \ facts.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_other"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_global_facts"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_biology.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_biology"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school biology.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_biology"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_chemistry.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_chemistry"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school chemistry.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_chemistry"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_computer_science.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_computer_science"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school computer science.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_computer_science"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_european_history.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_european_history"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school european history.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_humanities"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_european_history"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_geography.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_geography"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school geography.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_social_sciences"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_geography"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_government_and_politics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_government_and_politics"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school government and politics.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_social_sciences"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_government_and_politics"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_microeconomics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_microeconomics"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school microeconomics.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_social_sciences"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_microeconomics"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_physics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_physics"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school physics.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_physics"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_statistics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_statistics"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school statistics.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_statistics"
lm-evaluation/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_us_history.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_us_history"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school us history.\n\n"
4
+ "group": "mmlu_flan_n_shot_loglikelihood_humanities"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_us_history"