Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/_mmlu.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_anatomy.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_astronomy.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_business_ethics.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_chemistry.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_computer_science.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_mathematics.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_medicine.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_physics.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_computer_security.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_conceptual_physics.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_global_facts.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_biology.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_chemistry.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_government_and_politics.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_mathematics.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_physics.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_human_sexuality.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_international_law.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_jurisprudence.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_logical_fallacies.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_machine_learning.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_marketing.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_medical_genetics.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_miscellaneous.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_nutrition.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_philosophy.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_professional_accounting.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_professional_law.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_professional_psychology.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_public_relations.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_security_studies.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_us_foreign_policy.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/_mmlu.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_abstract_algebra.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_astronomy.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_business_ethics.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_biology.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_mathematics.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_medicine.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_computer_security.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_conceptual_physics.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_elementary_mathematics.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_formal_logic.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_global_facts.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_computer_science.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_geography.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_macroeconomics.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_mathematics.yaml +6 -0
- lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_microeconomics.yaml +6 -0
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/_mmlu.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
group: mmlu_flan_n_shot_generative
|
2 |
+
task:
|
3 |
+
- mmlu_flan_n_shot_generative_stem
|
4 |
+
- mmlu_flan_n_shot_generative_other
|
5 |
+
- mmlu_flan_n_shot_generative_social_sciences
|
6 |
+
- mmlu_flan_n_shot_generative_humanities
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_anatomy.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "anatomy"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about anatomy.\n\
|
3 |
+
\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_stem"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_anatomy"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_astronomy.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "astronomy"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about astronomy.\n\
|
3 |
+
\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_stem"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_astronomy"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_business_ethics.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "business_ethics"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about business\
|
3 |
+
\ ethics.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_other"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_business_ethics"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_chemistry.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "college_chemistry"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about college\
|
3 |
+
\ chemistry.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_stem"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_college_chemistry"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_computer_science.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "college_computer_science"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about college\
|
3 |
+
\ computer science.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_stem"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_college_computer_science"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_mathematics.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "college_mathematics"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about college\
|
3 |
+
\ mathematics.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_stem"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_college_mathematics"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_medicine.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "college_medicine"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about college\
|
3 |
+
\ medicine.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_other"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_college_medicine"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_physics.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "college_physics"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about college\
|
3 |
+
\ physics.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_stem"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_college_physics"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_computer_security.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "computer_security"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about computer\
|
3 |
+
\ security.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_stem"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_computer_security"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_conceptual_physics.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "conceptual_physics"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about conceptual\
|
3 |
+
\ physics.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_stem"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_conceptual_physics"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_global_facts.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "global_facts"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about global\
|
3 |
+
\ facts.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_other"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_global_facts"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_biology.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "high_school_biology"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about high\
|
3 |
+
\ school biology.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_stem"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_high_school_biology"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_chemistry.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "high_school_chemistry"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about high\
|
3 |
+
\ school chemistry.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_stem"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_high_school_chemistry"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_government_and_politics.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "high_school_government_and_politics"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about high\
|
3 |
+
\ school government and politics.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_social_sciences"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_high_school_government_and_politics"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_mathematics.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "high_school_mathematics"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about high\
|
3 |
+
\ school mathematics.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_stem"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_high_school_mathematics"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_physics.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "high_school_physics"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about high\
|
3 |
+
\ school physics.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_stem"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_high_school_physics"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_human_sexuality.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "human_sexuality"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about human\
|
3 |
+
\ sexuality.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_social_sciences"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_human_sexuality"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_international_law.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "international_law"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about international\
|
3 |
+
\ law.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_humanities"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_international_law"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_jurisprudence.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "jurisprudence"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about jurisprudence.\n\
|
3 |
+
\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_humanities"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_jurisprudence"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_logical_fallacies.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "logical_fallacies"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about logical\
|
3 |
+
\ fallacies.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_humanities"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_logical_fallacies"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_machine_learning.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "machine_learning"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about machine\
|
3 |
+
\ learning.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_stem"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_machine_learning"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_marketing.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "marketing"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about marketing.\n\
|
3 |
+
\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_other"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_marketing"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_medical_genetics.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "medical_genetics"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about medical\
|
3 |
+
\ genetics.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_other"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_medical_genetics"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_miscellaneous.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "miscellaneous"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about miscellaneous.\n\
|
3 |
+
\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_other"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_miscellaneous"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_nutrition.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "nutrition"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about nutrition.\n\
|
3 |
+
\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_other"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_nutrition"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_philosophy.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "philosophy"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about philosophy.\n\
|
3 |
+
\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_humanities"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_philosophy"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_professional_accounting.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "professional_accounting"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about professional\
|
3 |
+
\ accounting.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_other"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_professional_accounting"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_professional_law.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "professional_law"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about professional\
|
3 |
+
\ law.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_humanities"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_professional_law"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_professional_psychology.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "professional_psychology"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about professional\
|
3 |
+
\ psychology.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_social_sciences"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_professional_psychology"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_public_relations.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "public_relations"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about public\
|
3 |
+
\ relations.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_social_sciences"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_public_relations"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_security_studies.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "security_studies"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about security\
|
3 |
+
\ studies.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_social_sciences"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_security_studies"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_us_foreign_policy.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "us_foreign_policy"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about us\
|
3 |
+
\ foreign policy.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_generative_social_sciences"
|
5 |
+
"include": "_mmlu_flan_generative_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_generative_us_foreign_policy"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/_mmlu.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
group: mmlu_flan_n_shot_loglikelihood
|
2 |
+
task:
|
3 |
+
- mmlu_flan_n_shot_loglikelihood_stem
|
4 |
+
- mmlu_flan_n_shot_loglikelihood_other
|
5 |
+
- mmlu_flan_n_shot_loglikelihood_social_sciences
|
6 |
+
- mmlu_flan_n_shot_loglikelihood_humanities
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_abstract_algebra.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "abstract_algebra"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about abstract\
|
3 |
+
\ algebra.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_loglikelihood_stem"
|
5 |
+
"include": "_mmlu_flan_loglikelihood_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_loglikelihood_abstract_algebra"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_astronomy.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "astronomy"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about astronomy.\n\
|
3 |
+
\n"
|
4 |
+
"group": "mmlu_flan_n_shot_loglikelihood_stem"
|
5 |
+
"include": "_mmlu_flan_loglikelihood_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_loglikelihood_astronomy"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_business_ethics.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "business_ethics"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about business\
|
3 |
+
\ ethics.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_loglikelihood_other"
|
5 |
+
"include": "_mmlu_flan_loglikelihood_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_loglikelihood_business_ethics"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_biology.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "college_biology"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about college\
|
3 |
+
\ biology.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_loglikelihood_stem"
|
5 |
+
"include": "_mmlu_flan_loglikelihood_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_loglikelihood_college_biology"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_mathematics.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "college_mathematics"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about college\
|
3 |
+
\ mathematics.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_loglikelihood_stem"
|
5 |
+
"include": "_mmlu_flan_loglikelihood_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_loglikelihood_college_mathematics"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_medicine.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "college_medicine"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about college\
|
3 |
+
\ medicine.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_loglikelihood_other"
|
5 |
+
"include": "_mmlu_flan_loglikelihood_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_loglikelihood_college_medicine"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_computer_security.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "computer_security"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about computer\
|
3 |
+
\ security.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_loglikelihood_stem"
|
5 |
+
"include": "_mmlu_flan_loglikelihood_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_loglikelihood_computer_security"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_conceptual_physics.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "conceptual_physics"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about conceptual\
|
3 |
+
\ physics.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_loglikelihood_stem"
|
5 |
+
"include": "_mmlu_flan_loglikelihood_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_loglikelihood_conceptual_physics"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_elementary_mathematics.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "elementary_mathematics"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about elementary\
|
3 |
+
\ mathematics.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_loglikelihood_stem"
|
5 |
+
"include": "_mmlu_flan_loglikelihood_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_loglikelihood_elementary_mathematics"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_formal_logic.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "formal_logic"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about formal\
|
3 |
+
\ logic.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_loglikelihood_humanities"
|
5 |
+
"include": "_mmlu_flan_loglikelihood_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_loglikelihood_formal_logic"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_global_facts.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "global_facts"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about global\
|
3 |
+
\ facts.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_loglikelihood_other"
|
5 |
+
"include": "_mmlu_flan_loglikelihood_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_loglikelihood_global_facts"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_computer_science.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "high_school_computer_science"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about high\
|
3 |
+
\ school computer science.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_loglikelihood_stem"
|
5 |
+
"include": "_mmlu_flan_loglikelihood_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_loglikelihood_high_school_computer_science"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_geography.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "high_school_geography"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about high\
|
3 |
+
\ school geography.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_loglikelihood_social_sciences"
|
5 |
+
"include": "_mmlu_flan_loglikelihood_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_loglikelihood_high_school_geography"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_macroeconomics.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "high_school_macroeconomics"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about high\
|
3 |
+
\ school macroeconomics.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_loglikelihood_social_sciences"
|
5 |
+
"include": "_mmlu_flan_loglikelihood_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_loglikelihood_high_school_macroeconomics"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_mathematics.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "high_school_mathematics"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about high\
|
3 |
+
\ school mathematics.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_loglikelihood_stem"
|
5 |
+
"include": "_mmlu_flan_loglikelihood_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_loglikelihood_high_school_mathematics"
|
lm-evaluation/build/lib/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_microeconomics.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"dataset_name": "high_school_microeconomics"
|
2 |
+
"description": "The following are multiple choice questions (with answers) about high\
|
3 |
+
\ school microeconomics.\n\n"
|
4 |
+
"group": "mmlu_flan_n_shot_loglikelihood_social_sciences"
|
5 |
+
"include": "_mmlu_flan_loglikelihood_template_yaml"
|
6 |
+
"task": "mmlu_flan_n_shot_loglikelihood_high_school_microeconomics"
|