Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/authorship_verification.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/causal_judgement.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/code_line_description.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/conlang_translation.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/crass_ai.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/cryobiology_spanish.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/dark_humor_detection.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/discourse_marker_prediction.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/dyck_languages.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/emoji_movie.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/entailed_polarity_hindi.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/geometric_shapes.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/goal_step_wikihow.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/hindi_question_answering.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/hinglish_toxicity.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/human_organs_senses.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/identify_odd_metaphor.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/international_phonetic_alphabet_nli.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/international_phonetic_alphabet_transliterate.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/kanji_ascii.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/kannada.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/list_functions.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/logical_args.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/logical_deduction.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/logical_sequence.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/mathematical_induction.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/minute_mysteries_qa.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/misconceptions.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/moral_permissibility.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/natural_instructions.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/object_counting.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/operators.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/parsinlu_reading_comprehension.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/penguins_in_a_table.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/physical_intuition.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/polish_sequence_labeling.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/question_selection.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/riddle_sense.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/semantic_parsing_in_context_sparc.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/snarks.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/social_support.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/understanding_fables.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/unit_interpretation.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/which_wiki_edit.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/winowhy.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/crows_pairs/crows_pairs_english_age.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/crows_pairs/crows_pairs_english_disability.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/crows_pairs/crows_pairs_english_nationality.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/crows_pairs/crows_pairs_english_religion.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/crows_pairs/crows_pairs_french.yaml +3 -0
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/authorship_verification.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: authorship_verification_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_authorship_verification_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/causal_judgement.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: causal_judgment_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_causal_judgement_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/code_line_description.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: code_line_description_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_code_line_description_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/conlang_translation.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: conlang_translation_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_conlang_translation_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/crass_ai.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: crass_ai_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_crass_ai_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/cryobiology_spanish.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: cryobiology_spanish_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_cryobiology_spanish_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/dark_humor_detection.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: dark_humor_detection_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_dark_humor_detection_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/discourse_marker_prediction.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: discourse_marker_prediction_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_discourse_marker_prediction_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/dyck_languages.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: dyck_languages_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_dyck_languages_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/emoji_movie.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: emoji_movie_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_emoji_movie_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/entailed_polarity_hindi.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: entailed_polarity_hindi_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_entailed_polarity_hindi_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/geometric_shapes.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: geometric_shapes_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_geometric_shapes_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/goal_step_wikihow.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: goal_step_wikihow_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_goal_step_wikihow_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/hindi_question_answering.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: hindi_question_answering_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_hindi_question_answering_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/hinglish_toxicity.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: hinglish_toxicity_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_hinglish_toxicity_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/human_organs_senses.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: human_organs_senses_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_human_organs_senses_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/identify_odd_metaphor.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: identify_odd_metaphor_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_identify_odd_metaphor_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/international_phonetic_alphabet_nli.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: international_phonetic_alphabet_nli_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_international_phonetic_alphabet_nli_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/international_phonetic_alphabet_transliterate.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: international_phonetic_alphabet_transliterate_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_international_phonetic_alphabet_transliterate_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/kanji_ascii.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: kanji_ascii_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_kanji_ascii_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/kannada.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: kannada_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_kannada_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/list_functions.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: list_functions_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_list_functions_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/logical_args.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: logical_args_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_logical_args_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/logical_deduction.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: logical_deduction_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_logical_deduction_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/logical_sequence.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: logical_sequence_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_logical_sequence_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/mathematical_induction.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: mathematical_induction_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_mathematical_induction_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/minute_mysteries_qa.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: minute_mysteries_qa_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_minute_mysteries_qa_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/misconceptions.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: misconceptions_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_misconceptions_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/moral_permissibility.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: moral_permissibility_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_moral_permissibility_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/natural_instructions.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: natural_instructions_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_natural_instructions_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/object_counting.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: object_counting_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_object_counting_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/operators.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: operators_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_operators_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/parsinlu_reading_comprehension.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: parsinlu_reading_comprehension_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_parsinlu_reading_comprehension_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/penguins_in_a_table.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: penguins_in_a_table_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_penguins_in_a_table_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/physical_intuition.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: physical_intuition_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_physical_intuition_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/polish_sequence_labeling.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: polish_sequence_labeling_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_polish_sequence_labeling_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/question_selection.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: question_selection_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_question_selection_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/riddle_sense.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: riddle_sense_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_riddle_sense_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/semantic_parsing_in_context_sparc.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: semantic_parsing_in_context_sparc_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_semantic_parsing_in_context_sparc_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/snarks.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: snarks_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_snarks_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/social_support.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: social_support_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_social_support_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/understanding_fables.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: understanding_fables_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_understanding_fables_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/unit_interpretation.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: unit_interpretation_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_unit_interpretation_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/which_wiki_edit.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: which_wiki_edit_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_which_wiki_edit_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/winowhy.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: winowhy_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_winowhy_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/crows_pairs/crows_pairs_english_age.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
include: crows_pairs_english.yaml
|
2 |
+
task: crows_pairs_english_age
|
3 |
+
dataset_name: english
|
4 |
+
process_docs: !function utils.filter_age
|
lm-evaluation/build/lib/lm_eval/tasks/crows_pairs/crows_pairs_english_disability.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
include: crows_pairs_english.yaml
|
2 |
+
task: crows_pairs_english_disability
|
3 |
+
dataset_name: english
|
4 |
+
process_docs: !function utils.filter_disability
|
lm-evaluation/build/lib/lm_eval/tasks/crows_pairs/crows_pairs_english_nationality.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
include: crows_pairs_english.yaml
|
2 |
+
task: crows_pairs_english_nationality
|
3 |
+
dataset_name: english
|
4 |
+
process_docs: !function utils.filter_nationality
|
lm-evaluation/build/lib/lm_eval/tasks/crows_pairs/crows_pairs_english_religion.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
include: crows_pairs_english.yaml
|
2 |
+
task: crows_pairs_english_religion
|
3 |
+
dataset_name: english
|
4 |
+
process_docs: !function utils.filter_religion
|
lm-evaluation/build/lib/lm_eval/tasks/crows_pairs/crows_pairs_french.yaml
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
include: crows_pairs_english.yaml
|
2 |
+
task: crows_pairs_french
|
3 |
+
dataset_name: french
|