Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/abstract_narrative_understanding.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/anachronisms.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/analogical_similarity.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/analytic_entailment.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/arithmetic.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/ascii_word_recognition.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/bbq_lite_json.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/bridging_anaphora_resolution_barqa.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/causal_judgment.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/contextual_parametric_knowledge_conflicts.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/crash_blossom.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/cs_algorithms.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/date_understanding.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/disambiguation_qa.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/emojis_emotion_prediction.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/english_proverbs.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/entailed_polarity.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/epistemic_reasoning.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/evaluating_information_essentiality.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/fact_checker.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/fantasy_reasoning.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/formal_fallacies_syllogisms_negation.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/hindu_knowledge.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/hyperbaton.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/identify_math_theorems.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/implicit_relations.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/intent_recognition.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/irony_identification.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/key_value_maps.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/language_games.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/linguistics_puzzles.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/logic_grid_puzzle.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/metaphor_understanding.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/misconceptions_russian.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/movie_dialog_same_or_different.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/movie_recommendation.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/mult_data_wrangling.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/navigate.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/nonsense_words_grammar.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/novel_concepts.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/odd_one_out.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/persian_idioms.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/phrase_relatedness.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/physics.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/physics_questions.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/play_dialog_same_or_different.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/presuppositions_as_nli.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/qa_wikidata.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/reasoning_about_colored_objects.yaml +4 -0
- lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/repeat_copy_logic.yaml +4 -0
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/abstract_narrative_understanding.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: abstract_narrative_understanding_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_abstract_narrative_understanding_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/anachronisms.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: anachronisms_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_anachronisms_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/analogical_similarity.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: analogical_similarity_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_analogical_similarity_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/analytic_entailment.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: analytic_entailment_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_analytic_entailment_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/arithmetic.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: arithmetic_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_arithmetic_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/ascii_word_recognition.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: ascii_word_recognition_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_ascii_word_recognition_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/bbq_lite_json.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: bbq_lite_json_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_bbq_lite_json_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/bridging_anaphora_resolution_barqa.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: bridging_anaphora_resolution_barqa_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_bridging_anaphora_resolution_barqa_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/causal_judgment.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: causal_judgment_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_causal_judgment_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/contextual_parametric_knowledge_conflicts.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: contextual_parametric_knowledge_conflicts_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_contextual_parametric_knowledge_conflicts_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/crash_blossom.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: crash_blossom_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_crash_blossom_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/cs_algorithms.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: cs_algorithms_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_cs_algorithms_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/date_understanding.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: date_understanding_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_date_understanding_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/disambiguation_qa.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: disambiguation_qa_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_disambiguation_qa_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/emojis_emotion_prediction.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: emojis_emotion_prediction_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_emojis_emotion_prediction_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/english_proverbs.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: english_proverbs_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_english_proverbs_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/entailed_polarity.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: entailed_polarity_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_entailed_polarity_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/epistemic_reasoning.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: epistemic_reasoning_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_epistemic_reasoning_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/evaluating_information_essentiality.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: evaluating_information_essentiality_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_evaluating_information_essentiality_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/fact_checker.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: fact_checker_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_fact_checker_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/fantasy_reasoning.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: fantasy_reasoning_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_fantasy_reasoning_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/formal_fallacies_syllogisms_negation.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: formal_fallacies_syllogisms_negation_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_formal_fallacies_syllogisms_negation_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/hindu_knowledge.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: hindu_knowledge_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_hindu_knowledge_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/hyperbaton.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: hyperbaton_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_hyperbaton_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/identify_math_theorems.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: identify_math_theorems_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_identify_math_theorems_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/implicit_relations.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: implicit_relations_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_implicit_relations_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/intent_recognition.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: intent_recognition_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_intent_recognition_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/irony_identification.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: irony_identification_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_irony_identification_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/key_value_maps.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: key_value_maps_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_key_value_maps_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/language_games.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: language_games_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_language_games_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/linguistics_puzzles.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: linguistics_puzzles_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_linguistics_puzzles_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/logic_grid_puzzle.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: logic_grid_puzzle_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_logic_grid_puzzle_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/metaphor_understanding.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: metaphor_understanding_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_metaphor_understanding_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/misconceptions_russian.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: misconceptions_russian_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_misconceptions_russian_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/movie_dialog_same_or_different.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: movie_dialog_same_or_different_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_movie_dialog_same_or_different_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/movie_recommendation.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: movie_recommendation_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_movie_recommendation_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/mult_data_wrangling.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: mult_data_wrangling_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_mult_data_wrangling_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/navigate.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: navigate_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_navigate_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/nonsense_words_grammar.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: nonsense_words_grammar_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_nonsense_words_grammar_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/novel_concepts.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: novel_concepts_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_novel_concepts_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/odd_one_out.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: odd_one_out_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_odd_one_out_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/persian_idioms.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: persian_idioms_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_persian_idioms_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/phrase_relatedness.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: phrase_relatedness_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_phrase_relatedness_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/physics.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: physics_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_physics_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/physics_questions.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: physics_questions_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_physics_questions_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/play_dialog_same_or_different.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: play_dialog_same_or_different_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_play_dialog_same_or_different_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/presuppositions_as_nli.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: presuppositions_as_nli_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_presuppositions_as_nli_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/qa_wikidata.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: qa_wikidata_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_qa_wikidata_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/reasoning_about_colored_objects.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: reasoning_about_colored_objects_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_reasoning_about_colored_objects_multiple_choice
|
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/repeat_copy_logic.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: repeat_copy_logic_zero_shot
|
3 |
+
include: ../multiple_choice_template_yaml
|
4 |
+
task: bigbench_repeat_copy_logic_multiple_choice
|