applied-ai-018 commited on
Commit
8c47bdb
·
verified ·
1 Parent(s): b613cde

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/abstract_narrative_understanding.yaml +4 -0
  2. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/anachronisms.yaml +4 -0
  3. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/analogical_similarity.yaml +4 -0
  4. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/analytic_entailment.yaml +4 -0
  5. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/arithmetic.yaml +4 -0
  6. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/ascii_word_recognition.yaml +4 -0
  7. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/bbq_lite_json.yaml +4 -0
  8. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/bridging_anaphora_resolution_barqa.yaml +4 -0
  9. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/causal_judgment.yaml +4 -0
  10. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/contextual_parametric_knowledge_conflicts.yaml +4 -0
  11. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/crash_blossom.yaml +4 -0
  12. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/cs_algorithms.yaml +4 -0
  13. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/date_understanding.yaml +4 -0
  14. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/disambiguation_qa.yaml +4 -0
  15. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/emojis_emotion_prediction.yaml +4 -0
  16. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/english_proverbs.yaml +4 -0
  17. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/entailed_polarity.yaml +4 -0
  18. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/epistemic_reasoning.yaml +4 -0
  19. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/evaluating_information_essentiality.yaml +4 -0
  20. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/fact_checker.yaml +4 -0
  21. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/fantasy_reasoning.yaml +4 -0
  22. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/formal_fallacies_syllogisms_negation.yaml +4 -0
  23. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/hindu_knowledge.yaml +4 -0
  24. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/hyperbaton.yaml +4 -0
  25. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/identify_math_theorems.yaml +4 -0
  26. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/implicit_relations.yaml +4 -0
  27. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/intent_recognition.yaml +4 -0
  28. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/irony_identification.yaml +4 -0
  29. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/key_value_maps.yaml +4 -0
  30. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/language_games.yaml +4 -0
  31. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/linguistics_puzzles.yaml +4 -0
  32. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/logic_grid_puzzle.yaml +4 -0
  33. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/metaphor_understanding.yaml +4 -0
  34. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/misconceptions_russian.yaml +4 -0
  35. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/movie_dialog_same_or_different.yaml +4 -0
  36. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/movie_recommendation.yaml +4 -0
  37. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/mult_data_wrangling.yaml +4 -0
  38. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/navigate.yaml +4 -0
  39. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/nonsense_words_grammar.yaml +4 -0
  40. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/novel_concepts.yaml +4 -0
  41. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/odd_one_out.yaml +4 -0
  42. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/persian_idioms.yaml +4 -0
  43. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/phrase_relatedness.yaml +4 -0
  44. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/physics.yaml +4 -0
  45. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/physics_questions.yaml +4 -0
  46. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/play_dialog_same_or_different.yaml +4 -0
  47. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/presuppositions_as_nli.yaml +4 -0
  48. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/qa_wikidata.yaml +4 -0
  49. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/reasoning_about_colored_objects.yaml +4 -0
  50. lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/repeat_copy_logic.yaml +4 -0
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/abstract_narrative_understanding.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: abstract_narrative_understanding_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_abstract_narrative_understanding_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/anachronisms.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: anachronisms_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_anachronisms_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/analogical_similarity.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: analogical_similarity_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_analogical_similarity_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/analytic_entailment.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: analytic_entailment_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_analytic_entailment_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/arithmetic.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: arithmetic_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_arithmetic_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/ascii_word_recognition.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: ascii_word_recognition_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_ascii_word_recognition_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/bbq_lite_json.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: bbq_lite_json_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_bbq_lite_json_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/bridging_anaphora_resolution_barqa.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: bridging_anaphora_resolution_barqa_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_bridging_anaphora_resolution_barqa_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/causal_judgment.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: causal_judgment_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_causal_judgment_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/contextual_parametric_knowledge_conflicts.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: contextual_parametric_knowledge_conflicts_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_contextual_parametric_knowledge_conflicts_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/crash_blossom.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: crash_blossom_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_crash_blossom_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/cs_algorithms.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: cs_algorithms_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_cs_algorithms_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/date_understanding.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: date_understanding_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_date_understanding_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/disambiguation_qa.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: disambiguation_qa_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_disambiguation_qa_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/emojis_emotion_prediction.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: emojis_emotion_prediction_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_emojis_emotion_prediction_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/english_proverbs.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: english_proverbs_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_english_proverbs_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/entailed_polarity.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: entailed_polarity_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_entailed_polarity_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/epistemic_reasoning.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: epistemic_reasoning_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_epistemic_reasoning_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/evaluating_information_essentiality.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: evaluating_information_essentiality_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_evaluating_information_essentiality_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/fact_checker.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: fact_checker_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_fact_checker_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/fantasy_reasoning.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: fantasy_reasoning_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_fantasy_reasoning_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/formal_fallacies_syllogisms_negation.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: formal_fallacies_syllogisms_negation_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_formal_fallacies_syllogisms_negation_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/hindu_knowledge.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: hindu_knowledge_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_hindu_knowledge_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/hyperbaton.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: hyperbaton_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_hyperbaton_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/identify_math_theorems.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: identify_math_theorems_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_identify_math_theorems_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/implicit_relations.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: implicit_relations_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_implicit_relations_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/intent_recognition.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: intent_recognition_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_intent_recognition_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/irony_identification.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: irony_identification_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_irony_identification_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/key_value_maps.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: key_value_maps_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_key_value_maps_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/language_games.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: language_games_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_language_games_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/linguistics_puzzles.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: linguistics_puzzles_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_linguistics_puzzles_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/logic_grid_puzzle.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: logic_grid_puzzle_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_logic_grid_puzzle_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/metaphor_understanding.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: metaphor_understanding_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_metaphor_understanding_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/misconceptions_russian.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: misconceptions_russian_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_misconceptions_russian_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/movie_dialog_same_or_different.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: movie_dialog_same_or_different_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_movie_dialog_same_or_different_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/movie_recommendation.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: movie_recommendation_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_movie_recommendation_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/mult_data_wrangling.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: mult_data_wrangling_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_mult_data_wrangling_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/navigate.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: navigate_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_navigate_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/nonsense_words_grammar.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: nonsense_words_grammar_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_nonsense_words_grammar_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/novel_concepts.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: novel_concepts_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_novel_concepts_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/odd_one_out.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: odd_one_out_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_odd_one_out_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/persian_idioms.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: persian_idioms_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_persian_idioms_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/phrase_relatedness.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: phrase_relatedness_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_phrase_relatedness_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/physics.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: physics_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_physics_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/physics_questions.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: physics_questions_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_physics_questions_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/play_dialog_same_or_different.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: play_dialog_same_or_different_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_play_dialog_same_or_different_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/presuppositions_as_nli.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: presuppositions_as_nli_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_presuppositions_as_nli_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/qa_wikidata.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: qa_wikidata_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_qa_wikidata_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/reasoning_about_colored_objects.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: reasoning_about_colored_objects_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_reasoning_about_colored_objects_multiple_choice
lm-evaluation/build/lib/lm_eval/tasks/bigbench/multiple_choice/repeat_copy_logic.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: repeat_copy_logic_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_repeat_copy_logic_multiple_choice