applied-ai-018 commited on
Commit
a5954f5
·
verified ·
1 Parent(s): 99fe61a

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/abstract_narrative_understanding.yaml +4 -0
  2. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/auto_categorization.yaml +4 -0
  3. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/auto_debugging.yaml +4 -0
  4. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/bridging_anaphora_resolution_barqa.yaml +4 -0
  5. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/causal_judgment.yaml +4 -0
  6. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/checkmate_in_one.yaml +4 -0
  7. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cifar10_classification.yaml +4 -0
  8. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/color.yaml +4 -0
  9. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/crass_ai.yaml +4 -0
  10. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cryobiology_spanish.yaml +4 -0
  11. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cryptonite.yaml +4 -0
  12. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/dark_humor_detection.yaml +4 -0
  13. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/date_understanding.yaml +4 -0
  14. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/discourse_marker_prediction.yaml +4 -0
  15. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/dyck_languages.yaml +4 -0
  16. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/english_russian_proverbs.yaml +4 -0
  17. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/evaluating_information_essentiality.yaml +4 -0
  18. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/fact_checker.yaml +4 -0
  19. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/fantasy_reasoning.yaml +4 -0
  20. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/few_shot_nlg.yaml +4 -0
  21. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/gem.yaml +4 -0
  22. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/gender_inclusive_sentences_german.yaml +4 -0
  23. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/general_knowledge.yaml +4 -0
  24. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/geometric_shapes.yaml +4 -0
  25. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/goal_step_wikihow.yaml +4 -0
  26. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/gre_reading_comprehension.yaml +4 -0
  27. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hhh_alignment.yaml +4 -0
  28. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hindi_question_answering.yaml +4 -0
  29. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hindu_knowledge.yaml +4 -0
  30. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hinglish_toxicity.yaml +4 -0
  31. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/identify_odd_metaphor.yaml +4 -0
  32. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/implicatures.yaml +4 -0
  33. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/intersect_geometry.yaml +4 -0
  34. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/kannada.yaml +4 -0
  35. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/known_unknowns.yaml +4 -0
  36. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/language_games.yaml +4 -0
  37. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/linguistic_mappings.yaml +4 -0
  38. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/linguistics_puzzles.yaml +4 -0
  39. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/logic_grid_puzzle.yaml +4 -0
  40. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/logical_fallacy_detection.yaml +4 -0
  41. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/matrixshapes.yaml +4 -0
  42. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/metaphor_understanding.yaml +4 -0
  43. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/minute_mysteries_qa.yaml +4 -0
  44. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/misconceptions.yaml +4 -0
  45. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/misconceptions_russian.yaml +4 -0
  46. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/modified_arithmetic.yaml +4 -0
  47. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/moral_permissibility.yaml +4 -0
  48. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/movie_dialog_same_or_different.yaml +4 -0
  49. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/mult_data_wrangling.yaml +4 -0
  50. lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/multiemo.yaml +4 -0
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/abstract_narrative_understanding.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: abstract_narrative_understanding_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_abstract_narrative_understanding_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/auto_categorization.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: auto_categorization_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_auto_categorization_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/auto_debugging.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: auto_debugging_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_auto_debugging_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/bridging_anaphora_resolution_barqa.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: bridging_anaphora_resolution_barqa_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_bridging_anaphora_resolution_barqa_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/causal_judgment.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: causal_judgment_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_causal_judgment_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/checkmate_in_one.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: checkmate_in_one_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_checkmate_in_one_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cifar10_classification.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: cifar10_classification_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_cifar10_classification_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/color.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: color_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_color_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/crass_ai.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: crass_ai_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_crass_ai_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cryobiology_spanish.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: cryobiology_spanish_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_cryobiology_spanish_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cryptonite.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: cryptonite_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_cryptonite_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/dark_humor_detection.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: dark_humor_detection_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_dark_humor_detection_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/date_understanding.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: date_understanding_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_date_understanding_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/discourse_marker_prediction.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: discourse_marker_prediction_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_discourse_marker_prediction_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/dyck_languages.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: dyck_languages_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_dyck_languages_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/english_russian_proverbs.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: english_russian_proverbs_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_english_russian_proverbs_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/evaluating_information_essentiality.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: evaluating_information_essentiality_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_evaluating_information_essentiality_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/fact_checker.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: fact_checker_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_fact_checker_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/fantasy_reasoning.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: fantasy_reasoning_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_fantasy_reasoning_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/few_shot_nlg.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: few_shot_nlg_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_few_shot_nlg_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/gem.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: gem_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_gem_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/gender_inclusive_sentences_german.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: gender_inclusive_sentences_german_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_gender_inclusive_sentences_german_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/general_knowledge.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: general_knowledge_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_general_knowledge_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/geometric_shapes.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: geometric_shapes_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_geometric_shapes_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/goal_step_wikihow.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: goal_step_wikihow_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_goal_step_wikihow_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/gre_reading_comprehension.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: gre_reading_comprehension_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_gre_reading_comprehension_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hhh_alignment.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: hhh_alignment_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_hhh_alignment_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hindi_question_answering.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: hindi_question_answering_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_hindi_question_answering_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hindu_knowledge.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: hindu_knowledge_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_hindu_knowledge_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hinglish_toxicity.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: hinglish_toxicity_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_hinglish_toxicity_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/identify_odd_metaphor.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: identify_odd_metaphor_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_identify_odd_metaphor_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/implicatures.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: implicatures_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_implicatures_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/intersect_geometry.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: intersect_geometry_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_intersect_geometry_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/kannada.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: kannada_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_kannada_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/known_unknowns.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: known_unknowns_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_known_unknowns_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/language_games.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: language_games_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_language_games_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/linguistic_mappings.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: linguistic_mappings_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_linguistic_mappings_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/linguistics_puzzles.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: linguistics_puzzles_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_linguistics_puzzles_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/logic_grid_puzzle.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: logic_grid_puzzle_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_logic_grid_puzzle_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/logical_fallacy_detection.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: logical_fallacy_detection_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_logical_fallacy_detection_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/matrixshapes.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: matrixshapes_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_matrixshapes_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/metaphor_understanding.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: metaphor_understanding_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_metaphor_understanding_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/minute_mysteries_qa.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: minute_mysteries_qa_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_minute_mysteries_qa_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/misconceptions.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: misconceptions_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_misconceptions_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/misconceptions_russian.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: misconceptions_russian_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_misconceptions_russian_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/modified_arithmetic.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: modified_arithmetic_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_modified_arithmetic_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/moral_permissibility.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: moral_permissibility_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_moral_permissibility_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/movie_dialog_same_or_different.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: movie_dialog_same_or_different_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_movie_dialog_same_or_different_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/mult_data_wrangling.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: mult_data_wrangling_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_mult_data_wrangling_multiple_choice
lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/multiemo.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: multiemo_zero_shot
3
+ include: ../multiple_choice_template_yaml
4
+ task: bigbench_multiemo_multiple_choice