applied-ai-018 commited on
Commit
ed614f4
·
verified ·
1 Parent(s): 8db832d

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/analogical_similarity.yaml +4 -0
  2. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/analytic_entailment.yaml +4 -0
  3. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/ascii_word_recognition.yaml +4 -0
  4. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/auto_categorization.yaml +4 -0
  5. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cause_and_effect.yaml +4 -0
  6. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/checkmate_in_one.yaml +4 -0
  7. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/chess_state_tracking.yaml +4 -0
  8. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/chinese_remainder_theorem.yaml +4 -0
  9. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/codenames.yaml +4 -0
  10. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/color.yaml +4 -0
  11. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/common_morpheme.yaml +4 -0
  12. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conceptual_combinations.yaml +4 -0
  13. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conlang_translation.yaml +4 -0
  14. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/contextual_parametric_knowledge_conflicts.yaml +4 -0
  15. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/crash_blossom.yaml +4 -0
  16. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cryptonite.yaml +4 -0
  17. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cs_algorithms.yaml +4 -0
  18. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/date_understanding.yaml +4 -0
  19. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/disfl_qa.yaml +4 -0
  20. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/elementary_math_qa.yaml +4 -0
  21. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/english_proverbs.yaml +4 -0
  22. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/english_russian_proverbs.yaml +4 -0
  23. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/entailed_polarity.yaml +4 -0
  24. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/epistemic_reasoning.yaml +4 -0
  25. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/fantasy_reasoning.yaml +4 -0
  26. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/few_shot_nlg.yaml +4 -0
  27. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/figure_of_speech_detection.yaml +4 -0
  28. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/gender_inclusive_sentences_german.yaml +4 -0
  29. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/general_knowledge.yaml +4 -0
  30. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/gre_reading_comprehension.yaml +4 -0
  31. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hhh_alignment.yaml +4 -0
  32. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hindu_knowledge.yaml +4 -0
  33. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hinglish_toxicity.yaml +4 -0
  34. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/human_organs_senses.yaml +4 -0
  35. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/identify_math_theorems.yaml +4 -0
  36. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/international_phonetic_alphabet_nli.yaml +4 -0
  37. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kanji_ascii.yaml +4 -0
  38. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kannada.yaml +4 -0
  39. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/key_value_maps.yaml +4 -0
  40. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/language_games.yaml +4 -0
  41. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/linguistic_mappings.yaml +4 -0
  42. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/list_functions.yaml +4 -0
  43. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_args.yaml +4 -0
  44. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_fallacy_detection.yaml +4 -0
  45. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_sequence.yaml +4 -0
  46. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/matrixshapes.yaml +4 -0
  47. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/minute_mysteries_qa.yaml +4 -0
  48. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/mnist_ascii.yaml +4 -0
  49. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/multiemo.yaml +4 -0
  50. lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/natural_instructions.yaml +4 -0
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/analogical_similarity.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: analogical_similarity_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_analogical_similarity_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/analytic_entailment.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: analytic_entailment_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_analytic_entailment_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/ascii_word_recognition.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: ascii_word_recognition_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_ascii_word_recognition_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/auto_categorization.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: auto_categorization_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_auto_categorization_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cause_and_effect.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: cause_and_effect_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_cause_and_effect_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/checkmate_in_one.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: checkmate_in_one_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_checkmate_in_one_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/chess_state_tracking.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: chess_state_tracking_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_chess_state_tracking_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/chinese_remainder_theorem.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: chinese_remainder_theorem_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_chinese_remainder_theorem_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/codenames.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: codenames_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_codenames_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/color.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: color_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_color_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/common_morpheme.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: common_morpheme_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_common_morpheme_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conceptual_combinations.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: conceptual_combinations_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_conceptual_combinations_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conlang_translation.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: conlang_translation_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_conlang_translation_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/contextual_parametric_knowledge_conflicts.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: contextual_parametric_knowledge_conflicts_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_contextual_parametric_knowledge_conflicts_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/crash_blossom.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: crash_blossom_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_crash_blossom_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cryptonite.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: cryptonite_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_cryptonite_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cs_algorithms.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: cs_algorithms_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_cs_algorithms_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/date_understanding.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: date_understanding_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_date_understanding_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/disfl_qa.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: disfl_qa_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_disfl_qa_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/elementary_math_qa.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: elementary_math_qa_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_elementary_math_qa_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/english_proverbs.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: english_proverbs_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_english_proverbs_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/english_russian_proverbs.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: english_russian_proverbs_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_english_russian_proverbs_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/entailed_polarity.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: entailed_polarity_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_entailed_polarity_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/epistemic_reasoning.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: epistemic_reasoning_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_epistemic_reasoning_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/fantasy_reasoning.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: fantasy_reasoning_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_fantasy_reasoning_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/few_shot_nlg.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: few_shot_nlg_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_few_shot_nlg_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/figure_of_speech_detection.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: figure_of_speech_detection_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_figure_of_speech_detection_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/gender_inclusive_sentences_german.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: gender_inclusive_sentences_german_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_gender_inclusive_sentences_german_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/general_knowledge.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: general_knowledge_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_general_knowledge_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/gre_reading_comprehension.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: gre_reading_comprehension_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_gre_reading_comprehension_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hhh_alignment.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: hhh_alignment_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_hhh_alignment_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hindu_knowledge.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: hindu_knowledge_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_hindu_knowledge_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hinglish_toxicity.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: hinglish_toxicity_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_hinglish_toxicity_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/human_organs_senses.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: human_organs_senses_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_human_organs_senses_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/identify_math_theorems.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: identify_math_theorems_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_identify_math_theorems_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/international_phonetic_alphabet_nli.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: international_phonetic_alphabet_nli_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_international_phonetic_alphabet_nli_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kanji_ascii.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: kanji_ascii_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_kanji_ascii_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kannada.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: kannada_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_kannada_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/key_value_maps.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: key_value_maps_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_key_value_maps_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/language_games.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: language_games_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_language_games_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/linguistic_mappings.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: linguistic_mappings_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_linguistic_mappings_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/list_functions.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: list_functions_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_list_functions_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_args.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: logical_args_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_logical_args_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_fallacy_detection.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: logical_fallacy_detection_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_logical_fallacy_detection_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_sequence.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: logical_sequence_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_logical_sequence_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/matrixshapes.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: matrixshapes_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_matrixshapes_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/minute_mysteries_qa.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: minute_mysteries_qa_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_minute_mysteries_qa_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/mnist_ascii.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: mnist_ascii_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_mnist_ascii_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/multiemo.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: multiemo_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_multiemo_generate_until
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/natural_instructions.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: natural_instructions_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_natural_instructions_generate_until