Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/analogical_similarity.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/analytic_entailment.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/ascii_word_recognition.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/auto_categorization.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cause_and_effect.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/checkmate_in_one.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/chess_state_tracking.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/chinese_remainder_theorem.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/codenames.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/color.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/common_morpheme.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conceptual_combinations.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conlang_translation.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/contextual_parametric_knowledge_conflicts.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/crash_blossom.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cryptonite.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cs_algorithms.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/date_understanding.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/disfl_qa.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/elementary_math_qa.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/english_proverbs.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/english_russian_proverbs.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/entailed_polarity.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/epistemic_reasoning.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/fantasy_reasoning.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/few_shot_nlg.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/figure_of_speech_detection.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/gender_inclusive_sentences_german.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/general_knowledge.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/gre_reading_comprehension.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hhh_alignment.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hindu_knowledge.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hinglish_toxicity.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/human_organs_senses.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/identify_math_theorems.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/international_phonetic_alphabet_nli.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kanji_ascii.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kannada.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/key_value_maps.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/language_games.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/linguistic_mappings.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/list_functions.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_args.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_fallacy_detection.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_sequence.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/matrixshapes.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/minute_mysteries_qa.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/mnist_ascii.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/multiemo.yaml +4 -0
- lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/natural_instructions.yaml +4 -0
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/analogical_similarity.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: analogical_similarity_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_analogical_similarity_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/analytic_entailment.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: analytic_entailment_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_analytic_entailment_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/ascii_word_recognition.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: ascii_word_recognition_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_ascii_word_recognition_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/auto_categorization.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: auto_categorization_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_auto_categorization_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cause_and_effect.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: cause_and_effect_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_cause_and_effect_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/checkmate_in_one.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: checkmate_in_one_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_checkmate_in_one_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/chess_state_tracking.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: chess_state_tracking_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_chess_state_tracking_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/chinese_remainder_theorem.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: chinese_remainder_theorem_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_chinese_remainder_theorem_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/codenames.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: codenames_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_codenames_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/color.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: color_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_color_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/common_morpheme.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: common_morpheme_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_common_morpheme_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conceptual_combinations.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: conceptual_combinations_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_conceptual_combinations_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conlang_translation.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: conlang_translation_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_conlang_translation_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/contextual_parametric_knowledge_conflicts.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: contextual_parametric_knowledge_conflicts_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_contextual_parametric_knowledge_conflicts_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/crash_blossom.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: crash_blossom_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_crash_blossom_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cryptonite.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: cryptonite_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_cryptonite_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cs_algorithms.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: cs_algorithms_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_cs_algorithms_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/date_understanding.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: date_understanding_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_date_understanding_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/disfl_qa.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: disfl_qa_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_disfl_qa_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/elementary_math_qa.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: elementary_math_qa_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_elementary_math_qa_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/english_proverbs.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: english_proverbs_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_english_proverbs_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/english_russian_proverbs.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: english_russian_proverbs_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_english_russian_proverbs_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/entailed_polarity.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: entailed_polarity_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_entailed_polarity_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/epistemic_reasoning.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: epistemic_reasoning_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_epistemic_reasoning_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/fantasy_reasoning.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: fantasy_reasoning_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_fantasy_reasoning_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/few_shot_nlg.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: few_shot_nlg_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_few_shot_nlg_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/figure_of_speech_detection.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: figure_of_speech_detection_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_figure_of_speech_detection_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/gender_inclusive_sentences_german.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: gender_inclusive_sentences_german_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_gender_inclusive_sentences_german_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/general_knowledge.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: general_knowledge_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_general_knowledge_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/gre_reading_comprehension.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: gre_reading_comprehension_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_gre_reading_comprehension_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hhh_alignment.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: hhh_alignment_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_hhh_alignment_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hindu_knowledge.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: hindu_knowledge_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_hindu_knowledge_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/hinglish_toxicity.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: hinglish_toxicity_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_hinglish_toxicity_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/human_organs_senses.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: human_organs_senses_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_human_organs_senses_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/identify_math_theorems.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: identify_math_theorems_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_identify_math_theorems_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/international_phonetic_alphabet_nli.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: international_phonetic_alphabet_nli_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_international_phonetic_alphabet_nli_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kanji_ascii.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: kanji_ascii_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_kanji_ascii_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kannada.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: kannada_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_kannada_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/key_value_maps.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: key_value_maps_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_key_value_maps_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/language_games.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: language_games_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_language_games_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/linguistic_mappings.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: linguistic_mappings_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_linguistic_mappings_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/list_functions.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: list_functions_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_list_functions_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_args.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: logical_args_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_logical_args_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_fallacy_detection.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: logical_fallacy_detection_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_logical_fallacy_detection_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_sequence.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: logical_sequence_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_logical_sequence_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/matrixshapes.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: matrixshapes_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_matrixshapes_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/minute_mysteries_qa.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: minute_mysteries_qa_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_minute_mysteries_qa_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/mnist_ascii.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: mnist_ascii_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_mnist_ascii_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/multiemo.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: multiemo_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_multiemo_generate_until
|
lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/natural_instructions.yaml
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Generated by utils.py
|
2 |
+
dataset_name: natural_instructions_zero_shot
|
3 |
+
include: ../generate_until_template_yaml
|
4 |
+
task: bigbench_natural_instructions_generate_until
|