Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/arithmetic.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/ascii_word_recognition.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/authorship_verification.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/auto_categorization.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/bbq_lite_json.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/causal_judgment.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/checkmate_in_one.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/code_line_description.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conceptual_combinations.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conlang_translation.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/contextual_parametric_knowledge_conflicts.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cs_algorithms.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/dark_humor_detection.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/disambiguation_qa.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/discourse_marker_prediction.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/emoji_movie.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/emojis_emotion_prediction.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/epistemic_reasoning.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/formal_fallacies_syllogisms_negation.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/general_knowledge.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/human_organs_senses.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/implicatures.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/international_phonetic_alphabet_transliterate.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/intersect_geometry.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kanji_ascii.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/linguistic_mappings.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/linguistics_puzzles.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/list_functions.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_args.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_deduction.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_sequence.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/metaphor_boolean.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/metaphor_understanding.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/minute_mysteries_qa.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/misconceptions.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/misconceptions_russian.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/mnist_ascii.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/movie_dialog_same_or_different.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/movie_recommendation.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/multiemo.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/novel_concepts.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/object_counting.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/paragraph_segmentation.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/parsinlu_qa.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/persian_idioms.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physical_intuition.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physics.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physics_questions.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/presuppositions_as_nli.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/qa_wikidata.yaml +4 -0
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/arithmetic.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: arithmetic_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_arithmetic_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/ascii_word_recognition.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: ascii_word_recognition_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_ascii_word_recognition_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/authorship_verification.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: authorship_verification_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_authorship_verification_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/auto_categorization.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: auto_categorization_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_auto_categorization_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/bbq_lite_json.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: bbq_lite_json_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_bbq_lite_json_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/causal_judgment.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: causal_judgment_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_causal_judgment_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/checkmate_in_one.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: checkmate_in_one_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_checkmate_in_one_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/code_line_description.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: code_line_description_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_code_line_description_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conceptual_combinations.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: conceptual_combinations_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_conceptual_combinations_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conlang_translation.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: conlang_translation_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_conlang_translation_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/contextual_parametric_knowledge_conflicts.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: contextual_parametric_knowledge_conflicts_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_contextual_parametric_knowledge_conflicts_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cs_algorithms.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: cs_algorithms_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_cs_algorithms_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/dark_humor_detection.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: dark_humor_detection_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_dark_humor_detection_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/disambiguation_qa.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: disambiguation_qa_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_disambiguation_qa_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/discourse_marker_prediction.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: discourse_marker_prediction_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_discourse_marker_prediction_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/emoji_movie.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: emoji_movie_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_emoji_movie_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/emojis_emotion_prediction.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: emojis_emotion_prediction_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_emojis_emotion_prediction_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/epistemic_reasoning.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: epistemic_reasoning_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_epistemic_reasoning_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/formal_fallacies_syllogisms_negation.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: formal_fallacies_syllogisms_negation_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_formal_fallacies_syllogisms_negation_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/general_knowledge.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: general_knowledge_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_general_knowledge_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/human_organs_senses.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: human_organs_senses_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_human_organs_senses_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/implicatures.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: implicatures_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_implicatures_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/international_phonetic_alphabet_transliterate.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: international_phonetic_alphabet_transliterate_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_international_phonetic_alphabet_transliterate_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/intersect_geometry.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: intersect_geometry_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_intersect_geometry_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kanji_ascii.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: kanji_ascii_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_kanji_ascii_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/linguistic_mappings.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: linguistic_mappings_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_linguistic_mappings_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/linguistics_puzzles.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: linguistics_puzzles_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_linguistics_puzzles_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/list_functions.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: list_functions_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_list_functions_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_args.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: logical_args_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_logical_args_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_deduction.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: logical_deduction_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_logical_deduction_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_sequence.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: logical_sequence_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_logical_sequence_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/metaphor_boolean.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: metaphor_boolean_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_metaphor_boolean_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/metaphor_understanding.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: metaphor_understanding_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_metaphor_understanding_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/minute_mysteries_qa.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: minute_mysteries_qa_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_minute_mysteries_qa_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/misconceptions.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: misconceptions_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_misconceptions_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/misconceptions_russian.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: misconceptions_russian_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_misconceptions_russian_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/mnist_ascii.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: mnist_ascii_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_mnist_ascii_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/movie_dialog_same_or_different.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: movie_dialog_same_or_different_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_movie_dialog_same_or_different_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/movie_recommendation.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: movie_recommendation_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_movie_recommendation_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/multiemo.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: multiemo_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_multiemo_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/novel_concepts.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: novel_concepts_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_novel_concepts_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/object_counting.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: object_counting_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_object_counting_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/paragraph_segmentation.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: paragraph_segmentation_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_paragraph_segmentation_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/parsinlu_qa.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: parsinlu_qa_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_parsinlu_qa_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/persian_idioms.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: persian_idioms_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_persian_idioms_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physical_intuition.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: physical_intuition_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_physical_intuition_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physics.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: physics_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_physics_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physics_questions.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: physics_questions_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_physics_questions_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/presuppositions_as_nli.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: presuppositions_as_nli_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_presuppositions_as_nli_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/qa_wikidata.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: qa_wikidata_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_qa_wikidata_generate_until
|