Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/analogical_similarity.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/elementary_math_qa.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/evaluating_information_essentiality.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/gre_reading_comprehension.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/key_value_maps.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/language_identification.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/moral_permissibility.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/periodic_elements.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/reasoning_about_colored_objects.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/analytic_entailment.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/arithmetic.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/authorship_verification.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/bbq_lite_json.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/causal_judgment.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/checkmate_in_one.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/code_line_description.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/conceptual_combinations.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/contextual_parametric_knowledge_conflicts.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cs_algorithms.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/dark_humor_detection.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/disambiguation_qa.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/discourse_marker_prediction.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/emoji_movie.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/emojis_emotion_prediction.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/empirical_judgments.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/entailed_polarity.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/entailed_polarity_hindi.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/epistemic_reasoning.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/figure_of_speech_detection.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/formal_fallacies_syllogisms_negation.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/general_knowledge.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hhh_alignment.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/human_organs_senses.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/implicatures.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/international_phonetic_alphabet_nli.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/intersect_geometry.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/kanji_ascii.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/known_unknowns.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/logical_args.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/logical_deduction.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/logical_sequence.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/metaphor_boolean.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/metaphor_understanding.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/misconceptions.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/misconceptions_russian.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/mnist_ascii.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/movie_dialog_same_or_different.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/movie_recommendation.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/multiemo.yaml +4 -0
- scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/novel_concepts.yaml +4 -0
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/analogical_similarity.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: analogical_similarity_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_analogical_similarity_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/elementary_math_qa.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: elementary_math_qa_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_elementary_math_qa_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/evaluating_information_essentiality.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: evaluating_information_essentiality_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_evaluating_information_essentiality_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/gre_reading_comprehension.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: gre_reading_comprehension_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_gre_reading_comprehension_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/key_value_maps.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: key_value_maps_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_key_value_maps_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/language_identification.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: language_identification_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_language_identification_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/moral_permissibility.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: moral_permissibility_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_moral_permissibility_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/periodic_elements.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: periodic_elements_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_periodic_elements_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/reasoning_about_colored_objects.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: reasoning_about_colored_objects_zero_shot
|
| 3 |
+
include: ../generate_until_template_yaml
|
| 4 |
+
task: bigbench_reasoning_about_colored_objects_generate_until
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/analytic_entailment.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: analytic_entailment_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_analytic_entailment_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/arithmetic.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: arithmetic_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_arithmetic_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/authorship_verification.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: authorship_verification_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_authorship_verification_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/bbq_lite_json.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: bbq_lite_json_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_bbq_lite_json_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/causal_judgment.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: causal_judgment_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_causal_judgment_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/checkmate_in_one.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: checkmate_in_one_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_checkmate_in_one_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/code_line_description.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: code_line_description_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_code_line_description_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/conceptual_combinations.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: conceptual_combinations_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_conceptual_combinations_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/contextual_parametric_knowledge_conflicts.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: contextual_parametric_knowledge_conflicts_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_contextual_parametric_knowledge_conflicts_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/cs_algorithms.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: cs_algorithms_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_cs_algorithms_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/dark_humor_detection.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: dark_humor_detection_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_dark_humor_detection_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/disambiguation_qa.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: disambiguation_qa_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_disambiguation_qa_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/discourse_marker_prediction.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: discourse_marker_prediction_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_discourse_marker_prediction_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/emoji_movie.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: emoji_movie_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_emoji_movie_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/emojis_emotion_prediction.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: emojis_emotion_prediction_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_emojis_emotion_prediction_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/empirical_judgments.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: empirical_judgments_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_empirical_judgments_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/entailed_polarity.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: entailed_polarity_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_entailed_polarity_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/entailed_polarity_hindi.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: entailed_polarity_hindi_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_entailed_polarity_hindi_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/epistemic_reasoning.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: epistemic_reasoning_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_epistemic_reasoning_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/figure_of_speech_detection.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: figure_of_speech_detection_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_figure_of_speech_detection_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/formal_fallacies_syllogisms_negation.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: formal_fallacies_syllogisms_negation_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_formal_fallacies_syllogisms_negation_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/general_knowledge.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: general_knowledge_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_general_knowledge_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/hhh_alignment.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: hhh_alignment_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_hhh_alignment_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/human_organs_senses.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: human_organs_senses_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_human_organs_senses_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/implicatures.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: implicatures_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_implicatures_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/international_phonetic_alphabet_nli.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: international_phonetic_alphabet_nli_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_international_phonetic_alphabet_nli_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/intersect_geometry.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: intersect_geometry_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_intersect_geometry_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/kanji_ascii.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: kanji_ascii_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_kanji_ascii_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/known_unknowns.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: known_unknowns_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_known_unknowns_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/logical_args.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: logical_args_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_logical_args_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/logical_deduction.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: logical_deduction_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_logical_deduction_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/logical_sequence.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: logical_sequence_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_logical_sequence_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/metaphor_boolean.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: metaphor_boolean_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_metaphor_boolean_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/metaphor_understanding.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: metaphor_understanding_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_metaphor_understanding_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/misconceptions.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: misconceptions_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_misconceptions_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/misconceptions_russian.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: misconceptions_russian_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_misconceptions_russian_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/mnist_ascii.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: mnist_ascii_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_mnist_ascii_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/movie_dialog_same_or_different.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: movie_dialog_same_or_different_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_movie_dialog_same_or_different_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/movie_recommendation.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: movie_recommendation_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_movie_recommendation_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/multiemo.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: multiemo_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_multiemo_multiple_choice
|
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/multiple_choice/novel_concepts.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Generated by utils.py
|
| 2 |
+
dataset_name: novel_concepts_zero_shot
|
| 3 |
+
include: ../multiple_choice_template_a_yaml
|
| 4 |
+
task: bigbench_novel_concepts_multiple_choice
|