koichi12 commited on
Commit
925eda0
·
verified ·
1 Parent(s): 7332364

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/arithmetic.yaml +4 -0
  2. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/ascii_word_recognition.yaml +4 -0
  3. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/authorship_verification.yaml +4 -0
  4. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/auto_categorization.yaml +4 -0
  5. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/bbq_lite_json.yaml +4 -0
  6. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/causal_judgment.yaml +4 -0
  7. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/checkmate_in_one.yaml +4 -0
  8. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/code_line_description.yaml +4 -0
  9. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conceptual_combinations.yaml +4 -0
  10. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conlang_translation.yaml +4 -0
  11. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/contextual_parametric_knowledge_conflicts.yaml +4 -0
  12. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cs_algorithms.yaml +4 -0
  13. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/dark_humor_detection.yaml +4 -0
  14. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/disambiguation_qa.yaml +4 -0
  15. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/discourse_marker_prediction.yaml +4 -0
  16. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/emoji_movie.yaml +4 -0
  17. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/emojis_emotion_prediction.yaml +4 -0
  18. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/epistemic_reasoning.yaml +4 -0
  19. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/formal_fallacies_syllogisms_negation.yaml +4 -0
  20. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/general_knowledge.yaml +4 -0
  21. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/human_organs_senses.yaml +4 -0
  22. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/implicatures.yaml +4 -0
  23. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/international_phonetic_alphabet_transliterate.yaml +4 -0
  24. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/intersect_geometry.yaml +4 -0
  25. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kanji_ascii.yaml +4 -0
  26. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/linguistic_mappings.yaml +4 -0
  27. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/linguistics_puzzles.yaml +4 -0
  28. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/list_functions.yaml +4 -0
  29. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_args.yaml +4 -0
  30. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_deduction.yaml +4 -0
  31. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_sequence.yaml +4 -0
  32. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/metaphor_boolean.yaml +4 -0
  33. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/metaphor_understanding.yaml +4 -0
  34. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/minute_mysteries_qa.yaml +4 -0
  35. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/misconceptions.yaml +4 -0
  36. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/misconceptions_russian.yaml +4 -0
  37. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/mnist_ascii.yaml +4 -0
  38. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/movie_dialog_same_or_different.yaml +4 -0
  39. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/movie_recommendation.yaml +4 -0
  40. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/multiemo.yaml +4 -0
  41. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/novel_concepts.yaml +4 -0
  42. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/object_counting.yaml +4 -0
  43. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/paragraph_segmentation.yaml +4 -0
  44. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/parsinlu_qa.yaml +4 -0
  45. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/persian_idioms.yaml +4 -0
  46. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physical_intuition.yaml +4 -0
  47. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physics.yaml +4 -0
  48. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physics_questions.yaml +4 -0
  49. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/presuppositions_as_nli.yaml +4 -0
  50. scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/qa_wikidata.yaml +4 -0
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/arithmetic.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: arithmetic_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_arithmetic_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/ascii_word_recognition.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: ascii_word_recognition_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_ascii_word_recognition_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/authorship_verification.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: authorship_verification_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_authorship_verification_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/auto_categorization.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: auto_categorization_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_auto_categorization_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/bbq_lite_json.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: bbq_lite_json_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_bbq_lite_json_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/causal_judgment.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: causal_judgment_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_causal_judgment_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/checkmate_in_one.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: checkmate_in_one_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_checkmate_in_one_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/code_line_description.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: code_line_description_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_code_line_description_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conceptual_combinations.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: conceptual_combinations_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_conceptual_combinations_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/conlang_translation.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: conlang_translation_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_conlang_translation_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/contextual_parametric_knowledge_conflicts.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: contextual_parametric_knowledge_conflicts_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_contextual_parametric_knowledge_conflicts_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/cs_algorithms.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: cs_algorithms_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_cs_algorithms_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/dark_humor_detection.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: dark_humor_detection_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_dark_humor_detection_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/disambiguation_qa.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: disambiguation_qa_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_disambiguation_qa_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/discourse_marker_prediction.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: discourse_marker_prediction_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_discourse_marker_prediction_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/emoji_movie.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: emoji_movie_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_emoji_movie_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/emojis_emotion_prediction.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: emojis_emotion_prediction_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_emojis_emotion_prediction_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/epistemic_reasoning.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: epistemic_reasoning_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_epistemic_reasoning_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/formal_fallacies_syllogisms_negation.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: formal_fallacies_syllogisms_negation_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_formal_fallacies_syllogisms_negation_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/general_knowledge.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: general_knowledge_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_general_knowledge_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/human_organs_senses.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: human_organs_senses_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_human_organs_senses_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/implicatures.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: implicatures_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_implicatures_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/international_phonetic_alphabet_transliterate.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: international_phonetic_alphabet_transliterate_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_international_phonetic_alphabet_transliterate_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/intersect_geometry.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: intersect_geometry_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_intersect_geometry_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/kanji_ascii.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: kanji_ascii_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_kanji_ascii_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/linguistic_mappings.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: linguistic_mappings_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_linguistic_mappings_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/linguistics_puzzles.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: linguistics_puzzles_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_linguistics_puzzles_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/list_functions.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: list_functions_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_list_functions_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_args.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: logical_args_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_logical_args_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_deduction.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: logical_deduction_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_logical_deduction_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/logical_sequence.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: logical_sequence_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_logical_sequence_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/metaphor_boolean.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: metaphor_boolean_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_metaphor_boolean_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/metaphor_understanding.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: metaphor_understanding_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_metaphor_understanding_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/minute_mysteries_qa.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: minute_mysteries_qa_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_minute_mysteries_qa_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/misconceptions.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: misconceptions_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_misconceptions_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/misconceptions_russian.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: misconceptions_russian_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_misconceptions_russian_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/mnist_ascii.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: mnist_ascii_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_mnist_ascii_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/movie_dialog_same_or_different.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: movie_dialog_same_or_different_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_movie_dialog_same_or_different_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/movie_recommendation.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: movie_recommendation_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_movie_recommendation_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/multiemo.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: multiemo_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_multiemo_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/novel_concepts.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: novel_concepts_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_novel_concepts_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/object_counting.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: object_counting_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_object_counting_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/paragraph_segmentation.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: paragraph_segmentation_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_paragraph_segmentation_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/parsinlu_qa.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: parsinlu_qa_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_parsinlu_qa_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/persian_idioms.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: persian_idioms_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_persian_idioms_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physical_intuition.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: physical_intuition_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_physical_intuition_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physics.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: physics_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_physics_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/physics_questions.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: physics_questions_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_physics_questions_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/presuppositions_as_nli.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: presuppositions_as_nli_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_presuppositions_as_nli_generate_until
scripts/yans/lm-evaluation-harness/lm_eval/tasks/bigbench/generate_until/qa_wikidata.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ # Generated by utils.py
2
+ dataset_name: qa_wikidata_zero_shot
3
+ include: ../generate_until_template_yaml
4
+ task: bigbench_qa_wikidata_generate_until