koichi12 commited on
Commit
e2efe71
·
verified ·
1 Parent(s): 8a96407

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/_mmlu_flan_generative_template_yaml +34 -0
  2. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_anatomy.yaml +6 -0
  3. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_astronomy.yaml +6 -0
  4. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_medicine.yaml +6 -0
  5. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_computer_security.yaml +6 -0
  6. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_conceptual_physics.yaml +6 -0
  7. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_econometrics.yaml +6 -0
  8. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_electrical_engineering.yaml +6 -0
  9. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_formal_logic.yaml +6 -0
  10. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_global_facts.yaml +6 -0
  11. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_chemistry.yaml +6 -0
  12. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_mathematics.yaml +6 -0
  13. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_microeconomics.yaml +6 -0
  14. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_psychology.yaml +6 -0
  15. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_statistics.yaml +6 -0
  16. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_international_law.yaml +6 -0
  17. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_jurisprudence.yaml +6 -0
  18. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_machine_learning.yaml +6 -0
  19. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_management.yaml +6 -0
  20. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_marketing.yaml +6 -0
  21. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_medical_genetics.yaml +6 -0
  22. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_miscellaneous.yaml +6 -0
  23. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_moral_scenarios.yaml +6 -0
  24. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_nutrition.yaml +6 -0
  25. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_prehistory.yaml +6 -0
  26. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_professional_psychology.yaml +6 -0
  27. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_security_studies.yaml +6 -0
  28. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_us_foreign_policy.yaml +6 -0
  29. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_world_religions.yaml +6 -0
  30. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/_mmlu.yaml +32 -0
  31. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/_mmlu_flan_loglikelihood_template_yaml +17 -0
  32. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_abstract_algebra.yaml +6 -0
  33. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_business_ethics.yaml +6 -0
  34. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_clinical_knowledge.yaml +6 -0
  35. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_biology.yaml +6 -0
  36. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_chemistry.yaml +6 -0
  37. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_computer_science.yaml +6 -0
  38. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_mathematics.yaml +6 -0
  39. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_physics.yaml +6 -0
  40. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_conceptual_physics.yaml +6 -0
  41. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_electrical_engineering.yaml +6 -0
  42. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_elementary_mathematics.yaml +6 -0
  43. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_global_facts.yaml +6 -0
  44. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_biology.yaml +6 -0
  45. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_chemistry.yaml +6 -0
  46. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_computer_science.yaml +6 -0
  47. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_european_history.yaml +6 -0
  48. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_geography.yaml +6 -0
  49. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_government_and_politics.yaml +6 -0
  50. scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_macroeconomics.yaml +6 -0
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/_mmlu_flan_generative_template_yaml ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ dataset_path: hails/mmlu_no_train # a copy of `cais/mmlu` with no auxiliary_train split
2
+ test_split: test
3
+ fewshot_split: dev
4
+ fewshot_config:
5
+ sampler: first_n
6
+ output_type: generate_until
7
+ doc_to_text: "Q: {{question.strip()}}\n(A) {{choices[0]}} (B) {{choices[1]}} (C) {{choices[2]}} (D) {{choices[3]}}\nA:"
8
+ doc_to_target: "{{['(A)', '(B)', '(C)', '(D)'][answer]}}"
9
+ filter_list:
10
+ - name: "strict-match"
11
+ filter:
12
+ - function: "take_first"
13
+ - name: "flexible-extract"
14
+ filter:
15
+ - function: !function utils.MultiChoiceRegexFilter
16
+ group_select: 0
17
+ regex_pattern: "(\\([A-Z]\\))"
18
+ ignore_case: true
19
+ ignore_punctuation: true
20
+ - function: "take_first"
21
+ generation_kwargs:
22
+ until:
23
+ - "</s>"
24
+ - "Q:"
25
+ - "<|im_end|>"
26
+ - "\n"
27
+ metric_list:
28
+ - metric: exact_match
29
+ aggregation: mean
30
+ higher_is_better: true
31
+ metadata:
32
+ version: 3.0
33
+ dataset_kwargs:
34
+ trust_remote_code: true
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_anatomy.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "anatomy"
2
+ "description": "The following are multiple choice questions (with answers) about anatomy.\n\
3
+ \n"
4
+ "tag": "mmlu_flan_n_shot_generative_stem"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_anatomy"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_astronomy.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "astronomy"
2
+ "description": "The following are multiple choice questions (with answers) about astronomy.\n\
3
+ \n"
4
+ "tag": "mmlu_flan_n_shot_generative_stem"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_astronomy"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_college_medicine.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "college_medicine"
2
+ "description": "The following are multiple choice questions (with answers) about college\
3
+ \ medicine.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_other"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_college_medicine"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_computer_security.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "computer_security"
2
+ "description": "The following are multiple choice questions (with answers) about computer\
3
+ \ security.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_stem"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_computer_security"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_conceptual_physics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "conceptual_physics"
2
+ "description": "The following are multiple choice questions (with answers) about conceptual\
3
+ \ physics.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_stem"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_conceptual_physics"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_econometrics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "econometrics"
2
+ "description": "The following are multiple choice questions (with answers) about econometrics.\n\
3
+ \n"
4
+ "tag": "mmlu_flan_n_shot_generative_social_sciences"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_econometrics"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_electrical_engineering.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "electrical_engineering"
2
+ "description": "The following are multiple choice questions (with answers) about electrical\
3
+ \ engineering.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_stem"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_electrical_engineering"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_formal_logic.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "formal_logic"
2
+ "description": "The following are multiple choice questions (with answers) about formal\
3
+ \ logic.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_humanities"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_formal_logic"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_global_facts.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "global_facts"
2
+ "description": "The following are multiple choice questions (with answers) about global\
3
+ \ facts.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_other"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_global_facts"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_chemistry.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_chemistry"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school chemistry.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_stem"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_high_school_chemistry"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_mathematics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_mathematics"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school mathematics.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_stem"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_high_school_mathematics"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_microeconomics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_microeconomics"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school microeconomics.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_social_sciences"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_high_school_microeconomics"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_psychology.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_psychology"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school psychology.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_social_sciences"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_high_school_psychology"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_high_school_statistics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_statistics"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school statistics.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_stem"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_high_school_statistics"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_international_law.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "international_law"
2
+ "description": "The following are multiple choice questions (with answers) about international\
3
+ \ law.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_humanities"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_international_law"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_jurisprudence.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "jurisprudence"
2
+ "description": "The following are multiple choice questions (with answers) about jurisprudence.\n\
3
+ \n"
4
+ "tag": "mmlu_flan_n_shot_generative_humanities"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_jurisprudence"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_machine_learning.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "machine_learning"
2
+ "description": "The following are multiple choice questions (with answers) about machine\
3
+ \ learning.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_stem"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_machine_learning"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_management.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "management"
2
+ "description": "The following are multiple choice questions (with answers) about management.\n\
3
+ \n"
4
+ "tag": "mmlu_flan_n_shot_generative_other"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_management"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_marketing.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "marketing"
2
+ "description": "The following are multiple choice questions (with answers) about marketing.\n\
3
+ \n"
4
+ "tag": "mmlu_flan_n_shot_generative_other"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_marketing"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_medical_genetics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "medical_genetics"
2
+ "description": "The following are multiple choice questions (with answers) about medical\
3
+ \ genetics.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_other"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_medical_genetics"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_miscellaneous.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "miscellaneous"
2
+ "description": "The following are multiple choice questions (with answers) about miscellaneous.\n\
3
+ \n"
4
+ "tag": "mmlu_flan_n_shot_generative_other"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_miscellaneous"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_moral_scenarios.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "moral_scenarios"
2
+ "description": "The following are multiple choice questions (with answers) about moral\
3
+ \ scenarios.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_humanities"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_moral_scenarios"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_nutrition.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "nutrition"
2
+ "description": "The following are multiple choice questions (with answers) about nutrition.\n\
3
+ \n"
4
+ "tag": "mmlu_flan_n_shot_generative_other"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_nutrition"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_prehistory.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "prehistory"
2
+ "description": "The following are multiple choice questions (with answers) about prehistory.\n\
3
+ \n"
4
+ "tag": "mmlu_flan_n_shot_generative_humanities"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_prehistory"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_professional_psychology.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "professional_psychology"
2
+ "description": "The following are multiple choice questions (with answers) about professional\
3
+ \ psychology.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_social_sciences"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_professional_psychology"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_security_studies.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "security_studies"
2
+ "description": "The following are multiple choice questions (with answers) about security\
3
+ \ studies.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_social_sciences"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_security_studies"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_us_foreign_policy.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "us_foreign_policy"
2
+ "description": "The following are multiple choice questions (with answers) about us\
3
+ \ foreign policy.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_social_sciences"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_us_foreign_policy"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/generative/mmlu_world_religions.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "world_religions"
2
+ "description": "The following are multiple choice questions (with answers) about world\
3
+ \ religions.\n\n"
4
+ "tag": "mmlu_flan_n_shot_generative_humanities"
5
+ "include": "_mmlu_flan_generative_template_yaml"
6
+ "task": "mmlu_flan_n_shot_generative_world_religions"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/_mmlu.yaml ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ group: mmlu_flan_n_shot_loglikelihood
2
+ group_alias: mmlu (flan style, loglikelihood)
3
+ task:
4
+ - group: stem
5
+ task:
6
+ - mmlu_flan_n_shot_loglikelihood_stem
7
+ aggregate_metric_list:
8
+ - metric: acc
9
+ weight_by_size: True
10
+ - group: other
11
+ task:
12
+ - mmlu_flan_n_shot_loglikelihood_other
13
+ aggregate_metric_list:
14
+ - metric: acc
15
+ weight_by_size: True
16
+ - group: social sciences
17
+ task:
18
+ - mmlu_flan_n_shot_loglikelihood_social_sciences
19
+ aggregate_metric_list:
20
+ - metric: acc
21
+ weight_by_size: True
22
+ - group: humanities
23
+ task:
24
+ - mmlu_flan_n_shot_loglikelihood_humanities
25
+ aggregate_metric_list:
26
+ - metric: acc
27
+ weight_by_size: True
28
+ aggregate_metric_list:
29
+ - metric: acc
30
+ weight_by_size: True
31
+ metadata:
32
+ version: 2
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/_mmlu_flan_loglikelihood_template_yaml ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ dataset_path: hails/mmlu_no_train # a copy of `cais/mmlu` with no auxiliary_train split
2
+ test_split: test
3
+ fewshot_split: dev
4
+ fewshot_config:
5
+ sampler: first_n
6
+ output_type: multiple_choice
7
+ doc_to_text: "Q: {{question.strip()}}\n(A) {{choices[0]}} (B) {{choices[1]}} (C) {{choices[2]}} (D) {{choices[3]}}\nA:"
8
+ doc_to_choice: ["(A)", "(B)", "(C)", "(D)"]
9
+ doc_to_target: answer
10
+ metric_list:
11
+ - metric: acc
12
+ aggregation: mean
13
+ higher_is_better: true
14
+ metadata:
15
+ version: 2.0
16
+ dataset_kwargs:
17
+ trust_remote_code: true
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_abstract_algebra.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "abstract_algebra"
2
+ "description": "The following are multiple choice questions (with answers) about abstract\
3
+ \ algebra.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_abstract_algebra"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_business_ethics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "business_ethics"
2
+ "description": "The following are multiple choice questions (with answers) about business\
3
+ \ ethics.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_other"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_business_ethics"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_clinical_knowledge.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "clinical_knowledge"
2
+ "description": "The following are multiple choice questions (with answers) about clinical\
3
+ \ knowledge.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_other"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_clinical_knowledge"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_biology.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "college_biology"
2
+ "description": "The following are multiple choice questions (with answers) about college\
3
+ \ biology.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_college_biology"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_chemistry.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "college_chemistry"
2
+ "description": "The following are multiple choice questions (with answers) about college\
3
+ \ chemistry.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_college_chemistry"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_computer_science.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "college_computer_science"
2
+ "description": "The following are multiple choice questions (with answers) about college\
3
+ \ computer science.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_college_computer_science"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_mathematics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "college_mathematics"
2
+ "description": "The following are multiple choice questions (with answers) about college\
3
+ \ mathematics.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_college_mathematics"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_college_physics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "college_physics"
2
+ "description": "The following are multiple choice questions (with answers) about college\
3
+ \ physics.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_college_physics"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_conceptual_physics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "conceptual_physics"
2
+ "description": "The following are multiple choice questions (with answers) about conceptual\
3
+ \ physics.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_conceptual_physics"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_electrical_engineering.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "electrical_engineering"
2
+ "description": "The following are multiple choice questions (with answers) about electrical\
3
+ \ engineering.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_electrical_engineering"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_elementary_mathematics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "elementary_mathematics"
2
+ "description": "The following are multiple choice questions (with answers) about elementary\
3
+ \ mathematics.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_elementary_mathematics"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_global_facts.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "global_facts"
2
+ "description": "The following are multiple choice questions (with answers) about global\
3
+ \ facts.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_other"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_global_facts"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_biology.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_biology"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school biology.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_biology"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_chemistry.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_chemistry"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school chemistry.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_chemistry"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_computer_science.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_computer_science"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school computer science.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_stem"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_computer_science"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_european_history.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_european_history"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school european history.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_humanities"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_european_history"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_geography.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_geography"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school geography.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_social_sciences"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_geography"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_government_and_politics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_government_and_politics"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school government and politics.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_social_sciences"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_government_and_politics"
scripts/yans/lm-evaluation-harness/lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/mmlu_high_school_macroeconomics.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ "dataset_name": "high_school_macroeconomics"
2
+ "description": "The following are multiple choice questions (with answers) about high\
3
+ \ school macroeconomics.\n\n"
4
+ "tag": "mmlu_flan_n_shot_loglikelihood_social_sciences"
5
+ "include": "_mmlu_flan_loglikelihood_template_yaml"
6
+ "task": "mmlu_flan_n_shot_loglikelihood_high_school_macroeconomics"