yuhengtu commited on
Commit
8c500ac
·
verified ·
1 Parent(s): 2169461

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +3 -0
  2. mmlu_pythia-1.4b-step2000/costs.json +1 -0
  3. mmlu_pythia-1.4b-step2000/groups.json +2411 -0
  4. mmlu_pythia-1.4b-step2000/groups/json/mmlu_abstract_algebra_mmlu_abstract_algebra_subject:abstract_algebra.json +145 -0
  5. mmlu_pythia-1.4b-step2000/groups/json/mmlu_anatomy_mmlu_anatomy_subject:anatomy.json +145 -0
  6. mmlu_pythia-1.4b-step2000/groups/json/mmlu_astronomy_mmlu_astronomy_subject:astronomy.json +145 -0
  7. mmlu_pythia-1.4b-step2000/groups/json/mmlu_business_ethics_mmlu_business_ethics_subject:business_ethics.json +145 -0
  8. mmlu_pythia-1.4b-step2000/groups/json/mmlu_clinical_knowledge_mmlu_clinical_knowledge_subject:clinical_knowledge.json +145 -0
  9. mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_biology_mmlu_college_biology_subject:college_biology.json +145 -0
  10. mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_chemistry_mmlu_college_chemistry_subject:college_chemistry.json +145 -0
  11. mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_computer_science_mmlu_college_computer_science_subject:college_computer_science.json +145 -0
  12. mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_mathematics_mmlu_college_mathematics_subject:college_mathematics.json +145 -0
  13. mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_medicine_mmlu_college_medicine_subject:college_medicine.json +145 -0
  14. mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_physics_mmlu_college_physics_subject:college_physics.json +145 -0
  15. mmlu_pythia-1.4b-step2000/groups/json/mmlu_computer_security_mmlu_computer_security_subject:computer_security.json +145 -0
  16. mmlu_pythia-1.4b-step2000/groups/json/mmlu_conceptual_physics_mmlu_conceptual_physics_subject:conceptual_physics.json +145 -0
  17. mmlu_pythia-1.4b-step2000/groups/json/mmlu_econometrics_mmlu_econometrics_subject:econometrics.json +145 -0
  18. mmlu_pythia-1.4b-step2000/groups/json/mmlu_electrical_engineering_mmlu_electrical_engineering_subject:electrical_engineering.json +145 -0
  19. mmlu_pythia-1.4b-step2000/groups/json/mmlu_elementary_mathematics_mmlu_elementary_mathematics_subject:elementary_mathematics.json +145 -0
  20. mmlu_pythia-1.4b-step2000/groups/json/mmlu_formal_logic_mmlu_formal_logic_subject:formal_logic.json +145 -0
  21. mmlu_pythia-1.4b-step2000/groups/json/mmlu_global_facts_mmlu_global_facts_subject:global_facts.json +145 -0
  22. mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_biology_mmlu_high_school_biology_subject:high_school_biology.json +145 -0
  23. mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_chemistry_mmlu_high_school_chemistry_subject:high_school_chemistry.json +145 -0
  24. mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_computer_science_mmlu_high_school_computer_science_subject:high_school_computer_science.json +145 -0
  25. mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_european_history_mmlu_high_school_european_history_subject:high_school_european_history.json +145 -0
  26. mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_geography_mmlu_high_school_geography_subject:high_school_geography.json +145 -0
  27. mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_government_and_politics_mmlu_high_school_government_and_politics_subject:high_school_government_and_politics.json +145 -0
  28. mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_macroeconomics_mmlu_high_school_macroeconomics_subject:high_school_macroeconomics.json +145 -0
  29. mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_mathematics_mmlu_high_school_mathematics_subject:high_school_mathematics.json +145 -0
  30. mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_microeconomics_mmlu_high_school_microeconomics_subject:high_school_microeconomics.json +145 -0
  31. mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_physics_mmlu_high_school_physics_subject:high_school_physics.json +145 -0
  32. mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_psychology_mmlu_high_school_psychology_subject:high_school_psychology.json +145 -0
  33. mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_statistics_mmlu_high_school_statistics_subject:high_school_statistics.json +145 -0
  34. mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_us_history_mmlu_high_school_us_history_subject:high_school_us_history.json +145 -0
  35. mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_world_history_mmlu_high_school_world_history_subject:high_school_world_history.json +145 -0
  36. mmlu_pythia-1.4b-step2000/groups/json/mmlu_human_aging_mmlu_human_aging_subject:human_aging.json +145 -0
  37. mmlu_pythia-1.4b-step2000/groups/json/mmlu_human_sexuality_mmlu_human_sexuality_subject:human_sexuality.json +145 -0
  38. mmlu_pythia-1.4b-step2000/groups/json/mmlu_international_law_mmlu_international_law_subject:international_law.json +145 -0
  39. mmlu_pythia-1.4b-step2000/groups/json/mmlu_jurisprudence_mmlu_jurisprudence_subject:jurisprudence.json +145 -0
  40. mmlu_pythia-1.4b-step2000/groups/json/mmlu_logical_fallacies_mmlu_logical_fallacies_subject:logical_fallacies.json +145 -0
  41. mmlu_pythia-1.4b-step2000/groups/json/mmlu_machine_learning_mmlu_machine_learning_subject:machine_learning.json +145 -0
  42. mmlu_pythia-1.4b-step2000/groups/json/mmlu_management_mmlu_management_subject:management.json +145 -0
  43. mmlu_pythia-1.4b-step2000/groups/json/mmlu_marketing_mmlu_marketing_subject:marketing.json +145 -0
  44. mmlu_pythia-1.4b-step2000/groups/json/mmlu_medical_genetics_mmlu_medical_genetics_subject:medical_genetics.json +145 -0
  45. mmlu_pythia-1.4b-step2000/groups/json/mmlu_miscellaneous_mmlu_miscellaneous_subject:miscellaneous.json +145 -0
  46. mmlu_pythia-1.4b-step2000/groups/json/mmlu_mmlu.json +554 -0
  47. mmlu_pythia-1.4b-step2000/groups/json/mmlu_mmlu_subject:abstract_algebra.json +145 -0
  48. mmlu_pythia-1.4b-step2000/groups/json/mmlu_mmlu_subject:anatomy.json +145 -0
  49. mmlu_pythia-1.4b-step2000/groups/json/mmlu_mmlu_subject:astronomy.json +145 -0
  50. mmlu_pythia-1.4b-step2000/groups/json/mmlu_mmlu_subject:business_ethics.json +145 -0
.gitattributes CHANGED
@@ -1285,3 +1285,6 @@ classic_pythia-2.8b-step2/imdb:model=EleutherAI_pythia-2.8b,data_augmentation=ca
1285
  classic_pythia-2.8b-step2/imdb:model=EleutherAI_pythia-2.8b,data_augmentation=canonical/per_instance_stats.json filter=lfs diff=lfs merge=lfs -text
1286
  classic_pythia-2.8b-step2/imdb:model=EleutherAI_pythia-2.8b,data_augmentation=canonical/scenario_state.json filter=lfs diff=lfs merge=lfs -text
1287
  classic_pythia-2.8b-step2/runs.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
1285
  classic_pythia-2.8b-step2/imdb:model=EleutherAI_pythia-2.8b,data_augmentation=canonical/per_instance_stats.json filter=lfs diff=lfs merge=lfs -text
1286
  classic_pythia-2.8b-step2/imdb:model=EleutherAI_pythia-2.8b,data_augmentation=canonical/scenario_state.json filter=lfs diff=lfs merge=lfs -text
1287
  classic_pythia-2.8b-step2/runs.json filter=lfs diff=lfs merge=lfs -text
1288
+ mmlu_pythia-1.4b-step2000/mmlu:subject=professional_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_law/display_requests.json filter=lfs diff=lfs merge=lfs -text
1289
+ mmlu_pythia-1.4b-step2000/mmlu:subject=professional_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_law/per_instance_stats.json filter=lfs diff=lfs merge=lfs -text
1290
+ mmlu_pythia-1.4b-step2000/mmlu:subject=professional_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_law/scenario_state.json filter=lfs diff=lfs merge=lfs -text
mmlu_pythia-1.4b-step2000/costs.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {}
mmlu_pythia-1.4b-step2000/groups.json ADDED
@@ -0,0 +1,2411 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "title": "All Scenarios",
4
+ "header": [
5
+ {
6
+ "value": "Group",
7
+ "markdown": false,
8
+ "metadata": {}
9
+ },
10
+ {
11
+ "value": "Description",
12
+ "markdown": false,
13
+ "metadata": {}
14
+ },
15
+ {
16
+ "value": "Adaptation method",
17
+ "description": "Adaptation strategy (e.g., generation)",
18
+ "markdown": false,
19
+ "metadata": {}
20
+ },
21
+ {
22
+ "value": "# instances",
23
+ "description": "Number of instances evaluated on",
24
+ "markdown": false,
25
+ "metadata": {}
26
+ },
27
+ {
28
+ "value": "# references",
29
+ "description": "Number of references provided per instance",
30
+ "markdown": false,
31
+ "metadata": {}
32
+ },
33
+ {
34
+ "value": "# prompt tokens",
35
+ "description": "Total number of prompt tokens",
36
+ "markdown": false,
37
+ "metadata": {}
38
+ },
39
+ {
40
+ "value": "# completion tokens",
41
+ "description": "Total number of completion tokens",
42
+ "markdown": false,
43
+ "metadata": {}
44
+ },
45
+ {
46
+ "value": "# models",
47
+ "description": "Number of models we're evaluating",
48
+ "markdown": false,
49
+ "metadata": {}
50
+ }
51
+ ],
52
+ "rows": [
53
+ [
54
+ {
55
+ "value": "MMLU Subjects",
56
+ "href": "?group=mmlu_subjects",
57
+ "markdown": false
58
+ },
59
+ {
60
+ "value": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).",
61
+ "markdown": true
62
+ },
63
+ {
64
+ "value": "multiple_choice_joint",
65
+ "markdown": false
66
+ },
67
+ {
68
+ "value": 246.35087719298247,
69
+ "description": "min=100, mean=246.351, max=1534, sum=28084 (114)",
70
+ "markdown": false
71
+ },
72
+ {
73
+ "value": 4.0,
74
+ "description": "min=4, mean=4, max=4, sum=1368 (342)",
75
+ "markdown": false
76
+ },
77
+ {
78
+ "value": 204487.2446201024,
79
+ "description": "min=285.766, mean=597.916, max=1816.758, sum=204487.245 (342)",
80
+ "markdown": false
81
+ },
82
+ {
83
+ "value": 342.0,
84
+ "description": "min=1, mean=1, max=1, sum=342 (342)",
85
+ "markdown": false
86
+ },
87
+ {
88
+ "value": 1,
89
+ "markdown": false
90
+ }
91
+ ]
92
+ ],
93
+ "links": []
94
+ },
95
+ {
96
+ "title": "Scenarios",
97
+ "header": [
98
+ {
99
+ "value": "Group",
100
+ "markdown": false,
101
+ "metadata": {}
102
+ },
103
+ {
104
+ "value": "Description",
105
+ "markdown": false,
106
+ "metadata": {}
107
+ },
108
+ {
109
+ "value": "Adaptation method",
110
+ "description": "Adaptation strategy (e.g., generation)",
111
+ "markdown": false,
112
+ "metadata": {}
113
+ },
114
+ {
115
+ "value": "# instances",
116
+ "description": "Number of instances evaluated on",
117
+ "markdown": false,
118
+ "metadata": {}
119
+ },
120
+ {
121
+ "value": "# references",
122
+ "description": "Number of references provided per instance",
123
+ "markdown": false,
124
+ "metadata": {}
125
+ },
126
+ {
127
+ "value": "# prompt tokens",
128
+ "description": "Total number of prompt tokens",
129
+ "markdown": false,
130
+ "metadata": {}
131
+ },
132
+ {
133
+ "value": "# completion tokens",
134
+ "description": "Total number of completion tokens",
135
+ "markdown": false,
136
+ "metadata": {}
137
+ },
138
+ {
139
+ "value": "# models",
140
+ "description": "Number of models we're evaluating",
141
+ "markdown": false,
142
+ "metadata": {}
143
+ }
144
+ ],
145
+ "rows": [
146
+ [
147
+ {
148
+ "value": "Massive Multitask Language Understanding (MMLU) All Subjects",
149
+ "href": "?group=mmlu",
150
+ "markdown": false
151
+ },
152
+ {
153
+ "value": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).",
154
+ "markdown": true
155
+ },
156
+ {
157
+ "value": "multiple_choice_joint",
158
+ "markdown": false
159
+ },
160
+ {
161
+ "value": 246.35087719298247,
162
+ "description": "min=100, mean=246.351, max=1534, sum=14042 (57)",
163
+ "markdown": false
164
+ },
165
+ {
166
+ "value": 4.0,
167
+ "description": "min=4, mean=4, max=4, sum=684 (171)",
168
+ "markdown": false
169
+ },
170
+ {
171
+ "value": 102243.62231005121,
172
+ "description": "min=285.766, mean=597.916, max=1816.758, sum=102243.622 (171)",
173
+ "markdown": false
174
+ },
175
+ {
176
+ "value": 171.0,
177
+ "description": "min=1, mean=1, max=1, sum=171 (171)",
178
+ "markdown": false
179
+ },
180
+ {
181
+ "value": 1,
182
+ "markdown": false
183
+ }
184
+ ],
185
+ [
186
+ {
187
+ "value": "Abstract Algebra",
188
+ "href": "?group=mmlu_abstract_algebra",
189
+ "markdown": false
190
+ },
191
+ {
192
+ "value": "The abstract algebra subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
193
+ "markdown": true
194
+ },
195
+ {
196
+ "value": "multiple_choice_joint",
197
+ "markdown": false
198
+ },
199
+ {
200
+ "value": 100.0,
201
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
202
+ "markdown": false
203
+ },
204
+ {
205
+ "value": 4.0,
206
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
207
+ "markdown": false
208
+ },
209
+ {
210
+ "value": 1076.28,
211
+ "description": "min=358.76, mean=358.76, max=358.76, sum=1076.28 (3)",
212
+ "markdown": false
213
+ },
214
+ {
215
+ "value": 3.0,
216
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
217
+ "markdown": false
218
+ },
219
+ {
220
+ "value": 1,
221
+ "markdown": false
222
+ }
223
+ ],
224
+ [
225
+ {
226
+ "value": "Anatomy",
227
+ "href": "?group=mmlu_anatomy",
228
+ "markdown": false
229
+ },
230
+ {
231
+ "value": "The anatomy subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
232
+ "markdown": true
233
+ },
234
+ {
235
+ "value": "multiple_choice_joint",
236
+ "markdown": false
237
+ },
238
+ {
239
+ "value": 135.0,
240
+ "description": "min=135, mean=135, max=135, sum=135 (1)",
241
+ "markdown": false
242
+ },
243
+ {
244
+ "value": 4.0,
245
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
246
+ "markdown": false
247
+ },
248
+ {
249
+ "value": 1056.088888888889,
250
+ "description": "min=352.03, mean=352.03, max=352.03, sum=1056.089 (3)",
251
+ "markdown": false
252
+ },
253
+ {
254
+ "value": 3.0,
255
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
256
+ "markdown": false
257
+ },
258
+ {
259
+ "value": 1,
260
+ "markdown": false
261
+ }
262
+ ],
263
+ [
264
+ {
265
+ "value": "College Chemistry",
266
+ "href": "?group=mmlu_college_chemistry",
267
+ "markdown": false
268
+ },
269
+ {
270
+ "value": "The college chemistry subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
271
+ "markdown": true
272
+ },
273
+ {
274
+ "value": "multiple_choice_joint",
275
+ "markdown": false
276
+ },
277
+ {
278
+ "value": 100.0,
279
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
280
+ "markdown": false
281
+ },
282
+ {
283
+ "value": 4.0,
284
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
285
+ "markdown": false
286
+ },
287
+ {
288
+ "value": 1607.5500000000002,
289
+ "description": "min=535.85, mean=535.85, max=535.85, sum=1607.55 (3)",
290
+ "markdown": false
291
+ },
292
+ {
293
+ "value": 3.0,
294
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
295
+ "markdown": false
296
+ },
297
+ {
298
+ "value": 1,
299
+ "markdown": false
300
+ }
301
+ ],
302
+ [
303
+ {
304
+ "value": "Computer Security",
305
+ "href": "?group=mmlu_computer_security",
306
+ "markdown": false
307
+ },
308
+ {
309
+ "value": "The computer security subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
310
+ "markdown": true
311
+ },
312
+ {
313
+ "value": "multiple_choice_joint",
314
+ "markdown": false
315
+ },
316
+ {
317
+ "value": 100.0,
318
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
319
+ "markdown": false
320
+ },
321
+ {
322
+ "value": 4.0,
323
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
324
+ "markdown": false
325
+ },
326
+ {
327
+ "value": 1164.57,
328
+ "description": "min=388.19, mean=388.19, max=388.19, sum=1164.57 (3)",
329
+ "markdown": false
330
+ },
331
+ {
332
+ "value": 3.0,
333
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
334
+ "markdown": false
335
+ },
336
+ {
337
+ "value": 1,
338
+ "markdown": false
339
+ }
340
+ ],
341
+ [
342
+ {
343
+ "value": "Econometrics",
344
+ "href": "?group=mmlu_econometrics",
345
+ "markdown": false
346
+ },
347
+ {
348
+ "value": "The econometrics subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
349
+ "markdown": true
350
+ },
351
+ {
352
+ "value": "multiple_choice_joint",
353
+ "markdown": false
354
+ },
355
+ {
356
+ "value": 114.0,
357
+ "description": "min=114, mean=114, max=114, sum=114 (1)",
358
+ "markdown": false
359
+ },
360
+ {
361
+ "value": 4.0,
362
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
363
+ "markdown": false
364
+ },
365
+ {
366
+ "value": 1838.3947368421052,
367
+ "description": "min=612.798, mean=612.798, max=612.798, sum=1838.395 (3)",
368
+ "markdown": false
369
+ },
370
+ {
371
+ "value": 3.0,
372
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
373
+ "markdown": false
374
+ },
375
+ {
376
+ "value": 1,
377
+ "markdown": false
378
+ }
379
+ ],
380
+ [
381
+ {
382
+ "value": "Global Facts",
383
+ "href": "?group=mmlu_global_facts",
384
+ "markdown": false
385
+ },
386
+ {
387
+ "value": "The global facts subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
388
+ "markdown": true
389
+ },
390
+ {
391
+ "value": "multiple_choice_joint",
392
+ "markdown": false
393
+ },
394
+ {
395
+ "value": 100.0,
396
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
397
+ "markdown": false
398
+ },
399
+ {
400
+ "value": 4.0,
401
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
402
+ "markdown": false
403
+ },
404
+ {
405
+ "value": 1201.74,
406
+ "description": "min=400.58, mean=400.58, max=400.58, sum=1201.74 (3)",
407
+ "markdown": false
408
+ },
409
+ {
410
+ "value": 3.0,
411
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
412
+ "markdown": false
413
+ },
414
+ {
415
+ "value": 1,
416
+ "markdown": false
417
+ }
418
+ ],
419
+ [
420
+ {
421
+ "value": "Jurisprudence",
422
+ "href": "?group=mmlu_jurisprudence",
423
+ "markdown": false
424
+ },
425
+ {
426
+ "value": "The jurisprudence subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
427
+ "markdown": true
428
+ },
429
+ {
430
+ "value": "multiple_choice_joint",
431
+ "markdown": false
432
+ },
433
+ {
434
+ "value": 108.0,
435
+ "description": "min=108, mean=108, max=108, sum=108 (1)",
436
+ "markdown": false
437
+ },
438
+ {
439
+ "value": 4.0,
440
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
441
+ "markdown": false
442
+ },
443
+ {
444
+ "value": 1262.5833333333333,
445
+ "description": "min=420.861, mean=420.861, max=420.861, sum=1262.583 (3)",
446
+ "markdown": false
447
+ },
448
+ {
449
+ "value": 3.0,
450
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
451
+ "markdown": false
452
+ },
453
+ {
454
+ "value": 1,
455
+ "markdown": false
456
+ }
457
+ ],
458
+ [
459
+ {
460
+ "value": "Philosophy",
461
+ "href": "?group=mmlu_philosophy",
462
+ "markdown": false
463
+ },
464
+ {
465
+ "value": "The philosophy subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
466
+ "markdown": true
467
+ },
468
+ {
469
+ "value": "multiple_choice_joint",
470
+ "markdown": false
471
+ },
472
+ {
473
+ "value": 311.0,
474
+ "description": "min=311, mean=311, max=311, sum=311 (1)",
475
+ "markdown": false
476
+ },
477
+ {
478
+ "value": 4.0,
479
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
480
+ "markdown": false
481
+ },
482
+ {
483
+ "value": 1035.829581993569,
484
+ "description": "min=345.277, mean=345.277, max=345.277, sum=1035.83 (3)",
485
+ "markdown": false
486
+ },
487
+ {
488
+ "value": 3.0,
489
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
490
+ "markdown": false
491
+ },
492
+ {
493
+ "value": 1,
494
+ "markdown": false
495
+ }
496
+ ],
497
+ [
498
+ {
499
+ "value": "Professional Medicine",
500
+ "href": "?group=mmlu_professional_medicine",
501
+ "markdown": false
502
+ },
503
+ {
504
+ "value": "The professional medicine subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
505
+ "markdown": true
506
+ },
507
+ {
508
+ "value": "multiple_choice_joint",
509
+ "markdown": false
510
+ },
511
+ {
512
+ "value": 272.0,
513
+ "description": "min=272, mean=272, max=272, sum=272 (1)",
514
+ "markdown": false
515
+ },
516
+ {
517
+ "value": 4.0,
518
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
519
+ "markdown": false
520
+ },
521
+ {
522
+ "value": 3242.6470588235297,
523
+ "description": "min=1080.882, mean=1080.882, max=1080.882, sum=3242.647 (3)",
524
+ "markdown": false
525
+ },
526
+ {
527
+ "value": 3.0,
528
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
529
+ "markdown": false
530
+ },
531
+ {
532
+ "value": 1,
533
+ "markdown": false
534
+ }
535
+ ],
536
+ [
537
+ {
538
+ "value": "Us Foreign Policy",
539
+ "href": "?group=mmlu_us_foreign_policy",
540
+ "markdown": false
541
+ },
542
+ {
543
+ "value": "The us foreign policy subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
544
+ "markdown": true
545
+ },
546
+ {
547
+ "value": "multiple_choice_joint",
548
+ "markdown": false
549
+ },
550
+ {
551
+ "value": 100.0,
552
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
553
+ "markdown": false
554
+ },
555
+ {
556
+ "value": 4.0,
557
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
558
+ "markdown": false
559
+ },
560
+ {
561
+ "value": 1332.24,
562
+ "description": "min=444.08, mean=444.08, max=444.08, sum=1332.24 (3)",
563
+ "markdown": false
564
+ },
565
+ {
566
+ "value": 3.0,
567
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
568
+ "markdown": false
569
+ },
570
+ {
571
+ "value": 1,
572
+ "markdown": false
573
+ }
574
+ ],
575
+ [
576
+ {
577
+ "value": "Astronomy",
578
+ "href": "?group=mmlu_astronomy",
579
+ "markdown": false
580
+ },
581
+ {
582
+ "value": "The astronomy subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
583
+ "markdown": true
584
+ },
585
+ {
586
+ "value": "multiple_choice_joint",
587
+ "markdown": false
588
+ },
589
+ {
590
+ "value": 152.0,
591
+ "description": "min=152, mean=152, max=152, sum=152 (1)",
592
+ "markdown": false
593
+ },
594
+ {
595
+ "value": 4.0,
596
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
597
+ "markdown": false
598
+ },
599
+ {
600
+ "value": 1795.4605263157894,
601
+ "description": "min=598.487, mean=598.487, max=598.487, sum=1795.461 (3)",
602
+ "markdown": false
603
+ },
604
+ {
605
+ "value": 3.0,
606
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
607
+ "markdown": false
608
+ },
609
+ {
610
+ "value": 1,
611
+ "markdown": false
612
+ }
613
+ ],
614
+ [
615
+ {
616
+ "value": "Business Ethics",
617
+ "href": "?group=mmlu_business_ethics",
618
+ "markdown": false
619
+ },
620
+ {
621
+ "value": "The business ethics subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
622
+ "markdown": true
623
+ },
624
+ {
625
+ "value": "multiple_choice_joint",
626
+ "markdown": false
627
+ },
628
+ {
629
+ "value": 100.0,
630
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
631
+ "markdown": false
632
+ },
633
+ {
634
+ "value": 4.0,
635
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
636
+ "markdown": false
637
+ },
638
+ {
639
+ "value": 1755.1499999999999,
640
+ "description": "min=585.05, mean=585.05, max=585.05, sum=1755.15 (3)",
641
+ "markdown": false
642
+ },
643
+ {
644
+ "value": 3.0,
645
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
646
+ "markdown": false
647
+ },
648
+ {
649
+ "value": 1,
650
+ "markdown": false
651
+ }
652
+ ],
653
+ [
654
+ {
655
+ "value": "Clinical Knowledge",
656
+ "href": "?group=mmlu_clinical_knowledge",
657
+ "markdown": false
658
+ },
659
+ {
660
+ "value": "The clinical knowledge subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
661
+ "markdown": true
662
+ },
663
+ {
664
+ "value": "multiple_choice_joint",
665
+ "markdown": false
666
+ },
667
+ {
668
+ "value": 265.0,
669
+ "description": "min=265, mean=265, max=265, sum=265 (1)",
670
+ "markdown": false
671
+ },
672
+ {
673
+ "value": 4.0,
674
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
675
+ "markdown": false
676
+ },
677
+ {
678
+ "value": 1205.7509433962264,
679
+ "description": "min=401.917, mean=401.917, max=401.917, sum=1205.751 (3)",
680
+ "markdown": false
681
+ },
682
+ {
683
+ "value": 3.0,
684
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
685
+ "markdown": false
686
+ },
687
+ {
688
+ "value": 1,
689
+ "markdown": false
690
+ }
691
+ ],
692
+ [
693
+ {
694
+ "value": "College Biology",
695
+ "href": "?group=mmlu_college_biology",
696
+ "markdown": false
697
+ },
698
+ {
699
+ "value": "The college biology subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
700
+ "markdown": true
701
+ },
702
+ {
703
+ "value": "multiple_choice_joint",
704
+ "markdown": false
705
+ },
706
+ {
707
+ "value": 144.0,
708
+ "description": "min=144, mean=144, max=144, sum=144 (1)",
709
+ "markdown": false
710
+ },
711
+ {
712
+ "value": 4.0,
713
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
714
+ "markdown": false
715
+ },
716
+ {
717
+ "value": 1410.9583333333335,
718
+ "description": "min=470.319, mean=470.319, max=470.319, sum=1410.958 (3)",
719
+ "markdown": false
720
+ },
721
+ {
722
+ "value": 3.0,
723
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
724
+ "markdown": false
725
+ },
726
+ {
727
+ "value": 1,
728
+ "markdown": false
729
+ }
730
+ ],
731
+ [
732
+ {
733
+ "value": "College Computer Science",
734
+ "href": "?group=mmlu_college_computer_science",
735
+ "markdown": false
736
+ },
737
+ {
738
+ "value": "The college computer science subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
739
+ "markdown": true
740
+ },
741
+ {
742
+ "value": "multiple_choice_joint",
743
+ "markdown": false
744
+ },
745
+ {
746
+ "value": 100.0,
747
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
748
+ "markdown": false
749
+ },
750
+ {
751
+ "value": 4.0,
752
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
753
+ "markdown": false
754
+ },
755
+ {
756
+ "value": 2528.67,
757
+ "description": "min=842.89, mean=842.89, max=842.89, sum=2528.67 (3)",
758
+ "markdown": false
759
+ },
760
+ {
761
+ "value": 3.0,
762
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
763
+ "markdown": false
764
+ },
765
+ {
766
+ "value": 1,
767
+ "markdown": false
768
+ }
769
+ ],
770
+ [
771
+ {
772
+ "value": "College Mathematics",
773
+ "href": "?group=mmlu_college_mathematics",
774
+ "markdown": false
775
+ },
776
+ {
777
+ "value": "The college mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
778
+ "markdown": true
779
+ },
780
+ {
781
+ "value": "multiple_choice_joint",
782
+ "markdown": false
783
+ },
784
+ {
785
+ "value": 100.0,
786
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
787
+ "markdown": false
788
+ },
789
+ {
790
+ "value": 4.0,
791
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
792
+ "markdown": false
793
+ },
794
+ {
795
+ "value": 1778.46,
796
+ "description": "min=592.82, mean=592.82, max=592.82, sum=1778.46 (3)",
797
+ "markdown": false
798
+ },
799
+ {
800
+ "value": 3.0,
801
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
802
+ "markdown": false
803
+ },
804
+ {
805
+ "value": 1,
806
+ "markdown": false
807
+ }
808
+ ],
809
+ [
810
+ {
811
+ "value": "College Medicine",
812
+ "href": "?group=mmlu_college_medicine",
813
+ "markdown": false
814
+ },
815
+ {
816
+ "value": "The college medicine subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
817
+ "markdown": true
818
+ },
819
+ {
820
+ "value": "multiple_choice_joint",
821
+ "markdown": false
822
+ },
823
+ {
824
+ "value": 173.0,
825
+ "description": "min=173, mean=173, max=173, sum=173 (1)",
826
+ "markdown": false
827
+ },
828
+ {
829
+ "value": 4.0,
830
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
831
+ "markdown": false
832
+ },
833
+ {
834
+ "value": 1558.1271676300578,
835
+ "description": "min=519.376, mean=519.376, max=519.376, sum=1558.127 (3)",
836
+ "markdown": false
837
+ },
838
+ {
839
+ "value": 3.0,
840
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
841
+ "markdown": false
842
+ },
843
+ {
844
+ "value": 1,
845
+ "markdown": false
846
+ }
847
+ ],
848
+ [
849
+ {
850
+ "value": "College Physics",
851
+ "href": "?group=mmlu_college_physics",
852
+ "markdown": false
853
+ },
854
+ {
855
+ "value": "The college physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
856
+ "markdown": true
857
+ },
858
+ {
859
+ "value": "multiple_choice_joint",
860
+ "markdown": false
861
+ },
862
+ {
863
+ "value": 102.0,
864
+ "description": "min=102, mean=102, max=102, sum=102 (1)",
865
+ "markdown": false
866
+ },
867
+ {
868
+ "value": 4.0,
869
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
870
+ "markdown": false
871
+ },
872
+ {
873
+ "value": 1429.9705882352941,
874
+ "description": "min=476.657, mean=476.657, max=476.657, sum=1429.971 (3)",
875
+ "markdown": false
876
+ },
877
+ {
878
+ "value": 3.0,
879
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
880
+ "markdown": false
881
+ },
882
+ {
883
+ "value": 1,
884
+ "markdown": false
885
+ }
886
+ ],
887
+ [
888
+ {
889
+ "value": "Conceptual Physics",
890
+ "href": "?group=mmlu_conceptual_physics",
891
+ "markdown": false
892
+ },
893
+ {
894
+ "value": "The conceptual physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
895
+ "markdown": true
896
+ },
897
+ {
898
+ "value": "multiple_choice_joint",
899
+ "markdown": false
900
+ },
901
+ {
902
+ "value": 235.0,
903
+ "description": "min=235, mean=235, max=235, sum=235 (1)",
904
+ "markdown": false
905
+ },
906
+ {
907
+ "value": 4.0,
908
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
909
+ "markdown": false
910
+ },
911
+ {
912
+ "value": 933.931914893617,
913
+ "description": "min=311.311, mean=311.311, max=311.311, sum=933.932 (3)",
914
+ "markdown": false
915
+ },
916
+ {
917
+ "value": 3.0,
918
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
919
+ "markdown": false
920
+ },
921
+ {
922
+ "value": 1,
923
+ "markdown": false
924
+ }
925
+ ],
926
+ [
927
+ {
928
+ "value": "Electrical Engineering",
929
+ "href": "?group=mmlu_electrical_engineering",
930
+ "markdown": false
931
+ },
932
+ {
933
+ "value": "The electrical engineering subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
934
+ "markdown": true
935
+ },
936
+ {
937
+ "value": "multiple_choice_joint",
938
+ "markdown": false
939
+ },
940
+ {
941
+ "value": 145.0,
942
+ "description": "min=145, mean=145, max=145, sum=145 (1)",
943
+ "markdown": false
944
+ },
945
+ {
946
+ "value": 4.0,
947
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
948
+ "markdown": false
949
+ },
950
+ {
951
+ "value": 1274.544827586207,
952
+ "description": "min=424.848, mean=424.848, max=424.848, sum=1274.545 (3)",
953
+ "markdown": false
954
+ },
955
+ {
956
+ "value": 3.0,
957
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
958
+ "markdown": false
959
+ },
960
+ {
961
+ "value": 1,
962
+ "markdown": false
963
+ }
964
+ ],
965
+ [
966
+ {
967
+ "value": "Elementary Mathematics",
968
+ "href": "?group=mmlu_elementary_mathematics",
969
+ "markdown": false
970
+ },
971
+ {
972
+ "value": "The elementary mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
973
+ "markdown": true
974
+ },
975
+ {
976
+ "value": "multiple_choice_joint",
977
+ "markdown": false
978
+ },
979
+ {
980
+ "value": 378.0,
981
+ "description": "min=378, mean=378, max=378, sum=378 (1)",
982
+ "markdown": false
983
+ },
984
+ {
985
+ "value": 4.0,
986
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
987
+ "markdown": false
988
+ },
989
+ {
990
+ "value": 1515.2142857142858,
991
+ "description": "min=505.071, mean=505.071, max=505.071, sum=1515.214 (3)",
992
+ "markdown": false
993
+ },
994
+ {
995
+ "value": 3.0,
996
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
997
+ "markdown": false
998
+ },
999
+ {
1000
+ "value": 1,
1001
+ "markdown": false
1002
+ }
1003
+ ],
1004
+ [
1005
+ {
1006
+ "value": "Formal Logic",
1007
+ "href": "?group=mmlu_formal_logic",
1008
+ "markdown": false
1009
+ },
1010
+ {
1011
+ "value": "The formal logic subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1012
+ "markdown": true
1013
+ },
1014
+ {
1015
+ "value": "multiple_choice_joint",
1016
+ "markdown": false
1017
+ },
1018
+ {
1019
+ "value": 126.0,
1020
+ "description": "min=126, mean=126, max=126, sum=126 (1)",
1021
+ "markdown": false
1022
+ },
1023
+ {
1024
+ "value": 4.0,
1025
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1026
+ "markdown": false
1027
+ },
1028
+ {
1029
+ "value": 1960.7857142857142,
1030
+ "description": "min=653.595, mean=653.595, max=653.595, sum=1960.786 (3)",
1031
+ "markdown": false
1032
+ },
1033
+ {
1034
+ "value": 3.0,
1035
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1036
+ "markdown": false
1037
+ },
1038
+ {
1039
+ "value": 1,
1040
+ "markdown": false
1041
+ }
1042
+ ],
1043
+ [
1044
+ {
1045
+ "value": "High School Biology",
1046
+ "href": "?group=mmlu_high_school_biology",
1047
+ "markdown": false
1048
+ },
1049
+ {
1050
+ "value": "The high school biology subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1051
+ "markdown": true
1052
+ },
1053
+ {
1054
+ "value": "multiple_choice_joint",
1055
+ "markdown": false
1056
+ },
1057
+ {
1058
+ "value": 310.0,
1059
+ "description": "min=310, mean=310, max=310, sum=310 (1)",
1060
+ "markdown": false
1061
+ },
1062
+ {
1063
+ "value": 4.0,
1064
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1065
+ "markdown": false
1066
+ },
1067
+ {
1068
+ "value": 1541.7967741935481,
1069
+ "description": "min=513.932, mean=513.932, max=513.932, sum=1541.797 (3)",
1070
+ "markdown": false
1071
+ },
1072
+ {
1073
+ "value": 3.0,
1074
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1075
+ "markdown": false
1076
+ },
1077
+ {
1078
+ "value": 1,
1079
+ "markdown": false
1080
+ }
1081
+ ],
1082
+ [
1083
+ {
1084
+ "value": "High School Chemistry",
1085
+ "href": "?group=mmlu_high_school_chemistry",
1086
+ "markdown": false
1087
+ },
1088
+ {
1089
+ "value": "The high school chemistry subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1090
+ "markdown": true
1091
+ },
1092
+ {
1093
+ "value": "multiple_choice_joint",
1094
+ "markdown": false
1095
+ },
1096
+ {
1097
+ "value": 203.0,
1098
+ "description": "min=203, mean=203, max=203, sum=203 (1)",
1099
+ "markdown": false
1100
+ },
1101
+ {
1102
+ "value": 4.0,
1103
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1104
+ "markdown": false
1105
+ },
1106
+ {
1107
+ "value": 1439.527093596059,
1108
+ "description": "min=479.842, mean=479.842, max=479.842, sum=1439.527 (3)",
1109
+ "markdown": false
1110
+ },
1111
+ {
1112
+ "value": 3.0,
1113
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1114
+ "markdown": false
1115
+ },
1116
+ {
1117
+ "value": 1,
1118
+ "markdown": false
1119
+ }
1120
+ ],
1121
+ [
1122
+ {
1123
+ "value": "High School Computer Science",
1124
+ "href": "?group=mmlu_high_school_computer_science",
1125
+ "markdown": false
1126
+ },
1127
+ {
1128
+ "value": "The high school computer science subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1129
+ "markdown": true
1130
+ },
1131
+ {
1132
+ "value": "multiple_choice_joint",
1133
+ "markdown": false
1134
+ },
1135
+ {
1136
+ "value": 100.0,
1137
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
1138
+ "markdown": false
1139
+ },
1140
+ {
1141
+ "value": 4.0,
1142
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1143
+ "markdown": false
1144
+ },
1145
+ {
1146
+ "value": 2668.17,
1147
+ "description": "min=889.39, mean=889.39, max=889.39, sum=2668.17 (3)",
1148
+ "markdown": false
1149
+ },
1150
+ {
1151
+ "value": 3.0,
1152
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1153
+ "markdown": false
1154
+ },
1155
+ {
1156
+ "value": 1,
1157
+ "markdown": false
1158
+ }
1159
+ ],
1160
+ [
1161
+ {
1162
+ "value": "High School European History",
1163
+ "href": "?group=mmlu_high_school_european_history",
1164
+ "markdown": false
1165
+ },
1166
+ {
1167
+ "value": "The high school european history subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1168
+ "markdown": true
1169
+ },
1170
+ {
1171
+ "value": "multiple_choice_joint",
1172
+ "markdown": false
1173
+ },
1174
+ {
1175
+ "value": 165.0,
1176
+ "description": "min=165, mean=165, max=165, sum=165 (1)",
1177
+ "markdown": false
1178
+ },
1179
+ {
1180
+ "value": 4.0,
1181
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1182
+ "markdown": false
1183
+ },
1184
+ {
1185
+ "value": 5450.272727272727,
1186
+ "description": "min=1816.758, mean=1816.758, max=1816.758, sum=5450.273 (3)",
1187
+ "markdown": false
1188
+ },
1189
+ {
1190
+ "value": 3.0,
1191
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1192
+ "markdown": false
1193
+ },
1194
+ {
1195
+ "value": 1,
1196
+ "markdown": false
1197
+ }
1198
+ ],
1199
+ [
1200
+ {
1201
+ "value": "High School Geography",
1202
+ "href": "?group=mmlu_high_school_geography",
1203
+ "markdown": false
1204
+ },
1205
+ {
1206
+ "value": "The high school geography subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1207
+ "markdown": true
1208
+ },
1209
+ {
1210
+ "value": "multiple_choice_joint",
1211
+ "markdown": false
1212
+ },
1213
+ {
1214
+ "value": 198.0,
1215
+ "description": "min=198, mean=198, max=198, sum=198 (1)",
1216
+ "markdown": false
1217
+ },
1218
+ {
1219
+ "value": 4.0,
1220
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1221
+ "markdown": false
1222
+ },
1223
+ {
1224
+ "value": 1200.2727272727273,
1225
+ "description": "min=400.091, mean=400.091, max=400.091, sum=1200.273 (3)",
1226
+ "markdown": false
1227
+ },
1228
+ {
1229
+ "value": 3.0,
1230
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1231
+ "markdown": false
1232
+ },
1233
+ {
1234
+ "value": 1,
1235
+ "markdown": false
1236
+ }
1237
+ ],
1238
+ [
1239
+ {
1240
+ "value": "High School Government And Politics",
1241
+ "href": "?group=mmlu_high_school_government_and_politics",
1242
+ "markdown": false
1243
+ },
1244
+ {
1245
+ "value": "The high school government and politics subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1246
+ "markdown": true
1247
+ },
1248
+ {
1249
+ "value": "multiple_choice_joint",
1250
+ "markdown": false
1251
+ },
1252
+ {
1253
+ "value": 193.0,
1254
+ "description": "min=193, mean=193, max=193, sum=193 (1)",
1255
+ "markdown": false
1256
+ },
1257
+ {
1258
+ "value": 4.0,
1259
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1260
+ "markdown": false
1261
+ },
1262
+ {
1263
+ "value": 1448.2849740932643,
1264
+ "description": "min=482.762, mean=482.762, max=482.762, sum=1448.285 (3)",
1265
+ "markdown": false
1266
+ },
1267
+ {
1268
+ "value": 3.0,
1269
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1270
+ "markdown": false
1271
+ },
1272
+ {
1273
+ "value": 1,
1274
+ "markdown": false
1275
+ }
1276
+ ],
1277
+ [
1278
+ {
1279
+ "value": "High School Macroeconomics",
1280
+ "href": "?group=mmlu_high_school_macroeconomics",
1281
+ "markdown": false
1282
+ },
1283
+ {
1284
+ "value": "The high school macroeconomics subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1285
+ "markdown": true
1286
+ },
1287
+ {
1288
+ "value": "multiple_choice_joint",
1289
+ "markdown": false
1290
+ },
1291
+ {
1292
+ "value": 390.0,
1293
+ "description": "min=390, mean=390, max=390, sum=390 (1)",
1294
+ "markdown": false
1295
+ },
1296
+ {
1297
+ "value": 4.0,
1298
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1299
+ "markdown": false
1300
+ },
1301
+ {
1302
+ "value": 1177.0538461538463,
1303
+ "description": "min=392.351, mean=392.351, max=392.351, sum=1177.054 (3)",
1304
+ "markdown": false
1305
+ },
1306
+ {
1307
+ "value": 3.0,
1308
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1309
+ "markdown": false
1310
+ },
1311
+ {
1312
+ "value": 1,
1313
+ "markdown": false
1314
+ }
1315
+ ],
1316
+ [
1317
+ {
1318
+ "value": "High School Mathematics",
1319
+ "href": "?group=mmlu_high_school_mathematics",
1320
+ "markdown": false
1321
+ },
1322
+ {
1323
+ "value": "The high school mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1324
+ "markdown": true
1325
+ },
1326
+ {
1327
+ "value": "multiple_choice_joint",
1328
+ "markdown": false
1329
+ },
1330
+ {
1331
+ "value": 270.0,
1332
+ "description": "min=270, mean=270, max=270, sum=270 (1)",
1333
+ "markdown": false
1334
+ },
1335
+ {
1336
+ "value": 4.0,
1337
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1338
+ "markdown": false
1339
+ },
1340
+ {
1341
+ "value": 1520.0666666666666,
1342
+ "description": "min=506.689, mean=506.689, max=506.689, sum=1520.067 (3)",
1343
+ "markdown": false
1344
+ },
1345
+ {
1346
+ "value": 3.0,
1347
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1348
+ "markdown": false
1349
+ },
1350
+ {
1351
+ "value": 1,
1352
+ "markdown": false
1353
+ }
1354
+ ],
1355
+ [
1356
+ {
1357
+ "value": "High School Microeconomics",
1358
+ "href": "?group=mmlu_high_school_microeconomics",
1359
+ "markdown": false
1360
+ },
1361
+ {
1362
+ "value": "The high school microeconomics subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1363
+ "markdown": true
1364
+ },
1365
+ {
1366
+ "value": "multiple_choice_joint",
1367
+ "markdown": false
1368
+ },
1369
+ {
1370
+ "value": 238.0,
1371
+ "description": "min=238, mean=238, max=238, sum=238 (1)",
1372
+ "markdown": false
1373
+ },
1374
+ {
1375
+ "value": 4.0,
1376
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1377
+ "markdown": false
1378
+ },
1379
+ {
1380
+ "value": 1233.7058823529412,
1381
+ "description": "min=411.235, mean=411.235, max=411.235, sum=1233.706 (3)",
1382
+ "markdown": false
1383
+ },
1384
+ {
1385
+ "value": 3.0,
1386
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1387
+ "markdown": false
1388
+ },
1389
+ {
1390
+ "value": 1,
1391
+ "markdown": false
1392
+ }
1393
+ ],
1394
+ [
1395
+ {
1396
+ "value": "High School Physics",
1397
+ "href": "?group=mmlu_high_school_physics",
1398
+ "markdown": false
1399
+ },
1400
+ {
1401
+ "value": "The high school physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1402
+ "markdown": true
1403
+ },
1404
+ {
1405
+ "value": "multiple_choice_joint",
1406
+ "markdown": false
1407
+ },
1408
+ {
1409
+ "value": 151.0,
1410
+ "description": "min=151, mean=151, max=151, sum=151 (1)",
1411
+ "markdown": false
1412
+ },
1413
+ {
1414
+ "value": 4.0,
1415
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1416
+ "markdown": false
1417
+ },
1418
+ {
1419
+ "value": 1646.1854304635763,
1420
+ "description": "min=548.728, mean=548.728, max=548.728, sum=1646.185 (3)",
1421
+ "markdown": false
1422
+ },
1423
+ {
1424
+ "value": 3.0,
1425
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1426
+ "markdown": false
1427
+ },
1428
+ {
1429
+ "value": 1,
1430
+ "markdown": false
1431
+ }
1432
+ ],
1433
+ [
1434
+ {
1435
+ "value": "High School Psychology",
1436
+ "href": "?group=mmlu_high_school_psychology",
1437
+ "markdown": false
1438
+ },
1439
+ {
1440
+ "value": "The high school psychology subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1441
+ "markdown": true
1442
+ },
1443
+ {
1444
+ "value": "multiple_choice_joint",
1445
+ "markdown": false
1446
+ },
1447
+ {
1448
+ "value": 545.0,
1449
+ "description": "min=545, mean=545, max=545, sum=545 (1)",
1450
+ "markdown": false
1451
+ },
1452
+ {
1453
+ "value": 4.0,
1454
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1455
+ "markdown": false
1456
+ },
1457
+ {
1458
+ "value": 1544.377981651376,
1459
+ "description": "min=514.793, mean=514.793, max=514.793, sum=1544.378 (3)",
1460
+ "markdown": false
1461
+ },
1462
+ {
1463
+ "value": 3.0,
1464
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1465
+ "markdown": false
1466
+ },
1467
+ {
1468
+ "value": 1,
1469
+ "markdown": false
1470
+ }
1471
+ ],
1472
+ [
1473
+ {
1474
+ "value": "High School Statistics",
1475
+ "href": "?group=mmlu_high_school_statistics",
1476
+ "markdown": false
1477
+ },
1478
+ {
1479
+ "value": "The high school statistics subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1480
+ "markdown": true
1481
+ },
1482
+ {
1483
+ "value": "multiple_choice_joint",
1484
+ "markdown": false
1485
+ },
1486
+ {
1487
+ "value": 216.0,
1488
+ "description": "min=216, mean=216, max=216, sum=216 (1)",
1489
+ "markdown": false
1490
+ },
1491
+ {
1492
+ "value": 4.0,
1493
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1494
+ "markdown": false
1495
+ },
1496
+ {
1497
+ "value": 2389.8194444444443,
1498
+ "description": "min=796.606, mean=796.606, max=796.606, sum=2389.819 (3)",
1499
+ "markdown": false
1500
+ },
1501
+ {
1502
+ "value": 3.0,
1503
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1504
+ "markdown": false
1505
+ },
1506
+ {
1507
+ "value": 1,
1508
+ "markdown": false
1509
+ }
1510
+ ],
1511
+ [
1512
+ {
1513
+ "value": "High School US History",
1514
+ "href": "?group=mmlu_high_school_us_history",
1515
+ "markdown": false
1516
+ },
1517
+ {
1518
+ "value": "The high school us history subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1519
+ "markdown": true
1520
+ },
1521
+ {
1522
+ "value": "multiple_choice_joint",
1523
+ "markdown": false
1524
+ },
1525
+ {
1526
+ "value": 204.0,
1527
+ "description": "min=204, mean=204, max=204, sum=204 (1)",
1528
+ "markdown": false
1529
+ },
1530
+ {
1531
+ "value": 4.0,
1532
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1533
+ "markdown": false
1534
+ },
1535
+ {
1536
+ "value": 5365.161764705883,
1537
+ "description": "min=1788.387, mean=1788.387, max=1788.387, sum=5365.162 (3)",
1538
+ "markdown": false
1539
+ },
1540
+ {
1541
+ "value": 3.0,
1542
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1543
+ "markdown": false
1544
+ },
1545
+ {
1546
+ "value": 1,
1547
+ "markdown": false
1548
+ }
1549
+ ],
1550
+ [
1551
+ {
1552
+ "value": "High School World History",
1553
+ "href": "?group=mmlu_high_school_world_history",
1554
+ "markdown": false
1555
+ },
1556
+ {
1557
+ "value": "The high school world history subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1558
+ "markdown": true
1559
+ },
1560
+ {
1561
+ "value": "multiple_choice_joint",
1562
+ "markdown": false
1563
+ },
1564
+ {
1565
+ "value": 237.0,
1566
+ "description": "min=237, mean=237, max=237, sum=237 (1)",
1567
+ "markdown": false
1568
+ },
1569
+ {
1570
+ "value": 4.0,
1571
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1572
+ "markdown": false
1573
+ },
1574
+ {
1575
+ "value": 4384.329113924051,
1576
+ "description": "min=1461.443, mean=1461.443, max=1461.443, sum=4384.329 (3)",
1577
+ "markdown": false
1578
+ },
1579
+ {
1580
+ "value": 3.0,
1581
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1582
+ "markdown": false
1583
+ },
1584
+ {
1585
+ "value": 1,
1586
+ "markdown": false
1587
+ }
1588
+ ],
1589
+ [
1590
+ {
1591
+ "value": "Human Aging",
1592
+ "href": "?group=mmlu_human_aging",
1593
+ "markdown": false
1594
+ },
1595
+ {
1596
+ "value": "The human aging subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1597
+ "markdown": true
1598
+ },
1599
+ {
1600
+ "value": "multiple_choice_joint",
1601
+ "markdown": false
1602
+ },
1603
+ {
1604
+ "value": 223.0,
1605
+ "description": "min=223, mean=223, max=223, sum=223 (1)",
1606
+ "markdown": false
1607
+ },
1608
+ {
1609
+ "value": 4.0,
1610
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1611
+ "markdown": false
1612
+ },
1613
+ {
1614
+ "value": 971.0717488789237,
1615
+ "description": "min=323.691, mean=323.691, max=323.691, sum=971.072 (3)",
1616
+ "markdown": false
1617
+ },
1618
+ {
1619
+ "value": 3.0,
1620
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1621
+ "markdown": false
1622
+ },
1623
+ {
1624
+ "value": 1,
1625
+ "markdown": false
1626
+ }
1627
+ ],
1628
+ [
1629
+ {
1630
+ "value": "Human Sexuality",
1631
+ "href": "?group=mmlu_human_sexuality",
1632
+ "markdown": false
1633
+ },
1634
+ {
1635
+ "value": "The human sexuality subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1636
+ "markdown": true
1637
+ },
1638
+ {
1639
+ "value": "multiple_choice_joint",
1640
+ "markdown": false
1641
+ },
1642
+ {
1643
+ "value": 131.0,
1644
+ "description": "min=131, mean=131, max=131, sum=131 (1)",
1645
+ "markdown": false
1646
+ },
1647
+ {
1648
+ "value": 4.0,
1649
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1650
+ "markdown": false
1651
+ },
1652
+ {
1653
+ "value": 1066.0534351145038,
1654
+ "description": "min=355.351, mean=355.351, max=355.351, sum=1066.053 (3)",
1655
+ "markdown": false
1656
+ },
1657
+ {
1658
+ "value": 3.0,
1659
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1660
+ "markdown": false
1661
+ },
1662
+ {
1663
+ "value": 1,
1664
+ "markdown": false
1665
+ }
1666
+ ],
1667
+ [
1668
+ {
1669
+ "value": "International Law",
1670
+ "href": "?group=mmlu_international_law",
1671
+ "markdown": false
1672
+ },
1673
+ {
1674
+ "value": "The international law subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1675
+ "markdown": true
1676
+ },
1677
+ {
1678
+ "value": "multiple_choice_joint",
1679
+ "markdown": false
1680
+ },
1681
+ {
1682
+ "value": 121.0,
1683
+ "description": "min=121, mean=121, max=121, sum=121 (1)",
1684
+ "markdown": false
1685
+ },
1686
+ {
1687
+ "value": 4.0,
1688
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1689
+ "markdown": false
1690
+ },
1691
+ {
1692
+ "value": 1951.115702479339,
1693
+ "description": "min=650.372, mean=650.372, max=650.372, sum=1951.116 (3)",
1694
+ "markdown": false
1695
+ },
1696
+ {
1697
+ "value": 3.0,
1698
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1699
+ "markdown": false
1700
+ },
1701
+ {
1702
+ "value": 1,
1703
+ "markdown": false
1704
+ }
1705
+ ],
1706
+ [
1707
+ {
1708
+ "value": "Logical Fallacies",
1709
+ "href": "?group=mmlu_logical_fallacies",
1710
+ "markdown": false
1711
+ },
1712
+ {
1713
+ "value": "The logical fallacies subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1714
+ "markdown": true
1715
+ },
1716
+ {
1717
+ "value": "multiple_choice_joint",
1718
+ "markdown": false
1719
+ },
1720
+ {
1721
+ "value": 163.0,
1722
+ "description": "min=163, mean=163, max=163, sum=163 (1)",
1723
+ "markdown": false
1724
+ },
1725
+ {
1726
+ "value": 4.0,
1727
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1728
+ "markdown": false
1729
+ },
1730
+ {
1731
+ "value": 1376.484662576687,
1732
+ "description": "min=458.828, mean=458.828, max=458.828, sum=1376.485 (3)",
1733
+ "markdown": false
1734
+ },
1735
+ {
1736
+ "value": 3.0,
1737
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1738
+ "markdown": false
1739
+ },
1740
+ {
1741
+ "value": 1,
1742
+ "markdown": false
1743
+ }
1744
+ ],
1745
+ [
1746
+ {
1747
+ "value": "Machine Learning",
1748
+ "href": "?group=mmlu_machine_learning",
1749
+ "markdown": false
1750
+ },
1751
+ {
1752
+ "value": "The machine learning subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1753
+ "markdown": true
1754
+ },
1755
+ {
1756
+ "value": "multiple_choice_joint",
1757
+ "markdown": false
1758
+ },
1759
+ {
1760
+ "value": 112.0,
1761
+ "description": "min=112, mean=112, max=112, sum=112 (1)",
1762
+ "markdown": false
1763
+ },
1764
+ {
1765
+ "value": 4.0,
1766
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1767
+ "markdown": false
1768
+ },
1769
+ {
1770
+ "value": 1983.6428571428569,
1771
+ "description": "min=661.214, mean=661.214, max=661.214, sum=1983.643 (3)",
1772
+ "markdown": false
1773
+ },
1774
+ {
1775
+ "value": 3.0,
1776
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1777
+ "markdown": false
1778
+ },
1779
+ {
1780
+ "value": 1,
1781
+ "markdown": false
1782
+ }
1783
+ ],
1784
+ [
1785
+ {
1786
+ "value": "Management",
1787
+ "href": "?group=mmlu_management",
1788
+ "markdown": false
1789
+ },
1790
+ {
1791
+ "value": "The management subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1792
+ "markdown": true
1793
+ },
1794
+ {
1795
+ "value": "multiple_choice_joint",
1796
+ "markdown": false
1797
+ },
1798
+ {
1799
+ "value": 103.0,
1800
+ "description": "min=103, mean=103, max=103, sum=103 (1)",
1801
+ "markdown": false
1802
+ },
1803
+ {
1804
+ "value": 4.0,
1805
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1806
+ "markdown": false
1807
+ },
1808
+ {
1809
+ "value": 894.1456310679612,
1810
+ "description": "min=298.049, mean=298.049, max=298.049, sum=894.146 (3)",
1811
+ "markdown": false
1812
+ },
1813
+ {
1814
+ "value": 3.0,
1815
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1816
+ "markdown": false
1817
+ },
1818
+ {
1819
+ "value": 1,
1820
+ "markdown": false
1821
+ }
1822
+ ],
1823
+ [
1824
+ {
1825
+ "value": "Marketing",
1826
+ "href": "?group=mmlu_marketing",
1827
+ "markdown": false
1828
+ },
1829
+ {
1830
+ "value": "The marketing subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1831
+ "markdown": true
1832
+ },
1833
+ {
1834
+ "value": "multiple_choice_joint",
1835
+ "markdown": false
1836
+ },
1837
+ {
1838
+ "value": 234.0,
1839
+ "description": "min=234, mean=234, max=234, sum=234 (1)",
1840
+ "markdown": false
1841
+ },
1842
+ {
1843
+ "value": 4.0,
1844
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1845
+ "markdown": false
1846
+ },
1847
+ {
1848
+ "value": 1320.3076923076924,
1849
+ "description": "min=440.103, mean=440.103, max=440.103, sum=1320.308 (3)",
1850
+ "markdown": false
1851
+ },
1852
+ {
1853
+ "value": 3.0,
1854
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1855
+ "markdown": false
1856
+ },
1857
+ {
1858
+ "value": 1,
1859
+ "markdown": false
1860
+ }
1861
+ ],
1862
+ [
1863
+ {
1864
+ "value": "Medical Genetics",
1865
+ "href": "?group=mmlu_medical_genetics",
1866
+ "markdown": false
1867
+ },
1868
+ {
1869
+ "value": "The medical genetics subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1870
+ "markdown": true
1871
+ },
1872
+ {
1873
+ "value": "multiple_choice_joint",
1874
+ "markdown": false
1875
+ },
1876
+ {
1877
+ "value": 100.0,
1878
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
1879
+ "markdown": false
1880
+ },
1881
+ {
1882
+ "value": 4.0,
1883
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1884
+ "markdown": false
1885
+ },
1886
+ {
1887
+ "value": 1021.44,
1888
+ "description": "min=340.48, mean=340.48, max=340.48, sum=1021.44 (3)",
1889
+ "markdown": false
1890
+ },
1891
+ {
1892
+ "value": 3.0,
1893
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1894
+ "markdown": false
1895
+ },
1896
+ {
1897
+ "value": 1,
1898
+ "markdown": false
1899
+ }
1900
+ ],
1901
+ [
1902
+ {
1903
+ "value": "Miscellaneous",
1904
+ "href": "?group=mmlu_miscellaneous",
1905
+ "markdown": false
1906
+ },
1907
+ {
1908
+ "value": "The miscellaneous subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1909
+ "markdown": true
1910
+ },
1911
+ {
1912
+ "value": "multiple_choice_joint",
1913
+ "markdown": false
1914
+ },
1915
+ {
1916
+ "value": 783.0,
1917
+ "description": "min=783, mean=783, max=783, sum=783 (1)",
1918
+ "markdown": false
1919
+ },
1920
+ {
1921
+ "value": 4.0,
1922
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1923
+ "markdown": false
1924
+ },
1925
+ {
1926
+ "value": 961.3295019157088,
1927
+ "description": "min=320.443, mean=320.443, max=320.443, sum=961.33 (3)",
1928
+ "markdown": false
1929
+ },
1930
+ {
1931
+ "value": 3.0,
1932
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1933
+ "markdown": false
1934
+ },
1935
+ {
1936
+ "value": 1,
1937
+ "markdown": false
1938
+ }
1939
+ ],
1940
+ [
1941
+ {
1942
+ "value": "Moral Disputes",
1943
+ "href": "?group=mmlu_moral_disputes",
1944
+ "markdown": false
1945
+ },
1946
+ {
1947
+ "value": "The moral disputes subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1948
+ "markdown": true
1949
+ },
1950
+ {
1951
+ "value": "multiple_choice_joint",
1952
+ "markdown": false
1953
+ },
1954
+ {
1955
+ "value": 346.0,
1956
+ "description": "min=346, mean=346, max=346, sum=346 (1)",
1957
+ "markdown": false
1958
+ },
1959
+ {
1960
+ "value": 4.0,
1961
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
1962
+ "markdown": false
1963
+ },
1964
+ {
1965
+ "value": 1506.728323699422,
1966
+ "description": "min=502.243, mean=502.243, max=502.243, sum=1506.728 (3)",
1967
+ "markdown": false
1968
+ },
1969
+ {
1970
+ "value": 3.0,
1971
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
1972
+ "markdown": false
1973
+ },
1974
+ {
1975
+ "value": 1,
1976
+ "markdown": false
1977
+ }
1978
+ ],
1979
+ [
1980
+ {
1981
+ "value": "Moral Scenarios",
1982
+ "href": "?group=mmlu_moral_scenarios",
1983
+ "markdown": false
1984
+ },
1985
+ {
1986
+ "value": "The moral scenarios subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
1987
+ "markdown": true
1988
+ },
1989
+ {
1990
+ "value": "multiple_choice_joint",
1991
+ "markdown": false
1992
+ },
1993
+ {
1994
+ "value": 895.0,
1995
+ "description": "min=895, mean=895, max=895, sum=895 (1)",
1996
+ "markdown": false
1997
+ },
1998
+ {
1999
+ "value": 4.0,
2000
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
2001
+ "markdown": false
2002
+ },
2003
+ {
2004
+ "value": 2003.5843575418994,
2005
+ "description": "min=667.861, mean=667.861, max=667.861, sum=2003.584 (3)",
2006
+ "markdown": false
2007
+ },
2008
+ {
2009
+ "value": 3.0,
2010
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
2011
+ "markdown": false
2012
+ },
2013
+ {
2014
+ "value": 1,
2015
+ "markdown": false
2016
+ }
2017
+ ],
2018
+ [
2019
+ {
2020
+ "value": "Nutrition",
2021
+ "href": "?group=mmlu_nutrition",
2022
+ "markdown": false
2023
+ },
2024
+ {
2025
+ "value": "The nutrition subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
2026
+ "markdown": true
2027
+ },
2028
+ {
2029
+ "value": "multiple_choice_joint",
2030
+ "markdown": false
2031
+ },
2032
+ {
2033
+ "value": 306.0,
2034
+ "description": "min=306, mean=306, max=306, sum=306 (1)",
2035
+ "markdown": false
2036
+ },
2037
+ {
2038
+ "value": 4.0,
2039
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
2040
+ "markdown": false
2041
+ },
2042
+ {
2043
+ "value": 1737.3823529411766,
2044
+ "description": "min=579.127, mean=579.127, max=579.127, sum=1737.382 (3)",
2045
+ "markdown": false
2046
+ },
2047
+ {
2048
+ "value": 3.0,
2049
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
2050
+ "markdown": false
2051
+ },
2052
+ {
2053
+ "value": 1,
2054
+ "markdown": false
2055
+ }
2056
+ ],
2057
+ [
2058
+ {
2059
+ "value": "Prehistory",
2060
+ "href": "?group=mmlu_prehistory",
2061
+ "markdown": false
2062
+ },
2063
+ {
2064
+ "value": "The prehistory subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
2065
+ "markdown": true
2066
+ },
2067
+ {
2068
+ "value": "multiple_choice_joint",
2069
+ "markdown": false
2070
+ },
2071
+ {
2072
+ "value": 324.0,
2073
+ "description": "min=324, mean=324, max=324, sum=324 (1)",
2074
+ "markdown": false
2075
+ },
2076
+ {
2077
+ "value": 4.0,
2078
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
2079
+ "markdown": false
2080
+ },
2081
+ {
2082
+ "value": 1605.453703703704,
2083
+ "description": "min=535.151, mean=535.151, max=535.151, sum=1605.454 (3)",
2084
+ "markdown": false
2085
+ },
2086
+ {
2087
+ "value": 3.0,
2088
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
2089
+ "markdown": false
2090
+ },
2091
+ {
2092
+ "value": 1,
2093
+ "markdown": false
2094
+ }
2095
+ ],
2096
+ [
2097
+ {
2098
+ "value": "Professional Accounting",
2099
+ "href": "?group=mmlu_professional_accounting",
2100
+ "markdown": false
2101
+ },
2102
+ {
2103
+ "value": "The professional accounting subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
2104
+ "markdown": true
2105
+ },
2106
+ {
2107
+ "value": "multiple_choice_joint",
2108
+ "markdown": false
2109
+ },
2110
+ {
2111
+ "value": 282.0,
2112
+ "description": "min=282, mean=282, max=282, sum=282 (1)",
2113
+ "markdown": false
2114
+ },
2115
+ {
2116
+ "value": 4.0,
2117
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
2118
+ "markdown": false
2119
+ },
2120
+ {
2121
+ "value": 1982.7659574468084,
2122
+ "description": "min=660.922, mean=660.922, max=660.922, sum=1982.766 (3)",
2123
+ "markdown": false
2124
+ },
2125
+ {
2126
+ "value": 3.0,
2127
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
2128
+ "markdown": false
2129
+ },
2130
+ {
2131
+ "value": 1,
2132
+ "markdown": false
2133
+ }
2134
+ ],
2135
+ [
2136
+ {
2137
+ "value": "Professional Law",
2138
+ "href": "?group=mmlu_professional_law",
2139
+ "markdown": false
2140
+ },
2141
+ {
2142
+ "value": "The professional law subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
2143
+ "markdown": true
2144
+ },
2145
+ {
2146
+ "value": "multiple_choice_joint",
2147
+ "markdown": false
2148
+ },
2149
+ {
2150
+ "value": 1534.0,
2151
+ "description": "min=1534, mean=1534, max=1534, sum=1534 (1)",
2152
+ "markdown": false
2153
+ },
2154
+ {
2155
+ "value": 4.0,
2156
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
2157
+ "markdown": false
2158
+ },
2159
+ {
2160
+ "value": 4963.298565840939,
2161
+ "description": "min=1654.433, mean=1654.433, max=1654.433, sum=4963.299 (3)",
2162
+ "markdown": false
2163
+ },
2164
+ {
2165
+ "value": 3.0,
2166
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
2167
+ "markdown": false
2168
+ },
2169
+ {
2170
+ "value": 1,
2171
+ "markdown": false
2172
+ }
2173
+ ],
2174
+ [
2175
+ {
2176
+ "value": "Professional Psychology",
2177
+ "href": "?group=mmlu_professional_psychology",
2178
+ "markdown": false
2179
+ },
2180
+ {
2181
+ "value": "The professional psychology subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
2182
+ "markdown": true
2183
+ },
2184
+ {
2185
+ "value": "multiple_choice_joint",
2186
+ "markdown": false
2187
+ },
2188
+ {
2189
+ "value": 612.0,
2190
+ "description": "min=612, mean=612, max=612, sum=612 (1)",
2191
+ "markdown": false
2192
+ },
2193
+ {
2194
+ "value": 4.0,
2195
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
2196
+ "markdown": false
2197
+ },
2198
+ {
2199
+ "value": 1772.6176470588234,
2200
+ "description": "min=590.873, mean=590.873, max=590.873, sum=1772.618 (3)",
2201
+ "markdown": false
2202
+ },
2203
+ {
2204
+ "value": 3.0,
2205
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
2206
+ "markdown": false
2207
+ },
2208
+ {
2209
+ "value": 1,
2210
+ "markdown": false
2211
+ }
2212
+ ],
2213
+ [
2214
+ {
2215
+ "value": "Public Relations",
2216
+ "href": "?group=mmlu_public_relations",
2217
+ "markdown": false
2218
+ },
2219
+ {
2220
+ "value": "The public relations subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
2221
+ "markdown": true
2222
+ },
2223
+ {
2224
+ "value": "multiple_choice_joint",
2225
+ "markdown": false
2226
+ },
2227
+ {
2228
+ "value": 110.0,
2229
+ "description": "min=110, mean=110, max=110, sum=110 (1)",
2230
+ "markdown": false
2231
+ },
2232
+ {
2233
+ "value": 4.0,
2234
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
2235
+ "markdown": false
2236
+ },
2237
+ {
2238
+ "value": 1268.9454545454546,
2239
+ "description": "min=422.982, mean=422.982, max=422.982, sum=1268.945 (3)",
2240
+ "markdown": false
2241
+ },
2242
+ {
2243
+ "value": 3.0,
2244
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
2245
+ "markdown": false
2246
+ },
2247
+ {
2248
+ "value": 1,
2249
+ "markdown": false
2250
+ }
2251
+ ],
2252
+ [
2253
+ {
2254
+ "value": "Security Studies",
2255
+ "href": "?group=mmlu_security_studies",
2256
+ "markdown": false
2257
+ },
2258
+ {
2259
+ "value": "The security studies subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
2260
+ "markdown": true
2261
+ },
2262
+ {
2263
+ "value": "multiple_choice_joint",
2264
+ "markdown": false
2265
+ },
2266
+ {
2267
+ "value": 245.0,
2268
+ "description": "min=245, mean=245, max=245, sum=245 (1)",
2269
+ "markdown": false
2270
+ },
2271
+ {
2272
+ "value": 4.0,
2273
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
2274
+ "markdown": false
2275
+ },
2276
+ {
2277
+ "value": 3621.171428571429,
2278
+ "description": "min=1207.057, mean=1207.057, max=1207.057, sum=3621.171 (3)",
2279
+ "markdown": false
2280
+ },
2281
+ {
2282
+ "value": 3.0,
2283
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
2284
+ "markdown": false
2285
+ },
2286
+ {
2287
+ "value": 1,
2288
+ "markdown": false
2289
+ }
2290
+ ],
2291
+ [
2292
+ {
2293
+ "value": "Sociology",
2294
+ "href": "?group=mmlu_sociology",
2295
+ "markdown": false
2296
+ },
2297
+ {
2298
+ "value": "The sociology subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
2299
+ "markdown": true
2300
+ },
2301
+ {
2302
+ "value": "multiple_choice_joint",
2303
+ "markdown": false
2304
+ },
2305
+ {
2306
+ "value": 201.0,
2307
+ "description": "min=201, mean=201, max=201, sum=201 (1)",
2308
+ "markdown": false
2309
+ },
2310
+ {
2311
+ "value": 4.0,
2312
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
2313
+ "markdown": false
2314
+ },
2315
+ {
2316
+ "value": 1356.0597014925372,
2317
+ "description": "min=452.02, mean=452.02, max=452.02, sum=1356.06 (3)",
2318
+ "markdown": false
2319
+ },
2320
+ {
2321
+ "value": 3.0,
2322
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
2323
+ "markdown": false
2324
+ },
2325
+ {
2326
+ "value": 1,
2327
+ "markdown": false
2328
+ }
2329
+ ],
2330
+ [
2331
+ {
2332
+ "value": "Virology",
2333
+ "href": "?group=mmlu_virology",
2334
+ "markdown": false
2335
+ },
2336
+ {
2337
+ "value": "The virology subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
2338
+ "markdown": true
2339
+ },
2340
+ {
2341
+ "value": "multiple_choice_joint",
2342
+ "markdown": false
2343
+ },
2344
+ {
2345
+ "value": 166.0,
2346
+ "description": "min=166, mean=166, max=166, sum=166 (1)",
2347
+ "markdown": false
2348
+ },
2349
+ {
2350
+ "value": 4.0,
2351
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
2352
+ "markdown": false
2353
+ },
2354
+ {
2355
+ "value": 1048.7530120481927,
2356
+ "description": "min=349.584, mean=349.584, max=349.584, sum=1048.753 (3)",
2357
+ "markdown": false
2358
+ },
2359
+ {
2360
+ "value": 3.0,
2361
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
2362
+ "markdown": false
2363
+ },
2364
+ {
2365
+ "value": 1,
2366
+ "markdown": false
2367
+ }
2368
+ ],
2369
+ [
2370
+ {
2371
+ "value": "World Religions",
2372
+ "href": "?group=mmlu_world_religions",
2373
+ "markdown": false
2374
+ },
2375
+ {
2376
+ "value": "The world religions subject in the Massive Multitask Language Understanding (MMLU) benchmark.",
2377
+ "markdown": true
2378
+ },
2379
+ {
2380
+ "value": "multiple_choice_joint",
2381
+ "markdown": false
2382
+ },
2383
+ {
2384
+ "value": 171.0,
2385
+ "description": "min=171, mean=171, max=171, sum=171 (1)",
2386
+ "markdown": false
2387
+ },
2388
+ {
2389
+ "value": 4.0,
2390
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
2391
+ "markdown": false
2392
+ },
2393
+ {
2394
+ "value": 857.2982456140351,
2395
+ "description": "min=285.766, mean=285.766, max=285.766, sum=857.298 (3)",
2396
+ "markdown": false
2397
+ },
2398
+ {
2399
+ "value": 3.0,
2400
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
2401
+ "markdown": false
2402
+ },
2403
+ {
2404
+ "value": 1,
2405
+ "markdown": false
2406
+ }
2407
+ ]
2408
+ ],
2409
+ "links": []
2410
+ }
2411
+ ]
mmlu_pythia-1.4b-step2000/groups/json/mmlu_abstract_algebra_mmlu_abstract_algebra_subject:abstract_algebra.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: abstract_algebra",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The abstract algebra subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Abstract Algebra"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The abstract algebra subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Abstract Algebra"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The abstract algebra subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Abstract Algebra"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The abstract algebra subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Abstract Algebra"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The abstract algebra subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Abstract Algebra"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The abstract algebra subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Abstract Algebra"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The abstract algebra subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Abstract Algebra"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_abstract_algebra&subgroup=subject%3A%20abstract_algebra&runSpecs=%5B%22mmlu%3Asubject%3Dabstract_algebra%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_abstract_algebra%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_abstract_algebra"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.23,
88
+ "description": "min=0.23, mean=0.23, max=0.23, sum=0.23 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.06573188066482544,
96
+ "description": "min=0.066, mean=0.066, max=0.066, sum=0.066 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 100.0,
104
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 358.76,
122
+ "description": "min=358.76, mean=358.76, max=358.76, sum=358.76 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_abstract_algebra_mmlu_abstract_algebra_subject:abstract_algebra.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_abstract_algebra_mmlu_abstract_algebra_subject:abstract_algebra.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_abstract_algebra_subject:abstract_algebra"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_anatomy_mmlu_anatomy_subject:anatomy.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: anatomy",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The anatomy subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Anatomy"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The anatomy subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Anatomy"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The anatomy subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Anatomy"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The anatomy subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Anatomy"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The anatomy subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Anatomy"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The anatomy subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Anatomy"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The anatomy subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Anatomy"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_anatomy&subgroup=subject%3A%20anatomy&runSpecs=%5B%22mmlu%3Asubject%3Danatomy%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_anatomy%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=anatomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_anatomy"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.2962962962962963,
88
+ "description": "min=0.296, mean=0.296, max=0.296, sum=0.296 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.06830824922632288,
96
+ "description": "min=0.068, mean=0.068, max=0.068, sum=0.068 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 135.0,
104
+ "description": "min=135, mean=135, max=135, sum=135 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 352.02962962962965,
122
+ "description": "min=352.03, mean=352.03, max=352.03, sum=352.03 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_anatomy_mmlu_anatomy_subject:anatomy.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_anatomy_mmlu_anatomy_subject:anatomy.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_anatomy_subject:anatomy"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_astronomy_mmlu_astronomy_subject:astronomy.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: astronomy",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The astronomy subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Astronomy"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The astronomy subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Astronomy"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The astronomy subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Astronomy"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The astronomy subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Astronomy"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The astronomy subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Astronomy"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The astronomy subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Astronomy"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The astronomy subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Astronomy"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_astronomy&subgroup=subject%3A%20astronomy&runSpecs=%5B%22mmlu%3Asubject%3Dastronomy%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_astronomy%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=astronomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_astronomy"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.23684210526315788,
88
+ "description": "min=0.237, mean=0.237, max=0.237, sum=0.237 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.10126333958224247,
96
+ "description": "min=0.101, mean=0.101, max=0.101, sum=0.101 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 152.0,
104
+ "description": "min=152, mean=152, max=152, sum=152 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 598.4868421052631,
122
+ "description": "min=598.487, mean=598.487, max=598.487, sum=598.487 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_astronomy_mmlu_astronomy_subject:astronomy.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_astronomy_mmlu_astronomy_subject:astronomy.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_astronomy_subject:astronomy"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_business_ethics_mmlu_business_ethics_subject:business_ethics.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: business_ethics",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The business ethics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Business Ethics"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The business ethics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Business Ethics"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The business ethics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Business Ethics"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The business ethics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Business Ethics"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The business ethics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Business Ethics"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The business ethics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Business Ethics"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The business ethics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Business Ethics"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_business_ethics&subgroup=subject%3A%20business_ethics&runSpecs=%5B%22mmlu%3Asubject%3Dbusiness_ethics%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_business_ethics%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=business_ethics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_business_ethics"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.29,
88
+ "description": "min=0.29, mean=0.29, max=0.29, sum=0.29 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.09688833236694336,
96
+ "description": "min=0.097, mean=0.097, max=0.097, sum=0.097 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 100.0,
104
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 585.05,
122
+ "description": "min=585.05, mean=585.05, max=585.05, sum=585.05 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_business_ethics_mmlu_business_ethics_subject:business_ethics.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_business_ethics_mmlu_business_ethics_subject:business_ethics.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_business_ethics_subject:business_ethics"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_clinical_knowledge_mmlu_clinical_knowledge_subject:clinical_knowledge.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: clinical_knowledge",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The clinical knowledge subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Clinical Knowledge"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The clinical knowledge subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Clinical Knowledge"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The clinical knowledge subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Clinical Knowledge"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The clinical knowledge subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Clinical Knowledge"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The clinical knowledge subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Clinical Knowledge"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The clinical knowledge subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Clinical Knowledge"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The clinical knowledge subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Clinical Knowledge"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_clinical_knowledge&subgroup=subject%3A%20clinical_knowledge&runSpecs=%5B%22mmlu%3Asubject%3Dclinical_knowledge%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_clinical_knowledge%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=clinical_knowledge,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_clinical_knowledge"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.2037735849056604,
88
+ "description": "min=0.204, mean=0.204, max=0.204, sum=0.204 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.0775772769496126,
96
+ "description": "min=0.078, mean=0.078, max=0.078, sum=0.078 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 265.0,
104
+ "description": "min=265, mean=265, max=265, sum=265 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 401.9169811320755,
122
+ "description": "min=401.917, mean=401.917, max=401.917, sum=401.917 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_clinical_knowledge_mmlu_clinical_knowledge_subject:clinical_knowledge.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_clinical_knowledge_mmlu_clinical_knowledge_subject:clinical_knowledge.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_clinical_knowledge_subject:clinical_knowledge"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_biology_mmlu_college_biology_subject:college_biology.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: college_biology",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The college biology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "College Biology"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The college biology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "College Biology"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The college biology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "College Biology"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The college biology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "College Biology"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The college biology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "College Biology"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The college biology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "College Biology"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The college biology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "College Biology"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_college_biology&subgroup=subject%3A%20college_biology&runSpecs=%5B%22mmlu%3Asubject%3Dcollege_biology%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_college_biology%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=college_biology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_biology"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.3333333333333333,
88
+ "description": "min=0.333, mean=0.333, max=0.333, sum=0.333 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.08110955854256947,
96
+ "description": "min=0.081, mean=0.081, max=0.081, sum=0.081 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 144.0,
104
+ "description": "min=144, mean=144, max=144, sum=144 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 470.31944444444446,
122
+ "description": "min=470.319, mean=470.319, max=470.319, sum=470.319 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_college_biology_mmlu_college_biology_subject:college_biology.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_biology_mmlu_college_biology_subject:college_biology.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_college_biology_subject:college_biology"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_chemistry_mmlu_college_chemistry_subject:college_chemistry.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: college_chemistry",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The college chemistry subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "College Chemistry"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The college chemistry subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "College Chemistry"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The college chemistry subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "College Chemistry"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The college chemistry subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "College Chemistry"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The college chemistry subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "College Chemistry"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The college chemistry subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "College Chemistry"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The college chemistry subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "College Chemistry"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_college_chemistry&subgroup=subject%3A%20college_chemistry&runSpecs=%5B%22mmlu%3Asubject%3Dcollege_chemistry%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_college_chemistry%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_chemistry"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.26,
88
+ "description": "min=0.26, mean=0.26, max=0.26, sum=0.26 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.08982811689376831,
96
+ "description": "min=0.09, mean=0.09, max=0.09, sum=0.09 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 100.0,
104
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 535.85,
122
+ "description": "min=535.85, mean=535.85, max=535.85, sum=535.85 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_college_chemistry_mmlu_college_chemistry_subject:college_chemistry.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_chemistry_mmlu_college_chemistry_subject:college_chemistry.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_college_chemistry_subject:college_chemistry"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_computer_science_mmlu_college_computer_science_subject:college_computer_science.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: college_computer_science",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The college computer science subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "College Computer Science"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The college computer science subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "College Computer Science"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The college computer science subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "College Computer Science"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The college computer science subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "College Computer Science"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The college computer science subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "College Computer Science"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The college computer science subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "College Computer Science"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The college computer science subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "College Computer Science"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_college_computer_science&subgroup=subject%3A%20college_computer_science&runSpecs=%5B%22mmlu%3Asubject%3Dcollege_computer_science%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_college_computer_science%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=college_computer_science,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_computer_science"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.35,
88
+ "description": "min=0.35, mean=0.35, max=0.35, sum=0.35 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.13751029014587401,
96
+ "description": "min=0.138, mean=0.138, max=0.138, sum=0.138 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 100.0,
104
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 842.89,
122
+ "description": "min=842.89, mean=842.89, max=842.89, sum=842.89 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_college_computer_science_mmlu_college_computer_science_subject:college_computer_science.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_computer_science_mmlu_college_computer_science_subject:college_computer_science.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_college_computer_science_subject:college_computer_science"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_mathematics_mmlu_college_mathematics_subject:college_mathematics.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: college_mathematics",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The college mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "College Mathematics"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The college mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "College Mathematics"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The college mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "College Mathematics"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The college mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "College Mathematics"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The college mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "College Mathematics"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The college mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "College Mathematics"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The college mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "College Mathematics"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_college_mathematics&subgroup=subject%3A%20college_mathematics&runSpecs=%5B%22mmlu%3Asubject%3Dcollege_mathematics%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_college_mathematics%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=college_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_mathematics"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.2,
88
+ "description": "min=0.2, mean=0.2, max=0.2, sum=0.2 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.10322931528091431,
96
+ "description": "min=0.103, mean=0.103, max=0.103, sum=0.103 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 100.0,
104
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 592.82,
122
+ "description": "min=592.82, mean=592.82, max=592.82, sum=592.82 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_college_mathematics_mmlu_college_mathematics_subject:college_mathematics.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_mathematics_mmlu_college_mathematics_subject:college_mathematics.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_college_mathematics_subject:college_mathematics"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_medicine_mmlu_college_medicine_subject:college_medicine.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: college_medicine",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The college medicine subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "College Medicine"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The college medicine subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "College Medicine"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The college medicine subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "College Medicine"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The college medicine subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "College Medicine"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The college medicine subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "College Medicine"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The college medicine subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "College Medicine"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The college medicine subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "College Medicine"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_college_medicine&subgroup=subject%3A%20college_medicine&runSpecs=%5B%22mmlu%3Asubject%3Dcollege_medicine%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_college_medicine%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=college_medicine,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_medicine"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.23699421965317918,
88
+ "description": "min=0.237, mean=0.237, max=0.237, sum=0.237 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.08715281045505766,
96
+ "description": "min=0.087, mean=0.087, max=0.087, sum=0.087 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 173.0,
104
+ "description": "min=173, mean=173, max=173, sum=173 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 519.3757225433526,
122
+ "description": "min=519.376, mean=519.376, max=519.376, sum=519.376 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_college_medicine_mmlu_college_medicine_subject:college_medicine.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_medicine_mmlu_college_medicine_subject:college_medicine.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_college_medicine_subject:college_medicine"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_physics_mmlu_college_physics_subject:college_physics.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: college_physics",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The college physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "College Physics"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The college physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "College Physics"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The college physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "College Physics"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The college physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "College Physics"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The college physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "College Physics"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The college physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "College Physics"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The college physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "College Physics"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_college_physics&subgroup=subject%3A%20college_physics&runSpecs=%5B%22mmlu%3Asubject%3Dcollege_physics%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_college_physics%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=college_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_physics"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.22549019607843138,
88
+ "description": "min=0.225, mean=0.225, max=0.225, sum=0.225 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.08214688534830131,
96
+ "description": "min=0.082, mean=0.082, max=0.082, sum=0.082 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 102.0,
104
+ "description": "min=102, mean=102, max=102, sum=102 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 476.65686274509807,
122
+ "description": "min=476.657, mean=476.657, max=476.657, sum=476.657 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_college_physics_mmlu_college_physics_subject:college_physics.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_college_physics_mmlu_college_physics_subject:college_physics.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_college_physics_subject:college_physics"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_computer_security_mmlu_computer_security_subject:computer_security.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: computer_security",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The computer security subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Computer Security"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The computer security subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Computer Security"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The computer security subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Computer Security"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The computer security subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Computer Security"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The computer security subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Computer Security"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The computer security subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Computer Security"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The computer security subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Computer Security"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_computer_security&subgroup=subject%3A%20computer_security&runSpecs=%5B%22mmlu%3Asubject%3Dcomputer_security%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_computer_security%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_computer_security"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.22,
88
+ "description": "min=0.22, mean=0.22, max=0.22, sum=0.22 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.07067869663238525,
96
+ "description": "min=0.071, mean=0.071, max=0.071, sum=0.071 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 100.0,
104
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 388.19,
122
+ "description": "min=388.19, mean=388.19, max=388.19, sum=388.19 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_computer_security_mmlu_computer_security_subject:computer_security.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_computer_security_mmlu_computer_security_subject:computer_security.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_computer_security_subject:computer_security"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_conceptual_physics_mmlu_conceptual_physics_subject:conceptual_physics.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: conceptual_physics",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The conceptual physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Conceptual Physics"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The conceptual physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Conceptual Physics"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The conceptual physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Conceptual Physics"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The conceptual physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Conceptual Physics"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The conceptual physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Conceptual Physics"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The conceptual physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Conceptual Physics"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The conceptual physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Conceptual Physics"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_conceptual_physics&subgroup=subject%3A%20conceptual_physics&runSpecs=%5B%22mmlu%3Asubject%3Dconceptual_physics%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_conceptual_physics%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=conceptual_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_conceptual_physics"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.2170212765957447,
88
+ "description": "min=0.217, mean=0.217, max=0.217, sum=0.217 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.06520833461842639,
96
+ "description": "min=0.065, mean=0.065, max=0.065, sum=0.065 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 235.0,
104
+ "description": "min=235, mean=235, max=235, sum=235 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 311.31063829787234,
122
+ "description": "min=311.311, mean=311.311, max=311.311, sum=311.311 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_conceptual_physics_mmlu_conceptual_physics_subject:conceptual_physics.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_conceptual_physics_mmlu_conceptual_physics_subject:conceptual_physics.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_conceptual_physics_subject:conceptual_physics"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_econometrics_mmlu_econometrics_subject:econometrics.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: econometrics",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The econometrics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Econometrics"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The econometrics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Econometrics"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The econometrics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Econometrics"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The econometrics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Econometrics"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The econometrics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Econometrics"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The econometrics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Econometrics"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The econometrics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Econometrics"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_econometrics&subgroup=subject%3A%20econometrics&runSpecs=%5B%22mmlu%3Asubject%3Deconometrics%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_econometrics%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_econometrics"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.24561403508771928,
88
+ "description": "min=0.246, mean=0.246, max=0.246, sum=0.246 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.1076611740547314,
96
+ "description": "min=0.108, mean=0.108, max=0.108, sum=0.108 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 114.0,
104
+ "description": "min=114, mean=114, max=114, sum=114 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 612.7982456140351,
122
+ "description": "min=612.798, mean=612.798, max=612.798, sum=612.798 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_econometrics_mmlu_econometrics_subject:econometrics.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_econometrics_mmlu_econometrics_subject:econometrics.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_econometrics_subject:econometrics"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_electrical_engineering_mmlu_electrical_engineering_subject:electrical_engineering.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: electrical_engineering",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The electrical engineering subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Electrical Engineering"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The electrical engineering subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Electrical Engineering"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The electrical engineering subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Electrical Engineering"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The electrical engineering subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Electrical Engineering"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The electrical engineering subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Electrical Engineering"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The electrical engineering subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Electrical Engineering"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The electrical engineering subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Electrical Engineering"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_electrical_engineering&subgroup=subject%3A%20electrical_engineering&runSpecs=%5B%22mmlu%3Asubject%3Delectrical_engineering%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_electrical_engineering%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=electrical_engineering,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_electrical_engineering"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.2827586206896552,
88
+ "description": "min=0.283, mean=0.283, max=0.283, sum=0.283 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.07903624896345468,
96
+ "description": "min=0.079, mean=0.079, max=0.079, sum=0.079 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 145.0,
104
+ "description": "min=145, mean=145, max=145, sum=145 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 424.848275862069,
122
+ "description": "min=424.848, mean=424.848, max=424.848, sum=424.848 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_electrical_engineering_mmlu_electrical_engineering_subject:electrical_engineering.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_electrical_engineering_mmlu_electrical_engineering_subject:electrical_engineering.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_electrical_engineering_subject:electrical_engineering"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_elementary_mathematics_mmlu_elementary_mathematics_subject:elementary_mathematics.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: elementary_mathematics",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The elementary mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Elementary Mathematics"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The elementary mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Elementary Mathematics"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The elementary mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Elementary Mathematics"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The elementary mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Elementary Mathematics"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The elementary mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Elementary Mathematics"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The elementary mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Elementary Mathematics"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The elementary mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Elementary Mathematics"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_elementary_mathematics&subgroup=subject%3A%20elementary_mathematics&runSpecs=%5B%22mmlu%3Asubject%3Delementary_mathematics%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_elementary_mathematics%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=elementary_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_elementary_mathematics"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.2328042328042328,
88
+ "description": "min=0.233, mean=0.233, max=0.233, sum=0.233 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.08565072975461445,
96
+ "description": "min=0.086, mean=0.086, max=0.086, sum=0.086 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 378.0,
104
+ "description": "min=378, mean=378, max=378, sum=378 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 505.07142857142856,
122
+ "description": "min=505.071, mean=505.071, max=505.071, sum=505.071 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_elementary_mathematics_mmlu_elementary_mathematics_subject:elementary_mathematics.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_elementary_mathematics_mmlu_elementary_mathematics_subject:elementary_mathematics.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_elementary_mathematics_subject:elementary_mathematics"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_formal_logic_mmlu_formal_logic_subject:formal_logic.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: formal_logic",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The formal logic subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Formal Logic"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The formal logic subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Formal Logic"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The formal logic subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Formal Logic"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The formal logic subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Formal Logic"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The formal logic subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Formal Logic"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The formal logic subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Formal Logic"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The formal logic subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Formal Logic"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_formal_logic&subgroup=subject%3A%20formal_logic&runSpecs=%5B%22mmlu%3Asubject%3Dformal_logic%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_formal_logic%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=formal_logic,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_formal_logic"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.23015873015873015,
88
+ "description": "min=0.23, mean=0.23, max=0.23, sum=0.23 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.11266928627377465,
96
+ "description": "min=0.113, mean=0.113, max=0.113, sum=0.113 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 126.0,
104
+ "description": "min=126, mean=126, max=126, sum=126 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 653.5952380952381,
122
+ "description": "min=653.595, mean=653.595, max=653.595, sum=653.595 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_formal_logic_mmlu_formal_logic_subject:formal_logic.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_formal_logic_mmlu_formal_logic_subject:formal_logic.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_formal_logic_subject:formal_logic"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_global_facts_mmlu_global_facts_subject:global_facts.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: global_facts",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The global facts subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Global Facts"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The global facts subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Global Facts"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The global facts subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Global Facts"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The global facts subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Global Facts"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The global facts subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Global Facts"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The global facts subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Global Facts"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The global facts subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Global Facts"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_global_facts&subgroup=subject%3A%20global_facts&runSpecs=%5B%22mmlu%3Asubject%3Dglobal_facts%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_global_facts%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=global_facts,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_global_facts"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.28,
88
+ "description": "min=0.28, mean=0.28, max=0.28, sum=0.28 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.07499703168869018,
96
+ "description": "min=0.075, mean=0.075, max=0.075, sum=0.075 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 100.0,
104
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 400.58,
122
+ "description": "min=400.58, mean=400.58, max=400.58, sum=400.58 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_global_facts_mmlu_global_facts_subject:global_facts.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_global_facts_mmlu_global_facts_subject:global_facts.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_global_facts_subject:global_facts"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_biology_mmlu_high_school_biology_subject:high_school_biology.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: high_school_biology",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The high school biology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "High School Biology"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The high school biology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "High School Biology"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The high school biology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "High School Biology"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The high school biology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "High School Biology"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The high school biology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "High School Biology"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The high school biology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "High School Biology"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The high school biology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "High School Biology"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_high_school_biology&subgroup=subject%3A%20high_school_biology&runSpecs=%5B%22mmlu%3Asubject%3Dhigh_school_biology%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_high_school_biology%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=high_school_biology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_biology"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.21935483870967742,
88
+ "description": "min=0.219, mean=0.219, max=0.219, sum=0.219 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.08429911059717979,
96
+ "description": "min=0.084, mean=0.084, max=0.084, sum=0.084 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 310.0,
104
+ "description": "min=310, mean=310, max=310, sum=310 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 513.9322580645161,
122
+ "description": "min=513.932, mean=513.932, max=513.932, sum=513.932 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_high_school_biology_mmlu_high_school_biology_subject:high_school_biology.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_biology_mmlu_high_school_biology_subject:high_school_biology.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_high_school_biology_subject:high_school_biology"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_chemistry_mmlu_high_school_chemistry_subject:high_school_chemistry.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: high_school_chemistry",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The high school chemistry subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "High School Chemistry"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The high school chemistry subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "High School Chemistry"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The high school chemistry subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "High School Chemistry"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The high school chemistry subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "High School Chemistry"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The high school chemistry subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "High School Chemistry"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The high school chemistry subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "High School Chemistry"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The high school chemistry subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "High School Chemistry"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_high_school_chemistry&subgroup=subject%3A%20high_school_chemistry&runSpecs=%5B%22mmlu%3Asubject%3Dhigh_school_chemistry%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_high_school_chemistry%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=high_school_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_chemistry"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.27586206896551724,
88
+ "description": "min=0.276, mean=0.276, max=0.276, sum=0.276 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.08463767714101106,
96
+ "description": "min=0.085, mean=0.085, max=0.085, sum=0.085 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 203.0,
104
+ "description": "min=203, mean=203, max=203, sum=203 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 479.8423645320197,
122
+ "description": "min=479.842, mean=479.842, max=479.842, sum=479.842 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_high_school_chemistry_mmlu_high_school_chemistry_subject:high_school_chemistry.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_chemistry_mmlu_high_school_chemistry_subject:high_school_chemistry.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_high_school_chemistry_subject:high_school_chemistry"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_computer_science_mmlu_high_school_computer_science_subject:high_school_computer_science.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: high_school_computer_science",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The high school computer science subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "High School Computer Science"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The high school computer science subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "High School Computer Science"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The high school computer science subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "High School Computer Science"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The high school computer science subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "High School Computer Science"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The high school computer science subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "High School Computer Science"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The high school computer science subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "High School Computer Science"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The high school computer science subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "High School Computer Science"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_high_school_computer_science&subgroup=subject%3A%20high_school_computer_science&runSpecs=%5B%22mmlu%3Asubject%3Dhigh_school_computer_science%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_high_school_computer_science%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=high_school_computer_science,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_computer_science"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.17,
88
+ "description": "min=0.17, mean=0.17, max=0.17, sum=0.17 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.14540541887283326,
96
+ "description": "min=0.145, mean=0.145, max=0.145, sum=0.145 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 100.0,
104
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 889.39,
122
+ "description": "min=889.39, mean=889.39, max=889.39, sum=889.39 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_high_school_computer_science_mmlu_high_school_computer_science_subject:high_school_computer_science.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_computer_science_mmlu_high_school_computer_science_subject:high_school_computer_science.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_high_school_computer_science_subject:high_school_computer_science"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_european_history_mmlu_high_school_european_history_subject:high_school_european_history.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: high_school_european_history",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The high school european history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "High School European History"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The high school european history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "High School European History"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The high school european history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "High School European History"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The high school european history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "High School European History"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The high school european history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "High School European History"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The high school european history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "High School European History"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The high school european history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "High School European History"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_high_school_european_history&subgroup=subject%3A%20high_school_european_history&runSpecs=%5B%22mmlu%3Asubject%3Dhigh_school_european_history%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_high_school_european_history%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=high_school_european_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_european_history"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.20606060606060606,
88
+ "description": "min=0.206, mean=0.206, max=0.206, sum=0.206 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.3232730460889412,
96
+ "description": "min=0.323, mean=0.323, max=0.323, sum=0.323 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 165.0,
104
+ "description": "min=165, mean=165, max=165, sum=165 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 2.909090909090909,
110
+ "description": "min=2.909, mean=2.909, max=2.909, sum=2.909 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 1816.7575757575758,
122
+ "description": "min=1816.758, mean=1816.758, max=1816.758, sum=1816.758 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_high_school_european_history_mmlu_high_school_european_history_subject:high_school_european_history.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_european_history_mmlu_high_school_european_history_subject:high_school_european_history.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_high_school_european_history_subject:high_school_european_history"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_geography_mmlu_high_school_geography_subject:high_school_geography.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: high_school_geography",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The high school geography subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "High School Geography"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The high school geography subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "High School Geography"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The high school geography subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "High School Geography"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The high school geography subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "High School Geography"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The high school geography subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "High School Geography"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The high school geography subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "High School Geography"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The high school geography subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "High School Geography"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_high_school_geography&subgroup=subject%3A%20high_school_geography&runSpecs=%5B%22mmlu%3Asubject%3Dhigh_school_geography%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_high_school_geography%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=high_school_geography,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_geography"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.2828282828282828,
88
+ "description": "min=0.283, mean=0.283, max=0.283, sum=0.283 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.07503375019690003,
96
+ "description": "min=0.075, mean=0.075, max=0.075, sum=0.075 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 198.0,
104
+ "description": "min=198, mean=198, max=198, sum=198 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 400.09090909090907,
122
+ "description": "min=400.091, mean=400.091, max=400.091, sum=400.091 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_high_school_geography_mmlu_high_school_geography_subject:high_school_geography.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_geography_mmlu_high_school_geography_subject:high_school_geography.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_high_school_geography_subject:high_school_geography"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_government_and_politics_mmlu_high_school_government_and_politics_subject:high_school_government_and_politics.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: high_school_government_and_politics",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The high school government and politics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "High School Government And Politics"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The high school government and politics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "High School Government And Politics"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The high school government and politics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "High School Government And Politics"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The high school government and politics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "High School Government And Politics"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The high school government and politics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "High School Government And Politics"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The high school government and politics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "High School Government And Politics"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The high school government and politics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "High School Government And Politics"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_high_school_government_and_politics&subgroup=subject%3A%20high_school_government_and_politics&runSpecs=%5B%22mmlu%3Asubject%3Dhigh_school_government_and_politics%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_high_school_government_and_politics%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=high_school_government_and_politics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_government_and_politics"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.2849740932642487,
88
+ "description": "min=0.285, mean=0.285, max=0.285, sum=0.285 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.08498252858769709,
96
+ "description": "min=0.085, mean=0.085, max=0.085, sum=0.085 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 193.0,
104
+ "description": "min=193, mean=193, max=193, sum=193 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 482.7616580310881,
122
+ "description": "min=482.762, mean=482.762, max=482.762, sum=482.762 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_high_school_government_and_politics_mmlu_high_school_government_and_politics_subject:high_school_government_and_politics.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_government_and_politics_mmlu_high_school_government_and_politics_subject:high_school_government_and_politics.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_high_school_government_and_politics_subject:high_school_government_and_politics"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_macroeconomics_mmlu_high_school_macroeconomics_subject:high_school_macroeconomics.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: high_school_macroeconomics",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The high school macroeconomics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "High School Macroeconomics"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The high school macroeconomics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "High School Macroeconomics"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The high school macroeconomics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "High School Macroeconomics"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The high school macroeconomics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "High School Macroeconomics"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The high school macroeconomics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "High School Macroeconomics"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The high school macroeconomics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "High School Macroeconomics"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The high school macroeconomics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "High School Macroeconomics"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_high_school_macroeconomics&subgroup=subject%3A%20high_school_macroeconomics&runSpecs=%5B%22mmlu%3Asubject%3Dhigh_school_macroeconomics%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_high_school_macroeconomics%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=high_school_macroeconomics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_macroeconomics"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.3384615384615385,
88
+ "description": "min=0.338, mean=0.338, max=0.338, sum=0.338 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.07281008500319261,
96
+ "description": "min=0.073, mean=0.073, max=0.073, sum=0.073 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 390.0,
104
+ "description": "min=390, mean=390, max=390, sum=390 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 392.35128205128206,
122
+ "description": "min=392.351, mean=392.351, max=392.351, sum=392.351 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_high_school_macroeconomics_mmlu_high_school_macroeconomics_subject:high_school_macroeconomics.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_macroeconomics_mmlu_high_school_macroeconomics_subject:high_school_macroeconomics.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_high_school_macroeconomics_subject:high_school_macroeconomics"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_mathematics_mmlu_high_school_mathematics_subject:high_school_mathematics.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: high_school_mathematics",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The high school mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "High School Mathematics"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The high school mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "High School Mathematics"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The high school mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "High School Mathematics"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The high school mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "High School Mathematics"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The high school mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "High School Mathematics"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The high school mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "High School Mathematics"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The high school mathematics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "High School Mathematics"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_high_school_mathematics&subgroup=subject%3A%20high_school_mathematics&runSpecs=%5B%22mmlu%3Asubject%3Dhigh_school_mathematics%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_high_school_mathematics%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=high_school_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_mathematics"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.2518518518518518,
88
+ "description": "min=0.252, mean=0.252, max=0.252, sum=0.252 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.08458723138879846,
96
+ "description": "min=0.085, mean=0.085, max=0.085, sum=0.085 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 270.0,
104
+ "description": "min=270, mean=270, max=270, sum=270 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 506.68888888888887,
122
+ "description": "min=506.689, mean=506.689, max=506.689, sum=506.689 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_high_school_mathematics_mmlu_high_school_mathematics_subject:high_school_mathematics.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_mathematics_mmlu_high_school_mathematics_subject:high_school_mathematics.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_high_school_mathematics_subject:high_school_mathematics"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_microeconomics_mmlu_high_school_microeconomics_subject:high_school_microeconomics.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: high_school_microeconomics",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The high school microeconomics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "High School Microeconomics"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The high school microeconomics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "High School Microeconomics"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The high school microeconomics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "High School Microeconomics"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The high school microeconomics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "High School Microeconomics"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The high school microeconomics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "High School Microeconomics"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The high school microeconomics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "High School Microeconomics"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The high school microeconomics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "High School Microeconomics"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_high_school_microeconomics&subgroup=subject%3A%20high_school_microeconomics&runSpecs=%5B%22mmlu%3Asubject%3Dhigh_school_microeconomics%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_high_school_microeconomics%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=high_school_microeconomics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_microeconomics"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.3445378151260504,
88
+ "description": "min=0.345, mean=0.345, max=0.345, sum=0.345 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.07571322076460894,
96
+ "description": "min=0.076, mean=0.076, max=0.076, sum=0.076 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 238.0,
104
+ "description": "min=238, mean=238, max=238, sum=238 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 411.2352941176471,
122
+ "description": "min=411.235, mean=411.235, max=411.235, sum=411.235 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_high_school_microeconomics_mmlu_high_school_microeconomics_subject:high_school_microeconomics.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_microeconomics_mmlu_high_school_microeconomics_subject:high_school_microeconomics.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_high_school_microeconomics_subject:high_school_microeconomics"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_physics_mmlu_high_school_physics_subject:high_school_physics.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: high_school_physics",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The high school physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "High School Physics"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The high school physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "High School Physics"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The high school physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "High School Physics"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The high school physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "High School Physics"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The high school physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "High School Physics"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The high school physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "High School Physics"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The high school physics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "High School Physics"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_high_school_physics&subgroup=subject%3A%20high_school_physics&runSpecs=%5B%22mmlu%3Asubject%3Dhigh_school_physics%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_high_school_physics%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=high_school_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_physics"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.31125827814569534,
88
+ "description": "min=0.311, mean=0.311, max=0.311, sum=0.311 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.09131200739879482,
96
+ "description": "min=0.091, mean=0.091, max=0.091, sum=0.091 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 151.0,
104
+ "description": "min=151, mean=151, max=151, sum=151 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 548.7284768211921,
122
+ "description": "min=548.728, mean=548.728, max=548.728, sum=548.728 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_high_school_physics_mmlu_high_school_physics_subject:high_school_physics.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_physics_mmlu_high_school_physics_subject:high_school_physics.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_high_school_physics_subject:high_school_physics"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_psychology_mmlu_high_school_psychology_subject:high_school_psychology.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: high_school_psychology",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The high school psychology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "High School Psychology"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The high school psychology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "High School Psychology"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The high school psychology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "High School Psychology"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The high school psychology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "High School Psychology"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The high school psychology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "High School Psychology"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The high school psychology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "High School Psychology"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The high school psychology subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "High School Psychology"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_high_school_psychology&subgroup=subject%3A%20high_school_psychology&runSpecs=%5B%22mmlu%3Asubject%3Dhigh_school_psychology%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_high_school_psychology%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=high_school_psychology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_psychology"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.26238532110091745,
88
+ "description": "min=0.262, mean=0.262, max=0.262, sum=0.262 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.08488267058626227,
96
+ "description": "min=0.085, mean=0.085, max=0.085, sum=0.085 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 545.0,
104
+ "description": "min=545, mean=545, max=545, sum=545 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 514.7926605504587,
122
+ "description": "min=514.793, mean=514.793, max=514.793, sum=514.793 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_high_school_psychology_mmlu_high_school_psychology_subject:high_school_psychology.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_psychology_mmlu_high_school_psychology_subject:high_school_psychology.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_high_school_psychology_subject:high_school_psychology"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_statistics_mmlu_high_school_statistics_subject:high_school_statistics.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: high_school_statistics",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The high school statistics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "High School Statistics"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The high school statistics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "High School Statistics"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The high school statistics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "High School Statistics"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The high school statistics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "High School Statistics"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The high school statistics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "High School Statistics"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The high school statistics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "High School Statistics"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The high school statistics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "High School Statistics"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_high_school_statistics&subgroup=subject%3A%20high_school_statistics&runSpecs=%5B%22mmlu%3Asubject%3Dhigh_school_statistics%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_high_school_statistics%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=high_school_statistics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_statistics"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.4351851851851852,
88
+ "description": "min=0.435, mean=0.435, max=0.435, sum=0.435 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.13173586792416042,
96
+ "description": "min=0.132, mean=0.132, max=0.132, sum=0.132 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 216.0,
104
+ "description": "min=216, mean=216, max=216, sum=216 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 796.6064814814815,
122
+ "description": "min=796.606, mean=796.606, max=796.606, sum=796.606 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_high_school_statistics_mmlu_high_school_statistics_subject:high_school_statistics.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_statistics_mmlu_high_school_statistics_subject:high_school_statistics.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_high_school_statistics_subject:high_school_statistics"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_us_history_mmlu_high_school_us_history_subject:high_school_us_history.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: high_school_us_history",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The high school us history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "High School US History"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The high school us history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "High School US History"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The high school us history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "High School US History"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The high school us history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "High School US History"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The high school us history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "High School US History"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The high school us history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "High School US History"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The high school us history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "High School US History"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_high_school_us_history&subgroup=subject%3A%20high_school_us_history&runSpecs=%5B%22mmlu%3Asubject%3Dhigh_school_us_history%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_high_school_us_history%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=high_school_us_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_us_history"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.23529411764705882,
88
+ "description": "min=0.235, mean=0.235, max=0.235, sum=0.235 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.3188782766753552,
96
+ "description": "min=0.319, mean=0.319, max=0.319, sum=0.319 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 204.0,
104
+ "description": "min=204, mean=204, max=204, sum=204 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 4.0,
110
+ "description": "min=4, mean=4, max=4, sum=4 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 1788.387254901961,
122
+ "description": "min=1788.387, mean=1788.387, max=1788.387, sum=1788.387 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_high_school_us_history_mmlu_high_school_us_history_subject:high_school_us_history.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_us_history_mmlu_high_school_us_history_subject:high_school_us_history.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_high_school_us_history_subject:high_school_us_history"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_world_history_mmlu_high_school_world_history_subject:high_school_world_history.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: high_school_world_history",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The high school world history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "High School World History"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The high school world history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "High School World History"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The high school world history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "High School World History"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The high school world history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "High School World History"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The high school world history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "High School World History"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The high school world history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "High School World History"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The high school world history subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "High School World History"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_high_school_world_history&subgroup=subject%3A%20high_school_world_history&runSpecs=%5B%22mmlu%3Asubject%3Dhigh_school_world_history%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_high_school_world_history%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=high_school_world_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_world_history"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.18143459915611815,
88
+ "description": "min=0.181, mean=0.181, max=0.181, sum=0.181 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.2615184371481465,
96
+ "description": "min=0.262, mean=0.262, max=0.262, sum=0.262 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 237.0,
104
+ "description": "min=237, mean=237, max=237, sum=237 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 1461.4430379746836,
122
+ "description": "min=1461.443, mean=1461.443, max=1461.443, sum=1461.443 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_high_school_world_history_mmlu_high_school_world_history_subject:high_school_world_history.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_high_school_world_history_mmlu_high_school_world_history_subject:high_school_world_history.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_high_school_world_history_subject:high_school_world_history"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_human_aging_mmlu_human_aging_subject:human_aging.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: human_aging",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The human aging subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Human Aging"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The human aging subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Human Aging"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The human aging subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Human Aging"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The human aging subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Human Aging"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The human aging subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Human Aging"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The human aging subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Human Aging"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The human aging subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Human Aging"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_human_aging&subgroup=subject%3A%20human_aging&runSpecs=%5B%22mmlu%3Asubject%3Dhuman_aging%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_human_aging%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=human_aging,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_human_aging"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.16143497757847533,
88
+ "description": "min=0.161, mean=0.161, max=0.161, sum=0.161 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.06565283766776457,
96
+ "description": "min=0.066, mean=0.066, max=0.066, sum=0.066 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 223.0,
104
+ "description": "min=223, mean=223, max=223, sum=223 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 323.69058295964123,
122
+ "description": "min=323.691, mean=323.691, max=323.691, sum=323.691 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_human_aging_mmlu_human_aging_subject:human_aging.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_human_aging_mmlu_human_aging_subject:human_aging.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_human_aging_subject:human_aging"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_human_sexuality_mmlu_human_sexuality_subject:human_sexuality.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: human_sexuality",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The human sexuality subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Human Sexuality"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The human sexuality subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Human Sexuality"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The human sexuality subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Human Sexuality"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The human sexuality subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Human Sexuality"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The human sexuality subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Human Sexuality"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The human sexuality subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Human Sexuality"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The human sexuality subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Human Sexuality"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_human_sexuality&subgroup=subject%3A%20human_sexuality&runSpecs=%5B%22mmlu%3Asubject%3Dhuman_sexuality%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_human_sexuality%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=human_sexuality,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_human_sexuality"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.2595419847328244,
88
+ "description": "min=0.26, mean=0.26, max=0.26, sum=0.26 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.06619657268960967,
96
+ "description": "min=0.066, mean=0.066, max=0.066, sum=0.066 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 131.0,
104
+ "description": "min=131, mean=131, max=131, sum=131 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 355.35114503816794,
122
+ "description": "min=355.351, mean=355.351, max=355.351, sum=355.351 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_human_sexuality_mmlu_human_sexuality_subject:human_sexuality.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_human_sexuality_mmlu_human_sexuality_subject:human_sexuality.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_human_sexuality_subject:human_sexuality"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_international_law_mmlu_international_law_subject:international_law.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: international_law",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The international law subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "International Law"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The international law subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "International Law"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The international law subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "International Law"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The international law subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "International Law"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The international law subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "International Law"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The international law subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "International Law"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The international law subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "International Law"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_international_law&subgroup=subject%3A%20international_law&runSpecs=%5B%22mmlu%3Asubject%3Dinternational_law%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_international_law%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=international_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_international_law"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.256198347107438,
88
+ "description": "min=0.256, mean=0.256, max=0.256, sum=0.256 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.10936137073296161,
96
+ "description": "min=0.109, mean=0.109, max=0.109, sum=0.109 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 121.0,
104
+ "description": "min=121, mean=121, max=121, sum=121 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 650.3719008264463,
122
+ "description": "min=650.372, mean=650.372, max=650.372, sum=650.372 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_international_law_mmlu_international_law_subject:international_law.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_international_law_mmlu_international_law_subject:international_law.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_international_law_subject:international_law"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_jurisprudence_mmlu_jurisprudence_subject:jurisprudence.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: jurisprudence",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The jurisprudence subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Jurisprudence"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The jurisprudence subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Jurisprudence"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The jurisprudence subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Jurisprudence"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The jurisprudence subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Jurisprudence"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The jurisprudence subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Jurisprudence"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The jurisprudence subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Jurisprudence"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The jurisprudence subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Jurisprudence"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_jurisprudence&subgroup=subject%3A%20jurisprudence&runSpecs=%5B%22mmlu%3Asubject%3Djurisprudence%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_jurisprudence%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=jurisprudence,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_jurisprudence"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.21296296296296297,
88
+ "description": "min=0.213, mean=0.213, max=0.213, sum=0.213 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.07681376183474506,
96
+ "description": "min=0.077, mean=0.077, max=0.077, sum=0.077 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 108.0,
104
+ "description": "min=108, mean=108, max=108, sum=108 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 420.8611111111111,
122
+ "description": "min=420.861, mean=420.861, max=420.861, sum=420.861 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_jurisprudence_mmlu_jurisprudence_subject:jurisprudence.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_jurisprudence_mmlu_jurisprudence_subject:jurisprudence.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_jurisprudence_subject:jurisprudence"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_logical_fallacies_mmlu_logical_fallacies_subject:logical_fallacies.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: logical_fallacies",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The logical fallacies subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Logical Fallacies"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The logical fallacies subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Logical Fallacies"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The logical fallacies subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Logical Fallacies"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The logical fallacies subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Logical Fallacies"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The logical fallacies subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Logical Fallacies"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The logical fallacies subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Logical Fallacies"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The logical fallacies subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Logical Fallacies"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_logical_fallacies&subgroup=subject%3A%20logical_fallacies&runSpecs=%5B%22mmlu%3Asubject%3Dlogical_fallacies%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_logical_fallacies%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=logical_fallacies,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_logical_fallacies"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.22699386503067484,
88
+ "description": "min=0.227, mean=0.227, max=0.227, sum=0.227 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.08394306715280732,
96
+ "description": "min=0.084, mean=0.084, max=0.084, sum=0.084 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 163.0,
104
+ "description": "min=163, mean=163, max=163, sum=163 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 458.8282208588957,
122
+ "description": "min=458.828, mean=458.828, max=458.828, sum=458.828 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_logical_fallacies_mmlu_logical_fallacies_subject:logical_fallacies.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_logical_fallacies_mmlu_logical_fallacies_subject:logical_fallacies.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_logical_fallacies_subject:logical_fallacies"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_machine_learning_mmlu_machine_learning_subject:machine_learning.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: machine_learning",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The machine learning subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Machine Learning"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The machine learning subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Machine Learning"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The machine learning subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Machine Learning"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The machine learning subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Machine Learning"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The machine learning subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Machine Learning"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The machine learning subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Machine Learning"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The machine learning subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Machine Learning"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_machine_learning&subgroup=subject%3A%20machine_learning&runSpecs=%5B%22mmlu%3Asubject%3Dmachine_learning%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_machine_learning%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=machine_learning,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_machine_learning"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.21428571428571427,
88
+ "description": "min=0.214, mean=0.214, max=0.214, sum=0.214 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.11298963214669909,
96
+ "description": "min=0.113, mean=0.113, max=0.113, sum=0.113 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 112.0,
104
+ "description": "min=112, mean=112, max=112, sum=112 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 661.2142857142857,
122
+ "description": "min=661.214, mean=661.214, max=661.214, sum=661.214 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_machine_learning_mmlu_machine_learning_subject:machine_learning.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_machine_learning_mmlu_machine_learning_subject:machine_learning.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_machine_learning_subject:machine_learning"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_management_mmlu_management_subject:management.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: management",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The management subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Management"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The management subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Management"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The management subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Management"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The management subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Management"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The management subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Management"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The management subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Management"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The management subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Management"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_management&subgroup=subject%3A%20management&runSpecs=%5B%22mmlu%3Asubject%3Dmanagement%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_management%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=management,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_management"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.17475728155339806,
88
+ "description": "min=0.175, mean=0.175, max=0.175, sum=0.175 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.06428557923696573,
96
+ "description": "min=0.064, mean=0.064, max=0.064, sum=0.064 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 103.0,
104
+ "description": "min=103, mean=103, max=103, sum=103 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 298.0485436893204,
122
+ "description": "min=298.049, mean=298.049, max=298.049, sum=298.049 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_management_mmlu_management_subject:management.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_management_mmlu_management_subject:management.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_management_subject:management"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_marketing_mmlu_marketing_subject:marketing.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: marketing",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The marketing subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Marketing"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The marketing subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Marketing"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The marketing subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Marketing"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The marketing subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Marketing"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The marketing subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Marketing"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The marketing subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Marketing"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The marketing subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Marketing"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_marketing&subgroup=subject%3A%20marketing&runSpecs=%5B%22mmlu%3Asubject%3Dmarketing%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_marketing%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=marketing,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_marketing"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.23931623931623933,
88
+ "description": "min=0.239, mean=0.239, max=0.239, sum=0.239 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.08168192704518636,
96
+ "description": "min=0.082, mean=0.082, max=0.082, sum=0.082 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 234.0,
104
+ "description": "min=234, mean=234, max=234, sum=234 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 440.1025641025641,
122
+ "description": "min=440.103, mean=440.103, max=440.103, sum=440.103 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_marketing_mmlu_marketing_subject:marketing.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_marketing_mmlu_marketing_subject:marketing.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_marketing_subject:marketing"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_medical_genetics_mmlu_medical_genetics_subject:medical_genetics.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: medical_genetics",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The medical genetics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Medical Genetics"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The medical genetics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Medical Genetics"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The medical genetics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Medical Genetics"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The medical genetics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Medical Genetics"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The medical genetics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Medical Genetics"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The medical genetics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Medical Genetics"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The medical genetics subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Medical Genetics"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_medical_genetics&subgroup=subject%3A%20medical_genetics&runSpecs=%5B%22mmlu%3Asubject%3Dmedical_genetics%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_medical_genetics%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=medical_genetics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_medical_genetics"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.31,
88
+ "description": "min=0.31, mean=0.31, max=0.31, sum=0.31 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.06659618139266968,
96
+ "description": "min=0.067, mean=0.067, max=0.067, sum=0.067 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 100.0,
104
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 340.48,
122
+ "description": "min=340.48, mean=340.48, max=340.48, sum=340.48 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_medical_genetics_mmlu_medical_genetics_subject:medical_genetics.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_medical_genetics_mmlu_medical_genetics_subject:medical_genetics.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_medical_genetics_subject:medical_genetics"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_miscellaneous_mmlu_miscellaneous_subject:miscellaneous.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: miscellaneous",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The miscellaneous subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "Miscellaneous"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The miscellaneous subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "Miscellaneous"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The miscellaneous subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "Miscellaneous"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The miscellaneous subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "Miscellaneous"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The miscellaneous subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "Miscellaneous"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The miscellaneous subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "Miscellaneous"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The miscellaneous subject in the Massive Multitask Language Understanding (MMLU) benchmark.\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "Miscellaneous"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu_miscellaneous&subgroup=subject%3A%20miscellaneous&runSpecs=%5B%22mmlu%3Asubject%3Dmiscellaneous%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_miscellaneous%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=miscellaneous,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_miscellaneous"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.26436781609195403,
88
+ "description": "min=0.264, mean=0.264, max=0.264, sum=0.264 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.06466218734000682,
96
+ "description": "min=0.065, mean=0.065, max=0.065, sum=0.065 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 783.0,
104
+ "description": "min=783, mean=783, max=783, sum=783 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 320.4431673052363,
122
+ "description": "min=320.443, mean=320.443, max=320.443, sum=320.443 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_miscellaneous_mmlu_miscellaneous_subject:miscellaneous.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_miscellaneous_mmlu_miscellaneous_subject:miscellaneous.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_miscellaneous_subject:miscellaneous"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_mmlu.json ADDED
@@ -0,0 +1,554 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "Massive Multitask Language Understanding (MMLU) All Subjects",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "MMLU All Subjects"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "MMLU All Subjects"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "MMLU All Subjects"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "MMLU All Subjects"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "MMLU All Subjects"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "MMLU All Subjects"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "MMLU All Subjects"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "markdown": false
81
+ },
82
+ {
83
+ "value": 0.2514631519798282,
84
+ "description": "min=0.161, mean=0.251, max=0.438, sum=14.333 (57)",
85
+ "style": {
86
+ "font-weight": "bold"
87
+ },
88
+ "markdown": false,
89
+ "run_spec_names": [
90
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_abstract_algebra",
91
+ "mmlu:subject=anatomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_anatomy",
92
+ "mmlu:subject=astronomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_astronomy",
93
+ "mmlu:subject=business_ethics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_business_ethics",
94
+ "mmlu:subject=clinical_knowledge,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_clinical_knowledge",
95
+ "mmlu:subject=college_biology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_biology",
96
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_chemistry",
97
+ "mmlu:subject=college_computer_science,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_computer_science",
98
+ "mmlu:subject=college_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_mathematics",
99
+ "mmlu:subject=college_medicine,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_medicine",
100
+ "mmlu:subject=college_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_physics",
101
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_computer_security",
102
+ "mmlu:subject=conceptual_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_conceptual_physics",
103
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_econometrics",
104
+ "mmlu:subject=electrical_engineering,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_electrical_engineering",
105
+ "mmlu:subject=elementary_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_elementary_mathematics",
106
+ "mmlu:subject=formal_logic,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_formal_logic",
107
+ "mmlu:subject=global_facts,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_global_facts",
108
+ "mmlu:subject=high_school_biology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_biology",
109
+ "mmlu:subject=high_school_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_chemistry",
110
+ "mmlu:subject=high_school_computer_science,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_computer_science",
111
+ "mmlu:subject=high_school_european_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_european_history",
112
+ "mmlu:subject=high_school_geography,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_geography",
113
+ "mmlu:subject=high_school_government_and_politics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_government_and_politics",
114
+ "mmlu:subject=high_school_macroeconomics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_macroeconomics",
115
+ "mmlu:subject=high_school_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_mathematics",
116
+ "mmlu:subject=high_school_microeconomics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_microeconomics",
117
+ "mmlu:subject=high_school_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_physics",
118
+ "mmlu:subject=high_school_psychology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_psychology",
119
+ "mmlu:subject=high_school_statistics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_statistics",
120
+ "mmlu:subject=high_school_us_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_us_history",
121
+ "mmlu:subject=high_school_world_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_world_history",
122
+ "mmlu:subject=human_aging,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_human_aging",
123
+ "mmlu:subject=human_sexuality,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_human_sexuality",
124
+ "mmlu:subject=international_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_international_law",
125
+ "mmlu:subject=jurisprudence,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_jurisprudence",
126
+ "mmlu:subject=logical_fallacies,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_logical_fallacies",
127
+ "mmlu:subject=machine_learning,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_machine_learning",
128
+ "mmlu:subject=management,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_management",
129
+ "mmlu:subject=marketing,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_marketing",
130
+ "mmlu:subject=medical_genetics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_medical_genetics",
131
+ "mmlu:subject=miscellaneous,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_miscellaneous",
132
+ "mmlu:subject=moral_disputes,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_moral_disputes",
133
+ "mmlu:subject=moral_scenarios,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_moral_scenarios",
134
+ "mmlu:subject=nutrition,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_nutrition",
135
+ "mmlu:subject=philosophy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_philosophy",
136
+ "mmlu:subject=prehistory,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_prehistory",
137
+ "mmlu:subject=professional_accounting,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_accounting",
138
+ "mmlu:subject=professional_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_law",
139
+ "mmlu:subject=professional_medicine,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_medicine",
140
+ "mmlu:subject=professional_psychology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_psychology",
141
+ "mmlu:subject=public_relations,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_public_relations",
142
+ "mmlu:subject=security_studies,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_security_studies",
143
+ "mmlu:subject=sociology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_sociology",
144
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_us_foreign_policy",
145
+ "mmlu:subject=virology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_virology",
146
+ "mmlu:subject=world_religions,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_world_religions"
147
+ ]
148
+ },
149
+ {
150
+ "value": 0.1056762656473459,
151
+ "description": "min=0.061, mean=0.106, max=0.323, sum=6.024 (57)",
152
+ "style": {
153
+ "font-weight": "bold"
154
+ },
155
+ "markdown": false,
156
+ "run_spec_names": [
157
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_abstract_algebra",
158
+ "mmlu:subject=anatomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_anatomy",
159
+ "mmlu:subject=astronomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_astronomy",
160
+ "mmlu:subject=business_ethics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_business_ethics",
161
+ "mmlu:subject=clinical_knowledge,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_clinical_knowledge",
162
+ "mmlu:subject=college_biology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_biology",
163
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_chemistry",
164
+ "mmlu:subject=college_computer_science,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_computer_science",
165
+ "mmlu:subject=college_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_mathematics",
166
+ "mmlu:subject=college_medicine,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_medicine",
167
+ "mmlu:subject=college_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_physics",
168
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_computer_security",
169
+ "mmlu:subject=conceptual_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_conceptual_physics",
170
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_econometrics",
171
+ "mmlu:subject=electrical_engineering,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_electrical_engineering",
172
+ "mmlu:subject=elementary_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_elementary_mathematics",
173
+ "mmlu:subject=formal_logic,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_formal_logic",
174
+ "mmlu:subject=global_facts,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_global_facts",
175
+ "mmlu:subject=high_school_biology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_biology",
176
+ "mmlu:subject=high_school_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_chemistry",
177
+ "mmlu:subject=high_school_computer_science,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_computer_science",
178
+ "mmlu:subject=high_school_european_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_european_history",
179
+ "mmlu:subject=high_school_geography,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_geography",
180
+ "mmlu:subject=high_school_government_and_politics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_government_and_politics",
181
+ "mmlu:subject=high_school_macroeconomics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_macroeconomics",
182
+ "mmlu:subject=high_school_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_mathematics",
183
+ "mmlu:subject=high_school_microeconomics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_microeconomics",
184
+ "mmlu:subject=high_school_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_physics",
185
+ "mmlu:subject=high_school_psychology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_psychology",
186
+ "mmlu:subject=high_school_statistics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_statistics",
187
+ "mmlu:subject=high_school_us_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_us_history",
188
+ "mmlu:subject=high_school_world_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_world_history",
189
+ "mmlu:subject=human_aging,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_human_aging",
190
+ "mmlu:subject=human_sexuality,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_human_sexuality",
191
+ "mmlu:subject=international_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_international_law",
192
+ "mmlu:subject=jurisprudence,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_jurisprudence",
193
+ "mmlu:subject=logical_fallacies,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_logical_fallacies",
194
+ "mmlu:subject=machine_learning,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_machine_learning",
195
+ "mmlu:subject=management,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_management",
196
+ "mmlu:subject=marketing,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_marketing",
197
+ "mmlu:subject=medical_genetics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_medical_genetics",
198
+ "mmlu:subject=miscellaneous,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_miscellaneous",
199
+ "mmlu:subject=moral_disputes,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_moral_disputes",
200
+ "mmlu:subject=moral_scenarios,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_moral_scenarios",
201
+ "mmlu:subject=nutrition,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_nutrition",
202
+ "mmlu:subject=philosophy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_philosophy",
203
+ "mmlu:subject=prehistory,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_prehistory",
204
+ "mmlu:subject=professional_accounting,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_accounting",
205
+ "mmlu:subject=professional_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_law",
206
+ "mmlu:subject=professional_medicine,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_medicine",
207
+ "mmlu:subject=professional_psychology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_psychology",
208
+ "mmlu:subject=public_relations,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_public_relations",
209
+ "mmlu:subject=security_studies,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_security_studies",
210
+ "mmlu:subject=sociology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_sociology",
211
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_us_foreign_policy",
212
+ "mmlu:subject=virology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_virology",
213
+ "mmlu:subject=world_religions,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_world_religions"
214
+ ]
215
+ },
216
+ {
217
+ "value": 246.35087719298247,
218
+ "description": "min=100, mean=246.351, max=1534, sum=14042 (57)",
219
+ "style": {},
220
+ "markdown": false,
221
+ "run_spec_names": [
222
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_abstract_algebra",
223
+ "mmlu:subject=anatomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_anatomy",
224
+ "mmlu:subject=astronomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_astronomy",
225
+ "mmlu:subject=business_ethics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_business_ethics",
226
+ "mmlu:subject=clinical_knowledge,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_clinical_knowledge",
227
+ "mmlu:subject=college_biology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_biology",
228
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_chemistry",
229
+ "mmlu:subject=college_computer_science,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_computer_science",
230
+ "mmlu:subject=college_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_mathematics",
231
+ "mmlu:subject=college_medicine,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_medicine",
232
+ "mmlu:subject=college_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_physics",
233
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_computer_security",
234
+ "mmlu:subject=conceptual_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_conceptual_physics",
235
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_econometrics",
236
+ "mmlu:subject=electrical_engineering,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_electrical_engineering",
237
+ "mmlu:subject=elementary_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_elementary_mathematics",
238
+ "mmlu:subject=formal_logic,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_formal_logic",
239
+ "mmlu:subject=global_facts,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_global_facts",
240
+ "mmlu:subject=high_school_biology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_biology",
241
+ "mmlu:subject=high_school_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_chemistry",
242
+ "mmlu:subject=high_school_computer_science,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_computer_science",
243
+ "mmlu:subject=high_school_european_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_european_history",
244
+ "mmlu:subject=high_school_geography,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_geography",
245
+ "mmlu:subject=high_school_government_and_politics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_government_and_politics",
246
+ "mmlu:subject=high_school_macroeconomics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_macroeconomics",
247
+ "mmlu:subject=high_school_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_mathematics",
248
+ "mmlu:subject=high_school_microeconomics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_microeconomics",
249
+ "mmlu:subject=high_school_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_physics",
250
+ "mmlu:subject=high_school_psychology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_psychology",
251
+ "mmlu:subject=high_school_statistics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_statistics",
252
+ "mmlu:subject=high_school_us_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_us_history",
253
+ "mmlu:subject=high_school_world_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_world_history",
254
+ "mmlu:subject=human_aging,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_human_aging",
255
+ "mmlu:subject=human_sexuality,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_human_sexuality",
256
+ "mmlu:subject=international_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_international_law",
257
+ "mmlu:subject=jurisprudence,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_jurisprudence",
258
+ "mmlu:subject=logical_fallacies,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_logical_fallacies",
259
+ "mmlu:subject=machine_learning,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_machine_learning",
260
+ "mmlu:subject=management,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_management",
261
+ "mmlu:subject=marketing,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_marketing",
262
+ "mmlu:subject=medical_genetics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_medical_genetics",
263
+ "mmlu:subject=miscellaneous,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_miscellaneous",
264
+ "mmlu:subject=moral_disputes,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_moral_disputes",
265
+ "mmlu:subject=moral_scenarios,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_moral_scenarios",
266
+ "mmlu:subject=nutrition,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_nutrition",
267
+ "mmlu:subject=philosophy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_philosophy",
268
+ "mmlu:subject=prehistory,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_prehistory",
269
+ "mmlu:subject=professional_accounting,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_accounting",
270
+ "mmlu:subject=professional_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_law",
271
+ "mmlu:subject=professional_medicine,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_medicine",
272
+ "mmlu:subject=professional_psychology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_psychology",
273
+ "mmlu:subject=public_relations,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_public_relations",
274
+ "mmlu:subject=security_studies,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_security_studies",
275
+ "mmlu:subject=sociology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_sociology",
276
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_us_foreign_policy",
277
+ "mmlu:subject=virology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_virology",
278
+ "mmlu:subject=world_religions,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_world_religions"
279
+ ]
280
+ },
281
+ {
282
+ "value": 4.945727778020373,
283
+ "description": "min=2.909, mean=4.946, max=5, sum=281.906 (57)",
284
+ "style": {},
285
+ "markdown": false,
286
+ "run_spec_names": [
287
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_abstract_algebra",
288
+ "mmlu:subject=anatomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_anatomy",
289
+ "mmlu:subject=astronomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_astronomy",
290
+ "mmlu:subject=business_ethics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_business_ethics",
291
+ "mmlu:subject=clinical_knowledge,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_clinical_knowledge",
292
+ "mmlu:subject=college_biology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_biology",
293
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_chemistry",
294
+ "mmlu:subject=college_computer_science,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_computer_science",
295
+ "mmlu:subject=college_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_mathematics",
296
+ "mmlu:subject=college_medicine,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_medicine",
297
+ "mmlu:subject=college_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_physics",
298
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_computer_security",
299
+ "mmlu:subject=conceptual_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_conceptual_physics",
300
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_econometrics",
301
+ "mmlu:subject=electrical_engineering,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_electrical_engineering",
302
+ "mmlu:subject=elementary_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_elementary_mathematics",
303
+ "mmlu:subject=formal_logic,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_formal_logic",
304
+ "mmlu:subject=global_facts,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_global_facts",
305
+ "mmlu:subject=high_school_biology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_biology",
306
+ "mmlu:subject=high_school_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_chemistry",
307
+ "mmlu:subject=high_school_computer_science,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_computer_science",
308
+ "mmlu:subject=high_school_european_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_european_history",
309
+ "mmlu:subject=high_school_geography,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_geography",
310
+ "mmlu:subject=high_school_government_and_politics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_government_and_politics",
311
+ "mmlu:subject=high_school_macroeconomics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_macroeconomics",
312
+ "mmlu:subject=high_school_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_mathematics",
313
+ "mmlu:subject=high_school_microeconomics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_microeconomics",
314
+ "mmlu:subject=high_school_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_physics",
315
+ "mmlu:subject=high_school_psychology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_psychology",
316
+ "mmlu:subject=high_school_statistics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_statistics",
317
+ "mmlu:subject=high_school_us_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_us_history",
318
+ "mmlu:subject=high_school_world_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_world_history",
319
+ "mmlu:subject=human_aging,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_human_aging",
320
+ "mmlu:subject=human_sexuality,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_human_sexuality",
321
+ "mmlu:subject=international_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_international_law",
322
+ "mmlu:subject=jurisprudence,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_jurisprudence",
323
+ "mmlu:subject=logical_fallacies,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_logical_fallacies",
324
+ "mmlu:subject=machine_learning,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_machine_learning",
325
+ "mmlu:subject=management,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_management",
326
+ "mmlu:subject=marketing,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_marketing",
327
+ "mmlu:subject=medical_genetics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_medical_genetics",
328
+ "mmlu:subject=miscellaneous,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_miscellaneous",
329
+ "mmlu:subject=moral_disputes,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_moral_disputes",
330
+ "mmlu:subject=moral_scenarios,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_moral_scenarios",
331
+ "mmlu:subject=nutrition,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_nutrition",
332
+ "mmlu:subject=philosophy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_philosophy",
333
+ "mmlu:subject=prehistory,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_prehistory",
334
+ "mmlu:subject=professional_accounting,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_accounting",
335
+ "mmlu:subject=professional_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_law",
336
+ "mmlu:subject=professional_medicine,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_medicine",
337
+ "mmlu:subject=professional_psychology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_psychology",
338
+ "mmlu:subject=public_relations,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_public_relations",
339
+ "mmlu:subject=security_studies,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_security_studies",
340
+ "mmlu:subject=sociology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_sociology",
341
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_us_foreign_policy",
342
+ "mmlu:subject=virology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_virology",
343
+ "mmlu:subject=world_religions,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_world_religions"
344
+ ]
345
+ },
346
+ {
347
+ "value": 0.0,
348
+ "description": "min=0, mean=0, max=0, sum=0 (57)",
349
+ "style": {},
350
+ "markdown": false,
351
+ "run_spec_names": [
352
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_abstract_algebra",
353
+ "mmlu:subject=anatomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_anatomy",
354
+ "mmlu:subject=astronomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_astronomy",
355
+ "mmlu:subject=business_ethics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_business_ethics",
356
+ "mmlu:subject=clinical_knowledge,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_clinical_knowledge",
357
+ "mmlu:subject=college_biology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_biology",
358
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_chemistry",
359
+ "mmlu:subject=college_computer_science,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_computer_science",
360
+ "mmlu:subject=college_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_mathematics",
361
+ "mmlu:subject=college_medicine,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_medicine",
362
+ "mmlu:subject=college_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_physics",
363
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_computer_security",
364
+ "mmlu:subject=conceptual_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_conceptual_physics",
365
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_econometrics",
366
+ "mmlu:subject=electrical_engineering,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_electrical_engineering",
367
+ "mmlu:subject=elementary_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_elementary_mathematics",
368
+ "mmlu:subject=formal_logic,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_formal_logic",
369
+ "mmlu:subject=global_facts,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_global_facts",
370
+ "mmlu:subject=high_school_biology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_biology",
371
+ "mmlu:subject=high_school_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_chemistry",
372
+ "mmlu:subject=high_school_computer_science,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_computer_science",
373
+ "mmlu:subject=high_school_european_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_european_history",
374
+ "mmlu:subject=high_school_geography,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_geography",
375
+ "mmlu:subject=high_school_government_and_politics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_government_and_politics",
376
+ "mmlu:subject=high_school_macroeconomics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_macroeconomics",
377
+ "mmlu:subject=high_school_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_mathematics",
378
+ "mmlu:subject=high_school_microeconomics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_microeconomics",
379
+ "mmlu:subject=high_school_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_physics",
380
+ "mmlu:subject=high_school_psychology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_psychology",
381
+ "mmlu:subject=high_school_statistics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_statistics",
382
+ "mmlu:subject=high_school_us_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_us_history",
383
+ "mmlu:subject=high_school_world_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_world_history",
384
+ "mmlu:subject=human_aging,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_human_aging",
385
+ "mmlu:subject=human_sexuality,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_human_sexuality",
386
+ "mmlu:subject=international_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_international_law",
387
+ "mmlu:subject=jurisprudence,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_jurisprudence",
388
+ "mmlu:subject=logical_fallacies,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_logical_fallacies",
389
+ "mmlu:subject=machine_learning,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_machine_learning",
390
+ "mmlu:subject=management,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_management",
391
+ "mmlu:subject=marketing,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_marketing",
392
+ "mmlu:subject=medical_genetics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_medical_genetics",
393
+ "mmlu:subject=miscellaneous,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_miscellaneous",
394
+ "mmlu:subject=moral_disputes,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_moral_disputes",
395
+ "mmlu:subject=moral_scenarios,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_moral_scenarios",
396
+ "mmlu:subject=nutrition,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_nutrition",
397
+ "mmlu:subject=philosophy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_philosophy",
398
+ "mmlu:subject=prehistory,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_prehistory",
399
+ "mmlu:subject=professional_accounting,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_accounting",
400
+ "mmlu:subject=professional_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_law",
401
+ "mmlu:subject=professional_medicine,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_medicine",
402
+ "mmlu:subject=professional_psychology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_psychology",
403
+ "mmlu:subject=public_relations,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_public_relations",
404
+ "mmlu:subject=security_studies,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_security_studies",
405
+ "mmlu:subject=sociology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_sociology",
406
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_us_foreign_policy",
407
+ "mmlu:subject=virology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_virology",
408
+ "mmlu:subject=world_religions,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_world_religions"
409
+ ]
410
+ },
411
+ {
412
+ "value": 597.9159199418197,
413
+ "description": "min=285.766, mean=597.916, max=1816.758, sum=34081.207 (57)",
414
+ "style": {},
415
+ "markdown": false,
416
+ "run_spec_names": [
417
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_abstract_algebra",
418
+ "mmlu:subject=anatomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_anatomy",
419
+ "mmlu:subject=astronomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_astronomy",
420
+ "mmlu:subject=business_ethics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_business_ethics",
421
+ "mmlu:subject=clinical_knowledge,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_clinical_knowledge",
422
+ "mmlu:subject=college_biology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_biology",
423
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_chemistry",
424
+ "mmlu:subject=college_computer_science,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_computer_science",
425
+ "mmlu:subject=college_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_mathematics",
426
+ "mmlu:subject=college_medicine,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_medicine",
427
+ "mmlu:subject=college_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_physics",
428
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_computer_security",
429
+ "mmlu:subject=conceptual_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_conceptual_physics",
430
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_econometrics",
431
+ "mmlu:subject=electrical_engineering,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_electrical_engineering",
432
+ "mmlu:subject=elementary_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_elementary_mathematics",
433
+ "mmlu:subject=formal_logic,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_formal_logic",
434
+ "mmlu:subject=global_facts,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_global_facts",
435
+ "mmlu:subject=high_school_biology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_biology",
436
+ "mmlu:subject=high_school_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_chemistry",
437
+ "mmlu:subject=high_school_computer_science,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_computer_science",
438
+ "mmlu:subject=high_school_european_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_european_history",
439
+ "mmlu:subject=high_school_geography,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_geography",
440
+ "mmlu:subject=high_school_government_and_politics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_government_and_politics",
441
+ "mmlu:subject=high_school_macroeconomics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_macroeconomics",
442
+ "mmlu:subject=high_school_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_mathematics",
443
+ "mmlu:subject=high_school_microeconomics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_microeconomics",
444
+ "mmlu:subject=high_school_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_physics",
445
+ "mmlu:subject=high_school_psychology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_psychology",
446
+ "mmlu:subject=high_school_statistics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_statistics",
447
+ "mmlu:subject=high_school_us_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_us_history",
448
+ "mmlu:subject=high_school_world_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_world_history",
449
+ "mmlu:subject=human_aging,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_human_aging",
450
+ "mmlu:subject=human_sexuality,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_human_sexuality",
451
+ "mmlu:subject=international_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_international_law",
452
+ "mmlu:subject=jurisprudence,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_jurisprudence",
453
+ "mmlu:subject=logical_fallacies,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_logical_fallacies",
454
+ "mmlu:subject=machine_learning,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_machine_learning",
455
+ "mmlu:subject=management,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_management",
456
+ "mmlu:subject=marketing,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_marketing",
457
+ "mmlu:subject=medical_genetics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_medical_genetics",
458
+ "mmlu:subject=miscellaneous,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_miscellaneous",
459
+ "mmlu:subject=moral_disputes,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_moral_disputes",
460
+ "mmlu:subject=moral_scenarios,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_moral_scenarios",
461
+ "mmlu:subject=nutrition,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_nutrition",
462
+ "mmlu:subject=philosophy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_philosophy",
463
+ "mmlu:subject=prehistory,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_prehistory",
464
+ "mmlu:subject=professional_accounting,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_accounting",
465
+ "mmlu:subject=professional_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_law",
466
+ "mmlu:subject=professional_medicine,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_medicine",
467
+ "mmlu:subject=professional_psychology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_psychology",
468
+ "mmlu:subject=public_relations,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_public_relations",
469
+ "mmlu:subject=security_studies,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_security_studies",
470
+ "mmlu:subject=sociology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_sociology",
471
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_us_foreign_policy",
472
+ "mmlu:subject=virology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_virology",
473
+ "mmlu:subject=world_religions,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_world_religions"
474
+ ]
475
+ },
476
+ {
477
+ "value": 1.0,
478
+ "description": "min=1, mean=1, max=1, sum=57 (57)",
479
+ "style": {},
480
+ "markdown": false,
481
+ "run_spec_names": [
482
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_abstract_algebra",
483
+ "mmlu:subject=anatomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_anatomy",
484
+ "mmlu:subject=astronomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_astronomy",
485
+ "mmlu:subject=business_ethics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_business_ethics",
486
+ "mmlu:subject=clinical_knowledge,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_clinical_knowledge",
487
+ "mmlu:subject=college_biology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_biology",
488
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_chemistry",
489
+ "mmlu:subject=college_computer_science,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_computer_science",
490
+ "mmlu:subject=college_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_mathematics",
491
+ "mmlu:subject=college_medicine,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_medicine",
492
+ "mmlu:subject=college_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_college_physics",
493
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_computer_security",
494
+ "mmlu:subject=conceptual_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_conceptual_physics",
495
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_econometrics",
496
+ "mmlu:subject=electrical_engineering,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_electrical_engineering",
497
+ "mmlu:subject=elementary_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_elementary_mathematics",
498
+ "mmlu:subject=formal_logic,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_formal_logic",
499
+ "mmlu:subject=global_facts,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_global_facts",
500
+ "mmlu:subject=high_school_biology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_biology",
501
+ "mmlu:subject=high_school_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_chemistry",
502
+ "mmlu:subject=high_school_computer_science,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_computer_science",
503
+ "mmlu:subject=high_school_european_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_european_history",
504
+ "mmlu:subject=high_school_geography,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_geography",
505
+ "mmlu:subject=high_school_government_and_politics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_government_and_politics",
506
+ "mmlu:subject=high_school_macroeconomics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_macroeconomics",
507
+ "mmlu:subject=high_school_mathematics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_mathematics",
508
+ "mmlu:subject=high_school_microeconomics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_microeconomics",
509
+ "mmlu:subject=high_school_physics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_physics",
510
+ "mmlu:subject=high_school_psychology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_psychology",
511
+ "mmlu:subject=high_school_statistics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_statistics",
512
+ "mmlu:subject=high_school_us_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_us_history",
513
+ "mmlu:subject=high_school_world_history,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_high_school_world_history",
514
+ "mmlu:subject=human_aging,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_human_aging",
515
+ "mmlu:subject=human_sexuality,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_human_sexuality",
516
+ "mmlu:subject=international_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_international_law",
517
+ "mmlu:subject=jurisprudence,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_jurisprudence",
518
+ "mmlu:subject=logical_fallacies,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_logical_fallacies",
519
+ "mmlu:subject=machine_learning,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_machine_learning",
520
+ "mmlu:subject=management,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_management",
521
+ "mmlu:subject=marketing,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_marketing",
522
+ "mmlu:subject=medical_genetics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_medical_genetics",
523
+ "mmlu:subject=miscellaneous,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_miscellaneous",
524
+ "mmlu:subject=moral_disputes,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_moral_disputes",
525
+ "mmlu:subject=moral_scenarios,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_moral_scenarios",
526
+ "mmlu:subject=nutrition,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_nutrition",
527
+ "mmlu:subject=philosophy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_philosophy",
528
+ "mmlu:subject=prehistory,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_prehistory",
529
+ "mmlu:subject=professional_accounting,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_accounting",
530
+ "mmlu:subject=professional_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_law",
531
+ "mmlu:subject=professional_medicine,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_medicine",
532
+ "mmlu:subject=professional_psychology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_psychology",
533
+ "mmlu:subject=public_relations,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_public_relations",
534
+ "mmlu:subject=security_studies,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_security_studies",
535
+ "mmlu:subject=sociology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_sociology",
536
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_us_foreign_policy",
537
+ "mmlu:subject=virology,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_virology",
538
+ "mmlu:subject=world_religions,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_world_religions"
539
+ ]
540
+ }
541
+ ]
542
+ ],
543
+ "links": [
544
+ {
545
+ "text": "LaTeX",
546
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_mmlu.tex"
547
+ },
548
+ {
549
+ "text": "JSON",
550
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_mmlu.json"
551
+ }
552
+ ],
553
+ "name": "mmlu"
554
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_mmlu_subject:abstract_algebra.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: abstract_algebra",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "MMLU All Subjects"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "MMLU All Subjects"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "MMLU All Subjects"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "MMLU All Subjects"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "MMLU All Subjects"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "MMLU All Subjects"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "MMLU All Subjects"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu&subgroup=subject%3A%20abstract_algebra&runSpecs=%5B%22mmlu%3Asubject%3Dabstract_algebra%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_abstract_algebra%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_abstract_algebra"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.23,
88
+ "description": "min=0.23, mean=0.23, max=0.23, sum=0.23 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.06573188066482544,
96
+ "description": "min=0.066, mean=0.066, max=0.066, sum=0.066 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 100.0,
104
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 358.76,
122
+ "description": "min=358.76, mean=358.76, max=358.76, sum=358.76 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_mmlu_subject:abstract_algebra.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_mmlu_subject:abstract_algebra.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_subject:abstract_algebra"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_mmlu_subject:anatomy.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: anatomy",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "MMLU All Subjects"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "MMLU All Subjects"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "MMLU All Subjects"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "MMLU All Subjects"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "MMLU All Subjects"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "MMLU All Subjects"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "MMLU All Subjects"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu&subgroup=subject%3A%20anatomy&runSpecs=%5B%22mmlu%3Asubject%3Danatomy%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_anatomy%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=anatomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_anatomy"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.2962962962962963,
88
+ "description": "min=0.296, mean=0.296, max=0.296, sum=0.296 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.06830824922632288,
96
+ "description": "min=0.068, mean=0.068, max=0.068, sum=0.068 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 135.0,
104
+ "description": "min=135, mean=135, max=135, sum=135 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 352.02962962962965,
122
+ "description": "min=352.03, mean=352.03, max=352.03, sum=352.03 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_mmlu_subject:anatomy.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_mmlu_subject:anatomy.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_subject:anatomy"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_mmlu_subject:astronomy.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: astronomy",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "MMLU All Subjects"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "MMLU All Subjects"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "MMLU All Subjects"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "MMLU All Subjects"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "MMLU All Subjects"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "MMLU All Subjects"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "MMLU All Subjects"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu&subgroup=subject%3A%20astronomy&runSpecs=%5B%22mmlu%3Asubject%3Dastronomy%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_astronomy%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=astronomy,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_astronomy"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.23684210526315788,
88
+ "description": "min=0.237, mean=0.237, max=0.237, sum=0.237 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.10126333958224247,
96
+ "description": "min=0.101, mean=0.101, max=0.101, sum=0.101 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 152.0,
104
+ "description": "min=152, mean=152, max=152, sum=152 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 598.4868421052631,
122
+ "description": "min=598.487, mean=598.487, max=598.487, sum=598.487 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_mmlu_subject:astronomy.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_mmlu_subject:astronomy.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_subject:astronomy"
145
+ }
mmlu_pythia-1.4b-step2000/groups/json/mmlu_mmlu_subject:business_ethics.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: business_ethics",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "MMLU All Subjects"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "MMLU All Subjects"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "MMLU All Subjects"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "MMLU All Subjects"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "MMLU All Subjects"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "MMLU All Subjects"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "MMLU All Subjects"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-1.4b",
79
+ "description": "",
80
+ "href": "?group=mmlu&subgroup=subject%3A%20business_ethics&runSpecs=%5B%22mmlu%3Asubject%3Dbusiness_ethics%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-1.4b%2Ceval_split%3Dtest%2Cgroups%3Dmmlu_business_ethics%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=business_ethics,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_business_ethics"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.29,
88
+ "description": "min=0.29, mean=0.29, max=0.29, sum=0.29 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.09688833236694336,
96
+ "description": "min=0.097, mean=0.097, max=0.097, sum=0.097 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 100.0,
104
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 585.05,
122
+ "description": "min=585.05, mean=585.05, max=585.05, sum=585.05 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/latex/mmlu_mmlu_subject:business_ethics.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/mmlu_pythia-1.4b-step2000/groups/json/mmlu_mmlu_subject:business_ethics.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_subject:business_ethics"
145
+ }