yuhengtu commited on
Commit
7419fbd
·
verified ·
1 Parent(s): 8c500ac

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. lite_pythia-2.8b-step5000/commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b/display_predictions.json +0 -0
  3. lite_pythia-2.8b-step5000/commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b/display_requests.json +0 -0
  4. lite_pythia-2.8b-step5000/commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b/instances.json +0 -0
  5. lite_pythia-2.8b-step5000/commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b/per_instance_stats.json +0 -0
  6. lite_pythia-2.8b-step5000/commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b/run_spec.json +74 -0
  7. lite_pythia-2.8b-step5000/commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b/scenario.json +9 -0
  8. lite_pythia-2.8b-step5000/commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b/scenario_state.json +0 -0
  9. lite_pythia-2.8b-step5000/commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b/stats.json +1538 -0
  10. lite_pythia-2.8b-step5000/costs.json +1 -0
  11. lite_pythia-2.8b-step5000/groups.json +491 -0
  12. lite_pythia-2.8b-step5000/groups/core_scenarios.json +1264 -0
  13. lite_pythia-2.8b-step5000/groups/gsm.json +147 -0
  14. lite_pythia-2.8b-step5000/groups/json/core_scenarios_accuracy.json +216 -0
  15. lite_pythia-2.8b-step5000/groups/json/core_scenarios_efficiency.json +216 -0
  16. lite_pythia-2.8b-step5000/groups/json/core_scenarios_general_information.json +830 -0
  17. lite_pythia-2.8b-step5000/groups/json/gsm_gsm_.json +145 -0
  18. lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench.json +190 -0
  19. lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:abercrombie.json +145 -0
  20. lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:corporate_lobbying.json +145 -0
  21. lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:function_of_decision_section.json +145 -0
  22. lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:international_citizenship_questions.json +145 -0
  23. lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:proa.json +145 -0
  24. lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu.json +190 -0
  25. lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:abstract_algebra.json +145 -0
  26. lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:college_chemistry.json +145 -0
  27. lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:computer_security.json +145 -0
  28. lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:econometrics.json +145 -0
  29. lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:us_foreign_policy.json +145 -0
  30. lite_pythia-2.8b-step5000/groups/json/openbookqa_openbookqa_.json +145 -0
  31. lite_pythia-2.8b-step5000/groups/latex/core_scenarios_accuracy.tex +12 -0
  32. lite_pythia-2.8b-step5000/groups/latex/core_scenarios_efficiency.tex +12 -0
  33. lite_pythia-2.8b-step5000/groups/latex/core_scenarios_general_information.tex +12 -0
  34. lite_pythia-2.8b-step5000/groups/latex/gsm_gsm_.tex +12 -0
  35. lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench.tex +12 -0
  36. lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:abercrombie.tex +12 -0
  37. lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:corporate_lobbying.tex +12 -0
  38. lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:function_of_decision_section.tex +12 -0
  39. lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:international_citizenship_questions.tex +12 -0
  40. lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:proa.tex +12 -0
  41. lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu.tex +12 -0
  42. lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:abstract_algebra.tex +12 -0
  43. lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:college_chemistry.tex +12 -0
  44. lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:computer_security.tex +12 -0
  45. lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:econometrics.tex +12 -0
  46. lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:us_foreign_policy.tex +12 -0
  47. lite_pythia-2.8b-step5000/groups/latex/openbookqa_openbookqa_.tex +12 -0
  48. lite_pythia-2.8b-step5000/groups/legalbench.json +917 -0
  49. lite_pythia-2.8b-step5000/groups/mmlu.json +917 -0
  50. lite_pythia-2.8b-step5000/groups/openbookqa.json +147 -0
.gitattributes CHANGED
@@ -1288,3 +1288,4 @@ classic_pythia-2.8b-step2/runs.json filter=lfs diff=lfs merge=lfs -text
1288
  mmlu_pythia-1.4b-step2000/mmlu:subject=professional_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_law/display_requests.json filter=lfs diff=lfs merge=lfs -text
1289
  mmlu_pythia-1.4b-step2000/mmlu:subject=professional_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_law/per_instance_stats.json filter=lfs diff=lfs merge=lfs -text
1290
  mmlu_pythia-1.4b-step2000/mmlu:subject=professional_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_law/scenario_state.json filter=lfs diff=lfs merge=lfs -text
 
 
1288
  mmlu_pythia-1.4b-step2000/mmlu:subject=professional_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_law/display_requests.json filter=lfs diff=lfs merge=lfs -text
1289
  mmlu_pythia-1.4b-step2000/mmlu:subject=professional_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_law/per_instance_stats.json filter=lfs diff=lfs merge=lfs -text
1290
  mmlu_pythia-1.4b-step2000/mmlu:subject=professional_law,method=multiple_choice_joint,model=EleutherAI_pythia-1.4b,eval_split=test,groups=mmlu_professional_law/scenario_state.json filter=lfs diff=lfs merge=lfs -text
1291
+ lite_pythia-2.8b-step5000/gsm:model=EleutherAI_pythia-2.8b/scenario_state.json filter=lfs diff=lfs merge=lfs -text
lite_pythia-2.8b-step5000/commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b/display_predictions.json ADDED
The diff for this file is too large to render. See raw diff
 
lite_pythia-2.8b-step5000/commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b/display_requests.json ADDED
The diff for this file is too large to render. See raw diff
 
lite_pythia-2.8b-step5000/commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b/instances.json ADDED
The diff for this file is too large to render. See raw diff
 
lite_pythia-2.8b-step5000/commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b/per_instance_stats.json ADDED
The diff for this file is too large to render. See raw diff
 
lite_pythia-2.8b-step5000/commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b/run_spec.json ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
3
+ "scenario_spec": {
4
+ "class_name": "helm.benchmark.scenarios.commonsense_scenario.OpenBookQA",
5
+ "args": {}
6
+ },
7
+ "adapter_spec": {
8
+ "method": "multiple_choice_joint",
9
+ "global_prefix": "",
10
+ "global_suffix": "",
11
+ "instructions": "The following are multiple choice questions (with answers) about common sense.\n",
12
+ "input_prefix": "Question: ",
13
+ "input_suffix": "\n",
14
+ "reference_prefix": "A. ",
15
+ "reference_suffix": "\n",
16
+ "chain_of_thought_prefix": "",
17
+ "chain_of_thought_suffix": "\n",
18
+ "output_prefix": "Answer: ",
19
+ "output_suffix": "\n",
20
+ "instance_prefix": "\n",
21
+ "substitutions": [],
22
+ "max_train_instances": 5,
23
+ "max_eval_instances": 1000,
24
+ "num_outputs": 5,
25
+ "num_train_trials": 1,
26
+ "num_trials": 1,
27
+ "sample_train": true,
28
+ "model_deployment": "EleutherAI/pythia-2.8b",
29
+ "model": "EleutherAI/pythia-2.8b",
30
+ "model_ability": 0.0,
31
+ "temperature": 0.0,
32
+ "max_tokens": 1,
33
+ "stop_sequences": [
34
+ "\n"
35
+ ],
36
+ "multi_label": false
37
+ },
38
+ "metric_specs": [
39
+ {
40
+ "class_name": "helm.benchmark.metrics.basic_metrics.BasicGenerationMetric",
41
+ "args": {
42
+ "names": [
43
+ "exact_match",
44
+ "quasi_exact_match",
45
+ "prefix_exact_match",
46
+ "quasi_prefix_exact_match"
47
+ ]
48
+ }
49
+ },
50
+ {
51
+ "class_name": "helm.benchmark.metrics.basic_metrics.BasicReferenceMetric",
52
+ "args": {}
53
+ },
54
+ {
55
+ "class_name": "helm.benchmark.metrics.basic_metrics.InstancesPerSplitMetric",
56
+ "args": {}
57
+ }
58
+ ],
59
+ "data_augmenter_spec": {
60
+ "perturbation_specs": [],
61
+ "should_augment_train_instances": false,
62
+ "should_include_original_train": false,
63
+ "should_skip_unchanged_train": false,
64
+ "should_augment_eval_instances": false,
65
+ "should_include_original_eval": false,
66
+ "should_skip_unchanged_eval": false,
67
+ "seeds_per_instance": 1
68
+ },
69
+ "groups": [
70
+ "openbookqa"
71
+ ],
72
+ "reeval_mode": false,
73
+ "reeval_max_samples": 50
74
+ }
lite_pythia-2.8b-step5000/commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b/scenario.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "openbookqa",
3
+ "description": "Benchmark from https://aclanthology.org/D18-1260.pdf.",
4
+ "tags": [
5
+ "knowledge",
6
+ "multiple_choice"
7
+ ],
8
+ "definition_path": "https://github.com/stanford-crfm/helm/blob/main/src/helm/benchmark/scenarios/commonsense_scenario.py"
9
+ }
lite_pythia-2.8b-step5000/commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b/scenario_state.json ADDED
The diff for this file is too large to render. See raw diff
 
lite_pythia-2.8b-step5000/commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b/stats.json ADDED
@@ -0,0 +1,1538 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "name": {
4
+ "name": "num_references",
5
+ "split": "test"
6
+ },
7
+ "count": 1,
8
+ "sum": 4.0,
9
+ "sum_squared": 16.0,
10
+ "min": 4.0,
11
+ "max": 4.0,
12
+ "mean": 4.0,
13
+ "variance": 0.0,
14
+ "stddev": 0.0
15
+ },
16
+ {
17
+ "name": {
18
+ "name": "num_train_trials",
19
+ "split": "test"
20
+ },
21
+ "count": 1,
22
+ "sum": 1.0,
23
+ "sum_squared": 1.0,
24
+ "min": 1.0,
25
+ "max": 1.0,
26
+ "mean": 1.0,
27
+ "variance": 0.0,
28
+ "stddev": 0.0
29
+ },
30
+ {
31
+ "name": {
32
+ "name": "num_prompt_tokens",
33
+ "split": "test"
34
+ },
35
+ "count": 1,
36
+ "sum": 251.556,
37
+ "sum_squared": 63280.421136000004,
38
+ "min": 251.556,
39
+ "max": 251.556,
40
+ "mean": 251.556,
41
+ "variance": 0.0,
42
+ "stddev": 0.0
43
+ },
44
+ {
45
+ "name": {
46
+ "name": "num_completion_tokens",
47
+ "split": "test"
48
+ },
49
+ "count": 1,
50
+ "sum": 1.0,
51
+ "sum_squared": 1.0,
52
+ "min": 1.0,
53
+ "max": 1.0,
54
+ "mean": 1.0,
55
+ "variance": 0.0,
56
+ "stddev": 0.0
57
+ },
58
+ {
59
+ "name": {
60
+ "name": "num_output_tokens",
61
+ "split": "test"
62
+ },
63
+ "count": 1,
64
+ "sum": 1.0,
65
+ "sum_squared": 1.0,
66
+ "min": 1.0,
67
+ "max": 1.0,
68
+ "mean": 1.0,
69
+ "variance": 0.0,
70
+ "stddev": 0.0
71
+ },
72
+ {
73
+ "name": {
74
+ "name": "training_co2_cost",
75
+ "split": "test"
76
+ },
77
+ "count": 0,
78
+ "sum": 0,
79
+ "sum_squared": 0
80
+ },
81
+ {
82
+ "name": {
83
+ "name": "training_energy_cost",
84
+ "split": "test"
85
+ },
86
+ "count": 0,
87
+ "sum": 0,
88
+ "sum_squared": 0
89
+ },
90
+ {
91
+ "name": {
92
+ "name": "inference_runtime",
93
+ "split": "test"
94
+ },
95
+ "count": 1,
96
+ "sum": 0.12883714818954467,
97
+ "sum_squared": 0.016599010753614692,
98
+ "min": 0.12883714818954467,
99
+ "max": 0.12883714818954467,
100
+ "mean": 0.12883714818954467,
101
+ "variance": 0.0,
102
+ "stddev": 0.0
103
+ },
104
+ {
105
+ "name": {
106
+ "name": "batch_size",
107
+ "split": "test"
108
+ },
109
+ "count": 1,
110
+ "sum": 1.0,
111
+ "sum_squared": 1.0,
112
+ "min": 1.0,
113
+ "max": 1.0,
114
+ "mean": 1.0,
115
+ "variance": 0.0,
116
+ "stddev": 0.0
117
+ },
118
+ {
119
+ "name": {
120
+ "name": "finish_reason_length",
121
+ "split": "test"
122
+ },
123
+ "count": 1,
124
+ "sum": 0.0,
125
+ "sum_squared": 0.0,
126
+ "min": 0.0,
127
+ "max": 0.0,
128
+ "mean": 0.0,
129
+ "variance": 0.0,
130
+ "stddev": 0.0
131
+ },
132
+ {
133
+ "name": {
134
+ "name": "finish_reason_stop",
135
+ "split": "test"
136
+ },
137
+ "count": 1,
138
+ "sum": 0.0,
139
+ "sum_squared": 0.0,
140
+ "min": 0.0,
141
+ "max": 0.0,
142
+ "mean": 0.0,
143
+ "variance": 0.0,
144
+ "stddev": 0.0
145
+ },
146
+ {
147
+ "name": {
148
+ "name": "finish_reason_endoftext",
149
+ "split": "test"
150
+ },
151
+ "count": 1,
152
+ "sum": 0.0,
153
+ "sum_squared": 0.0,
154
+ "min": 0.0,
155
+ "max": 0.0,
156
+ "mean": 0.0,
157
+ "variance": 0.0,
158
+ "stddev": 0.0
159
+ },
160
+ {
161
+ "name": {
162
+ "name": "finish_reason_unknown",
163
+ "split": "test"
164
+ },
165
+ "count": 1,
166
+ "sum": 1.0,
167
+ "sum_squared": 1.0,
168
+ "min": 1.0,
169
+ "max": 1.0,
170
+ "mean": 1.0,
171
+ "variance": 0.0,
172
+ "stddev": 0.0
173
+ },
174
+ {
175
+ "name": {
176
+ "name": "num_train_instances",
177
+ "split": "test"
178
+ },
179
+ "count": 1,
180
+ "sum": 5.0,
181
+ "sum_squared": 25.0,
182
+ "min": 5.0,
183
+ "max": 5.0,
184
+ "mean": 5.0,
185
+ "variance": 0.0,
186
+ "stddev": 0.0
187
+ },
188
+ {
189
+ "name": {
190
+ "name": "prompt_truncated",
191
+ "split": "test"
192
+ },
193
+ "count": 1,
194
+ "sum": 0.0,
195
+ "sum_squared": 0.0,
196
+ "min": 0.0,
197
+ "max": 0.0,
198
+ "mean": 0.0,
199
+ "variance": 0.0,
200
+ "stddev": 0.0
201
+ },
202
+ {
203
+ "name": {
204
+ "name": "max_prob",
205
+ "split": "test"
206
+ },
207
+ "count": 1,
208
+ "sum": 0.45501411941835707,
209
+ "sum_squared": 0.2070378488700629,
210
+ "min": 0.45501411941835707,
211
+ "max": 0.45501411941835707,
212
+ "mean": 0.45501411941835707,
213
+ "variance": 0.0,
214
+ "stddev": 0.0
215
+ },
216
+ {
217
+ "name": {
218
+ "name": "exact_match",
219
+ "split": "test"
220
+ },
221
+ "count": 1,
222
+ "sum": 0.22,
223
+ "sum_squared": 0.0484,
224
+ "min": 0.22,
225
+ "max": 0.22,
226
+ "mean": 0.22,
227
+ "variance": 0.0,
228
+ "stddev": 0.0
229
+ },
230
+ {
231
+ "name": {
232
+ "name": "exact_match@5",
233
+ "split": "test"
234
+ },
235
+ "count": 1,
236
+ "sum": 0.22,
237
+ "sum_squared": 0.0484,
238
+ "min": 0.22,
239
+ "max": 0.22,
240
+ "mean": 0.22,
241
+ "variance": 0.0,
242
+ "stddev": 0.0
243
+ },
244
+ {
245
+ "name": {
246
+ "name": "quasi_exact_match",
247
+ "split": "test"
248
+ },
249
+ "count": 1,
250
+ "sum": 0.22,
251
+ "sum_squared": 0.0484,
252
+ "min": 0.22,
253
+ "max": 0.22,
254
+ "mean": 0.22,
255
+ "variance": 0.0,
256
+ "stddev": 0.0
257
+ },
258
+ {
259
+ "name": {
260
+ "name": "quasi_exact_match@5",
261
+ "split": "test"
262
+ },
263
+ "count": 1,
264
+ "sum": 0.22,
265
+ "sum_squared": 0.0484,
266
+ "min": 0.22,
267
+ "max": 0.22,
268
+ "mean": 0.22,
269
+ "variance": 0.0,
270
+ "stddev": 0.0
271
+ },
272
+ {
273
+ "name": {
274
+ "name": "prefix_exact_match",
275
+ "split": "test"
276
+ },
277
+ "count": 1,
278
+ "sum": 0.22,
279
+ "sum_squared": 0.0484,
280
+ "min": 0.22,
281
+ "max": 0.22,
282
+ "mean": 0.22,
283
+ "variance": 0.0,
284
+ "stddev": 0.0
285
+ },
286
+ {
287
+ "name": {
288
+ "name": "prefix_exact_match@5",
289
+ "split": "test"
290
+ },
291
+ "count": 1,
292
+ "sum": 0.22,
293
+ "sum_squared": 0.0484,
294
+ "min": 0.22,
295
+ "max": 0.22,
296
+ "mean": 0.22,
297
+ "variance": 0.0,
298
+ "stddev": 0.0
299
+ },
300
+ {
301
+ "name": {
302
+ "name": "quasi_prefix_exact_match",
303
+ "split": "test"
304
+ },
305
+ "count": 1,
306
+ "sum": 0.22,
307
+ "sum_squared": 0.0484,
308
+ "min": 0.22,
309
+ "max": 0.22,
310
+ "mean": 0.22,
311
+ "variance": 0.0,
312
+ "stddev": 0.0
313
+ },
314
+ {
315
+ "name": {
316
+ "name": "quasi_prefix_exact_match@5",
317
+ "split": "test"
318
+ },
319
+ "count": 1,
320
+ "sum": 0.22,
321
+ "sum_squared": 0.0484,
322
+ "min": 0.22,
323
+ "max": 0.22,
324
+ "mean": 0.22,
325
+ "variance": 0.0,
326
+ "stddev": 0.0
327
+ },
328
+ {
329
+ "name": {
330
+ "name": "logprob",
331
+ "split": "test"
332
+ },
333
+ "count": 1,
334
+ "sum": -0.805416056394577,
335
+ "sum_squared": 0.6486950238981924,
336
+ "min": -0.805416056394577,
337
+ "max": -0.805416056394577,
338
+ "mean": -0.805416056394577,
339
+ "variance": 0.0,
340
+ "stddev": 0.0
341
+ },
342
+ {
343
+ "name": {
344
+ "name": "num_perplexity_tokens",
345
+ "split": "test"
346
+ },
347
+ "count": 1,
348
+ "sum": 1.0,
349
+ "sum_squared": 1.0,
350
+ "min": 1.0,
351
+ "max": 1.0,
352
+ "mean": 1.0,
353
+ "variance": 0.0,
354
+ "stddev": 0.0
355
+ },
356
+ {
357
+ "name": {
358
+ "name": "num_bytes",
359
+ "split": "test"
360
+ },
361
+ "count": 1,
362
+ "sum": 2.0,
363
+ "sum_squared": 4.0,
364
+ "min": 2.0,
365
+ "max": 2.0,
366
+ "mean": 2.0,
367
+ "variance": 0.0,
368
+ "stddev": 0.0
369
+ },
370
+ {
371
+ "name": {
372
+ "name": "perplexity",
373
+ "split": "test"
374
+ },
375
+ "count": 1,
376
+ "sum": 2.2376272843172513,
377
+ "sum_squared": 5.0069758635209975,
378
+ "min": 2.2376272843172513,
379
+ "max": 2.2376272843172513,
380
+ "mean": 2.2376272843172513,
381
+ "variance": 0.0,
382
+ "stddev": 0.0
383
+ },
384
+ {
385
+ "name": {
386
+ "name": "bits_per_byte",
387
+ "split": "test"
388
+ },
389
+ "count": 1,
390
+ "sum": 0.580984875206401,
391
+ "sum_squared": 0.3375434252185973,
392
+ "min": 0.580984875206401,
393
+ "max": 0.580984875206401,
394
+ "mean": 0.580984875206401,
395
+ "variance": 0.0,
396
+ "stddev": 0.0
397
+ },
398
+ {
399
+ "name": {
400
+ "name": "logprob_per_byte",
401
+ "split": "test"
402
+ },
403
+ "count": 1,
404
+ "sum": -0.4027080281972885,
405
+ "sum_squared": 0.1621737559745481,
406
+ "min": -0.4027080281972885,
407
+ "max": -0.4027080281972885,
408
+ "mean": -0.4027080281972885,
409
+ "variance": 0.0,
410
+ "stddev": 0.0
411
+ },
412
+ {
413
+ "name": {
414
+ "name": "ece_10_bin",
415
+ "split": "test"
416
+ },
417
+ "count": 1,
418
+ "sum": 0.23501411941835657,
419
+ "sum_squared": 0.05523163632598556,
420
+ "min": 0.23501411941835657,
421
+ "max": 0.23501411941835657,
422
+ "mean": 0.23501411941835657,
423
+ "variance": 0.0,
424
+ "stddev": 0.0
425
+ },
426
+ {
427
+ "name": {
428
+ "name": "ece_1_bin",
429
+ "split": "test"
430
+ },
431
+ "count": 1,
432
+ "sum": 0.23501411941835657,
433
+ "sum_squared": 0.05523163632598556,
434
+ "min": 0.23501411941835657,
435
+ "max": 0.23501411941835657,
436
+ "mean": 0.23501411941835657,
437
+ "variance": 0.0,
438
+ "stddev": 0.0
439
+ },
440
+ {
441
+ "name": {
442
+ "name": "selective_cov_acc_area",
443
+ "split": "test"
444
+ },
445
+ "count": 1,
446
+ "sum": 0.2114153671720523,
447
+ "sum_squared": 0.044696457476493694,
448
+ "min": 0.2114153671720523,
449
+ "max": 0.2114153671720523,
450
+ "mean": 0.2114153671720523,
451
+ "variance": 0.0,
452
+ "stddev": 0.0
453
+ },
454
+ {
455
+ "name": {
456
+ "name": "selective_acc@10",
457
+ "split": "test"
458
+ },
459
+ "count": 1,
460
+ "sum": 0.17647058823529413,
461
+ "sum_squared": 0.031141868512110732,
462
+ "min": 0.17647058823529413,
463
+ "max": 0.17647058823529413,
464
+ "mean": 0.17647058823529413,
465
+ "variance": 0.0,
466
+ "stddev": 0.0
467
+ },
468
+ {
469
+ "name": {
470
+ "name": "platt_coef",
471
+ "split": "test"
472
+ },
473
+ "count": 1,
474
+ "sum": -0.09815250868690073,
475
+ "sum_squared": 0.009633914961532123,
476
+ "min": -0.09815250868690073,
477
+ "max": -0.09815250868690073,
478
+ "mean": -0.09815250868690073,
479
+ "variance": 0.0,
480
+ "stddev": 0.0
481
+ },
482
+ {
483
+ "name": {
484
+ "name": "platt_intercept",
485
+ "split": "test"
486
+ },
487
+ "count": 1,
488
+ "sum": -1.2836296815308241,
489
+ "sum_squared": 1.647705159306925,
490
+ "min": -1.2836296815308241,
491
+ "max": -1.2836296815308241,
492
+ "mean": -1.2836296815308241,
493
+ "variance": 0.0,
494
+ "stddev": 0.0
495
+ },
496
+ {
497
+ "name": {
498
+ "name": "platt_ece_10_bin",
499
+ "split": "test"
500
+ },
501
+ "count": 1,
502
+ "sum": 0.03173473384942972,
503
+ "sum_squared": 0.0010070933324941406,
504
+ "min": 0.03173473384942972,
505
+ "max": 0.03173473384942972,
506
+ "mean": 0.03173473384942972,
507
+ "variance": 0.0,
508
+ "stddev": 0.0
509
+ },
510
+ {
511
+ "name": {
512
+ "name": "platt_ece_1_bin",
513
+ "split": "test"
514
+ },
515
+ "count": 1,
516
+ "sum": 7.75700809032831e-05,
517
+ "sum_squared": 6.017117451341886e-09,
518
+ "min": 7.75700809032831e-05,
519
+ "max": 7.75700809032831e-05,
520
+ "mean": 7.75700809032831e-05,
521
+ "variance": 0.0,
522
+ "stddev": 0.0
523
+ },
524
+ {
525
+ "name": {
526
+ "name": "num_references",
527
+ "split": "test",
528
+ "perturbation": {
529
+ "name": "robustness",
530
+ "robustness": true,
531
+ "fairness": false,
532
+ "computed_on": "worst"
533
+ }
534
+ },
535
+ "count": 1,
536
+ "sum": 4.0,
537
+ "sum_squared": 16.0,
538
+ "min": 4.0,
539
+ "max": 4.0,
540
+ "mean": 4.0,
541
+ "variance": 0.0,
542
+ "stddev": 0.0
543
+ },
544
+ {
545
+ "name": {
546
+ "name": "num_references",
547
+ "split": "test",
548
+ "perturbation": {
549
+ "name": "fairness",
550
+ "robustness": false,
551
+ "fairness": true,
552
+ "computed_on": "worst"
553
+ }
554
+ },
555
+ "count": 1,
556
+ "sum": 4.0,
557
+ "sum_squared": 16.0,
558
+ "min": 4.0,
559
+ "max": 4.0,
560
+ "mean": 4.0,
561
+ "variance": 0.0,
562
+ "stddev": 0.0
563
+ },
564
+ {
565
+ "name": {
566
+ "name": "num_train_trials",
567
+ "split": "test",
568
+ "perturbation": {
569
+ "name": "robustness",
570
+ "robustness": true,
571
+ "fairness": false,
572
+ "computed_on": "worst"
573
+ }
574
+ },
575
+ "count": 1,
576
+ "sum": 1.0,
577
+ "sum_squared": 1.0,
578
+ "min": 1.0,
579
+ "max": 1.0,
580
+ "mean": 1.0,
581
+ "variance": 0.0,
582
+ "stddev": 0.0
583
+ },
584
+ {
585
+ "name": {
586
+ "name": "num_train_trials",
587
+ "split": "test",
588
+ "perturbation": {
589
+ "name": "fairness",
590
+ "robustness": false,
591
+ "fairness": true,
592
+ "computed_on": "worst"
593
+ }
594
+ },
595
+ "count": 1,
596
+ "sum": 1.0,
597
+ "sum_squared": 1.0,
598
+ "min": 1.0,
599
+ "max": 1.0,
600
+ "mean": 1.0,
601
+ "variance": 0.0,
602
+ "stddev": 0.0
603
+ },
604
+ {
605
+ "name": {
606
+ "name": "num_prompt_tokens",
607
+ "split": "test",
608
+ "perturbation": {
609
+ "name": "robustness",
610
+ "robustness": true,
611
+ "fairness": false,
612
+ "computed_on": "worst"
613
+ }
614
+ },
615
+ "count": 1,
616
+ "sum": 251.556,
617
+ "sum_squared": 63280.421136000004,
618
+ "min": 251.556,
619
+ "max": 251.556,
620
+ "mean": 251.556,
621
+ "variance": 0.0,
622
+ "stddev": 0.0
623
+ },
624
+ {
625
+ "name": {
626
+ "name": "num_prompt_tokens",
627
+ "split": "test",
628
+ "perturbation": {
629
+ "name": "fairness",
630
+ "robustness": false,
631
+ "fairness": true,
632
+ "computed_on": "worst"
633
+ }
634
+ },
635
+ "count": 1,
636
+ "sum": 251.556,
637
+ "sum_squared": 63280.421136000004,
638
+ "min": 251.556,
639
+ "max": 251.556,
640
+ "mean": 251.556,
641
+ "variance": 0.0,
642
+ "stddev": 0.0
643
+ },
644
+ {
645
+ "name": {
646
+ "name": "num_completion_tokens",
647
+ "split": "test",
648
+ "perturbation": {
649
+ "name": "robustness",
650
+ "robustness": true,
651
+ "fairness": false,
652
+ "computed_on": "worst"
653
+ }
654
+ },
655
+ "count": 1,
656
+ "sum": 1.0,
657
+ "sum_squared": 1.0,
658
+ "min": 1.0,
659
+ "max": 1.0,
660
+ "mean": 1.0,
661
+ "variance": 0.0,
662
+ "stddev": 0.0
663
+ },
664
+ {
665
+ "name": {
666
+ "name": "num_completion_tokens",
667
+ "split": "test",
668
+ "perturbation": {
669
+ "name": "fairness",
670
+ "robustness": false,
671
+ "fairness": true,
672
+ "computed_on": "worst"
673
+ }
674
+ },
675
+ "count": 1,
676
+ "sum": 1.0,
677
+ "sum_squared": 1.0,
678
+ "min": 1.0,
679
+ "max": 1.0,
680
+ "mean": 1.0,
681
+ "variance": 0.0,
682
+ "stddev": 0.0
683
+ },
684
+ {
685
+ "name": {
686
+ "name": "num_output_tokens",
687
+ "split": "test",
688
+ "perturbation": {
689
+ "name": "robustness",
690
+ "robustness": true,
691
+ "fairness": false,
692
+ "computed_on": "worst"
693
+ }
694
+ },
695
+ "count": 1,
696
+ "sum": 1.0,
697
+ "sum_squared": 1.0,
698
+ "min": 1.0,
699
+ "max": 1.0,
700
+ "mean": 1.0,
701
+ "variance": 0.0,
702
+ "stddev": 0.0
703
+ },
704
+ {
705
+ "name": {
706
+ "name": "num_output_tokens",
707
+ "split": "test",
708
+ "perturbation": {
709
+ "name": "fairness",
710
+ "robustness": false,
711
+ "fairness": true,
712
+ "computed_on": "worst"
713
+ }
714
+ },
715
+ "count": 1,
716
+ "sum": 1.0,
717
+ "sum_squared": 1.0,
718
+ "min": 1.0,
719
+ "max": 1.0,
720
+ "mean": 1.0,
721
+ "variance": 0.0,
722
+ "stddev": 0.0
723
+ },
724
+ {
725
+ "name": {
726
+ "name": "inference_runtime",
727
+ "split": "test",
728
+ "perturbation": {
729
+ "name": "robustness",
730
+ "robustness": true,
731
+ "fairness": false,
732
+ "computed_on": "worst"
733
+ }
734
+ },
735
+ "count": 1,
736
+ "sum": 0.12883714818954467,
737
+ "sum_squared": 0.016599010753614692,
738
+ "min": 0.12883714818954467,
739
+ "max": 0.12883714818954467,
740
+ "mean": 0.12883714818954467,
741
+ "variance": 0.0,
742
+ "stddev": 0.0
743
+ },
744
+ {
745
+ "name": {
746
+ "name": "inference_runtime",
747
+ "split": "test",
748
+ "perturbation": {
749
+ "name": "fairness",
750
+ "robustness": false,
751
+ "fairness": true,
752
+ "computed_on": "worst"
753
+ }
754
+ },
755
+ "count": 1,
756
+ "sum": 0.12883714818954467,
757
+ "sum_squared": 0.016599010753614692,
758
+ "min": 0.12883714818954467,
759
+ "max": 0.12883714818954467,
760
+ "mean": 0.12883714818954467,
761
+ "variance": 0.0,
762
+ "stddev": 0.0
763
+ },
764
+ {
765
+ "name": {
766
+ "name": "batch_size",
767
+ "split": "test",
768
+ "perturbation": {
769
+ "name": "robustness",
770
+ "robustness": true,
771
+ "fairness": false,
772
+ "computed_on": "worst"
773
+ }
774
+ },
775
+ "count": 1,
776
+ "sum": 1.0,
777
+ "sum_squared": 1.0,
778
+ "min": 1.0,
779
+ "max": 1.0,
780
+ "mean": 1.0,
781
+ "variance": 0.0,
782
+ "stddev": 0.0
783
+ },
784
+ {
785
+ "name": {
786
+ "name": "batch_size",
787
+ "split": "test",
788
+ "perturbation": {
789
+ "name": "fairness",
790
+ "robustness": false,
791
+ "fairness": true,
792
+ "computed_on": "worst"
793
+ }
794
+ },
795
+ "count": 1,
796
+ "sum": 1.0,
797
+ "sum_squared": 1.0,
798
+ "min": 1.0,
799
+ "max": 1.0,
800
+ "mean": 1.0,
801
+ "variance": 0.0,
802
+ "stddev": 0.0
803
+ },
804
+ {
805
+ "name": {
806
+ "name": "finish_reason_length",
807
+ "split": "test",
808
+ "perturbation": {
809
+ "name": "robustness",
810
+ "robustness": true,
811
+ "fairness": false,
812
+ "computed_on": "worst"
813
+ }
814
+ },
815
+ "count": 1,
816
+ "sum": 0.0,
817
+ "sum_squared": 0.0,
818
+ "min": 0.0,
819
+ "max": 0.0,
820
+ "mean": 0.0,
821
+ "variance": 0.0,
822
+ "stddev": 0.0
823
+ },
824
+ {
825
+ "name": {
826
+ "name": "finish_reason_length",
827
+ "split": "test",
828
+ "perturbation": {
829
+ "name": "fairness",
830
+ "robustness": false,
831
+ "fairness": true,
832
+ "computed_on": "worst"
833
+ }
834
+ },
835
+ "count": 1,
836
+ "sum": 0.0,
837
+ "sum_squared": 0.0,
838
+ "min": 0.0,
839
+ "max": 0.0,
840
+ "mean": 0.0,
841
+ "variance": 0.0,
842
+ "stddev": 0.0
843
+ },
844
+ {
845
+ "name": {
846
+ "name": "finish_reason_stop",
847
+ "split": "test",
848
+ "perturbation": {
849
+ "name": "robustness",
850
+ "robustness": true,
851
+ "fairness": false,
852
+ "computed_on": "worst"
853
+ }
854
+ },
855
+ "count": 1,
856
+ "sum": 0.0,
857
+ "sum_squared": 0.0,
858
+ "min": 0.0,
859
+ "max": 0.0,
860
+ "mean": 0.0,
861
+ "variance": 0.0,
862
+ "stddev": 0.0
863
+ },
864
+ {
865
+ "name": {
866
+ "name": "finish_reason_stop",
867
+ "split": "test",
868
+ "perturbation": {
869
+ "name": "fairness",
870
+ "robustness": false,
871
+ "fairness": true,
872
+ "computed_on": "worst"
873
+ }
874
+ },
875
+ "count": 1,
876
+ "sum": 0.0,
877
+ "sum_squared": 0.0,
878
+ "min": 0.0,
879
+ "max": 0.0,
880
+ "mean": 0.0,
881
+ "variance": 0.0,
882
+ "stddev": 0.0
883
+ },
884
+ {
885
+ "name": {
886
+ "name": "finish_reason_endoftext",
887
+ "split": "test",
888
+ "perturbation": {
889
+ "name": "robustness",
890
+ "robustness": true,
891
+ "fairness": false,
892
+ "computed_on": "worst"
893
+ }
894
+ },
895
+ "count": 1,
896
+ "sum": 0.0,
897
+ "sum_squared": 0.0,
898
+ "min": 0.0,
899
+ "max": 0.0,
900
+ "mean": 0.0,
901
+ "variance": 0.0,
902
+ "stddev": 0.0
903
+ },
904
+ {
905
+ "name": {
906
+ "name": "finish_reason_endoftext",
907
+ "split": "test",
908
+ "perturbation": {
909
+ "name": "fairness",
910
+ "robustness": false,
911
+ "fairness": true,
912
+ "computed_on": "worst"
913
+ }
914
+ },
915
+ "count": 1,
916
+ "sum": 0.0,
917
+ "sum_squared": 0.0,
918
+ "min": 0.0,
919
+ "max": 0.0,
920
+ "mean": 0.0,
921
+ "variance": 0.0,
922
+ "stddev": 0.0
923
+ },
924
+ {
925
+ "name": {
926
+ "name": "finish_reason_unknown",
927
+ "split": "test",
928
+ "perturbation": {
929
+ "name": "robustness",
930
+ "robustness": true,
931
+ "fairness": false,
932
+ "computed_on": "worst"
933
+ }
934
+ },
935
+ "count": 1,
936
+ "sum": 1.0,
937
+ "sum_squared": 1.0,
938
+ "min": 1.0,
939
+ "max": 1.0,
940
+ "mean": 1.0,
941
+ "variance": 0.0,
942
+ "stddev": 0.0
943
+ },
944
+ {
945
+ "name": {
946
+ "name": "finish_reason_unknown",
947
+ "split": "test",
948
+ "perturbation": {
949
+ "name": "fairness",
950
+ "robustness": false,
951
+ "fairness": true,
952
+ "computed_on": "worst"
953
+ }
954
+ },
955
+ "count": 1,
956
+ "sum": 1.0,
957
+ "sum_squared": 1.0,
958
+ "min": 1.0,
959
+ "max": 1.0,
960
+ "mean": 1.0,
961
+ "variance": 0.0,
962
+ "stddev": 0.0
963
+ },
964
+ {
965
+ "name": {
966
+ "name": "num_train_instances",
967
+ "split": "test",
968
+ "perturbation": {
969
+ "name": "robustness",
970
+ "robustness": true,
971
+ "fairness": false,
972
+ "computed_on": "worst"
973
+ }
974
+ },
975
+ "count": 1,
976
+ "sum": 5.0,
977
+ "sum_squared": 25.0,
978
+ "min": 5.0,
979
+ "max": 5.0,
980
+ "mean": 5.0,
981
+ "variance": 0.0,
982
+ "stddev": 0.0
983
+ },
984
+ {
985
+ "name": {
986
+ "name": "num_train_instances",
987
+ "split": "test",
988
+ "perturbation": {
989
+ "name": "fairness",
990
+ "robustness": false,
991
+ "fairness": true,
992
+ "computed_on": "worst"
993
+ }
994
+ },
995
+ "count": 1,
996
+ "sum": 5.0,
997
+ "sum_squared": 25.0,
998
+ "min": 5.0,
999
+ "max": 5.0,
1000
+ "mean": 5.0,
1001
+ "variance": 0.0,
1002
+ "stddev": 0.0
1003
+ },
1004
+ {
1005
+ "name": {
1006
+ "name": "prompt_truncated",
1007
+ "split": "test",
1008
+ "perturbation": {
1009
+ "name": "robustness",
1010
+ "robustness": true,
1011
+ "fairness": false,
1012
+ "computed_on": "worst"
1013
+ }
1014
+ },
1015
+ "count": 1,
1016
+ "sum": 0.0,
1017
+ "sum_squared": 0.0,
1018
+ "min": 0.0,
1019
+ "max": 0.0,
1020
+ "mean": 0.0,
1021
+ "variance": 0.0,
1022
+ "stddev": 0.0
1023
+ },
1024
+ {
1025
+ "name": {
1026
+ "name": "prompt_truncated",
1027
+ "split": "test",
1028
+ "perturbation": {
1029
+ "name": "fairness",
1030
+ "robustness": false,
1031
+ "fairness": true,
1032
+ "computed_on": "worst"
1033
+ }
1034
+ },
1035
+ "count": 1,
1036
+ "sum": 0.0,
1037
+ "sum_squared": 0.0,
1038
+ "min": 0.0,
1039
+ "max": 0.0,
1040
+ "mean": 0.0,
1041
+ "variance": 0.0,
1042
+ "stddev": 0.0
1043
+ },
1044
+ {
1045
+ "name": {
1046
+ "name": "max_prob",
1047
+ "split": "test",
1048
+ "perturbation": {
1049
+ "name": "robustness",
1050
+ "robustness": true,
1051
+ "fairness": false,
1052
+ "computed_on": "worst"
1053
+ }
1054
+ },
1055
+ "count": 1,
1056
+ "sum": 0.45501411941835707,
1057
+ "sum_squared": 0.2070378488700629,
1058
+ "min": 0.45501411941835707,
1059
+ "max": 0.45501411941835707,
1060
+ "mean": 0.45501411941835707,
1061
+ "variance": 0.0,
1062
+ "stddev": 0.0
1063
+ },
1064
+ {
1065
+ "name": {
1066
+ "name": "max_prob",
1067
+ "split": "test",
1068
+ "perturbation": {
1069
+ "name": "fairness",
1070
+ "robustness": false,
1071
+ "fairness": true,
1072
+ "computed_on": "worst"
1073
+ }
1074
+ },
1075
+ "count": 1,
1076
+ "sum": 0.45501411941835707,
1077
+ "sum_squared": 0.2070378488700629,
1078
+ "min": 0.45501411941835707,
1079
+ "max": 0.45501411941835707,
1080
+ "mean": 0.45501411941835707,
1081
+ "variance": 0.0,
1082
+ "stddev": 0.0
1083
+ },
1084
+ {
1085
+ "name": {
1086
+ "name": "exact_match",
1087
+ "split": "test",
1088
+ "perturbation": {
1089
+ "name": "robustness",
1090
+ "robustness": true,
1091
+ "fairness": false,
1092
+ "computed_on": "worst"
1093
+ }
1094
+ },
1095
+ "count": 1,
1096
+ "sum": 0.22,
1097
+ "sum_squared": 0.0484,
1098
+ "min": 0.22,
1099
+ "max": 0.22,
1100
+ "mean": 0.22,
1101
+ "variance": 0.0,
1102
+ "stddev": 0.0
1103
+ },
1104
+ {
1105
+ "name": {
1106
+ "name": "exact_match",
1107
+ "split": "test",
1108
+ "perturbation": {
1109
+ "name": "fairness",
1110
+ "robustness": false,
1111
+ "fairness": true,
1112
+ "computed_on": "worst"
1113
+ }
1114
+ },
1115
+ "count": 1,
1116
+ "sum": 0.22,
1117
+ "sum_squared": 0.0484,
1118
+ "min": 0.22,
1119
+ "max": 0.22,
1120
+ "mean": 0.22,
1121
+ "variance": 0.0,
1122
+ "stddev": 0.0
1123
+ },
1124
+ {
1125
+ "name": {
1126
+ "name": "exact_match@5",
1127
+ "split": "test",
1128
+ "perturbation": {
1129
+ "name": "robustness",
1130
+ "robustness": true,
1131
+ "fairness": false,
1132
+ "computed_on": "worst"
1133
+ }
1134
+ },
1135
+ "count": 1,
1136
+ "sum": 0.22,
1137
+ "sum_squared": 0.0484,
1138
+ "min": 0.22,
1139
+ "max": 0.22,
1140
+ "mean": 0.22,
1141
+ "variance": 0.0,
1142
+ "stddev": 0.0
1143
+ },
1144
+ {
1145
+ "name": {
1146
+ "name": "exact_match@5",
1147
+ "split": "test",
1148
+ "perturbation": {
1149
+ "name": "fairness",
1150
+ "robustness": false,
1151
+ "fairness": true,
1152
+ "computed_on": "worst"
1153
+ }
1154
+ },
1155
+ "count": 1,
1156
+ "sum": 0.22,
1157
+ "sum_squared": 0.0484,
1158
+ "min": 0.22,
1159
+ "max": 0.22,
1160
+ "mean": 0.22,
1161
+ "variance": 0.0,
1162
+ "stddev": 0.0
1163
+ },
1164
+ {
1165
+ "name": {
1166
+ "name": "quasi_exact_match",
1167
+ "split": "test",
1168
+ "perturbation": {
1169
+ "name": "robustness",
1170
+ "robustness": true,
1171
+ "fairness": false,
1172
+ "computed_on": "worst"
1173
+ }
1174
+ },
1175
+ "count": 1,
1176
+ "sum": 0.22,
1177
+ "sum_squared": 0.0484,
1178
+ "min": 0.22,
1179
+ "max": 0.22,
1180
+ "mean": 0.22,
1181
+ "variance": 0.0,
1182
+ "stddev": 0.0
1183
+ },
1184
+ {
1185
+ "name": {
1186
+ "name": "quasi_exact_match",
1187
+ "split": "test",
1188
+ "perturbation": {
1189
+ "name": "fairness",
1190
+ "robustness": false,
1191
+ "fairness": true,
1192
+ "computed_on": "worst"
1193
+ }
1194
+ },
1195
+ "count": 1,
1196
+ "sum": 0.22,
1197
+ "sum_squared": 0.0484,
1198
+ "min": 0.22,
1199
+ "max": 0.22,
1200
+ "mean": 0.22,
1201
+ "variance": 0.0,
1202
+ "stddev": 0.0
1203
+ },
1204
+ {
1205
+ "name": {
1206
+ "name": "quasi_exact_match@5",
1207
+ "split": "test",
1208
+ "perturbation": {
1209
+ "name": "robustness",
1210
+ "robustness": true,
1211
+ "fairness": false,
1212
+ "computed_on": "worst"
1213
+ }
1214
+ },
1215
+ "count": 1,
1216
+ "sum": 0.22,
1217
+ "sum_squared": 0.0484,
1218
+ "min": 0.22,
1219
+ "max": 0.22,
1220
+ "mean": 0.22,
1221
+ "variance": 0.0,
1222
+ "stddev": 0.0
1223
+ },
1224
+ {
1225
+ "name": {
1226
+ "name": "quasi_exact_match@5",
1227
+ "split": "test",
1228
+ "perturbation": {
1229
+ "name": "fairness",
1230
+ "robustness": false,
1231
+ "fairness": true,
1232
+ "computed_on": "worst"
1233
+ }
1234
+ },
1235
+ "count": 1,
1236
+ "sum": 0.22,
1237
+ "sum_squared": 0.0484,
1238
+ "min": 0.22,
1239
+ "max": 0.22,
1240
+ "mean": 0.22,
1241
+ "variance": 0.0,
1242
+ "stddev": 0.0
1243
+ },
1244
+ {
1245
+ "name": {
1246
+ "name": "prefix_exact_match",
1247
+ "split": "test",
1248
+ "perturbation": {
1249
+ "name": "robustness",
1250
+ "robustness": true,
1251
+ "fairness": false,
1252
+ "computed_on": "worst"
1253
+ }
1254
+ },
1255
+ "count": 1,
1256
+ "sum": 0.22,
1257
+ "sum_squared": 0.0484,
1258
+ "min": 0.22,
1259
+ "max": 0.22,
1260
+ "mean": 0.22,
1261
+ "variance": 0.0,
1262
+ "stddev": 0.0
1263
+ },
1264
+ {
1265
+ "name": {
1266
+ "name": "prefix_exact_match",
1267
+ "split": "test",
1268
+ "perturbation": {
1269
+ "name": "fairness",
1270
+ "robustness": false,
1271
+ "fairness": true,
1272
+ "computed_on": "worst"
1273
+ }
1274
+ },
1275
+ "count": 1,
1276
+ "sum": 0.22,
1277
+ "sum_squared": 0.0484,
1278
+ "min": 0.22,
1279
+ "max": 0.22,
1280
+ "mean": 0.22,
1281
+ "variance": 0.0,
1282
+ "stddev": 0.0
1283
+ },
1284
+ {
1285
+ "name": {
1286
+ "name": "prefix_exact_match@5",
1287
+ "split": "test",
1288
+ "perturbation": {
1289
+ "name": "robustness",
1290
+ "robustness": true,
1291
+ "fairness": false,
1292
+ "computed_on": "worst"
1293
+ }
1294
+ },
1295
+ "count": 1,
1296
+ "sum": 0.22,
1297
+ "sum_squared": 0.0484,
1298
+ "min": 0.22,
1299
+ "max": 0.22,
1300
+ "mean": 0.22,
1301
+ "variance": 0.0,
1302
+ "stddev": 0.0
1303
+ },
1304
+ {
1305
+ "name": {
1306
+ "name": "prefix_exact_match@5",
1307
+ "split": "test",
1308
+ "perturbation": {
1309
+ "name": "fairness",
1310
+ "robustness": false,
1311
+ "fairness": true,
1312
+ "computed_on": "worst"
1313
+ }
1314
+ },
1315
+ "count": 1,
1316
+ "sum": 0.22,
1317
+ "sum_squared": 0.0484,
1318
+ "min": 0.22,
1319
+ "max": 0.22,
1320
+ "mean": 0.22,
1321
+ "variance": 0.0,
1322
+ "stddev": 0.0
1323
+ },
1324
+ {
1325
+ "name": {
1326
+ "name": "quasi_prefix_exact_match",
1327
+ "split": "test",
1328
+ "perturbation": {
1329
+ "name": "robustness",
1330
+ "robustness": true,
1331
+ "fairness": false,
1332
+ "computed_on": "worst"
1333
+ }
1334
+ },
1335
+ "count": 1,
1336
+ "sum": 0.22,
1337
+ "sum_squared": 0.0484,
1338
+ "min": 0.22,
1339
+ "max": 0.22,
1340
+ "mean": 0.22,
1341
+ "variance": 0.0,
1342
+ "stddev": 0.0
1343
+ },
1344
+ {
1345
+ "name": {
1346
+ "name": "quasi_prefix_exact_match",
1347
+ "split": "test",
1348
+ "perturbation": {
1349
+ "name": "fairness",
1350
+ "robustness": false,
1351
+ "fairness": true,
1352
+ "computed_on": "worst"
1353
+ }
1354
+ },
1355
+ "count": 1,
1356
+ "sum": 0.22,
1357
+ "sum_squared": 0.0484,
1358
+ "min": 0.22,
1359
+ "max": 0.22,
1360
+ "mean": 0.22,
1361
+ "variance": 0.0,
1362
+ "stddev": 0.0
1363
+ },
1364
+ {
1365
+ "name": {
1366
+ "name": "quasi_prefix_exact_match@5",
1367
+ "split": "test",
1368
+ "perturbation": {
1369
+ "name": "robustness",
1370
+ "robustness": true,
1371
+ "fairness": false,
1372
+ "computed_on": "worst"
1373
+ }
1374
+ },
1375
+ "count": 1,
1376
+ "sum": 0.22,
1377
+ "sum_squared": 0.0484,
1378
+ "min": 0.22,
1379
+ "max": 0.22,
1380
+ "mean": 0.22,
1381
+ "variance": 0.0,
1382
+ "stddev": 0.0
1383
+ },
1384
+ {
1385
+ "name": {
1386
+ "name": "quasi_prefix_exact_match@5",
1387
+ "split": "test",
1388
+ "perturbation": {
1389
+ "name": "fairness",
1390
+ "robustness": false,
1391
+ "fairness": true,
1392
+ "computed_on": "worst"
1393
+ }
1394
+ },
1395
+ "count": 1,
1396
+ "sum": 0.22,
1397
+ "sum_squared": 0.0484,
1398
+ "min": 0.22,
1399
+ "max": 0.22,
1400
+ "mean": 0.22,
1401
+ "variance": 0.0,
1402
+ "stddev": 0.0
1403
+ },
1404
+ {
1405
+ "name": {
1406
+ "name": "logprob",
1407
+ "split": "test",
1408
+ "perturbation": {
1409
+ "name": "robustness",
1410
+ "robustness": true,
1411
+ "fairness": false,
1412
+ "computed_on": "worst"
1413
+ }
1414
+ },
1415
+ "count": 1,
1416
+ "sum": -0.805416056394577,
1417
+ "sum_squared": 0.6486950238981924,
1418
+ "min": -0.805416056394577,
1419
+ "max": -0.805416056394577,
1420
+ "mean": -0.805416056394577,
1421
+ "variance": 0.0,
1422
+ "stddev": 0.0
1423
+ },
1424
+ {
1425
+ "name": {
1426
+ "name": "logprob",
1427
+ "split": "test",
1428
+ "perturbation": {
1429
+ "name": "fairness",
1430
+ "robustness": false,
1431
+ "fairness": true,
1432
+ "computed_on": "worst"
1433
+ }
1434
+ },
1435
+ "count": 1,
1436
+ "sum": -0.805416056394577,
1437
+ "sum_squared": 0.6486950238981924,
1438
+ "min": -0.805416056394577,
1439
+ "max": -0.805416056394577,
1440
+ "mean": -0.805416056394577,
1441
+ "variance": 0.0,
1442
+ "stddev": 0.0
1443
+ },
1444
+ {
1445
+ "name": {
1446
+ "name": "num_perplexity_tokens",
1447
+ "split": "test",
1448
+ "perturbation": {
1449
+ "name": "robustness",
1450
+ "robustness": true,
1451
+ "fairness": false,
1452
+ "computed_on": "worst"
1453
+ }
1454
+ },
1455
+ "count": 1,
1456
+ "sum": 1.0,
1457
+ "sum_squared": 1.0,
1458
+ "min": 1.0,
1459
+ "max": 1.0,
1460
+ "mean": 1.0,
1461
+ "variance": 0.0,
1462
+ "stddev": 0.0
1463
+ },
1464
+ {
1465
+ "name": {
1466
+ "name": "num_perplexity_tokens",
1467
+ "split": "test",
1468
+ "perturbation": {
1469
+ "name": "fairness",
1470
+ "robustness": false,
1471
+ "fairness": true,
1472
+ "computed_on": "worst"
1473
+ }
1474
+ },
1475
+ "count": 1,
1476
+ "sum": 1.0,
1477
+ "sum_squared": 1.0,
1478
+ "min": 1.0,
1479
+ "max": 1.0,
1480
+ "mean": 1.0,
1481
+ "variance": 0.0,
1482
+ "stddev": 0.0
1483
+ },
1484
+ {
1485
+ "name": {
1486
+ "name": "num_bytes",
1487
+ "split": "test",
1488
+ "perturbation": {
1489
+ "name": "robustness",
1490
+ "robustness": true,
1491
+ "fairness": false,
1492
+ "computed_on": "worst"
1493
+ }
1494
+ },
1495
+ "count": 1,
1496
+ "sum": 2.0,
1497
+ "sum_squared": 4.0,
1498
+ "min": 2.0,
1499
+ "max": 2.0,
1500
+ "mean": 2.0,
1501
+ "variance": 0.0,
1502
+ "stddev": 0.0
1503
+ },
1504
+ {
1505
+ "name": {
1506
+ "name": "num_bytes",
1507
+ "split": "test",
1508
+ "perturbation": {
1509
+ "name": "fairness",
1510
+ "robustness": false,
1511
+ "fairness": true,
1512
+ "computed_on": "worst"
1513
+ }
1514
+ },
1515
+ "count": 1,
1516
+ "sum": 2.0,
1517
+ "sum_squared": 4.0,
1518
+ "min": 2.0,
1519
+ "max": 2.0,
1520
+ "mean": 2.0,
1521
+ "variance": 0.0,
1522
+ "stddev": 0.0
1523
+ },
1524
+ {
1525
+ "name": {
1526
+ "name": "num_instances",
1527
+ "split": "test"
1528
+ },
1529
+ "count": 1,
1530
+ "sum": 500.0,
1531
+ "sum_squared": 250000.0,
1532
+ "min": 500.0,
1533
+ "max": 500.0,
1534
+ "mean": 500.0,
1535
+ "variance": 0.0,
1536
+ "stddev": 0.0
1537
+ }
1538
+ ]
lite_pythia-2.8b-step5000/costs.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {}
lite_pythia-2.8b-step5000/groups.json ADDED
@@ -0,0 +1,491 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "title": "All scenarios",
4
+ "header": [
5
+ {
6
+ "value": "Group",
7
+ "markdown": false,
8
+ "metadata": {}
9
+ },
10
+ {
11
+ "value": "Description",
12
+ "markdown": false,
13
+ "metadata": {}
14
+ },
15
+ {
16
+ "value": "Adaptation method",
17
+ "description": "Adaptation strategy (e.g., generation)",
18
+ "markdown": false,
19
+ "metadata": {}
20
+ },
21
+ {
22
+ "value": "# instances",
23
+ "description": "Number of instances evaluated on",
24
+ "markdown": false,
25
+ "metadata": {}
26
+ },
27
+ {
28
+ "value": "# references",
29
+ "description": "Number of references provided per instance",
30
+ "markdown": false,
31
+ "metadata": {}
32
+ },
33
+ {
34
+ "value": "# prompt tokens",
35
+ "description": "Total number of prompt tokens",
36
+ "markdown": false,
37
+ "metadata": {}
38
+ },
39
+ {
40
+ "value": "# completion tokens",
41
+ "description": "Total number of completion tokens",
42
+ "markdown": false,
43
+ "metadata": {}
44
+ },
45
+ {
46
+ "value": "# models",
47
+ "description": "Number of models we're evaluating",
48
+ "markdown": false,
49
+ "metadata": {}
50
+ }
51
+ ],
52
+ "rows": [
53
+ [
54
+ {
55
+ "value": "Core scenarios",
56
+ "href": "?group=core_scenarios",
57
+ "markdown": false
58
+ },
59
+ {
60
+ "value": "The scenarios where we evaluate all the models.",
61
+ "markdown": true
62
+ },
63
+ {
64
+ "value": "multiple_choice_joint, generation",
65
+ "markdown": false
66
+ },
67
+ {
68
+ "value": 242.0,
69
+ "description": "min=8, mean=242, max=1000, sum=4114 (17)",
70
+ "markdown": false
71
+ },
72
+ {
73
+ "value": 2.9411764705882355,
74
+ "description": "min=1, mean=2.941, max=4, sum=150 (51)",
75
+ "markdown": false
76
+ },
77
+ {
78
+ "value": 26153.916629538737,
79
+ "description": "min=206.779, mean=512.822, max=1497.455, sum=26153.917 (51)",
80
+ "markdown": false
81
+ },
82
+ {
83
+ "value": 563.0038753823058,
84
+ "description": "min=1, mean=11.039, max=168.459, sum=563.004 (51)",
85
+ "markdown": false
86
+ },
87
+ {
88
+ "value": 1,
89
+ "markdown": false
90
+ }
91
+ ]
92
+ ],
93
+ "links": []
94
+ },
95
+ {
96
+ "title": "Scenarios",
97
+ "header": [
98
+ {
99
+ "value": "Group",
100
+ "markdown": false,
101
+ "metadata": {}
102
+ },
103
+ {
104
+ "value": "Description",
105
+ "markdown": false,
106
+ "metadata": {}
107
+ },
108
+ {
109
+ "value": "Adaptation method",
110
+ "description": "Adaptation strategy (e.g., generation)",
111
+ "markdown": false,
112
+ "metadata": {}
113
+ },
114
+ {
115
+ "value": "# instances",
116
+ "description": "Number of instances evaluated on",
117
+ "markdown": false,
118
+ "metadata": {}
119
+ },
120
+ {
121
+ "value": "# references",
122
+ "description": "Number of references provided per instance",
123
+ "markdown": false,
124
+ "metadata": {}
125
+ },
126
+ {
127
+ "value": "# prompt tokens",
128
+ "description": "Total number of prompt tokens",
129
+ "markdown": false,
130
+ "metadata": {}
131
+ },
132
+ {
133
+ "value": "# completion tokens",
134
+ "description": "Total number of completion tokens",
135
+ "markdown": false,
136
+ "metadata": {}
137
+ },
138
+ {
139
+ "value": "# models",
140
+ "description": "Number of models we're evaluating",
141
+ "markdown": false,
142
+ "metadata": {}
143
+ }
144
+ ],
145
+ "rows": [
146
+ [
147
+ {
148
+ "value": "NarrativeQA",
149
+ "href": "?group=narrative_qa",
150
+ "markdown": false
151
+ },
152
+ {
153
+ "value": "The NarrativeQA benchmark for reading comprehension over narratives [(Ko\u010disk\u00fd et al., 2017)](https://aclanthology.org/Q18-1023/).",
154
+ "markdown": true
155
+ },
156
+ {
157
+ "value": "",
158
+ "markdown": false
159
+ },
160
+ {
161
+ "markdown": false
162
+ },
163
+ {
164
+ "markdown": false
165
+ },
166
+ {
167
+ "markdown": false
168
+ },
169
+ {
170
+ "markdown": false
171
+ },
172
+ {
173
+ "value": 0,
174
+ "markdown": false
175
+ }
176
+ ],
177
+ [
178
+ {
179
+ "value": "NaturalQuestions (closed-book)",
180
+ "href": "?group=natural_qa_closedbook",
181
+ "markdown": false
182
+ },
183
+ {
184
+ "value": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input does not include the Wikipedia page with the answer.",
185
+ "markdown": true
186
+ },
187
+ {
188
+ "value": "",
189
+ "markdown": false
190
+ },
191
+ {
192
+ "markdown": false
193
+ },
194
+ {
195
+ "markdown": false
196
+ },
197
+ {
198
+ "markdown": false
199
+ },
200
+ {
201
+ "markdown": false
202
+ },
203
+ {
204
+ "value": 0,
205
+ "markdown": false
206
+ }
207
+ ],
208
+ [
209
+ {
210
+ "value": "NaturalQuestions (open-book)",
211
+ "href": "?group=natural_qa_openbook_longans",
212
+ "markdown": false
213
+ },
214
+ {
215
+ "value": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input includes the Wikipedia page with the answer.",
216
+ "markdown": true
217
+ },
218
+ {
219
+ "value": "",
220
+ "markdown": false
221
+ },
222
+ {
223
+ "markdown": false
224
+ },
225
+ {
226
+ "markdown": false
227
+ },
228
+ {
229
+ "markdown": false
230
+ },
231
+ {
232
+ "markdown": false
233
+ },
234
+ {
235
+ "value": 0,
236
+ "markdown": false
237
+ }
238
+ ],
239
+ [
240
+ {
241
+ "value": "OpenbookQA",
242
+ "href": "?group=openbookqa",
243
+ "markdown": false
244
+ },
245
+ {
246
+ "value": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).",
247
+ "markdown": true
248
+ },
249
+ {
250
+ "value": "multiple_choice_joint",
251
+ "markdown": false
252
+ },
253
+ {
254
+ "value": 500.0,
255
+ "description": "min=500, mean=500, max=500, sum=500 (1)",
256
+ "markdown": false
257
+ },
258
+ {
259
+ "value": 4.0,
260
+ "description": "min=4, mean=4, max=4, sum=12 (3)",
261
+ "markdown": false
262
+ },
263
+ {
264
+ "value": 754.668,
265
+ "description": "min=251.556, mean=251.556, max=251.556, sum=754.668 (3)",
266
+ "markdown": false
267
+ },
268
+ {
269
+ "value": 3.0,
270
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
271
+ "markdown": false
272
+ },
273
+ {
274
+ "value": 1,
275
+ "markdown": false
276
+ }
277
+ ],
278
+ [
279
+ {
280
+ "value": "MMLU (Massive Multitask Language Understanding)",
281
+ "href": "?group=mmlu",
282
+ "markdown": false
283
+ },
284
+ {
285
+ "value": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).",
286
+ "markdown": true
287
+ },
288
+ {
289
+ "value": "multiple_choice_joint",
290
+ "markdown": false
291
+ },
292
+ {
293
+ "value": 56.7,
294
+ "description": "min=8, mean=56.7, max=114, sum=567 (10)",
295
+ "markdown": false
296
+ },
297
+ {
298
+ "value": 4.0,
299
+ "description": "min=4, mean=4, max=4, sum=120 (30)",
300
+ "markdown": false
301
+ },
302
+ {
303
+ "value": 14170.841555023922,
304
+ "description": "min=354.455, mean=472.361, max=612.798, sum=14170.842 (30)",
305
+ "markdown": false
306
+ },
307
+ {
308
+ "value": 30.0,
309
+ "description": "min=1, mean=1, max=1, sum=30 (30)",
310
+ "markdown": false
311
+ },
312
+ {
313
+ "value": 1,
314
+ "markdown": false
315
+ }
316
+ ],
317
+ [
318
+ {
319
+ "value": "GSM8K (Grade School Math)",
320
+ "href": "?group=gsm",
321
+ "markdown": false
322
+ },
323
+ {
324
+ "value": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).",
325
+ "markdown": true
326
+ },
327
+ {
328
+ "value": "generation",
329
+ "markdown": false
330
+ },
331
+ {
332
+ "value": 1000.0,
333
+ "description": "min=1000, mean=1000, max=1000, sum=1000 (1)",
334
+ "markdown": false
335
+ },
336
+ {
337
+ "value": 1.0,
338
+ "description": "min=1, mean=1, max=1, sum=3 (3)",
339
+ "markdown": false
340
+ },
341
+ {
342
+ "value": 2818.746,
343
+ "description": "min=939.582, mean=939.582, max=939.582, sum=2818.746 (3)",
344
+ "markdown": false
345
+ },
346
+ {
347
+ "value": 505.377,
348
+ "description": "min=168.459, mean=168.459, max=168.459, sum=505.377 (3)",
349
+ "markdown": false
350
+ },
351
+ {
352
+ "value": 1,
353
+ "markdown": false
354
+ }
355
+ ],
356
+ [
357
+ {
358
+ "value": "MATH",
359
+ "href": "?group=math_chain_of_thought",
360
+ "markdown": false
361
+ },
362
+ {
363
+ "value": "The MATH benchmark for measuring mathematical problem solving on competition math problems with chain-of-thought style reasoning [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2103.03874.pdf).",
364
+ "markdown": true
365
+ },
366
+ {
367
+ "value": "",
368
+ "markdown": false
369
+ },
370
+ {
371
+ "markdown": false
372
+ },
373
+ {
374
+ "markdown": false
375
+ },
376
+ {
377
+ "markdown": false
378
+ },
379
+ {
380
+ "markdown": false
381
+ },
382
+ {
383
+ "value": 0,
384
+ "markdown": false
385
+ }
386
+ ],
387
+ [
388
+ {
389
+ "value": "LegalBench",
390
+ "href": "?group=legalbench",
391
+ "markdown": false
392
+ },
393
+ {
394
+ "value": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).",
395
+ "markdown": true
396
+ },
397
+ {
398
+ "value": "generation",
399
+ "markdown": false
400
+ },
401
+ {
402
+ "value": 409.4,
403
+ "description": "min=95, mean=409.4, max=1000, sum=2047 (5)",
404
+ "markdown": false
405
+ },
406
+ {
407
+ "value": 1.0,
408
+ "description": "min=1, mean=1, max=1, sum=15 (15)",
409
+ "markdown": false
410
+ },
411
+ {
412
+ "value": 8409.66107451482,
413
+ "description": "min=206.779, mean=560.644, max=1497.455, sum=8409.661 (15)",
414
+ "markdown": false
415
+ },
416
+ {
417
+ "value": 24.626875382305506,
418
+ "description": "min=1, mean=1.642, max=4.124, sum=24.627 (15)",
419
+ "markdown": false
420
+ },
421
+ {
422
+ "value": 1,
423
+ "markdown": false
424
+ }
425
+ ],
426
+ [
427
+ {
428
+ "value": "MedQA",
429
+ "href": "?group=med_qa",
430
+ "markdown": false
431
+ },
432
+ {
433
+ "value": "MedQA is an open domain question answering dataset composed of questions from professional medical board exams ([Jin et al. 2020](https://arxiv.org/pdf/2009.13081.pdf)).",
434
+ "markdown": true
435
+ },
436
+ {
437
+ "value": "",
438
+ "markdown": false
439
+ },
440
+ {
441
+ "markdown": false
442
+ },
443
+ {
444
+ "markdown": false
445
+ },
446
+ {
447
+ "markdown": false
448
+ },
449
+ {
450
+ "markdown": false
451
+ },
452
+ {
453
+ "value": 0,
454
+ "markdown": false
455
+ }
456
+ ],
457
+ [
458
+ {
459
+ "value": "WMT 2014",
460
+ "href": "?group=wmt_14",
461
+ "markdown": false
462
+ },
463
+ {
464
+ "value": "WMT 2014 is a collection of machine translation datasets [(website)](https://www.statmt.org/wmt14/index.html).",
465
+ "markdown": true
466
+ },
467
+ {
468
+ "value": "",
469
+ "markdown": false
470
+ },
471
+ {
472
+ "markdown": false
473
+ },
474
+ {
475
+ "markdown": false
476
+ },
477
+ {
478
+ "markdown": false
479
+ },
480
+ {
481
+ "markdown": false
482
+ },
483
+ {
484
+ "value": 0,
485
+ "markdown": false
486
+ }
487
+ ]
488
+ ],
489
+ "links": []
490
+ }
491
+ ]
lite_pythia-2.8b-step5000/groups/core_scenarios.json ADDED
@@ -0,0 +1,1264 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "title": "Accuracy",
4
+ "header": [
5
+ {
6
+ "value": "Model",
7
+ "markdown": false,
8
+ "metadata": {}
9
+ },
10
+ {
11
+ "value": "Mean win rate",
12
+ "description": "How many models this model outperforms on average (over columns).",
13
+ "markdown": false,
14
+ "lower_is_better": false,
15
+ "metadata": {}
16
+ },
17
+ {
18
+ "value": "NarrativeQA - F1",
19
+ "description": "The NarrativeQA benchmark for reading comprehension over narratives [(Ko\u010disk\u00fd et al., 2017)](https://aclanthology.org/Q18-1023/).\n\nF1: Average F1 score in terms of word overlap between the model output and correct reference.",
20
+ "markdown": false,
21
+ "lower_is_better": false,
22
+ "metadata": {
23
+ "metric": "F1",
24
+ "run_group": "NarrativeQA"
25
+ }
26
+ },
27
+ {
28
+ "value": "NaturalQuestions (open-book) - F1",
29
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input includes the Wikipedia page with the answer.\n\nF1: Average F1 score in terms of word overlap between the model output and correct reference.",
30
+ "markdown": false,
31
+ "lower_is_better": false,
32
+ "metadata": {
33
+ "metric": "F1",
34
+ "run_group": "NaturalQuestions (open-book)"
35
+ }
36
+ },
37
+ {
38
+ "value": "NaturalQuestions (closed-book) - F1",
39
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input does not include the Wikipedia page with the answer.\n\nF1: Average F1 score in terms of word overlap between the model output and correct reference.",
40
+ "markdown": false,
41
+ "lower_is_better": false,
42
+ "metadata": {
43
+ "metric": "F1",
44
+ "run_group": "NaturalQuestions (closed-book)"
45
+ }
46
+ },
47
+ {
48
+ "value": "OpenbookQA - EM",
49
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
50
+ "markdown": false,
51
+ "lower_is_better": false,
52
+ "metadata": {
53
+ "metric": "EM",
54
+ "run_group": "OpenbookQA"
55
+ }
56
+ },
57
+ {
58
+ "value": "MMLU - EM",
59
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
60
+ "markdown": false,
61
+ "lower_is_better": false,
62
+ "metadata": {
63
+ "metric": "EM",
64
+ "run_group": "MMLU"
65
+ }
66
+ },
67
+ {
68
+ "value": "MATH - Equivalent (CoT)",
69
+ "description": "The MATH benchmark for measuring mathematical problem solving on competition math problems with chain-of-thought style reasoning [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2103.03874.pdf).\n\nEquivalent (CoT): Fraction of model outputs that are mathematically equivalent to the correct reference when using chain-of-thought prompting.",
70
+ "markdown": false,
71
+ "lower_is_better": false,
72
+ "metadata": {
73
+ "metric": "Equivalent (CoT)",
74
+ "run_group": "MATH"
75
+ }
76
+ },
77
+ {
78
+ "value": "GSM8K - EM",
79
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\nExact match (final number): Fraction of instances that the predicted output matches a correct reference exactly, ignoring text preceding the specified indicator.",
80
+ "markdown": false,
81
+ "lower_is_better": false,
82
+ "metadata": {
83
+ "metric": "EM",
84
+ "run_group": "GSM8K"
85
+ }
86
+ },
87
+ {
88
+ "value": "LegalBench - EM",
89
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.",
90
+ "markdown": false,
91
+ "lower_is_better": false,
92
+ "metadata": {
93
+ "metric": "EM",
94
+ "run_group": "LegalBench"
95
+ }
96
+ },
97
+ {
98
+ "value": "MedQA - EM",
99
+ "description": "MedQA is an open domain question answering dataset composed of questions from professional medical board exams ([Jin et al. 2020](https://arxiv.org/pdf/2009.13081.pdf)).\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.",
100
+ "markdown": false,
101
+ "lower_is_better": false,
102
+ "metadata": {
103
+ "metric": "EM",
104
+ "run_group": "MedQA"
105
+ }
106
+ },
107
+ {
108
+ "value": "WMT 2014 - BLEU-4",
109
+ "description": "WMT 2014 is a collection of machine translation datasets [(website)](https://www.statmt.org/wmt14/index.html).\n\nBLEU-4: Average BLEU score [(Papineni et al., 2002)](https://aclanthology.org/P02-1040/) based on 4-gram overlap.",
110
+ "markdown": false,
111
+ "lower_is_better": false,
112
+ "metadata": {
113
+ "metric": "BLEU-4",
114
+ "run_group": "WMT 2014"
115
+ }
116
+ }
117
+ ],
118
+ "rows": [
119
+ [
120
+ {
121
+ "value": "EleutherAI/pythia-2.8b",
122
+ "description": "",
123
+ "markdown": false
124
+ },
125
+ {
126
+ "markdown": false
127
+ },
128
+ {
129
+ "description": "No matching runs",
130
+ "markdown": false
131
+ },
132
+ {
133
+ "description": "No matching runs",
134
+ "markdown": false
135
+ },
136
+ {
137
+ "description": "No matching runs",
138
+ "markdown": false
139
+ },
140
+ {
141
+ "value": 0.22,
142
+ "description": "min=0.22, mean=0.22, max=0.22, sum=0.22 (1)",
143
+ "style": {
144
+ "font-weight": "bold"
145
+ },
146
+ "markdown": false,
147
+ "run_spec_names": [
148
+ "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
149
+ ]
150
+ },
151
+ {
152
+ "value": 0.27487719298245616,
153
+ "description": "min=0.25, mean=0.275, max=0.32, sum=1.374 (5)",
154
+ "style": {
155
+ "font-weight": "bold"
156
+ },
157
+ "markdown": false,
158
+ "run_spec_names": [
159
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
160
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
161
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
162
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
163
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
164
+ ]
165
+ },
166
+ {
167
+ "description": "No matching runs",
168
+ "markdown": false
169
+ },
170
+ {
171
+ "value": 0.01,
172
+ "description": "min=0.01, mean=0.01, max=0.01, sum=0.01 (1)",
173
+ "style": {
174
+ "font-weight": "bold"
175
+ },
176
+ "markdown": false,
177
+ "run_spec_names": [
178
+ "gsm:model=EleutherAI_pythia-2.8b"
179
+ ]
180
+ },
181
+ {
182
+ "value": 0.3189986232611502,
183
+ "description": "min=0.144, mean=0.319, max=0.619, sum=1.595 (5)",
184
+ "style": {
185
+ "font-weight": "bold"
186
+ },
187
+ "markdown": false,
188
+ "run_spec_names": [
189
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
190
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
191
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
192
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
193
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
194
+ ]
195
+ },
196
+ {
197
+ "description": "No matching runs",
198
+ "markdown": false
199
+ },
200
+ {
201
+ "description": "No matching runs",
202
+ "markdown": false
203
+ }
204
+ ]
205
+ ],
206
+ "links": [
207
+ {
208
+ "text": "LaTeX",
209
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/core_scenarios_accuracy.tex"
210
+ },
211
+ {
212
+ "text": "JSON",
213
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/core_scenarios_accuracy.json"
214
+ }
215
+ ],
216
+ "name": "accuracy"
217
+ },
218
+ {
219
+ "title": "Efficiency",
220
+ "header": [
221
+ {
222
+ "value": "Model",
223
+ "markdown": false,
224
+ "metadata": {}
225
+ },
226
+ {
227
+ "value": "Mean win rate",
228
+ "description": "How many models this model outperforms on average (over columns).",
229
+ "markdown": false,
230
+ "lower_is_better": false,
231
+ "metadata": {}
232
+ },
233
+ {
234
+ "value": "NarrativeQA - Observed inference time (s)",
235
+ "description": "The NarrativeQA benchmark for reading comprehension over narratives [(Ko\u010disk\u00fd et al., 2017)](https://aclanthology.org/Q18-1023/).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
236
+ "markdown": false,
237
+ "lower_is_better": true,
238
+ "metadata": {
239
+ "metric": "Observed inference time (s)",
240
+ "run_group": "NarrativeQA"
241
+ }
242
+ },
243
+ {
244
+ "value": "NaturalQuestions (open-book) - Observed inference time (s)",
245
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input includes the Wikipedia page with the answer.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
246
+ "markdown": false,
247
+ "lower_is_better": true,
248
+ "metadata": {
249
+ "metric": "Observed inference time (s)",
250
+ "run_group": "NaturalQuestions (open-book)"
251
+ }
252
+ },
253
+ {
254
+ "value": "NaturalQuestions (closed-book) - Observed inference time (s)",
255
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input does not include the Wikipedia page with the answer.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
256
+ "markdown": false,
257
+ "lower_is_better": true,
258
+ "metadata": {
259
+ "metric": "Observed inference time (s)",
260
+ "run_group": "NaturalQuestions (closed-book)"
261
+ }
262
+ },
263
+ {
264
+ "value": "OpenbookQA - Observed inference time (s)",
265
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
266
+ "markdown": false,
267
+ "lower_is_better": true,
268
+ "metadata": {
269
+ "metric": "Observed inference time (s)",
270
+ "run_group": "OpenbookQA"
271
+ }
272
+ },
273
+ {
274
+ "value": "MMLU - Observed inference time (s)",
275
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
276
+ "markdown": false,
277
+ "lower_is_better": true,
278
+ "metadata": {
279
+ "metric": "Observed inference time (s)",
280
+ "run_group": "MMLU"
281
+ }
282
+ },
283
+ {
284
+ "value": "MATH - Observed inference time (s)",
285
+ "description": "The MATH benchmark for measuring mathematical problem solving on competition math problems with chain-of-thought style reasoning [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2103.03874.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
286
+ "markdown": false,
287
+ "lower_is_better": true,
288
+ "metadata": {
289
+ "metric": "Observed inference time (s)",
290
+ "run_group": "MATH"
291
+ }
292
+ },
293
+ {
294
+ "value": "GSM8K - Observed inference time (s)",
295
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
296
+ "markdown": false,
297
+ "lower_is_better": true,
298
+ "metadata": {
299
+ "metric": "Observed inference time (s)",
300
+ "run_group": "GSM8K"
301
+ }
302
+ },
303
+ {
304
+ "value": "LegalBench - Observed inference time (s)",
305
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
306
+ "markdown": false,
307
+ "lower_is_better": true,
308
+ "metadata": {
309
+ "metric": "Observed inference time (s)",
310
+ "run_group": "LegalBench"
311
+ }
312
+ },
313
+ {
314
+ "value": "MedQA - Observed inference time (s)",
315
+ "description": "MedQA is an open domain question answering dataset composed of questions from professional medical board exams ([Jin et al. 2020](https://arxiv.org/pdf/2009.13081.pdf)).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
316
+ "markdown": false,
317
+ "lower_is_better": true,
318
+ "metadata": {
319
+ "metric": "Observed inference time (s)",
320
+ "run_group": "MedQA"
321
+ }
322
+ },
323
+ {
324
+ "value": "WMT 2014 - Observed inference time (s)",
325
+ "description": "WMT 2014 is a collection of machine translation datasets [(website)](https://www.statmt.org/wmt14/index.html).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
326
+ "markdown": false,
327
+ "lower_is_better": true,
328
+ "metadata": {
329
+ "metric": "Observed inference time (s)",
330
+ "run_group": "WMT 2014"
331
+ }
332
+ }
333
+ ],
334
+ "rows": [
335
+ [
336
+ {
337
+ "value": "EleutherAI/pythia-2.8b",
338
+ "description": "",
339
+ "markdown": false
340
+ },
341
+ {
342
+ "markdown": false
343
+ },
344
+ {
345
+ "description": "No matching runs",
346
+ "markdown": false
347
+ },
348
+ {
349
+ "description": "No matching runs",
350
+ "markdown": false
351
+ },
352
+ {
353
+ "description": "No matching runs",
354
+ "markdown": false
355
+ },
356
+ {
357
+ "value": 0.12883714818954467,
358
+ "description": "min=0.129, mean=0.129, max=0.129, sum=0.129 (1)",
359
+ "style": {
360
+ "font-weight": "bold"
361
+ },
362
+ "markdown": false,
363
+ "run_spec_names": [
364
+ "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
365
+ ]
366
+ },
367
+ {
368
+ "value": 0.22061018314696193,
369
+ "description": "min=0.173, mean=0.221, max=0.282, sum=1.103 (5)",
370
+ "style": {
371
+ "font-weight": "bold"
372
+ },
373
+ "markdown": false,
374
+ "run_spec_names": [
375
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
376
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
377
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
378
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
379
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
380
+ ]
381
+ },
382
+ {
383
+ "description": "No matching runs",
384
+ "markdown": false
385
+ },
386
+ {
387
+ "value": 2.5314217054843904,
388
+ "description": "min=2.531, mean=2.531, max=2.531, sum=2.531 (1)",
389
+ "style": {
390
+ "font-weight": "bold"
391
+ },
392
+ "markdown": false,
393
+ "run_spec_names": [
394
+ "gsm:model=EleutherAI_pythia-2.8b"
395
+ ]
396
+ },
397
+ {
398
+ "value": 0.3094366045219278,
399
+ "description": "min=0.141, mean=0.309, max=0.795, sum=1.547 (5)",
400
+ "style": {
401
+ "font-weight": "bold"
402
+ },
403
+ "markdown": false,
404
+ "run_spec_names": [
405
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
406
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
407
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
408
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
409
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
410
+ ]
411
+ },
412
+ {
413
+ "description": "No matching runs",
414
+ "markdown": false
415
+ },
416
+ {
417
+ "description": "No matching runs",
418
+ "markdown": false
419
+ }
420
+ ]
421
+ ],
422
+ "links": [
423
+ {
424
+ "text": "LaTeX",
425
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/core_scenarios_efficiency.tex"
426
+ },
427
+ {
428
+ "text": "JSON",
429
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/core_scenarios_efficiency.json"
430
+ }
431
+ ],
432
+ "name": "efficiency"
433
+ },
434
+ {
435
+ "title": "General information",
436
+ "header": [
437
+ {
438
+ "value": "Model",
439
+ "markdown": false,
440
+ "metadata": {}
441
+ },
442
+ {
443
+ "value": "Mean win rate",
444
+ "description": "How many models this model outperforms on average (over columns).",
445
+ "markdown": false,
446
+ "lower_is_better": false,
447
+ "metadata": {}
448
+ },
449
+ {
450
+ "value": "NarrativeQA - # eval",
451
+ "description": "The NarrativeQA benchmark for reading comprehension over narratives [(Ko\u010disk\u00fd et al., 2017)](https://aclanthology.org/Q18-1023/).\n\n# eval: Number of evaluation instances.",
452
+ "markdown": false,
453
+ "metadata": {
454
+ "metric": "# eval",
455
+ "run_group": "NarrativeQA"
456
+ }
457
+ },
458
+ {
459
+ "value": "NarrativeQA - # train",
460
+ "description": "The NarrativeQA benchmark for reading comprehension over narratives [(Ko\u010disk\u00fd et al., 2017)](https://aclanthology.org/Q18-1023/).\n\n# train: Number of training instances (e.g., in-context examples).",
461
+ "markdown": false,
462
+ "metadata": {
463
+ "metric": "# train",
464
+ "run_group": "NarrativeQA"
465
+ }
466
+ },
467
+ {
468
+ "value": "NarrativeQA - truncated",
469
+ "description": "The NarrativeQA benchmark for reading comprehension over narratives [(Ko\u010disk\u00fd et al., 2017)](https://aclanthology.org/Q18-1023/).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
470
+ "markdown": false,
471
+ "metadata": {
472
+ "metric": "truncated",
473
+ "run_group": "NarrativeQA"
474
+ }
475
+ },
476
+ {
477
+ "value": "NarrativeQA - # prompt tokens",
478
+ "description": "The NarrativeQA benchmark for reading comprehension over narratives [(Ko\u010disk\u00fd et al., 2017)](https://aclanthology.org/Q18-1023/).\n\n# prompt tokens: Number of tokens in the prompt.",
479
+ "markdown": false,
480
+ "metadata": {
481
+ "metric": "# prompt tokens",
482
+ "run_group": "NarrativeQA"
483
+ }
484
+ },
485
+ {
486
+ "value": "NarrativeQA - # output tokens",
487
+ "description": "The NarrativeQA benchmark for reading comprehension over narratives [(Ko\u010disk\u00fd et al., 2017)](https://aclanthology.org/Q18-1023/).\n\n# output tokens: Actual number of output tokens.",
488
+ "markdown": false,
489
+ "metadata": {
490
+ "metric": "# output tokens",
491
+ "run_group": "NarrativeQA"
492
+ }
493
+ },
494
+ {
495
+ "value": "NaturalQuestions (open-book) - # eval",
496
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input includes the Wikipedia page with the answer.\n\n# eval: Number of evaluation instances.",
497
+ "markdown": false,
498
+ "metadata": {
499
+ "metric": "# eval",
500
+ "run_group": "NaturalQuestions (open-book)"
501
+ }
502
+ },
503
+ {
504
+ "value": "NaturalQuestions (open-book) - # train",
505
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input includes the Wikipedia page with the answer.\n\n# train: Number of training instances (e.g., in-context examples).",
506
+ "markdown": false,
507
+ "metadata": {
508
+ "metric": "# train",
509
+ "run_group": "NaturalQuestions (open-book)"
510
+ }
511
+ },
512
+ {
513
+ "value": "NaturalQuestions (open-book) - truncated",
514
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input includes the Wikipedia page with the answer.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
515
+ "markdown": false,
516
+ "metadata": {
517
+ "metric": "truncated",
518
+ "run_group": "NaturalQuestions (open-book)"
519
+ }
520
+ },
521
+ {
522
+ "value": "NaturalQuestions (open-book) - # prompt tokens",
523
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input includes the Wikipedia page with the answer.\n\n# prompt tokens: Number of tokens in the prompt.",
524
+ "markdown": false,
525
+ "metadata": {
526
+ "metric": "# prompt tokens",
527
+ "run_group": "NaturalQuestions (open-book)"
528
+ }
529
+ },
530
+ {
531
+ "value": "NaturalQuestions (open-book) - # output tokens",
532
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input includes the Wikipedia page with the answer.\n\n# output tokens: Actual number of output tokens.",
533
+ "markdown": false,
534
+ "metadata": {
535
+ "metric": "# output tokens",
536
+ "run_group": "NaturalQuestions (open-book)"
537
+ }
538
+ },
539
+ {
540
+ "value": "NaturalQuestions (closed-book) - # eval",
541
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input does not include the Wikipedia page with the answer.\n\n# eval: Number of evaluation instances.",
542
+ "markdown": false,
543
+ "metadata": {
544
+ "metric": "# eval",
545
+ "run_group": "NaturalQuestions (closed-book)"
546
+ }
547
+ },
548
+ {
549
+ "value": "NaturalQuestions (closed-book) - # train",
550
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input does not include the Wikipedia page with the answer.\n\n# train: Number of training instances (e.g., in-context examples).",
551
+ "markdown": false,
552
+ "metadata": {
553
+ "metric": "# train",
554
+ "run_group": "NaturalQuestions (closed-book)"
555
+ }
556
+ },
557
+ {
558
+ "value": "NaturalQuestions (closed-book) - truncated",
559
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input does not include the Wikipedia page with the answer.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
560
+ "markdown": false,
561
+ "metadata": {
562
+ "metric": "truncated",
563
+ "run_group": "NaturalQuestions (closed-book)"
564
+ }
565
+ },
566
+ {
567
+ "value": "NaturalQuestions (closed-book) - # prompt tokens",
568
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input does not include the Wikipedia page with the answer.\n\n# prompt tokens: Number of tokens in the prompt.",
569
+ "markdown": false,
570
+ "metadata": {
571
+ "metric": "# prompt tokens",
572
+ "run_group": "NaturalQuestions (closed-book)"
573
+ }
574
+ },
575
+ {
576
+ "value": "NaturalQuestions (closed-book) - # output tokens",
577
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input does not include the Wikipedia page with the answer.\n\n# output tokens: Actual number of output tokens.",
578
+ "markdown": false,
579
+ "metadata": {
580
+ "metric": "# output tokens",
581
+ "run_group": "NaturalQuestions (closed-book)"
582
+ }
583
+ },
584
+ {
585
+ "value": "OpenbookQA - # eval",
586
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\n# eval: Number of evaluation instances.",
587
+ "markdown": false,
588
+ "metadata": {
589
+ "metric": "# eval",
590
+ "run_group": "OpenbookQA"
591
+ }
592
+ },
593
+ {
594
+ "value": "OpenbookQA - # train",
595
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\n# train: Number of training instances (e.g., in-context examples).",
596
+ "markdown": false,
597
+ "metadata": {
598
+ "metric": "# train",
599
+ "run_group": "OpenbookQA"
600
+ }
601
+ },
602
+ {
603
+ "value": "OpenbookQA - truncated",
604
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
605
+ "markdown": false,
606
+ "metadata": {
607
+ "metric": "truncated",
608
+ "run_group": "OpenbookQA"
609
+ }
610
+ },
611
+ {
612
+ "value": "OpenbookQA - # prompt tokens",
613
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\n# prompt tokens: Number of tokens in the prompt.",
614
+ "markdown": false,
615
+ "metadata": {
616
+ "metric": "# prompt tokens",
617
+ "run_group": "OpenbookQA"
618
+ }
619
+ },
620
+ {
621
+ "value": "OpenbookQA - # output tokens",
622
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\n# output tokens: Actual number of output tokens.",
623
+ "markdown": false,
624
+ "metadata": {
625
+ "metric": "# output tokens",
626
+ "run_group": "OpenbookQA"
627
+ }
628
+ },
629
+ {
630
+ "value": "MMLU - # eval",
631
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
632
+ "markdown": false,
633
+ "metadata": {
634
+ "metric": "# eval",
635
+ "run_group": "MMLU"
636
+ }
637
+ },
638
+ {
639
+ "value": "MMLU - # train",
640
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
641
+ "markdown": false,
642
+ "metadata": {
643
+ "metric": "# train",
644
+ "run_group": "MMLU"
645
+ }
646
+ },
647
+ {
648
+ "value": "MMLU - truncated",
649
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
650
+ "markdown": false,
651
+ "metadata": {
652
+ "metric": "truncated",
653
+ "run_group": "MMLU"
654
+ }
655
+ },
656
+ {
657
+ "value": "MMLU - # prompt tokens",
658
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
659
+ "markdown": false,
660
+ "metadata": {
661
+ "metric": "# prompt tokens",
662
+ "run_group": "MMLU"
663
+ }
664
+ },
665
+ {
666
+ "value": "MMLU - # output tokens",
667
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
668
+ "markdown": false,
669
+ "metadata": {
670
+ "metric": "# output tokens",
671
+ "run_group": "MMLU"
672
+ }
673
+ },
674
+ {
675
+ "value": "MATH - # eval",
676
+ "description": "The MATH benchmark for measuring mathematical problem solving on competition math problems with chain-of-thought style reasoning [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2103.03874.pdf).\n\n# eval: Number of evaluation instances.",
677
+ "markdown": false,
678
+ "metadata": {
679
+ "metric": "# eval",
680
+ "run_group": "MATH"
681
+ }
682
+ },
683
+ {
684
+ "value": "MATH - # train",
685
+ "description": "The MATH benchmark for measuring mathematical problem solving on competition math problems with chain-of-thought style reasoning [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2103.03874.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
686
+ "markdown": false,
687
+ "metadata": {
688
+ "metric": "# train",
689
+ "run_group": "MATH"
690
+ }
691
+ },
692
+ {
693
+ "value": "MATH - truncated",
694
+ "description": "The MATH benchmark for measuring mathematical problem solving on competition math problems with chain-of-thought style reasoning [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2103.03874.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
695
+ "markdown": false,
696
+ "metadata": {
697
+ "metric": "truncated",
698
+ "run_group": "MATH"
699
+ }
700
+ },
701
+ {
702
+ "value": "MATH - # prompt tokens",
703
+ "description": "The MATH benchmark for measuring mathematical problem solving on competition math problems with chain-of-thought style reasoning [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2103.03874.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
704
+ "markdown": false,
705
+ "metadata": {
706
+ "metric": "# prompt tokens",
707
+ "run_group": "MATH"
708
+ }
709
+ },
710
+ {
711
+ "value": "MATH - # output tokens",
712
+ "description": "The MATH benchmark for measuring mathematical problem solving on competition math problems with chain-of-thought style reasoning [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2103.03874.pdf).\n\n# output tokens: Actual number of output tokens.",
713
+ "markdown": false,
714
+ "metadata": {
715
+ "metric": "# output tokens",
716
+ "run_group": "MATH"
717
+ }
718
+ },
719
+ {
720
+ "value": "GSM8K - # eval",
721
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\n# eval: Number of evaluation instances.",
722
+ "markdown": false,
723
+ "metadata": {
724
+ "metric": "# eval",
725
+ "run_group": "GSM8K"
726
+ }
727
+ },
728
+ {
729
+ "value": "GSM8K - # train",
730
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
731
+ "markdown": false,
732
+ "metadata": {
733
+ "metric": "# train",
734
+ "run_group": "GSM8K"
735
+ }
736
+ },
737
+ {
738
+ "value": "GSM8K - truncated",
739
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
740
+ "markdown": false,
741
+ "metadata": {
742
+ "metric": "truncated",
743
+ "run_group": "GSM8K"
744
+ }
745
+ },
746
+ {
747
+ "value": "GSM8K - # prompt tokens",
748
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
749
+ "markdown": false,
750
+ "metadata": {
751
+ "metric": "# prompt tokens",
752
+ "run_group": "GSM8K"
753
+ }
754
+ },
755
+ {
756
+ "value": "GSM8K - # output tokens",
757
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\n# output tokens: Actual number of output tokens.",
758
+ "markdown": false,
759
+ "metadata": {
760
+ "metric": "# output tokens",
761
+ "run_group": "GSM8K"
762
+ }
763
+ },
764
+ {
765
+ "value": "LegalBench - # eval",
766
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# eval: Number of evaluation instances.",
767
+ "markdown": false,
768
+ "metadata": {
769
+ "metric": "# eval",
770
+ "run_group": "LegalBench"
771
+ }
772
+ },
773
+ {
774
+ "value": "LegalBench - # train",
775
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
776
+ "markdown": false,
777
+ "metadata": {
778
+ "metric": "# train",
779
+ "run_group": "LegalBench"
780
+ }
781
+ },
782
+ {
783
+ "value": "LegalBench - truncated",
784
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
785
+ "markdown": false,
786
+ "metadata": {
787
+ "metric": "truncated",
788
+ "run_group": "LegalBench"
789
+ }
790
+ },
791
+ {
792
+ "value": "LegalBench - # prompt tokens",
793
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
794
+ "markdown": false,
795
+ "metadata": {
796
+ "metric": "# prompt tokens",
797
+ "run_group": "LegalBench"
798
+ }
799
+ },
800
+ {
801
+ "value": "LegalBench - # output tokens",
802
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# output tokens: Actual number of output tokens.",
803
+ "markdown": false,
804
+ "metadata": {
805
+ "metric": "# output tokens",
806
+ "run_group": "LegalBench"
807
+ }
808
+ },
809
+ {
810
+ "value": "MedQA - # eval",
811
+ "description": "MedQA is an open domain question answering dataset composed of questions from professional medical board exams ([Jin et al. 2020](https://arxiv.org/pdf/2009.13081.pdf)).\n\n# eval: Number of evaluation instances.",
812
+ "markdown": false,
813
+ "metadata": {
814
+ "metric": "# eval",
815
+ "run_group": "MedQA"
816
+ }
817
+ },
818
+ {
819
+ "value": "MedQA - # train",
820
+ "description": "MedQA is an open domain question answering dataset composed of questions from professional medical board exams ([Jin et al. 2020](https://arxiv.org/pdf/2009.13081.pdf)).\n\n# train: Number of training instances (e.g., in-context examples).",
821
+ "markdown": false,
822
+ "metadata": {
823
+ "metric": "# train",
824
+ "run_group": "MedQA"
825
+ }
826
+ },
827
+ {
828
+ "value": "MedQA - truncated",
829
+ "description": "MedQA is an open domain question answering dataset composed of questions from professional medical board exams ([Jin et al. 2020](https://arxiv.org/pdf/2009.13081.pdf)).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
830
+ "markdown": false,
831
+ "metadata": {
832
+ "metric": "truncated",
833
+ "run_group": "MedQA"
834
+ }
835
+ },
836
+ {
837
+ "value": "MedQA - # prompt tokens",
838
+ "description": "MedQA is an open domain question answering dataset composed of questions from professional medical board exams ([Jin et al. 2020](https://arxiv.org/pdf/2009.13081.pdf)).\n\n# prompt tokens: Number of tokens in the prompt.",
839
+ "markdown": false,
840
+ "metadata": {
841
+ "metric": "# prompt tokens",
842
+ "run_group": "MedQA"
843
+ }
844
+ },
845
+ {
846
+ "value": "MedQA - # output tokens",
847
+ "description": "MedQA is an open domain question answering dataset composed of questions from professional medical board exams ([Jin et al. 2020](https://arxiv.org/pdf/2009.13081.pdf)).\n\n# output tokens: Actual number of output tokens.",
848
+ "markdown": false,
849
+ "metadata": {
850
+ "metric": "# output tokens",
851
+ "run_group": "MedQA"
852
+ }
853
+ },
854
+ {
855
+ "value": "WMT 2014 - # eval",
856
+ "description": "WMT 2014 is a collection of machine translation datasets [(website)](https://www.statmt.org/wmt14/index.html).\n\n# eval: Number of evaluation instances.",
857
+ "markdown": false,
858
+ "metadata": {
859
+ "metric": "# eval",
860
+ "run_group": "WMT 2014"
861
+ }
862
+ },
863
+ {
864
+ "value": "WMT 2014 - # train",
865
+ "description": "WMT 2014 is a collection of machine translation datasets [(website)](https://www.statmt.org/wmt14/index.html).\n\n# train: Number of training instances (e.g., in-context examples).",
866
+ "markdown": false,
867
+ "metadata": {
868
+ "metric": "# train",
869
+ "run_group": "WMT 2014"
870
+ }
871
+ },
872
+ {
873
+ "value": "WMT 2014 - truncated",
874
+ "description": "WMT 2014 is a collection of machine translation datasets [(website)](https://www.statmt.org/wmt14/index.html).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
875
+ "markdown": false,
876
+ "metadata": {
877
+ "metric": "truncated",
878
+ "run_group": "WMT 2014"
879
+ }
880
+ },
881
+ {
882
+ "value": "WMT 2014 - # prompt tokens",
883
+ "description": "WMT 2014 is a collection of machine translation datasets [(website)](https://www.statmt.org/wmt14/index.html).\n\n# prompt tokens: Number of tokens in the prompt.",
884
+ "markdown": false,
885
+ "metadata": {
886
+ "metric": "# prompt tokens",
887
+ "run_group": "WMT 2014"
888
+ }
889
+ },
890
+ {
891
+ "value": "WMT 2014 - # output tokens",
892
+ "description": "WMT 2014 is a collection of machine translation datasets [(website)](https://www.statmt.org/wmt14/index.html).\n\n# output tokens: Actual number of output tokens.",
893
+ "markdown": false,
894
+ "metadata": {
895
+ "metric": "# output tokens",
896
+ "run_group": "WMT 2014"
897
+ }
898
+ }
899
+ ],
900
+ "rows": [
901
+ [
902
+ {
903
+ "value": "EleutherAI/pythia-2.8b",
904
+ "description": "",
905
+ "markdown": false
906
+ },
907
+ {
908
+ "markdown": false
909
+ },
910
+ {
911
+ "description": "No matching runs",
912
+ "markdown": false
913
+ },
914
+ {
915
+ "description": "No matching runs",
916
+ "markdown": false
917
+ },
918
+ {
919
+ "description": "No matching runs",
920
+ "markdown": false
921
+ },
922
+ {
923
+ "description": "No matching runs",
924
+ "markdown": false
925
+ },
926
+ {
927
+ "description": "No matching runs",
928
+ "markdown": false
929
+ },
930
+ {
931
+ "description": "No matching runs",
932
+ "markdown": false
933
+ },
934
+ {
935
+ "description": "No matching runs",
936
+ "markdown": false
937
+ },
938
+ {
939
+ "description": "No matching runs",
940
+ "markdown": false
941
+ },
942
+ {
943
+ "description": "No matching runs",
944
+ "markdown": false
945
+ },
946
+ {
947
+ "description": "No matching runs",
948
+ "markdown": false
949
+ },
950
+ {
951
+ "description": "No matching runs",
952
+ "markdown": false
953
+ },
954
+ {
955
+ "description": "No matching runs",
956
+ "markdown": false
957
+ },
958
+ {
959
+ "description": "No matching runs",
960
+ "markdown": false
961
+ },
962
+ {
963
+ "description": "No matching runs",
964
+ "markdown": false
965
+ },
966
+ {
967
+ "description": "No matching runs",
968
+ "markdown": false
969
+ },
970
+ {
971
+ "value": 500.0,
972
+ "description": "min=500, mean=500, max=500, sum=500 (1)",
973
+ "style": {},
974
+ "markdown": false,
975
+ "run_spec_names": [
976
+ "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
977
+ ]
978
+ },
979
+ {
980
+ "value": 5.0,
981
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
982
+ "style": {},
983
+ "markdown": false,
984
+ "run_spec_names": [
985
+ "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
986
+ ]
987
+ },
988
+ {
989
+ "value": 0.0,
990
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
991
+ "style": {},
992
+ "markdown": false,
993
+ "run_spec_names": [
994
+ "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
995
+ ]
996
+ },
997
+ {
998
+ "value": 251.556,
999
+ "description": "min=251.556, mean=251.556, max=251.556, sum=251.556 (1)",
1000
+ "style": {},
1001
+ "markdown": false,
1002
+ "run_spec_names": [
1003
+ "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
1004
+ ]
1005
+ },
1006
+ {
1007
+ "value": 1.0,
1008
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
1009
+ "style": {},
1010
+ "markdown": false,
1011
+ "run_spec_names": [
1012
+ "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
1013
+ ]
1014
+ },
1015
+ {
1016
+ "value": 102.8,
1017
+ "description": "min=100, mean=102.8, max=114, sum=514 (5)",
1018
+ "style": {},
1019
+ "markdown": false,
1020
+ "run_spec_names": [
1021
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1022
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1023
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1024
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1025
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
1026
+ ]
1027
+ },
1028
+ {
1029
+ "value": 5.0,
1030
+ "description": "min=5, mean=5, max=5, sum=25 (5)",
1031
+ "style": {},
1032
+ "markdown": false,
1033
+ "run_spec_names": [
1034
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1035
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1036
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1037
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1038
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
1039
+ ]
1040
+ },
1041
+ {
1042
+ "value": 0.0,
1043
+ "description": "min=0, mean=0, max=0, sum=0 (5)",
1044
+ "style": {},
1045
+ "markdown": false,
1046
+ "run_spec_names": [
1047
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1048
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1049
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1050
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1051
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
1052
+ ]
1053
+ },
1054
+ {
1055
+ "value": 467.935649122807,
1056
+ "description": "min=358.76, mean=467.936, max=612.798, sum=2339.678 (5)",
1057
+ "style": {},
1058
+ "markdown": false,
1059
+ "run_spec_names": [
1060
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1061
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1062
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1063
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1064
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
1065
+ ]
1066
+ },
1067
+ {
1068
+ "value": 1.0,
1069
+ "description": "min=1, mean=1, max=1, sum=5 (5)",
1070
+ "style": {},
1071
+ "markdown": false,
1072
+ "run_spec_names": [
1073
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1074
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1075
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1076
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
1077
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
1078
+ ]
1079
+ },
1080
+ {
1081
+ "description": "No matching runs",
1082
+ "markdown": false
1083
+ },
1084
+ {
1085
+ "description": "No matching runs",
1086
+ "markdown": false
1087
+ },
1088
+ {
1089
+ "description": "No matching runs",
1090
+ "markdown": false
1091
+ },
1092
+ {
1093
+ "description": "No matching runs",
1094
+ "markdown": false
1095
+ },
1096
+ {
1097
+ "description": "No matching runs",
1098
+ "markdown": false
1099
+ },
1100
+ {
1101
+ "value": 1000.0,
1102
+ "description": "min=1000, mean=1000, max=1000, sum=1000 (1)",
1103
+ "style": {},
1104
+ "markdown": false,
1105
+ "run_spec_names": [
1106
+ "gsm:model=EleutherAI_pythia-2.8b"
1107
+ ]
1108
+ },
1109
+ {
1110
+ "value": 5.0,
1111
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
1112
+ "style": {},
1113
+ "markdown": false,
1114
+ "run_spec_names": [
1115
+ "gsm:model=EleutherAI_pythia-2.8b"
1116
+ ]
1117
+ },
1118
+ {
1119
+ "value": 0.0,
1120
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
1121
+ "style": {},
1122
+ "markdown": false,
1123
+ "run_spec_names": [
1124
+ "gsm:model=EleutherAI_pythia-2.8b"
1125
+ ]
1126
+ },
1127
+ {
1128
+ "value": 939.582,
1129
+ "description": "min=939.582, mean=939.582, max=939.582, sum=939.582 (1)",
1130
+ "style": {},
1131
+ "markdown": false,
1132
+ "run_spec_names": [
1133
+ "gsm:model=EleutherAI_pythia-2.8b"
1134
+ ]
1135
+ },
1136
+ {
1137
+ "value": 168.459,
1138
+ "description": "min=168.459, mean=168.459, max=168.459, sum=168.459 (1)",
1139
+ "style": {},
1140
+ "markdown": false,
1141
+ "run_spec_names": [
1142
+ "gsm:model=EleutherAI_pythia-2.8b"
1143
+ ]
1144
+ },
1145
+ {
1146
+ "value": 409.4,
1147
+ "description": "min=95, mean=409.4, max=1000, sum=2047 (5)",
1148
+ "style": {},
1149
+ "markdown": false,
1150
+ "run_spec_names": [
1151
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
1152
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
1153
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
1154
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
1155
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
1156
+ ]
1157
+ },
1158
+ {
1159
+ "value": 3.8604081632653062,
1160
+ "description": "min=0.302, mean=3.86, max=5, sum=19.302 (5)",
1161
+ "style": {},
1162
+ "markdown": false,
1163
+ "run_spec_names": [
1164
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
1165
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
1166
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
1167
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
1168
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
1169
+ ]
1170
+ },
1171
+ {
1172
+ "value": 0.002857142857142857,
1173
+ "description": "min=0, mean=0.003, max=0.014, sum=0.014 (5)",
1174
+ "style": {},
1175
+ "markdown": false,
1176
+ "run_spec_names": [
1177
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
1178
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
1179
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
1180
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
1181
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
1182
+ ]
1183
+ },
1184
+ {
1185
+ "value": 560.6440716343213,
1186
+ "description": "min=206.779, mean=560.644, max=1497.455, sum=2803.22 (5)",
1187
+ "style": {},
1188
+ "markdown": false,
1189
+ "run_spec_names": [
1190
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
1191
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
1192
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
1193
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
1194
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
1195
+ ]
1196
+ },
1197
+ {
1198
+ "value": 1.6417916921537006,
1199
+ "description": "min=1, mean=1.642, max=4.124, sum=8.209 (5)",
1200
+ "style": {},
1201
+ "markdown": false,
1202
+ "run_spec_names": [
1203
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
1204
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
1205
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
1206
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
1207
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
1208
+ ]
1209
+ },
1210
+ {
1211
+ "description": "No matching runs",
1212
+ "markdown": false
1213
+ },
1214
+ {
1215
+ "description": "No matching runs",
1216
+ "markdown": false
1217
+ },
1218
+ {
1219
+ "description": "No matching runs",
1220
+ "markdown": false
1221
+ },
1222
+ {
1223
+ "description": "No matching runs",
1224
+ "markdown": false
1225
+ },
1226
+ {
1227
+ "description": "No matching runs",
1228
+ "markdown": false
1229
+ },
1230
+ {
1231
+ "description": "No matching runs",
1232
+ "markdown": false
1233
+ },
1234
+ {
1235
+ "description": "No matching runs",
1236
+ "markdown": false
1237
+ },
1238
+ {
1239
+ "description": "No matching runs",
1240
+ "markdown": false
1241
+ },
1242
+ {
1243
+ "description": "No matching runs",
1244
+ "markdown": false
1245
+ },
1246
+ {
1247
+ "description": "No matching runs",
1248
+ "markdown": false
1249
+ }
1250
+ ]
1251
+ ],
1252
+ "links": [
1253
+ {
1254
+ "text": "LaTeX",
1255
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/core_scenarios_general_information.tex"
1256
+ },
1257
+ {
1258
+ "text": "JSON",
1259
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/core_scenarios_general_information.json"
1260
+ }
1261
+ ],
1262
+ "name": "general_information"
1263
+ }
1264
+ ]
lite_pythia-2.8b-step5000/groups/gsm.json ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "title": "",
4
+ "header": [
5
+ {
6
+ "value": "Model",
7
+ "markdown": false,
8
+ "metadata": {}
9
+ },
10
+ {
11
+ "value": "EM",
12
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\nExact match (final number): Fraction of instances that the predicted output matches a correct reference exactly, ignoring text preceding the specified indicator.",
13
+ "markdown": false,
14
+ "lower_is_better": false,
15
+ "metadata": {
16
+ "metric": "EM",
17
+ "run_group": "GSM8K"
18
+ }
19
+ },
20
+ {
21
+ "value": "Observed inference time (s)",
22
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
23
+ "markdown": false,
24
+ "lower_is_better": true,
25
+ "metadata": {
26
+ "metric": "Observed inference time (s)",
27
+ "run_group": "GSM8K"
28
+ }
29
+ },
30
+ {
31
+ "value": "# eval",
32
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\n# eval: Number of evaluation instances.",
33
+ "markdown": false,
34
+ "metadata": {
35
+ "metric": "# eval",
36
+ "run_group": "GSM8K"
37
+ }
38
+ },
39
+ {
40
+ "value": "# train",
41
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
42
+ "markdown": false,
43
+ "metadata": {
44
+ "metric": "# train",
45
+ "run_group": "GSM8K"
46
+ }
47
+ },
48
+ {
49
+ "value": "truncated",
50
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
51
+ "markdown": false,
52
+ "metadata": {
53
+ "metric": "truncated",
54
+ "run_group": "GSM8K"
55
+ }
56
+ },
57
+ {
58
+ "value": "# prompt tokens",
59
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
60
+ "markdown": false,
61
+ "metadata": {
62
+ "metric": "# prompt tokens",
63
+ "run_group": "GSM8K"
64
+ }
65
+ },
66
+ {
67
+ "value": "# output tokens",
68
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\n# output tokens: Actual number of output tokens.",
69
+ "markdown": false,
70
+ "metadata": {
71
+ "metric": "# output tokens",
72
+ "run_group": "GSM8K"
73
+ }
74
+ }
75
+ ],
76
+ "rows": [
77
+ [
78
+ {
79
+ "value": "EleutherAI/pythia-2.8b",
80
+ "description": "",
81
+ "href": "?group=gsm&subgroup=&runSpecs=%5B%22gsm%3Amodel%3DEleutherAI_pythia-2.8b%22%5D",
82
+ "markdown": false,
83
+ "run_spec_names": [
84
+ "gsm:model=EleutherAI_pythia-2.8b"
85
+ ]
86
+ },
87
+ {
88
+ "value": 0.01,
89
+ "description": "min=0.01, mean=0.01, max=0.01, sum=0.01 (1)",
90
+ "style": {
91
+ "font-weight": "bold"
92
+ },
93
+ "markdown": false
94
+ },
95
+ {
96
+ "value": 2.5314217054843904,
97
+ "description": "min=2.531, mean=2.531, max=2.531, sum=2.531 (1)",
98
+ "style": {
99
+ "font-weight": "bold"
100
+ },
101
+ "markdown": false
102
+ },
103
+ {
104
+ "value": 1000.0,
105
+ "description": "min=1000, mean=1000, max=1000, sum=1000 (1)",
106
+ "style": {},
107
+ "markdown": false
108
+ },
109
+ {
110
+ "value": 5.0,
111
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
112
+ "style": {},
113
+ "markdown": false
114
+ },
115
+ {
116
+ "value": 0.0,
117
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
118
+ "style": {},
119
+ "markdown": false
120
+ },
121
+ {
122
+ "value": 939.582,
123
+ "description": "min=939.582, mean=939.582, max=939.582, sum=939.582 (1)",
124
+ "style": {},
125
+ "markdown": false
126
+ },
127
+ {
128
+ "value": 168.459,
129
+ "description": "min=168.459, mean=168.459, max=168.459, sum=168.459 (1)",
130
+ "style": {},
131
+ "markdown": false
132
+ }
133
+ ]
134
+ ],
135
+ "links": [
136
+ {
137
+ "text": "LaTeX",
138
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/gsm_gsm_.tex"
139
+ },
140
+ {
141
+ "text": "JSON",
142
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/gsm_gsm_.json"
143
+ }
144
+ ],
145
+ "name": "gsm_"
146
+ }
147
+ ]
lite_pythia-2.8b-step5000/groups/json/core_scenarios_accuracy.json ADDED
@@ -0,0 +1,216 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "Accuracy",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "Mean win rate",
11
+ "description": "How many models this model outperforms on average (over columns).",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {}
15
+ },
16
+ {
17
+ "value": "NarrativeQA - F1",
18
+ "description": "The NarrativeQA benchmark for reading comprehension over narratives [(Ko\u010disk\u00fd et al., 2017)](https://aclanthology.org/Q18-1023/).\n\nF1: Average F1 score in terms of word overlap between the model output and correct reference.",
19
+ "markdown": false,
20
+ "lower_is_better": false,
21
+ "metadata": {
22
+ "metric": "F1",
23
+ "run_group": "NarrativeQA"
24
+ }
25
+ },
26
+ {
27
+ "value": "NaturalQuestions (open-book) - F1",
28
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input includes the Wikipedia page with the answer.\n\nF1: Average F1 score in terms of word overlap between the model output and correct reference.",
29
+ "markdown": false,
30
+ "lower_is_better": false,
31
+ "metadata": {
32
+ "metric": "F1",
33
+ "run_group": "NaturalQuestions (open-book)"
34
+ }
35
+ },
36
+ {
37
+ "value": "NaturalQuestions (closed-book) - F1",
38
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input does not include the Wikipedia page with the answer.\n\nF1: Average F1 score in terms of word overlap between the model output and correct reference.",
39
+ "markdown": false,
40
+ "lower_is_better": false,
41
+ "metadata": {
42
+ "metric": "F1",
43
+ "run_group": "NaturalQuestions (closed-book)"
44
+ }
45
+ },
46
+ {
47
+ "value": "OpenbookQA - EM",
48
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
49
+ "markdown": false,
50
+ "lower_is_better": false,
51
+ "metadata": {
52
+ "metric": "EM",
53
+ "run_group": "OpenbookQA"
54
+ }
55
+ },
56
+ {
57
+ "value": "MMLU - EM",
58
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
59
+ "markdown": false,
60
+ "lower_is_better": false,
61
+ "metadata": {
62
+ "metric": "EM",
63
+ "run_group": "MMLU"
64
+ }
65
+ },
66
+ {
67
+ "value": "MATH - Equivalent (CoT)",
68
+ "description": "The MATH benchmark for measuring mathematical problem solving on competition math problems with chain-of-thought style reasoning [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2103.03874.pdf).\n\nEquivalent (CoT): Fraction of model outputs that are mathematically equivalent to the correct reference when using chain-of-thought prompting.",
69
+ "markdown": false,
70
+ "lower_is_better": false,
71
+ "metadata": {
72
+ "metric": "Equivalent (CoT)",
73
+ "run_group": "MATH"
74
+ }
75
+ },
76
+ {
77
+ "value": "GSM8K - EM",
78
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\nExact match (final number): Fraction of instances that the predicted output matches a correct reference exactly, ignoring text preceding the specified indicator.",
79
+ "markdown": false,
80
+ "lower_is_better": false,
81
+ "metadata": {
82
+ "metric": "EM",
83
+ "run_group": "GSM8K"
84
+ }
85
+ },
86
+ {
87
+ "value": "LegalBench - EM",
88
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.",
89
+ "markdown": false,
90
+ "lower_is_better": false,
91
+ "metadata": {
92
+ "metric": "EM",
93
+ "run_group": "LegalBench"
94
+ }
95
+ },
96
+ {
97
+ "value": "MedQA - EM",
98
+ "description": "MedQA is an open domain question answering dataset composed of questions from professional medical board exams ([Jin et al. 2020](https://arxiv.org/pdf/2009.13081.pdf)).\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.",
99
+ "markdown": false,
100
+ "lower_is_better": false,
101
+ "metadata": {
102
+ "metric": "EM",
103
+ "run_group": "MedQA"
104
+ }
105
+ },
106
+ {
107
+ "value": "WMT 2014 - BLEU-4",
108
+ "description": "WMT 2014 is a collection of machine translation datasets [(website)](https://www.statmt.org/wmt14/index.html).\n\nBLEU-4: Average BLEU score [(Papineni et al., 2002)](https://aclanthology.org/P02-1040/) based on 4-gram overlap.",
109
+ "markdown": false,
110
+ "lower_is_better": false,
111
+ "metadata": {
112
+ "metric": "BLEU-4",
113
+ "run_group": "WMT 2014"
114
+ }
115
+ }
116
+ ],
117
+ "rows": [
118
+ [
119
+ {
120
+ "value": "EleutherAI/pythia-2.8b",
121
+ "description": "",
122
+ "markdown": false
123
+ },
124
+ {
125
+ "markdown": false
126
+ },
127
+ {
128
+ "description": "No matching runs",
129
+ "markdown": false
130
+ },
131
+ {
132
+ "description": "No matching runs",
133
+ "markdown": false
134
+ },
135
+ {
136
+ "description": "No matching runs",
137
+ "markdown": false
138
+ },
139
+ {
140
+ "value": 0.22,
141
+ "description": "min=0.22, mean=0.22, max=0.22, sum=0.22 (1)",
142
+ "style": {
143
+ "font-weight": "bold"
144
+ },
145
+ "markdown": false,
146
+ "run_spec_names": [
147
+ "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
148
+ ]
149
+ },
150
+ {
151
+ "value": 0.27487719298245616,
152
+ "description": "min=0.25, mean=0.275, max=0.32, sum=1.374 (5)",
153
+ "style": {
154
+ "font-weight": "bold"
155
+ },
156
+ "markdown": false,
157
+ "run_spec_names": [
158
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
159
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
160
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
161
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
162
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
163
+ ]
164
+ },
165
+ {
166
+ "description": "No matching runs",
167
+ "markdown": false
168
+ },
169
+ {
170
+ "value": 0.01,
171
+ "description": "min=0.01, mean=0.01, max=0.01, sum=0.01 (1)",
172
+ "style": {
173
+ "font-weight": "bold"
174
+ },
175
+ "markdown": false,
176
+ "run_spec_names": [
177
+ "gsm:model=EleutherAI_pythia-2.8b"
178
+ ]
179
+ },
180
+ {
181
+ "value": 0.3189986232611502,
182
+ "description": "min=0.144, mean=0.319, max=0.619, sum=1.595 (5)",
183
+ "style": {
184
+ "font-weight": "bold"
185
+ },
186
+ "markdown": false,
187
+ "run_spec_names": [
188
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
189
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
190
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
191
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
192
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
193
+ ]
194
+ },
195
+ {
196
+ "description": "No matching runs",
197
+ "markdown": false
198
+ },
199
+ {
200
+ "description": "No matching runs",
201
+ "markdown": false
202
+ }
203
+ ]
204
+ ],
205
+ "links": [
206
+ {
207
+ "text": "LaTeX",
208
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/core_scenarios_accuracy.tex"
209
+ },
210
+ {
211
+ "text": "JSON",
212
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/core_scenarios_accuracy.json"
213
+ }
214
+ ],
215
+ "name": "accuracy"
216
+ }
lite_pythia-2.8b-step5000/groups/json/core_scenarios_efficiency.json ADDED
@@ -0,0 +1,216 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "Efficiency",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "Mean win rate",
11
+ "description": "How many models this model outperforms on average (over columns).",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {}
15
+ },
16
+ {
17
+ "value": "NarrativeQA - Observed inference time (s)",
18
+ "description": "The NarrativeQA benchmark for reading comprehension over narratives [(Ko\u010disk\u00fd et al., 2017)](https://aclanthology.org/Q18-1023/).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
19
+ "markdown": false,
20
+ "lower_is_better": true,
21
+ "metadata": {
22
+ "metric": "Observed inference time (s)",
23
+ "run_group": "NarrativeQA"
24
+ }
25
+ },
26
+ {
27
+ "value": "NaturalQuestions (open-book) - Observed inference time (s)",
28
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input includes the Wikipedia page with the answer.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
29
+ "markdown": false,
30
+ "lower_is_better": true,
31
+ "metadata": {
32
+ "metric": "Observed inference time (s)",
33
+ "run_group": "NaturalQuestions (open-book)"
34
+ }
35
+ },
36
+ {
37
+ "value": "NaturalQuestions (closed-book) - Observed inference time (s)",
38
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input does not include the Wikipedia page with the answer.\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
39
+ "markdown": false,
40
+ "lower_is_better": true,
41
+ "metadata": {
42
+ "metric": "Observed inference time (s)",
43
+ "run_group": "NaturalQuestions (closed-book)"
44
+ }
45
+ },
46
+ {
47
+ "value": "OpenbookQA - Observed inference time (s)",
48
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
49
+ "markdown": false,
50
+ "lower_is_better": true,
51
+ "metadata": {
52
+ "metric": "Observed inference time (s)",
53
+ "run_group": "OpenbookQA"
54
+ }
55
+ },
56
+ {
57
+ "value": "MMLU - Observed inference time (s)",
58
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
59
+ "markdown": false,
60
+ "lower_is_better": true,
61
+ "metadata": {
62
+ "metric": "Observed inference time (s)",
63
+ "run_group": "MMLU"
64
+ }
65
+ },
66
+ {
67
+ "value": "MATH - Observed inference time (s)",
68
+ "description": "The MATH benchmark for measuring mathematical problem solving on competition math problems with chain-of-thought style reasoning [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2103.03874.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
69
+ "markdown": false,
70
+ "lower_is_better": true,
71
+ "metadata": {
72
+ "metric": "Observed inference time (s)",
73
+ "run_group": "MATH"
74
+ }
75
+ },
76
+ {
77
+ "value": "GSM8K - Observed inference time (s)",
78
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
79
+ "markdown": false,
80
+ "lower_is_better": true,
81
+ "metadata": {
82
+ "metric": "Observed inference time (s)",
83
+ "run_group": "GSM8K"
84
+ }
85
+ },
86
+ {
87
+ "value": "LegalBench - Observed inference time (s)",
88
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
89
+ "markdown": false,
90
+ "lower_is_better": true,
91
+ "metadata": {
92
+ "metric": "Observed inference time (s)",
93
+ "run_group": "LegalBench"
94
+ }
95
+ },
96
+ {
97
+ "value": "MedQA - Observed inference time (s)",
98
+ "description": "MedQA is an open domain question answering dataset composed of questions from professional medical board exams ([Jin et al. 2020](https://arxiv.org/pdf/2009.13081.pdf)).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
99
+ "markdown": false,
100
+ "lower_is_better": true,
101
+ "metadata": {
102
+ "metric": "Observed inference time (s)",
103
+ "run_group": "MedQA"
104
+ }
105
+ },
106
+ {
107
+ "value": "WMT 2014 - Observed inference time (s)",
108
+ "description": "WMT 2014 is a collection of machine translation datasets [(website)](https://www.statmt.org/wmt14/index.html).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
109
+ "markdown": false,
110
+ "lower_is_better": true,
111
+ "metadata": {
112
+ "metric": "Observed inference time (s)",
113
+ "run_group": "WMT 2014"
114
+ }
115
+ }
116
+ ],
117
+ "rows": [
118
+ [
119
+ {
120
+ "value": "EleutherAI/pythia-2.8b",
121
+ "description": "",
122
+ "markdown": false
123
+ },
124
+ {
125
+ "markdown": false
126
+ },
127
+ {
128
+ "description": "No matching runs",
129
+ "markdown": false
130
+ },
131
+ {
132
+ "description": "No matching runs",
133
+ "markdown": false
134
+ },
135
+ {
136
+ "description": "No matching runs",
137
+ "markdown": false
138
+ },
139
+ {
140
+ "value": 0.12883714818954467,
141
+ "description": "min=0.129, mean=0.129, max=0.129, sum=0.129 (1)",
142
+ "style": {
143
+ "font-weight": "bold"
144
+ },
145
+ "markdown": false,
146
+ "run_spec_names": [
147
+ "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
148
+ ]
149
+ },
150
+ {
151
+ "value": 0.22061018314696193,
152
+ "description": "min=0.173, mean=0.221, max=0.282, sum=1.103 (5)",
153
+ "style": {
154
+ "font-weight": "bold"
155
+ },
156
+ "markdown": false,
157
+ "run_spec_names": [
158
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
159
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
160
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
161
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
162
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
163
+ ]
164
+ },
165
+ {
166
+ "description": "No matching runs",
167
+ "markdown": false
168
+ },
169
+ {
170
+ "value": 2.5314217054843904,
171
+ "description": "min=2.531, mean=2.531, max=2.531, sum=2.531 (1)",
172
+ "style": {
173
+ "font-weight": "bold"
174
+ },
175
+ "markdown": false,
176
+ "run_spec_names": [
177
+ "gsm:model=EleutherAI_pythia-2.8b"
178
+ ]
179
+ },
180
+ {
181
+ "value": 0.3094366045219278,
182
+ "description": "min=0.141, mean=0.309, max=0.795, sum=1.547 (5)",
183
+ "style": {
184
+ "font-weight": "bold"
185
+ },
186
+ "markdown": false,
187
+ "run_spec_names": [
188
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
189
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
190
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
191
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
192
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
193
+ ]
194
+ },
195
+ {
196
+ "description": "No matching runs",
197
+ "markdown": false
198
+ },
199
+ {
200
+ "description": "No matching runs",
201
+ "markdown": false
202
+ }
203
+ ]
204
+ ],
205
+ "links": [
206
+ {
207
+ "text": "LaTeX",
208
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/core_scenarios_efficiency.tex"
209
+ },
210
+ {
211
+ "text": "JSON",
212
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/core_scenarios_efficiency.json"
213
+ }
214
+ ],
215
+ "name": "efficiency"
216
+ }
lite_pythia-2.8b-step5000/groups/json/core_scenarios_general_information.json ADDED
@@ -0,0 +1,830 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "General information",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "Mean win rate",
11
+ "description": "How many models this model outperforms on average (over columns).",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {}
15
+ },
16
+ {
17
+ "value": "NarrativeQA - # eval",
18
+ "description": "The NarrativeQA benchmark for reading comprehension over narratives [(Ko\u010disk\u00fd et al., 2017)](https://aclanthology.org/Q18-1023/).\n\n# eval: Number of evaluation instances.",
19
+ "markdown": false,
20
+ "metadata": {
21
+ "metric": "# eval",
22
+ "run_group": "NarrativeQA"
23
+ }
24
+ },
25
+ {
26
+ "value": "NarrativeQA - # train",
27
+ "description": "The NarrativeQA benchmark for reading comprehension over narratives [(Ko\u010disk\u00fd et al., 2017)](https://aclanthology.org/Q18-1023/).\n\n# train: Number of training instances (e.g., in-context examples).",
28
+ "markdown": false,
29
+ "metadata": {
30
+ "metric": "# train",
31
+ "run_group": "NarrativeQA"
32
+ }
33
+ },
34
+ {
35
+ "value": "NarrativeQA - truncated",
36
+ "description": "The NarrativeQA benchmark for reading comprehension over narratives [(Ko\u010disk\u00fd et al., 2017)](https://aclanthology.org/Q18-1023/).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
37
+ "markdown": false,
38
+ "metadata": {
39
+ "metric": "truncated",
40
+ "run_group": "NarrativeQA"
41
+ }
42
+ },
43
+ {
44
+ "value": "NarrativeQA - # prompt tokens",
45
+ "description": "The NarrativeQA benchmark for reading comprehension over narratives [(Ko\u010disk\u00fd et al., 2017)](https://aclanthology.org/Q18-1023/).\n\n# prompt tokens: Number of tokens in the prompt.",
46
+ "markdown": false,
47
+ "metadata": {
48
+ "metric": "# prompt tokens",
49
+ "run_group": "NarrativeQA"
50
+ }
51
+ },
52
+ {
53
+ "value": "NarrativeQA - # output tokens",
54
+ "description": "The NarrativeQA benchmark for reading comprehension over narratives [(Ko\u010disk\u00fd et al., 2017)](https://aclanthology.org/Q18-1023/).\n\n# output tokens: Actual number of output tokens.",
55
+ "markdown": false,
56
+ "metadata": {
57
+ "metric": "# output tokens",
58
+ "run_group": "NarrativeQA"
59
+ }
60
+ },
61
+ {
62
+ "value": "NaturalQuestions (open-book) - # eval",
63
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input includes the Wikipedia page with the answer.\n\n# eval: Number of evaluation instances.",
64
+ "markdown": false,
65
+ "metadata": {
66
+ "metric": "# eval",
67
+ "run_group": "NaturalQuestions (open-book)"
68
+ }
69
+ },
70
+ {
71
+ "value": "NaturalQuestions (open-book) - # train",
72
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input includes the Wikipedia page with the answer.\n\n# train: Number of training instances (e.g., in-context examples).",
73
+ "markdown": false,
74
+ "metadata": {
75
+ "metric": "# train",
76
+ "run_group": "NaturalQuestions (open-book)"
77
+ }
78
+ },
79
+ {
80
+ "value": "NaturalQuestions (open-book) - truncated",
81
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input includes the Wikipedia page with the answer.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
82
+ "markdown": false,
83
+ "metadata": {
84
+ "metric": "truncated",
85
+ "run_group": "NaturalQuestions (open-book)"
86
+ }
87
+ },
88
+ {
89
+ "value": "NaturalQuestions (open-book) - # prompt tokens",
90
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input includes the Wikipedia page with the answer.\n\n# prompt tokens: Number of tokens in the prompt.",
91
+ "markdown": false,
92
+ "metadata": {
93
+ "metric": "# prompt tokens",
94
+ "run_group": "NaturalQuestions (open-book)"
95
+ }
96
+ },
97
+ {
98
+ "value": "NaturalQuestions (open-book) - # output tokens",
99
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input includes the Wikipedia page with the answer.\n\n# output tokens: Actual number of output tokens.",
100
+ "markdown": false,
101
+ "metadata": {
102
+ "metric": "# output tokens",
103
+ "run_group": "NaturalQuestions (open-book)"
104
+ }
105
+ },
106
+ {
107
+ "value": "NaturalQuestions (closed-book) - # eval",
108
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input does not include the Wikipedia page with the answer.\n\n# eval: Number of evaluation instances.",
109
+ "markdown": false,
110
+ "metadata": {
111
+ "metric": "# eval",
112
+ "run_group": "NaturalQuestions (closed-book)"
113
+ }
114
+ },
115
+ {
116
+ "value": "NaturalQuestions (closed-book) - # train",
117
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input does not include the Wikipedia page with the answer.\n\n# train: Number of training instances (e.g., in-context examples).",
118
+ "markdown": false,
119
+ "metadata": {
120
+ "metric": "# train",
121
+ "run_group": "NaturalQuestions (closed-book)"
122
+ }
123
+ },
124
+ {
125
+ "value": "NaturalQuestions (closed-book) - truncated",
126
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input does not include the Wikipedia page with the answer.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
127
+ "markdown": false,
128
+ "metadata": {
129
+ "metric": "truncated",
130
+ "run_group": "NaturalQuestions (closed-book)"
131
+ }
132
+ },
133
+ {
134
+ "value": "NaturalQuestions (closed-book) - # prompt tokens",
135
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input does not include the Wikipedia page with the answer.\n\n# prompt tokens: Number of tokens in the prompt.",
136
+ "markdown": false,
137
+ "metadata": {
138
+ "metric": "# prompt tokens",
139
+ "run_group": "NaturalQuestions (closed-book)"
140
+ }
141
+ },
142
+ {
143
+ "value": "NaturalQuestions (closed-book) - # output tokens",
144
+ "description": "The NaturalQuestions [(Kwiatkowski et al., 2019)](https://aclanthology.org/Q19-1026/) benchmark for question answering based on naturally-occurring queries through Google Search. The input does not include the Wikipedia page with the answer.\n\n# output tokens: Actual number of output tokens.",
145
+ "markdown": false,
146
+ "metadata": {
147
+ "metric": "# output tokens",
148
+ "run_group": "NaturalQuestions (closed-book)"
149
+ }
150
+ },
151
+ {
152
+ "value": "OpenbookQA - # eval",
153
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\n# eval: Number of evaluation instances.",
154
+ "markdown": false,
155
+ "metadata": {
156
+ "metric": "# eval",
157
+ "run_group": "OpenbookQA"
158
+ }
159
+ },
160
+ {
161
+ "value": "OpenbookQA - # train",
162
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\n# train: Number of training instances (e.g., in-context examples).",
163
+ "markdown": false,
164
+ "metadata": {
165
+ "metric": "# train",
166
+ "run_group": "OpenbookQA"
167
+ }
168
+ },
169
+ {
170
+ "value": "OpenbookQA - truncated",
171
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
172
+ "markdown": false,
173
+ "metadata": {
174
+ "metric": "truncated",
175
+ "run_group": "OpenbookQA"
176
+ }
177
+ },
178
+ {
179
+ "value": "OpenbookQA - # prompt tokens",
180
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\n# prompt tokens: Number of tokens in the prompt.",
181
+ "markdown": false,
182
+ "metadata": {
183
+ "metric": "# prompt tokens",
184
+ "run_group": "OpenbookQA"
185
+ }
186
+ },
187
+ {
188
+ "value": "OpenbookQA - # output tokens",
189
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\n# output tokens: Actual number of output tokens.",
190
+ "markdown": false,
191
+ "metadata": {
192
+ "metric": "# output tokens",
193
+ "run_group": "OpenbookQA"
194
+ }
195
+ },
196
+ {
197
+ "value": "MMLU - # eval",
198
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
199
+ "markdown": false,
200
+ "metadata": {
201
+ "metric": "# eval",
202
+ "run_group": "MMLU"
203
+ }
204
+ },
205
+ {
206
+ "value": "MMLU - # train",
207
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
208
+ "markdown": false,
209
+ "metadata": {
210
+ "metric": "# train",
211
+ "run_group": "MMLU"
212
+ }
213
+ },
214
+ {
215
+ "value": "MMLU - truncated",
216
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
217
+ "markdown": false,
218
+ "metadata": {
219
+ "metric": "truncated",
220
+ "run_group": "MMLU"
221
+ }
222
+ },
223
+ {
224
+ "value": "MMLU - # prompt tokens",
225
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
226
+ "markdown": false,
227
+ "metadata": {
228
+ "metric": "# prompt tokens",
229
+ "run_group": "MMLU"
230
+ }
231
+ },
232
+ {
233
+ "value": "MMLU - # output tokens",
234
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
235
+ "markdown": false,
236
+ "metadata": {
237
+ "metric": "# output tokens",
238
+ "run_group": "MMLU"
239
+ }
240
+ },
241
+ {
242
+ "value": "MATH - # eval",
243
+ "description": "The MATH benchmark for measuring mathematical problem solving on competition math problems with chain-of-thought style reasoning [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2103.03874.pdf).\n\n# eval: Number of evaluation instances.",
244
+ "markdown": false,
245
+ "metadata": {
246
+ "metric": "# eval",
247
+ "run_group": "MATH"
248
+ }
249
+ },
250
+ {
251
+ "value": "MATH - # train",
252
+ "description": "The MATH benchmark for measuring mathematical problem solving on competition math problems with chain-of-thought style reasoning [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2103.03874.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
253
+ "markdown": false,
254
+ "metadata": {
255
+ "metric": "# train",
256
+ "run_group": "MATH"
257
+ }
258
+ },
259
+ {
260
+ "value": "MATH - truncated",
261
+ "description": "The MATH benchmark for measuring mathematical problem solving on competition math problems with chain-of-thought style reasoning [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2103.03874.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
262
+ "markdown": false,
263
+ "metadata": {
264
+ "metric": "truncated",
265
+ "run_group": "MATH"
266
+ }
267
+ },
268
+ {
269
+ "value": "MATH - # prompt tokens",
270
+ "description": "The MATH benchmark for measuring mathematical problem solving on competition math problems with chain-of-thought style reasoning [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2103.03874.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
271
+ "markdown": false,
272
+ "metadata": {
273
+ "metric": "# prompt tokens",
274
+ "run_group": "MATH"
275
+ }
276
+ },
277
+ {
278
+ "value": "MATH - # output tokens",
279
+ "description": "The MATH benchmark for measuring mathematical problem solving on competition math problems with chain-of-thought style reasoning [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2103.03874.pdf).\n\n# output tokens: Actual number of output tokens.",
280
+ "markdown": false,
281
+ "metadata": {
282
+ "metric": "# output tokens",
283
+ "run_group": "MATH"
284
+ }
285
+ },
286
+ {
287
+ "value": "GSM8K - # eval",
288
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\n# eval: Number of evaluation instances.",
289
+ "markdown": false,
290
+ "metadata": {
291
+ "metric": "# eval",
292
+ "run_group": "GSM8K"
293
+ }
294
+ },
295
+ {
296
+ "value": "GSM8K - # train",
297
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
298
+ "markdown": false,
299
+ "metadata": {
300
+ "metric": "# train",
301
+ "run_group": "GSM8K"
302
+ }
303
+ },
304
+ {
305
+ "value": "GSM8K - truncated",
306
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
307
+ "markdown": false,
308
+ "metadata": {
309
+ "metric": "truncated",
310
+ "run_group": "GSM8K"
311
+ }
312
+ },
313
+ {
314
+ "value": "GSM8K - # prompt tokens",
315
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
316
+ "markdown": false,
317
+ "metadata": {
318
+ "metric": "# prompt tokens",
319
+ "run_group": "GSM8K"
320
+ }
321
+ },
322
+ {
323
+ "value": "GSM8K - # output tokens",
324
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\n# output tokens: Actual number of output tokens.",
325
+ "markdown": false,
326
+ "metadata": {
327
+ "metric": "# output tokens",
328
+ "run_group": "GSM8K"
329
+ }
330
+ },
331
+ {
332
+ "value": "LegalBench - # eval",
333
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# eval: Number of evaluation instances.",
334
+ "markdown": false,
335
+ "metadata": {
336
+ "metric": "# eval",
337
+ "run_group": "LegalBench"
338
+ }
339
+ },
340
+ {
341
+ "value": "LegalBench - # train",
342
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
343
+ "markdown": false,
344
+ "metadata": {
345
+ "metric": "# train",
346
+ "run_group": "LegalBench"
347
+ }
348
+ },
349
+ {
350
+ "value": "LegalBench - truncated",
351
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
352
+ "markdown": false,
353
+ "metadata": {
354
+ "metric": "truncated",
355
+ "run_group": "LegalBench"
356
+ }
357
+ },
358
+ {
359
+ "value": "LegalBench - # prompt tokens",
360
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
361
+ "markdown": false,
362
+ "metadata": {
363
+ "metric": "# prompt tokens",
364
+ "run_group": "LegalBench"
365
+ }
366
+ },
367
+ {
368
+ "value": "LegalBench - # output tokens",
369
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# output tokens: Actual number of output tokens.",
370
+ "markdown": false,
371
+ "metadata": {
372
+ "metric": "# output tokens",
373
+ "run_group": "LegalBench"
374
+ }
375
+ },
376
+ {
377
+ "value": "MedQA - # eval",
378
+ "description": "MedQA is an open domain question answering dataset composed of questions from professional medical board exams ([Jin et al. 2020](https://arxiv.org/pdf/2009.13081.pdf)).\n\n# eval: Number of evaluation instances.",
379
+ "markdown": false,
380
+ "metadata": {
381
+ "metric": "# eval",
382
+ "run_group": "MedQA"
383
+ }
384
+ },
385
+ {
386
+ "value": "MedQA - # train",
387
+ "description": "MedQA is an open domain question answering dataset composed of questions from professional medical board exams ([Jin et al. 2020](https://arxiv.org/pdf/2009.13081.pdf)).\n\n# train: Number of training instances (e.g., in-context examples).",
388
+ "markdown": false,
389
+ "metadata": {
390
+ "metric": "# train",
391
+ "run_group": "MedQA"
392
+ }
393
+ },
394
+ {
395
+ "value": "MedQA - truncated",
396
+ "description": "MedQA is an open domain question answering dataset composed of questions from professional medical board exams ([Jin et al. 2020](https://arxiv.org/pdf/2009.13081.pdf)).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
397
+ "markdown": false,
398
+ "metadata": {
399
+ "metric": "truncated",
400
+ "run_group": "MedQA"
401
+ }
402
+ },
403
+ {
404
+ "value": "MedQA - # prompt tokens",
405
+ "description": "MedQA is an open domain question answering dataset composed of questions from professional medical board exams ([Jin et al. 2020](https://arxiv.org/pdf/2009.13081.pdf)).\n\n# prompt tokens: Number of tokens in the prompt.",
406
+ "markdown": false,
407
+ "metadata": {
408
+ "metric": "# prompt tokens",
409
+ "run_group": "MedQA"
410
+ }
411
+ },
412
+ {
413
+ "value": "MedQA - # output tokens",
414
+ "description": "MedQA is an open domain question answering dataset composed of questions from professional medical board exams ([Jin et al. 2020](https://arxiv.org/pdf/2009.13081.pdf)).\n\n# output tokens: Actual number of output tokens.",
415
+ "markdown": false,
416
+ "metadata": {
417
+ "metric": "# output tokens",
418
+ "run_group": "MedQA"
419
+ }
420
+ },
421
+ {
422
+ "value": "WMT 2014 - # eval",
423
+ "description": "WMT 2014 is a collection of machine translation datasets [(website)](https://www.statmt.org/wmt14/index.html).\n\n# eval: Number of evaluation instances.",
424
+ "markdown": false,
425
+ "metadata": {
426
+ "metric": "# eval",
427
+ "run_group": "WMT 2014"
428
+ }
429
+ },
430
+ {
431
+ "value": "WMT 2014 - # train",
432
+ "description": "WMT 2014 is a collection of machine translation datasets [(website)](https://www.statmt.org/wmt14/index.html).\n\n# train: Number of training instances (e.g., in-context examples).",
433
+ "markdown": false,
434
+ "metadata": {
435
+ "metric": "# train",
436
+ "run_group": "WMT 2014"
437
+ }
438
+ },
439
+ {
440
+ "value": "WMT 2014 - truncated",
441
+ "description": "WMT 2014 is a collection of machine translation datasets [(website)](https://www.statmt.org/wmt14/index.html).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
442
+ "markdown": false,
443
+ "metadata": {
444
+ "metric": "truncated",
445
+ "run_group": "WMT 2014"
446
+ }
447
+ },
448
+ {
449
+ "value": "WMT 2014 - # prompt tokens",
450
+ "description": "WMT 2014 is a collection of machine translation datasets [(website)](https://www.statmt.org/wmt14/index.html).\n\n# prompt tokens: Number of tokens in the prompt.",
451
+ "markdown": false,
452
+ "metadata": {
453
+ "metric": "# prompt tokens",
454
+ "run_group": "WMT 2014"
455
+ }
456
+ },
457
+ {
458
+ "value": "WMT 2014 - # output tokens",
459
+ "description": "WMT 2014 is a collection of machine translation datasets [(website)](https://www.statmt.org/wmt14/index.html).\n\n# output tokens: Actual number of output tokens.",
460
+ "markdown": false,
461
+ "metadata": {
462
+ "metric": "# output tokens",
463
+ "run_group": "WMT 2014"
464
+ }
465
+ }
466
+ ],
467
+ "rows": [
468
+ [
469
+ {
470
+ "value": "EleutherAI/pythia-2.8b",
471
+ "description": "",
472
+ "markdown": false
473
+ },
474
+ {
475
+ "markdown": false
476
+ },
477
+ {
478
+ "description": "No matching runs",
479
+ "markdown": false
480
+ },
481
+ {
482
+ "description": "No matching runs",
483
+ "markdown": false
484
+ },
485
+ {
486
+ "description": "No matching runs",
487
+ "markdown": false
488
+ },
489
+ {
490
+ "description": "No matching runs",
491
+ "markdown": false
492
+ },
493
+ {
494
+ "description": "No matching runs",
495
+ "markdown": false
496
+ },
497
+ {
498
+ "description": "No matching runs",
499
+ "markdown": false
500
+ },
501
+ {
502
+ "description": "No matching runs",
503
+ "markdown": false
504
+ },
505
+ {
506
+ "description": "No matching runs",
507
+ "markdown": false
508
+ },
509
+ {
510
+ "description": "No matching runs",
511
+ "markdown": false
512
+ },
513
+ {
514
+ "description": "No matching runs",
515
+ "markdown": false
516
+ },
517
+ {
518
+ "description": "No matching runs",
519
+ "markdown": false
520
+ },
521
+ {
522
+ "description": "No matching runs",
523
+ "markdown": false
524
+ },
525
+ {
526
+ "description": "No matching runs",
527
+ "markdown": false
528
+ },
529
+ {
530
+ "description": "No matching runs",
531
+ "markdown": false
532
+ },
533
+ {
534
+ "description": "No matching runs",
535
+ "markdown": false
536
+ },
537
+ {
538
+ "value": 500.0,
539
+ "description": "min=500, mean=500, max=500, sum=500 (1)",
540
+ "style": {},
541
+ "markdown": false,
542
+ "run_spec_names": [
543
+ "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
544
+ ]
545
+ },
546
+ {
547
+ "value": 5.0,
548
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
549
+ "style": {},
550
+ "markdown": false,
551
+ "run_spec_names": [
552
+ "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
553
+ ]
554
+ },
555
+ {
556
+ "value": 0.0,
557
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
558
+ "style": {},
559
+ "markdown": false,
560
+ "run_spec_names": [
561
+ "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
562
+ ]
563
+ },
564
+ {
565
+ "value": 251.556,
566
+ "description": "min=251.556, mean=251.556, max=251.556, sum=251.556 (1)",
567
+ "style": {},
568
+ "markdown": false,
569
+ "run_spec_names": [
570
+ "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
571
+ ]
572
+ },
573
+ {
574
+ "value": 1.0,
575
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
576
+ "style": {},
577
+ "markdown": false,
578
+ "run_spec_names": [
579
+ "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
580
+ ]
581
+ },
582
+ {
583
+ "value": 102.8,
584
+ "description": "min=100, mean=102.8, max=114, sum=514 (5)",
585
+ "style": {},
586
+ "markdown": false,
587
+ "run_spec_names": [
588
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
589
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
590
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
591
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
592
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
593
+ ]
594
+ },
595
+ {
596
+ "value": 5.0,
597
+ "description": "min=5, mean=5, max=5, sum=25 (5)",
598
+ "style": {},
599
+ "markdown": false,
600
+ "run_spec_names": [
601
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
602
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
603
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
604
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
605
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
606
+ ]
607
+ },
608
+ {
609
+ "value": 0.0,
610
+ "description": "min=0, mean=0, max=0, sum=0 (5)",
611
+ "style": {},
612
+ "markdown": false,
613
+ "run_spec_names": [
614
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
615
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
616
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
617
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
618
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
619
+ ]
620
+ },
621
+ {
622
+ "value": 467.935649122807,
623
+ "description": "min=358.76, mean=467.936, max=612.798, sum=2339.678 (5)",
624
+ "style": {},
625
+ "markdown": false,
626
+ "run_spec_names": [
627
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
628
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
629
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
630
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
631
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
632
+ ]
633
+ },
634
+ {
635
+ "value": 1.0,
636
+ "description": "min=1, mean=1, max=1, sum=5 (5)",
637
+ "style": {},
638
+ "markdown": false,
639
+ "run_spec_names": [
640
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
641
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
642
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
643
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
644
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
645
+ ]
646
+ },
647
+ {
648
+ "description": "No matching runs",
649
+ "markdown": false
650
+ },
651
+ {
652
+ "description": "No matching runs",
653
+ "markdown": false
654
+ },
655
+ {
656
+ "description": "No matching runs",
657
+ "markdown": false
658
+ },
659
+ {
660
+ "description": "No matching runs",
661
+ "markdown": false
662
+ },
663
+ {
664
+ "description": "No matching runs",
665
+ "markdown": false
666
+ },
667
+ {
668
+ "value": 1000.0,
669
+ "description": "min=1000, mean=1000, max=1000, sum=1000 (1)",
670
+ "style": {},
671
+ "markdown": false,
672
+ "run_spec_names": [
673
+ "gsm:model=EleutherAI_pythia-2.8b"
674
+ ]
675
+ },
676
+ {
677
+ "value": 5.0,
678
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
679
+ "style": {},
680
+ "markdown": false,
681
+ "run_spec_names": [
682
+ "gsm:model=EleutherAI_pythia-2.8b"
683
+ ]
684
+ },
685
+ {
686
+ "value": 0.0,
687
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
688
+ "style": {},
689
+ "markdown": false,
690
+ "run_spec_names": [
691
+ "gsm:model=EleutherAI_pythia-2.8b"
692
+ ]
693
+ },
694
+ {
695
+ "value": 939.582,
696
+ "description": "min=939.582, mean=939.582, max=939.582, sum=939.582 (1)",
697
+ "style": {},
698
+ "markdown": false,
699
+ "run_spec_names": [
700
+ "gsm:model=EleutherAI_pythia-2.8b"
701
+ ]
702
+ },
703
+ {
704
+ "value": 168.459,
705
+ "description": "min=168.459, mean=168.459, max=168.459, sum=168.459 (1)",
706
+ "style": {},
707
+ "markdown": false,
708
+ "run_spec_names": [
709
+ "gsm:model=EleutherAI_pythia-2.8b"
710
+ ]
711
+ },
712
+ {
713
+ "value": 409.4,
714
+ "description": "min=95, mean=409.4, max=1000, sum=2047 (5)",
715
+ "style": {},
716
+ "markdown": false,
717
+ "run_spec_names": [
718
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
719
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
720
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
721
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
722
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
723
+ ]
724
+ },
725
+ {
726
+ "value": 3.8604081632653062,
727
+ "description": "min=0.302, mean=3.86, max=5, sum=19.302 (5)",
728
+ "style": {},
729
+ "markdown": false,
730
+ "run_spec_names": [
731
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
732
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
733
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
734
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
735
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
736
+ ]
737
+ },
738
+ {
739
+ "value": 0.002857142857142857,
740
+ "description": "min=0, mean=0.003, max=0.014, sum=0.014 (5)",
741
+ "style": {},
742
+ "markdown": false,
743
+ "run_spec_names": [
744
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
745
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
746
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
747
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
748
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
749
+ ]
750
+ },
751
+ {
752
+ "value": 560.6440716343213,
753
+ "description": "min=206.779, mean=560.644, max=1497.455, sum=2803.22 (5)",
754
+ "style": {},
755
+ "markdown": false,
756
+ "run_spec_names": [
757
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
758
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
759
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
760
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
761
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
762
+ ]
763
+ },
764
+ {
765
+ "value": 1.6417916921537006,
766
+ "description": "min=1, mean=1.642, max=4.124, sum=8.209 (5)",
767
+ "style": {},
768
+ "markdown": false,
769
+ "run_spec_names": [
770
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
771
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
772
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
773
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
774
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
775
+ ]
776
+ },
777
+ {
778
+ "description": "No matching runs",
779
+ "markdown": false
780
+ },
781
+ {
782
+ "description": "No matching runs",
783
+ "markdown": false
784
+ },
785
+ {
786
+ "description": "No matching runs",
787
+ "markdown": false
788
+ },
789
+ {
790
+ "description": "No matching runs",
791
+ "markdown": false
792
+ },
793
+ {
794
+ "description": "No matching runs",
795
+ "markdown": false
796
+ },
797
+ {
798
+ "description": "No matching runs",
799
+ "markdown": false
800
+ },
801
+ {
802
+ "description": "No matching runs",
803
+ "markdown": false
804
+ },
805
+ {
806
+ "description": "No matching runs",
807
+ "markdown": false
808
+ },
809
+ {
810
+ "description": "No matching runs",
811
+ "markdown": false
812
+ },
813
+ {
814
+ "description": "No matching runs",
815
+ "markdown": false
816
+ }
817
+ ]
818
+ ],
819
+ "links": [
820
+ {
821
+ "text": "LaTeX",
822
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/core_scenarios_general_information.tex"
823
+ },
824
+ {
825
+ "text": "JSON",
826
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/core_scenarios_general_information.json"
827
+ }
828
+ ],
829
+ "name": "general_information"
830
+ }
lite_pythia-2.8b-step5000/groups/json/gsm_gsm_.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\nExact match (final number): Fraction of instances that the predicted output matches a correct reference exactly, ignoring text preceding the specified indicator.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "GSM8K"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "GSM8K"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "GSM8K"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "GSM8K"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "GSM8K"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "GSM8K"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The grade school math word problems dataset (GSM8K) for testing mathematical reasoning on grade-school math problems [(Cobbe et al., 2021)](https://arxiv.org/pdf/2110.14168.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "GSM8K"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-2.8b",
79
+ "description": "",
80
+ "href": "?group=gsm&subgroup=&runSpecs=%5B%22gsm%3Amodel%3DEleutherAI_pythia-2.8b%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "gsm:model=EleutherAI_pythia-2.8b"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.01,
88
+ "description": "min=0.01, mean=0.01, max=0.01, sum=0.01 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 2.5314217054843904,
96
+ "description": "min=2.531, mean=2.531, max=2.531, sum=2.531 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 1000.0,
104
+ "description": "min=1000, mean=1000, max=1000, sum=1000 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 939.582,
122
+ "description": "min=939.582, mean=939.582, max=939.582, sum=939.582 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 168.459,
128
+ "description": "min=168.459, mean=168.459, max=168.459, sum=168.459 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/gsm_gsm_.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/gsm_gsm_.json"
142
+ }
143
+ ],
144
+ "name": "gsm_"
145
+ }
lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench.json ADDED
@@ -0,0 +1,190 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "LegalBench",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "LegalBench"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "LegalBench"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "LegalBench"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "LegalBench"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "LegalBench"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "LegalBench"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "LegalBench"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-2.8b",
79
+ "description": "",
80
+ "markdown": false
81
+ },
82
+ {
83
+ "value": 0.3189986232611502,
84
+ "description": "min=0.144, mean=0.319, max=0.619, sum=1.595 (5)",
85
+ "style": {
86
+ "font-weight": "bold"
87
+ },
88
+ "markdown": false,
89
+ "run_spec_names": [
90
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
91
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
92
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
93
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
94
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
95
+ ]
96
+ },
97
+ {
98
+ "value": 0.3094366045219278,
99
+ "description": "min=0.141, mean=0.309, max=0.795, sum=1.547 (5)",
100
+ "style": {
101
+ "font-weight": "bold"
102
+ },
103
+ "markdown": false,
104
+ "run_spec_names": [
105
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
106
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
107
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
108
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
109
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
110
+ ]
111
+ },
112
+ {
113
+ "value": 409.4,
114
+ "description": "min=95, mean=409.4, max=1000, sum=2047 (5)",
115
+ "style": {},
116
+ "markdown": false,
117
+ "run_spec_names": [
118
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
119
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
120
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
121
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
122
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
123
+ ]
124
+ },
125
+ {
126
+ "value": 3.8604081632653062,
127
+ "description": "min=0.302, mean=3.86, max=5, sum=19.302 (5)",
128
+ "style": {},
129
+ "markdown": false,
130
+ "run_spec_names": [
131
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
132
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
133
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
134
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
135
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
136
+ ]
137
+ },
138
+ {
139
+ "value": 0.002857142857142857,
140
+ "description": "min=0, mean=0.003, max=0.014, sum=0.014 (5)",
141
+ "style": {},
142
+ "markdown": false,
143
+ "run_spec_names": [
144
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
145
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
146
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
147
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
148
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
149
+ ]
150
+ },
151
+ {
152
+ "value": 560.6440716343213,
153
+ "description": "min=206.779, mean=560.644, max=1497.455, sum=2803.22 (5)",
154
+ "style": {},
155
+ "markdown": false,
156
+ "run_spec_names": [
157
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
158
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
159
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
160
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
161
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
162
+ ]
163
+ },
164
+ {
165
+ "value": 1.6417916921537006,
166
+ "description": "min=1, mean=1.642, max=4.124, sum=8.209 (5)",
167
+ "style": {},
168
+ "markdown": false,
169
+ "run_spec_names": [
170
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
171
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
172
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
173
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
174
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
175
+ ]
176
+ }
177
+ ]
178
+ ],
179
+ "links": [
180
+ {
181
+ "text": "LaTeX",
182
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench.tex"
183
+ },
184
+ {
185
+ "text": "JSON",
186
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench.json"
187
+ }
188
+ ],
189
+ "name": "legalbench"
190
+ }
lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:abercrombie.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subset: abercrombie",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "LegalBench"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "LegalBench"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "LegalBench"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "LegalBench"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "LegalBench"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "LegalBench"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "LegalBench"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-2.8b",
79
+ "description": "",
80
+ "href": "?group=legalbench&subgroup=subset%3A%20abercrombie&runSpecs=%5B%22legalbench%3Asubset%3Dabercrombie%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.2,
88
+ "description": "min=0.2, mean=0.2, max=0.2, sum=0.2 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.1405831638135408,
96
+ "description": "min=0.141, mean=0.141, max=0.141, sum=0.141 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 95.0,
104
+ "description": "min=95, mean=95, max=95, sum=95 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 206.77894736842106,
122
+ "description": "min=206.779, mean=206.779, max=206.779, sum=206.779 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:abercrombie.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:abercrombie.json"
142
+ }
143
+ ],
144
+ "name": "legalbench_subset:abercrombie"
145
+ }
lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:corporate_lobbying.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subset: corporate_lobbying",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "LegalBench"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "LegalBench"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "LegalBench"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "LegalBench"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "LegalBench"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "LegalBench"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "LegalBench"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-2.8b",
79
+ "description": "",
80
+ "href": "?group=legalbench&subgroup=subset%3A%20corporate_lobbying&runSpecs=%5B%22legalbench%3Asubset%3Dcorporate_lobbying%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.2,
88
+ "description": "min=0.2, mean=0.2, max=0.2, sum=0.2 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.7949594166814065,
96
+ "description": "min=0.795, mean=0.795, max=0.795, sum=0.795 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 490.0,
104
+ "description": "min=490, mean=490, max=490, sum=490 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 0.3020408163265306,
110
+ "description": "min=0.302, mean=0.302, max=0.302, sum=0.302 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.014285714285714285,
116
+ "description": "min=0.014, mean=0.014, max=0.014, sum=0.014 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 1497.4551020408164,
122
+ "description": "min=1497.455, mean=1497.455, max=1497.455, sum=1497.455 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 4.124489795918367,
128
+ "description": "min=4.124, mean=4.124, max=4.124, sum=4.124 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:corporate_lobbying.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:corporate_lobbying.json"
142
+ }
143
+ ],
144
+ "name": "legalbench_subset:corporate_lobbying"
145
+ }
lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:function_of_decision_section.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subset: function_of_decision_section",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "LegalBench"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "LegalBench"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "LegalBench"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "LegalBench"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "LegalBench"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "LegalBench"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "LegalBench"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-2.8b",
79
+ "description": "",
80
+ "href": "?group=legalbench&subgroup=subset%3A%20function_of_decision_section&runSpecs=%5B%22legalbench%3Asubset%3Dfunction_of_decision_section%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.1444141689373297,
88
+ "description": "min=0.144, mean=0.144, max=0.144, sum=0.144 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.2725935534495424,
96
+ "description": "min=0.273, mean=0.273, max=0.273, sum=0.273 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 367.0,
104
+ "description": "min=367, mean=367, max=367, sum=367 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 514.9182561307902,
122
+ "description": "min=514.918, mean=514.918, max=514.918, sum=514.918 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0844686648501363,
128
+ "description": "min=1.084, mean=1.084, max=1.084, sum=1.084 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:function_of_decision_section.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:function_of_decision_section.json"
142
+ }
143
+ ],
144
+ "name": "legalbench_subset:function_of_decision_section"
145
+ }
lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:international_citizenship_questions.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subset: international_citizenship_questions",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "LegalBench"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "LegalBench"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "LegalBench"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "LegalBench"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "LegalBench"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "LegalBench"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "LegalBench"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-2.8b",
79
+ "description": "",
80
+ "href": "?group=legalbench&subgroup=subset%3A%20international_citizenship_questions&runSpecs=%5B%22legalbench%3Asubset%3Dinternational_citizenship_questions%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.619,
88
+ "description": "min=0.619, mean=0.619, max=0.619, sum=0.619 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.15224194407463074,
96
+ "description": "min=0.152, mean=0.152, max=0.152, sum=0.152 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 1000.0,
104
+ "description": "min=1000, mean=1000, max=1000, sum=1000 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 4.0,
110
+ "description": "min=4, mean=4, max=4, sum=4 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 250.447,
122
+ "description": "min=250.447, mean=250.447, max=250.447, sum=250.447 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:international_citizenship_questions.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:international_citizenship_questions.json"
142
+ }
143
+ ],
144
+ "name": "legalbench_subset:international_citizenship_questions"
145
+ }
lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:proa.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subset: proa",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "LegalBench"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "LegalBench"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "LegalBench"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "LegalBench"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "LegalBench"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "LegalBench"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "LegalBench"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-2.8b",
79
+ "description": "",
80
+ "href": "?group=legalbench&subgroup=subset%3A%20proa&runSpecs=%5B%22legalbench%3Asubset%3Dproa%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.43157894736842106,
88
+ "description": "min=0.432, mean=0.432, max=0.432, sum=0.432 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.18680494459051836,
96
+ "description": "min=0.187, mean=0.187, max=0.187, sum=0.187 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 95.0,
104
+ "description": "min=95, mean=95, max=95, sum=95 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 333.62105263157895,
122
+ "description": "min=333.621, mean=333.621, max=333.621, sum=333.621 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:proa.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:proa.json"
142
+ }
143
+ ],
144
+ "name": "legalbench_subset:proa"
145
+ }
lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu.json ADDED
@@ -0,0 +1,190 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "MMLU (Massive Multitask Language Understanding)",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "MMLU"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "MMLU"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "MMLU"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "MMLU"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "MMLU"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "MMLU"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "MMLU"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-2.8b",
79
+ "description": "",
80
+ "markdown": false
81
+ },
82
+ {
83
+ "value": 0.27487719298245616,
84
+ "description": "min=0.25, mean=0.275, max=0.32, sum=1.374 (5)",
85
+ "style": {
86
+ "font-weight": "bold"
87
+ },
88
+ "markdown": false,
89
+ "run_spec_names": [
90
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
91
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
92
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
93
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
94
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
95
+ ]
96
+ },
97
+ {
98
+ "value": 0.22061018314696193,
99
+ "description": "min=0.173, mean=0.221, max=0.282, sum=1.103 (5)",
100
+ "style": {
101
+ "font-weight": "bold"
102
+ },
103
+ "markdown": false,
104
+ "run_spec_names": [
105
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
106
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
107
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
108
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
109
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
110
+ ]
111
+ },
112
+ {
113
+ "value": 102.8,
114
+ "description": "min=100, mean=102.8, max=114, sum=514 (5)",
115
+ "style": {},
116
+ "markdown": false,
117
+ "run_spec_names": [
118
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
119
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
120
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
121
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
122
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
123
+ ]
124
+ },
125
+ {
126
+ "value": 5.0,
127
+ "description": "min=5, mean=5, max=5, sum=25 (5)",
128
+ "style": {},
129
+ "markdown": false,
130
+ "run_spec_names": [
131
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
132
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
133
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
134
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
135
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
136
+ ]
137
+ },
138
+ {
139
+ "value": 0.0,
140
+ "description": "min=0, mean=0, max=0, sum=0 (5)",
141
+ "style": {},
142
+ "markdown": false,
143
+ "run_spec_names": [
144
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
145
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
146
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
147
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
148
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
149
+ ]
150
+ },
151
+ {
152
+ "value": 467.935649122807,
153
+ "description": "min=358.76, mean=467.936, max=612.798, sum=2339.678 (5)",
154
+ "style": {},
155
+ "markdown": false,
156
+ "run_spec_names": [
157
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
158
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
159
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
160
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
161
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
162
+ ]
163
+ },
164
+ {
165
+ "value": 1.0,
166
+ "description": "min=1, mean=1, max=1, sum=5 (5)",
167
+ "style": {},
168
+ "markdown": false,
169
+ "run_spec_names": [
170
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
171
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
172
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
173
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
174
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
175
+ ]
176
+ }
177
+ ]
178
+ ],
179
+ "links": [
180
+ {
181
+ "text": "LaTeX",
182
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu.tex"
183
+ },
184
+ {
185
+ "text": "JSON",
186
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu.json"
187
+ }
188
+ ],
189
+ "name": "mmlu"
190
+ }
lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:abstract_algebra.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: abstract_algebra",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "MMLU"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "MMLU"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "MMLU"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "MMLU"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "MMLU"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "MMLU"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "MMLU"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-2.8b",
79
+ "description": "",
80
+ "href": "?group=mmlu&subgroup=subject%3A%20abstract_algebra&runSpecs=%5B%22mmlu%3Asubject%3Dabstract_algebra%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.27,
88
+ "description": "min=0.27, mean=0.27, max=0.27, sum=0.27 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.17254346132278442,
96
+ "description": "min=0.173, mean=0.173, max=0.173, sum=0.173 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 100.0,
104
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 358.76,
122
+ "description": "min=358.76, mean=358.76, max=358.76, sum=358.76 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:abstract_algebra.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:abstract_algebra.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_subject:abstract_algebra"
145
+ }
lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:college_chemistry.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: college_chemistry",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "MMLU"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "MMLU"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "MMLU"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "MMLU"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "MMLU"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "MMLU"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "MMLU"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-2.8b",
79
+ "description": "",
80
+ "href": "?group=mmlu&subgroup=subject%3A%20college_chemistry&runSpecs=%5B%22mmlu%3Asubject%3Dcollege_chemistry%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.32,
88
+ "description": "min=0.32, mean=0.32, max=0.32, sum=0.32 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.2506118679046631,
96
+ "description": "min=0.251, mean=0.251, max=0.251, sum=0.251 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 100.0,
104
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 535.85,
122
+ "description": "min=535.85, mean=535.85, max=535.85, sum=535.85 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:college_chemistry.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:college_chemistry.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_subject:college_chemistry"
145
+ }
lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:computer_security.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: computer_security",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "MMLU"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "MMLU"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "MMLU"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "MMLU"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "MMLU"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "MMLU"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "MMLU"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-2.8b",
79
+ "description": "",
80
+ "href": "?group=mmlu&subgroup=subject%3A%20computer_security&runSpecs=%5B%22mmlu%3Asubject%3Dcomputer_security%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.25,
88
+ "description": "min=0.25, mean=0.25, max=0.25, sum=0.25 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.18395774602890014,
96
+ "description": "min=0.184, mean=0.184, max=0.184, sum=0.184 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 100.0,
104
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 388.19,
122
+ "description": "min=388.19, mean=388.19, max=388.19, sum=388.19 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:computer_security.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:computer_security.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_subject:computer_security"
145
+ }
lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:econometrics.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: econometrics",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "MMLU"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "MMLU"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "MMLU"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "MMLU"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "MMLU"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "MMLU"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "MMLU"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-2.8b",
79
+ "description": "",
80
+ "href": "?group=mmlu&subgroup=subject%3A%20econometrics&runSpecs=%5B%22mmlu%3Asubject%3Deconometrics%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.2543859649122807,
88
+ "description": "min=0.254, mean=0.254, max=0.254, sum=0.254 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.28234320356134784,
96
+ "description": "min=0.282, mean=0.282, max=0.282, sum=0.282 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 114.0,
104
+ "description": "min=114, mean=114, max=114, sum=114 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 612.7982456140351,
122
+ "description": "min=612.798, mean=612.798, max=612.798, sum=612.798 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:econometrics.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:econometrics.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_subject:econometrics"
145
+ }
lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:us_foreign_policy.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "subject: us_foreign_policy",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "MMLU"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "MMLU"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "MMLU"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "MMLU"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "MMLU"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "MMLU"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "MMLU"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-2.8b",
79
+ "description": "",
80
+ "href": "?group=mmlu&subgroup=subject%3A%20us_foreign_policy&runSpecs=%5B%22mmlu%3Asubject%3Dus_foreign_policy%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.28,
88
+ "description": "min=0.28, mean=0.28, max=0.28, sum=0.28 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.21359463691711425,
96
+ "description": "min=0.214, mean=0.214, max=0.214, sum=0.214 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 100.0,
104
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 444.08,
122
+ "description": "min=444.08, mean=444.08, max=444.08, sum=444.08 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:us_foreign_policy.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:us_foreign_policy.json"
142
+ }
143
+ ],
144
+ "name": "mmlu_subject:us_foreign_policy"
145
+ }
lite_pythia-2.8b-step5000/groups/json/openbookqa_openbookqa_.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "title": "",
3
+ "header": [
4
+ {
5
+ "value": "Model",
6
+ "markdown": false,
7
+ "metadata": {}
8
+ },
9
+ {
10
+ "value": "EM",
11
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
12
+ "markdown": false,
13
+ "lower_is_better": false,
14
+ "metadata": {
15
+ "metric": "EM",
16
+ "run_group": "OpenbookQA"
17
+ }
18
+ },
19
+ {
20
+ "value": "Observed inference time (s)",
21
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
22
+ "markdown": false,
23
+ "lower_is_better": true,
24
+ "metadata": {
25
+ "metric": "Observed inference time (s)",
26
+ "run_group": "OpenbookQA"
27
+ }
28
+ },
29
+ {
30
+ "value": "# eval",
31
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\n# eval: Number of evaluation instances.",
32
+ "markdown": false,
33
+ "metadata": {
34
+ "metric": "# eval",
35
+ "run_group": "OpenbookQA"
36
+ }
37
+ },
38
+ {
39
+ "value": "# train",
40
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\n# train: Number of training instances (e.g., in-context examples).",
41
+ "markdown": false,
42
+ "metadata": {
43
+ "metric": "# train",
44
+ "run_group": "OpenbookQA"
45
+ }
46
+ },
47
+ {
48
+ "value": "truncated",
49
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
50
+ "markdown": false,
51
+ "metadata": {
52
+ "metric": "truncated",
53
+ "run_group": "OpenbookQA"
54
+ }
55
+ },
56
+ {
57
+ "value": "# prompt tokens",
58
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\n# prompt tokens: Number of tokens in the prompt.",
59
+ "markdown": false,
60
+ "metadata": {
61
+ "metric": "# prompt tokens",
62
+ "run_group": "OpenbookQA"
63
+ }
64
+ },
65
+ {
66
+ "value": "# output tokens",
67
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\n# output tokens: Actual number of output tokens.",
68
+ "markdown": false,
69
+ "metadata": {
70
+ "metric": "# output tokens",
71
+ "run_group": "OpenbookQA"
72
+ }
73
+ }
74
+ ],
75
+ "rows": [
76
+ [
77
+ {
78
+ "value": "EleutherAI/pythia-2.8b",
79
+ "description": "",
80
+ "href": "?group=openbookqa&subgroup=&runSpecs=%5B%22commonsense%3Adataset%3Dopenbookqa%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
81
+ "markdown": false,
82
+ "run_spec_names": [
83
+ "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
84
+ ]
85
+ },
86
+ {
87
+ "value": 0.22,
88
+ "description": "min=0.22, mean=0.22, max=0.22, sum=0.22 (1)",
89
+ "style": {
90
+ "font-weight": "bold"
91
+ },
92
+ "markdown": false
93
+ },
94
+ {
95
+ "value": 0.12883714818954467,
96
+ "description": "min=0.129, mean=0.129, max=0.129, sum=0.129 (1)",
97
+ "style": {
98
+ "font-weight": "bold"
99
+ },
100
+ "markdown": false
101
+ },
102
+ {
103
+ "value": 500.0,
104
+ "description": "min=500, mean=500, max=500, sum=500 (1)",
105
+ "style": {},
106
+ "markdown": false
107
+ },
108
+ {
109
+ "value": 5.0,
110
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
111
+ "style": {},
112
+ "markdown": false
113
+ },
114
+ {
115
+ "value": 0.0,
116
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
117
+ "style": {},
118
+ "markdown": false
119
+ },
120
+ {
121
+ "value": 251.556,
122
+ "description": "min=251.556, mean=251.556, max=251.556, sum=251.556 (1)",
123
+ "style": {},
124
+ "markdown": false
125
+ },
126
+ {
127
+ "value": 1.0,
128
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
129
+ "style": {},
130
+ "markdown": false
131
+ }
132
+ ]
133
+ ],
134
+ "links": [
135
+ {
136
+ "text": "LaTeX",
137
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/openbookqa_openbookqa_.tex"
138
+ },
139
+ {
140
+ "text": "JSON",
141
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/openbookqa_openbookqa_.json"
142
+ }
143
+ ],
144
+ "name": "openbookqa_"
145
+ }
lite_pythia-2.8b-step5000/groups/latex/core_scenarios_accuracy.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrr}
4
+ \toprule
5
+ Model & OpenbookQA - EM & MMLU - EM & GSM8K - EM & LegalBench - EM \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 0.22 & 0.27487719298245616 & 0.01 & 0.3189986232611502 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for accuracy (core_scenarios)}
11
+ \label{fig:accuracy (core_scenarios)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/latex/core_scenarios_efficiency.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrr}
4
+ \toprule
5
+ Model & OpenbookQA - Observed inference time (s) & MMLU - Observed inference time (s) & GSM8K - Observed inference time (s) & LegalBench - Observed inference time (s) \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 0.12883714818954467 & 0.22061018314696193 & 2.5314217054843904 & 0.3094366045219278 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for efficiency (core_scenarios)}
11
+ \label{fig:efficiency (core_scenarios)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/latex/core_scenarios_general_information.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrrrrrrrrrrrrrrrrrr}
4
+ \toprule
5
+ Model & OpenbookQA - # eval & OpenbookQA - # train & OpenbookQA - truncated & OpenbookQA - # prompt tokens & OpenbookQA - # output tokens & MMLU - # eval & MMLU - # train & MMLU - truncated & MMLU - # prompt tokens & MMLU - # output tokens & GSM8K - # eval & GSM8K - # train & GSM8K - truncated & GSM8K - # prompt tokens & GSM8K - # output tokens & LegalBench - # eval & LegalBench - # train & LegalBench - truncated & LegalBench - # prompt tokens & LegalBench - # output tokens \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 500.0 & 5.0 & & 251.556 & 1.0 & 102.8 & 5.0 & & 467.935649122807 & 1.0 & 1000.0 & 5.0 & & 939.582 & 168.459 & 409.4 & 3.8604081632653062 & 0.002857142857142857 & 560.6440716343213 & 1.6417916921537006 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for general_information (core_scenarios)}
11
+ \label{fig:general_information (core_scenarios)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/latex/gsm_gsm_.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrrrrr}
4
+ \toprule
5
+ Model & EM & Observed inference time (s) & # eval & # train & truncated & # prompt tokens & # output tokens \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 0.01 & 2.5314217054843904 & 1000.0 & 5.0 & & 939.582 & 168.459 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for gsm_ (gsm)}
11
+ \label{fig:gsm_ (gsm)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrrrrr}
4
+ \toprule
5
+ Model & EM & Observed inference time (s) & # eval & # train & truncated & # prompt tokens & # output tokens \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 0.3189986232611502 & 0.3094366045219278 & 409.4 & 3.8604081632653062 & 0.002857142857142857 & 560.6440716343213 & 1.6417916921537006 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for legalbench (legalbench)}
11
+ \label{fig:legalbench (legalbench)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:abercrombie.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrrrrr}
4
+ \toprule
5
+ Model & EM & Observed inference time (s) & # eval & # train & truncated & # prompt tokens & # output tokens \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 0.2 & 0.1405831638135408 & 95.0 & 5.0 & & 206.77894736842106 & 1.0 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for legalbench_subset:abercrombie (legalbench)}
11
+ \label{fig:legalbench_subset:abercrombie (legalbench)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:corporate_lobbying.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrrrrr}
4
+ \toprule
5
+ Model & EM & Observed inference time (s) & # eval & # train & truncated & # prompt tokens & # output tokens \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 0.2 & 0.7949594166814065 & 490.0 & 0.3020408163265306 & 0.014285714285714285 & 1497.4551020408164 & 4.124489795918367 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for legalbench_subset:corporate_lobbying (legalbench)}
11
+ \label{fig:legalbench_subset:corporate_lobbying (legalbench)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:function_of_decision_section.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrrrrr}
4
+ \toprule
5
+ Model & EM & Observed inference time (s) & # eval & # train & truncated & # prompt tokens & # output tokens \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 0.1444141689373297 & 0.2725935534495424 & 367.0 & 5.0 & & 514.9182561307902 & 1.0844686648501363 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for legalbench_subset:function_of_decision_section (legalbench)}
11
+ \label{fig:legalbench_subset:function_of_decision_section (legalbench)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:international_citizenship_questions.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrrrrr}
4
+ \toprule
5
+ Model & EM & Observed inference time (s) & # eval & # train & truncated & # prompt tokens & # output tokens \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 0.619 & 0.15224194407463074 & 1000.0 & 4.0 & & 250.447 & 1.0 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for legalbench_subset:international_citizenship_questions (legalbench)}
11
+ \label{fig:legalbench_subset:international_citizenship_questions (legalbench)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:proa.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrrrrr}
4
+ \toprule
5
+ Model & EM & Observed inference time (s) & # eval & # train & truncated & # prompt tokens & # output tokens \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 0.43157894736842106 & 0.18680494459051836 & 95.0 & 5.0 & & 333.62105263157895 & 1.0 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for legalbench_subset:proa (legalbench)}
11
+ \label{fig:legalbench_subset:proa (legalbench)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrrrrr}
4
+ \toprule
5
+ Model & EM & Observed inference time (s) & # eval & # train & truncated & # prompt tokens & # output tokens \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 0.27487719298245616 & 0.22061018314696193 & 102.8 & 5.0 & & 467.935649122807 & 1.0 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for mmlu (mmlu)}
11
+ \label{fig:mmlu (mmlu)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:abstract_algebra.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrrrrr}
4
+ \toprule
5
+ Model & EM & Observed inference time (s) & # eval & # train & truncated & # prompt tokens & # output tokens \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 0.27 & 0.17254346132278442 & 100.0 & 5.0 & & 358.76 & 1.0 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for mmlu_subject:abstract_algebra (mmlu)}
11
+ \label{fig:mmlu_subject:abstract_algebra (mmlu)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:college_chemistry.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrrrrr}
4
+ \toprule
5
+ Model & EM & Observed inference time (s) & # eval & # train & truncated & # prompt tokens & # output tokens \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 0.32 & 0.2506118679046631 & 100.0 & 5.0 & & 535.85 & 1.0 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for mmlu_subject:college_chemistry (mmlu)}
11
+ \label{fig:mmlu_subject:college_chemistry (mmlu)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:computer_security.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrrrrr}
4
+ \toprule
5
+ Model & EM & Observed inference time (s) & # eval & # train & truncated & # prompt tokens & # output tokens \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 0.25 & 0.18395774602890014 & 100.0 & 5.0 & & 388.19 & 1.0 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for mmlu_subject:computer_security (mmlu)}
11
+ \label{fig:mmlu_subject:computer_security (mmlu)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:econometrics.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrrrrr}
4
+ \toprule
5
+ Model & EM & Observed inference time (s) & # eval & # train & truncated & # prompt tokens & # output tokens \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 0.2543859649122807 & 0.28234320356134784 & 114.0 & 5.0 & & 612.7982456140351 & 1.0 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for mmlu_subject:econometrics (mmlu)}
11
+ \label{fig:mmlu_subject:econometrics (mmlu)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:us_foreign_policy.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrrrrr}
4
+ \toprule
5
+ Model & EM & Observed inference time (s) & # eval & # train & truncated & # prompt tokens & # output tokens \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 0.28 & 0.21359463691711425 & 100.0 & 5.0 & & 444.08 & 1.0 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for mmlu_subject:us_foreign_policy (mmlu)}
11
+ \label{fig:mmlu_subject:us_foreign_policy (mmlu)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/latex/openbookqa_openbookqa_.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ \begin{table*}[htp]
2
+ \resizebox{\textwidth}{!}{
3
+ \begin{tabular}{lrrrrrrr}
4
+ \toprule
5
+ Model & EM & Observed inference time (s) & # eval & # train & truncated & # prompt tokens & # output tokens \\
6
+ \midrule
7
+ EleutherAI/pythia-2.8b & 0.22 & 0.12883714818954467 & 500.0 & 5.0 & & 251.556 & 1.0 \\
8
+ \bottomrule
9
+ \end{tabular}}
10
+ \caption{Results for openbookqa_ (openbookqa)}
11
+ \label{fig:openbookqa_ (openbookqa)}
12
+ \end{table*}
lite_pythia-2.8b-step5000/groups/legalbench.json ADDED
@@ -0,0 +1,917 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "title": "LegalBench",
4
+ "header": [
5
+ {
6
+ "value": "Model",
7
+ "markdown": false,
8
+ "metadata": {}
9
+ },
10
+ {
11
+ "value": "EM",
12
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.",
13
+ "markdown": false,
14
+ "lower_is_better": false,
15
+ "metadata": {
16
+ "metric": "EM",
17
+ "run_group": "LegalBench"
18
+ }
19
+ },
20
+ {
21
+ "value": "Observed inference time (s)",
22
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
23
+ "markdown": false,
24
+ "lower_is_better": true,
25
+ "metadata": {
26
+ "metric": "Observed inference time (s)",
27
+ "run_group": "LegalBench"
28
+ }
29
+ },
30
+ {
31
+ "value": "# eval",
32
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# eval: Number of evaluation instances.",
33
+ "markdown": false,
34
+ "metadata": {
35
+ "metric": "# eval",
36
+ "run_group": "LegalBench"
37
+ }
38
+ },
39
+ {
40
+ "value": "# train",
41
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
42
+ "markdown": false,
43
+ "metadata": {
44
+ "metric": "# train",
45
+ "run_group": "LegalBench"
46
+ }
47
+ },
48
+ {
49
+ "value": "truncated",
50
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
51
+ "markdown": false,
52
+ "metadata": {
53
+ "metric": "truncated",
54
+ "run_group": "LegalBench"
55
+ }
56
+ },
57
+ {
58
+ "value": "# prompt tokens",
59
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
60
+ "markdown": false,
61
+ "metadata": {
62
+ "metric": "# prompt tokens",
63
+ "run_group": "LegalBench"
64
+ }
65
+ },
66
+ {
67
+ "value": "# output tokens",
68
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# output tokens: Actual number of output tokens.",
69
+ "markdown": false,
70
+ "metadata": {
71
+ "metric": "# output tokens",
72
+ "run_group": "LegalBench"
73
+ }
74
+ }
75
+ ],
76
+ "rows": [
77
+ [
78
+ {
79
+ "value": "EleutherAI/pythia-2.8b",
80
+ "description": "",
81
+ "markdown": false
82
+ },
83
+ {
84
+ "value": 0.3189986232611502,
85
+ "description": "min=0.144, mean=0.319, max=0.619, sum=1.595 (5)",
86
+ "style": {
87
+ "font-weight": "bold"
88
+ },
89
+ "markdown": false,
90
+ "run_spec_names": [
91
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
92
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
93
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
94
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
95
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
96
+ ]
97
+ },
98
+ {
99
+ "value": 0.3094366045219278,
100
+ "description": "min=0.141, mean=0.309, max=0.795, sum=1.547 (5)",
101
+ "style": {
102
+ "font-weight": "bold"
103
+ },
104
+ "markdown": false,
105
+ "run_spec_names": [
106
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
107
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
108
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
109
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
110
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
111
+ ]
112
+ },
113
+ {
114
+ "value": 409.4,
115
+ "description": "min=95, mean=409.4, max=1000, sum=2047 (5)",
116
+ "style": {},
117
+ "markdown": false,
118
+ "run_spec_names": [
119
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
120
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
121
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
122
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
123
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
124
+ ]
125
+ },
126
+ {
127
+ "value": 3.8604081632653062,
128
+ "description": "min=0.302, mean=3.86, max=5, sum=19.302 (5)",
129
+ "style": {},
130
+ "markdown": false,
131
+ "run_spec_names": [
132
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
133
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
134
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
135
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
136
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
137
+ ]
138
+ },
139
+ {
140
+ "value": 0.002857142857142857,
141
+ "description": "min=0, mean=0.003, max=0.014, sum=0.014 (5)",
142
+ "style": {},
143
+ "markdown": false,
144
+ "run_spec_names": [
145
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
146
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
147
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
148
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
149
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
150
+ ]
151
+ },
152
+ {
153
+ "value": 560.6440716343213,
154
+ "description": "min=206.779, mean=560.644, max=1497.455, sum=2803.22 (5)",
155
+ "style": {},
156
+ "markdown": false,
157
+ "run_spec_names": [
158
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
159
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
160
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
161
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
162
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
163
+ ]
164
+ },
165
+ {
166
+ "value": 1.6417916921537006,
167
+ "description": "min=1, mean=1.642, max=4.124, sum=8.209 (5)",
168
+ "style": {},
169
+ "markdown": false,
170
+ "run_spec_names": [
171
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b",
172
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b",
173
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b",
174
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b",
175
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
176
+ ]
177
+ }
178
+ ]
179
+ ],
180
+ "links": [
181
+ {
182
+ "text": "LaTeX",
183
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench.tex"
184
+ },
185
+ {
186
+ "text": "JSON",
187
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench.json"
188
+ }
189
+ ],
190
+ "name": "legalbench"
191
+ },
192
+ {
193
+ "title": "subset: abercrombie",
194
+ "header": [
195
+ {
196
+ "value": "Model",
197
+ "markdown": false,
198
+ "metadata": {}
199
+ },
200
+ {
201
+ "value": "EM",
202
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.",
203
+ "markdown": false,
204
+ "lower_is_better": false,
205
+ "metadata": {
206
+ "metric": "EM",
207
+ "run_group": "LegalBench"
208
+ }
209
+ },
210
+ {
211
+ "value": "Observed inference time (s)",
212
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
213
+ "markdown": false,
214
+ "lower_is_better": true,
215
+ "metadata": {
216
+ "metric": "Observed inference time (s)",
217
+ "run_group": "LegalBench"
218
+ }
219
+ },
220
+ {
221
+ "value": "# eval",
222
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# eval: Number of evaluation instances.",
223
+ "markdown": false,
224
+ "metadata": {
225
+ "metric": "# eval",
226
+ "run_group": "LegalBench"
227
+ }
228
+ },
229
+ {
230
+ "value": "# train",
231
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
232
+ "markdown": false,
233
+ "metadata": {
234
+ "metric": "# train",
235
+ "run_group": "LegalBench"
236
+ }
237
+ },
238
+ {
239
+ "value": "truncated",
240
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
241
+ "markdown": false,
242
+ "metadata": {
243
+ "metric": "truncated",
244
+ "run_group": "LegalBench"
245
+ }
246
+ },
247
+ {
248
+ "value": "# prompt tokens",
249
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
250
+ "markdown": false,
251
+ "metadata": {
252
+ "metric": "# prompt tokens",
253
+ "run_group": "LegalBench"
254
+ }
255
+ },
256
+ {
257
+ "value": "# output tokens",
258
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# output tokens: Actual number of output tokens.",
259
+ "markdown": false,
260
+ "metadata": {
261
+ "metric": "# output tokens",
262
+ "run_group": "LegalBench"
263
+ }
264
+ }
265
+ ],
266
+ "rows": [
267
+ [
268
+ {
269
+ "value": "EleutherAI/pythia-2.8b",
270
+ "description": "",
271
+ "href": "?group=legalbench&subgroup=subset%3A%20abercrombie&runSpecs=%5B%22legalbench%3Asubset%3Dabercrombie%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
272
+ "markdown": false,
273
+ "run_spec_names": [
274
+ "legalbench:subset=abercrombie,model=EleutherAI_pythia-2.8b"
275
+ ]
276
+ },
277
+ {
278
+ "value": 0.2,
279
+ "description": "min=0.2, mean=0.2, max=0.2, sum=0.2 (1)",
280
+ "style": {
281
+ "font-weight": "bold"
282
+ },
283
+ "markdown": false
284
+ },
285
+ {
286
+ "value": 0.1405831638135408,
287
+ "description": "min=0.141, mean=0.141, max=0.141, sum=0.141 (1)",
288
+ "style": {
289
+ "font-weight": "bold"
290
+ },
291
+ "markdown": false
292
+ },
293
+ {
294
+ "value": 95.0,
295
+ "description": "min=95, mean=95, max=95, sum=95 (1)",
296
+ "style": {},
297
+ "markdown": false
298
+ },
299
+ {
300
+ "value": 5.0,
301
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
302
+ "style": {},
303
+ "markdown": false
304
+ },
305
+ {
306
+ "value": 0.0,
307
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
308
+ "style": {},
309
+ "markdown": false
310
+ },
311
+ {
312
+ "value": 206.77894736842106,
313
+ "description": "min=206.779, mean=206.779, max=206.779, sum=206.779 (1)",
314
+ "style": {},
315
+ "markdown": false
316
+ },
317
+ {
318
+ "value": 1.0,
319
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
320
+ "style": {},
321
+ "markdown": false
322
+ }
323
+ ]
324
+ ],
325
+ "links": [
326
+ {
327
+ "text": "LaTeX",
328
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:abercrombie.tex"
329
+ },
330
+ {
331
+ "text": "JSON",
332
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:abercrombie.json"
333
+ }
334
+ ],
335
+ "name": "legalbench_subset:abercrombie"
336
+ },
337
+ {
338
+ "title": "subset: corporate_lobbying",
339
+ "header": [
340
+ {
341
+ "value": "Model",
342
+ "markdown": false,
343
+ "metadata": {}
344
+ },
345
+ {
346
+ "value": "EM",
347
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.",
348
+ "markdown": false,
349
+ "lower_is_better": false,
350
+ "metadata": {
351
+ "metric": "EM",
352
+ "run_group": "LegalBench"
353
+ }
354
+ },
355
+ {
356
+ "value": "Observed inference time (s)",
357
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
358
+ "markdown": false,
359
+ "lower_is_better": true,
360
+ "metadata": {
361
+ "metric": "Observed inference time (s)",
362
+ "run_group": "LegalBench"
363
+ }
364
+ },
365
+ {
366
+ "value": "# eval",
367
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# eval: Number of evaluation instances.",
368
+ "markdown": false,
369
+ "metadata": {
370
+ "metric": "# eval",
371
+ "run_group": "LegalBench"
372
+ }
373
+ },
374
+ {
375
+ "value": "# train",
376
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
377
+ "markdown": false,
378
+ "metadata": {
379
+ "metric": "# train",
380
+ "run_group": "LegalBench"
381
+ }
382
+ },
383
+ {
384
+ "value": "truncated",
385
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
386
+ "markdown": false,
387
+ "metadata": {
388
+ "metric": "truncated",
389
+ "run_group": "LegalBench"
390
+ }
391
+ },
392
+ {
393
+ "value": "# prompt tokens",
394
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
395
+ "markdown": false,
396
+ "metadata": {
397
+ "metric": "# prompt tokens",
398
+ "run_group": "LegalBench"
399
+ }
400
+ },
401
+ {
402
+ "value": "# output tokens",
403
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# output tokens: Actual number of output tokens.",
404
+ "markdown": false,
405
+ "metadata": {
406
+ "metric": "# output tokens",
407
+ "run_group": "LegalBench"
408
+ }
409
+ }
410
+ ],
411
+ "rows": [
412
+ [
413
+ {
414
+ "value": "EleutherAI/pythia-2.8b",
415
+ "description": "",
416
+ "href": "?group=legalbench&subgroup=subset%3A%20corporate_lobbying&runSpecs=%5B%22legalbench%3Asubset%3Dcorporate_lobbying%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
417
+ "markdown": false,
418
+ "run_spec_names": [
419
+ "legalbench:subset=corporate_lobbying,model=EleutherAI_pythia-2.8b"
420
+ ]
421
+ },
422
+ {
423
+ "value": 0.2,
424
+ "description": "min=0.2, mean=0.2, max=0.2, sum=0.2 (1)",
425
+ "style": {
426
+ "font-weight": "bold"
427
+ },
428
+ "markdown": false
429
+ },
430
+ {
431
+ "value": 0.7949594166814065,
432
+ "description": "min=0.795, mean=0.795, max=0.795, sum=0.795 (1)",
433
+ "style": {
434
+ "font-weight": "bold"
435
+ },
436
+ "markdown": false
437
+ },
438
+ {
439
+ "value": 490.0,
440
+ "description": "min=490, mean=490, max=490, sum=490 (1)",
441
+ "style": {},
442
+ "markdown": false
443
+ },
444
+ {
445
+ "value": 0.3020408163265306,
446
+ "description": "min=0.302, mean=0.302, max=0.302, sum=0.302 (1)",
447
+ "style": {},
448
+ "markdown": false
449
+ },
450
+ {
451
+ "value": 0.014285714285714285,
452
+ "description": "min=0.014, mean=0.014, max=0.014, sum=0.014 (1)",
453
+ "style": {},
454
+ "markdown": false
455
+ },
456
+ {
457
+ "value": 1497.4551020408164,
458
+ "description": "min=1497.455, mean=1497.455, max=1497.455, sum=1497.455 (1)",
459
+ "style": {},
460
+ "markdown": false
461
+ },
462
+ {
463
+ "value": 4.124489795918367,
464
+ "description": "min=4.124, mean=4.124, max=4.124, sum=4.124 (1)",
465
+ "style": {},
466
+ "markdown": false
467
+ }
468
+ ]
469
+ ],
470
+ "links": [
471
+ {
472
+ "text": "LaTeX",
473
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:corporate_lobbying.tex"
474
+ },
475
+ {
476
+ "text": "JSON",
477
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:corporate_lobbying.json"
478
+ }
479
+ ],
480
+ "name": "legalbench_subset:corporate_lobbying"
481
+ },
482
+ {
483
+ "title": "subset: function_of_decision_section",
484
+ "header": [
485
+ {
486
+ "value": "Model",
487
+ "markdown": false,
488
+ "metadata": {}
489
+ },
490
+ {
491
+ "value": "EM",
492
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.",
493
+ "markdown": false,
494
+ "lower_is_better": false,
495
+ "metadata": {
496
+ "metric": "EM",
497
+ "run_group": "LegalBench"
498
+ }
499
+ },
500
+ {
501
+ "value": "Observed inference time (s)",
502
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
503
+ "markdown": false,
504
+ "lower_is_better": true,
505
+ "metadata": {
506
+ "metric": "Observed inference time (s)",
507
+ "run_group": "LegalBench"
508
+ }
509
+ },
510
+ {
511
+ "value": "# eval",
512
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# eval: Number of evaluation instances.",
513
+ "markdown": false,
514
+ "metadata": {
515
+ "metric": "# eval",
516
+ "run_group": "LegalBench"
517
+ }
518
+ },
519
+ {
520
+ "value": "# train",
521
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
522
+ "markdown": false,
523
+ "metadata": {
524
+ "metric": "# train",
525
+ "run_group": "LegalBench"
526
+ }
527
+ },
528
+ {
529
+ "value": "truncated",
530
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
531
+ "markdown": false,
532
+ "metadata": {
533
+ "metric": "truncated",
534
+ "run_group": "LegalBench"
535
+ }
536
+ },
537
+ {
538
+ "value": "# prompt tokens",
539
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
540
+ "markdown": false,
541
+ "metadata": {
542
+ "metric": "# prompt tokens",
543
+ "run_group": "LegalBench"
544
+ }
545
+ },
546
+ {
547
+ "value": "# output tokens",
548
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# output tokens: Actual number of output tokens.",
549
+ "markdown": false,
550
+ "metadata": {
551
+ "metric": "# output tokens",
552
+ "run_group": "LegalBench"
553
+ }
554
+ }
555
+ ],
556
+ "rows": [
557
+ [
558
+ {
559
+ "value": "EleutherAI/pythia-2.8b",
560
+ "description": "",
561
+ "href": "?group=legalbench&subgroup=subset%3A%20function_of_decision_section&runSpecs=%5B%22legalbench%3Asubset%3Dfunction_of_decision_section%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
562
+ "markdown": false,
563
+ "run_spec_names": [
564
+ "legalbench:subset=function_of_decision_section,model=EleutherAI_pythia-2.8b"
565
+ ]
566
+ },
567
+ {
568
+ "value": 0.1444141689373297,
569
+ "description": "min=0.144, mean=0.144, max=0.144, sum=0.144 (1)",
570
+ "style": {
571
+ "font-weight": "bold"
572
+ },
573
+ "markdown": false
574
+ },
575
+ {
576
+ "value": 0.2725935534495424,
577
+ "description": "min=0.273, mean=0.273, max=0.273, sum=0.273 (1)",
578
+ "style": {
579
+ "font-weight": "bold"
580
+ },
581
+ "markdown": false
582
+ },
583
+ {
584
+ "value": 367.0,
585
+ "description": "min=367, mean=367, max=367, sum=367 (1)",
586
+ "style": {},
587
+ "markdown": false
588
+ },
589
+ {
590
+ "value": 5.0,
591
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
592
+ "style": {},
593
+ "markdown": false
594
+ },
595
+ {
596
+ "value": 0.0,
597
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
598
+ "style": {},
599
+ "markdown": false
600
+ },
601
+ {
602
+ "value": 514.9182561307902,
603
+ "description": "min=514.918, mean=514.918, max=514.918, sum=514.918 (1)",
604
+ "style": {},
605
+ "markdown": false
606
+ },
607
+ {
608
+ "value": 1.0844686648501363,
609
+ "description": "min=1.084, mean=1.084, max=1.084, sum=1.084 (1)",
610
+ "style": {},
611
+ "markdown": false
612
+ }
613
+ ]
614
+ ],
615
+ "links": [
616
+ {
617
+ "text": "LaTeX",
618
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:function_of_decision_section.tex"
619
+ },
620
+ {
621
+ "text": "JSON",
622
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:function_of_decision_section.json"
623
+ }
624
+ ],
625
+ "name": "legalbench_subset:function_of_decision_section"
626
+ },
627
+ {
628
+ "title": "subset: international_citizenship_questions",
629
+ "header": [
630
+ {
631
+ "value": "Model",
632
+ "markdown": false,
633
+ "metadata": {}
634
+ },
635
+ {
636
+ "value": "EM",
637
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.",
638
+ "markdown": false,
639
+ "lower_is_better": false,
640
+ "metadata": {
641
+ "metric": "EM",
642
+ "run_group": "LegalBench"
643
+ }
644
+ },
645
+ {
646
+ "value": "Observed inference time (s)",
647
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
648
+ "markdown": false,
649
+ "lower_is_better": true,
650
+ "metadata": {
651
+ "metric": "Observed inference time (s)",
652
+ "run_group": "LegalBench"
653
+ }
654
+ },
655
+ {
656
+ "value": "# eval",
657
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# eval: Number of evaluation instances.",
658
+ "markdown": false,
659
+ "metadata": {
660
+ "metric": "# eval",
661
+ "run_group": "LegalBench"
662
+ }
663
+ },
664
+ {
665
+ "value": "# train",
666
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
667
+ "markdown": false,
668
+ "metadata": {
669
+ "metric": "# train",
670
+ "run_group": "LegalBench"
671
+ }
672
+ },
673
+ {
674
+ "value": "truncated",
675
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
676
+ "markdown": false,
677
+ "metadata": {
678
+ "metric": "truncated",
679
+ "run_group": "LegalBench"
680
+ }
681
+ },
682
+ {
683
+ "value": "# prompt tokens",
684
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
685
+ "markdown": false,
686
+ "metadata": {
687
+ "metric": "# prompt tokens",
688
+ "run_group": "LegalBench"
689
+ }
690
+ },
691
+ {
692
+ "value": "# output tokens",
693
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# output tokens: Actual number of output tokens.",
694
+ "markdown": false,
695
+ "metadata": {
696
+ "metric": "# output tokens",
697
+ "run_group": "LegalBench"
698
+ }
699
+ }
700
+ ],
701
+ "rows": [
702
+ [
703
+ {
704
+ "value": "EleutherAI/pythia-2.8b",
705
+ "description": "",
706
+ "href": "?group=legalbench&subgroup=subset%3A%20international_citizenship_questions&runSpecs=%5B%22legalbench%3Asubset%3Dinternational_citizenship_questions%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
707
+ "markdown": false,
708
+ "run_spec_names": [
709
+ "legalbench:subset=international_citizenship_questions,model=EleutherAI_pythia-2.8b"
710
+ ]
711
+ },
712
+ {
713
+ "value": 0.619,
714
+ "description": "min=0.619, mean=0.619, max=0.619, sum=0.619 (1)",
715
+ "style": {
716
+ "font-weight": "bold"
717
+ },
718
+ "markdown": false
719
+ },
720
+ {
721
+ "value": 0.15224194407463074,
722
+ "description": "min=0.152, mean=0.152, max=0.152, sum=0.152 (1)",
723
+ "style": {
724
+ "font-weight": "bold"
725
+ },
726
+ "markdown": false
727
+ },
728
+ {
729
+ "value": 1000.0,
730
+ "description": "min=1000, mean=1000, max=1000, sum=1000 (1)",
731
+ "style": {},
732
+ "markdown": false
733
+ },
734
+ {
735
+ "value": 4.0,
736
+ "description": "min=4, mean=4, max=4, sum=4 (1)",
737
+ "style": {},
738
+ "markdown": false
739
+ },
740
+ {
741
+ "value": 0.0,
742
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
743
+ "style": {},
744
+ "markdown": false
745
+ },
746
+ {
747
+ "value": 250.447,
748
+ "description": "min=250.447, mean=250.447, max=250.447, sum=250.447 (1)",
749
+ "style": {},
750
+ "markdown": false
751
+ },
752
+ {
753
+ "value": 1.0,
754
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
755
+ "style": {},
756
+ "markdown": false
757
+ }
758
+ ]
759
+ ],
760
+ "links": [
761
+ {
762
+ "text": "LaTeX",
763
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:international_citizenship_questions.tex"
764
+ },
765
+ {
766
+ "text": "JSON",
767
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:international_citizenship_questions.json"
768
+ }
769
+ ],
770
+ "name": "legalbench_subset:international_citizenship_questions"
771
+ },
772
+ {
773
+ "title": "subset: proa",
774
+ "header": [
775
+ {
776
+ "value": "Model",
777
+ "markdown": false,
778
+ "metadata": {}
779
+ },
780
+ {
781
+ "value": "EM",
782
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.",
783
+ "markdown": false,
784
+ "lower_is_better": false,
785
+ "metadata": {
786
+ "metric": "EM",
787
+ "run_group": "LegalBench"
788
+ }
789
+ },
790
+ {
791
+ "value": "Observed inference time (s)",
792
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
793
+ "markdown": false,
794
+ "lower_is_better": true,
795
+ "metadata": {
796
+ "metric": "Observed inference time (s)",
797
+ "run_group": "LegalBench"
798
+ }
799
+ },
800
+ {
801
+ "value": "# eval",
802
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# eval: Number of evaluation instances.",
803
+ "markdown": false,
804
+ "metadata": {
805
+ "metric": "# eval",
806
+ "run_group": "LegalBench"
807
+ }
808
+ },
809
+ {
810
+ "value": "# train",
811
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
812
+ "markdown": false,
813
+ "metadata": {
814
+ "metric": "# train",
815
+ "run_group": "LegalBench"
816
+ }
817
+ },
818
+ {
819
+ "value": "truncated",
820
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
821
+ "markdown": false,
822
+ "metadata": {
823
+ "metric": "truncated",
824
+ "run_group": "LegalBench"
825
+ }
826
+ },
827
+ {
828
+ "value": "# prompt tokens",
829
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
830
+ "markdown": false,
831
+ "metadata": {
832
+ "metric": "# prompt tokens",
833
+ "run_group": "LegalBench"
834
+ }
835
+ },
836
+ {
837
+ "value": "# output tokens",
838
+ "description": "LegalBench is a large collaboratively constructed benchmark of legal reasoning tasks [(Guha et al, 2023)](https://arxiv.org/pdf/2308.11462.pdf).\n\n# output tokens: Actual number of output tokens.",
839
+ "markdown": false,
840
+ "metadata": {
841
+ "metric": "# output tokens",
842
+ "run_group": "LegalBench"
843
+ }
844
+ }
845
+ ],
846
+ "rows": [
847
+ [
848
+ {
849
+ "value": "EleutherAI/pythia-2.8b",
850
+ "description": "",
851
+ "href": "?group=legalbench&subgroup=subset%3A%20proa&runSpecs=%5B%22legalbench%3Asubset%3Dproa%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
852
+ "markdown": false,
853
+ "run_spec_names": [
854
+ "legalbench:subset=proa,model=EleutherAI_pythia-2.8b"
855
+ ]
856
+ },
857
+ {
858
+ "value": 0.43157894736842106,
859
+ "description": "min=0.432, mean=0.432, max=0.432, sum=0.432 (1)",
860
+ "style": {
861
+ "font-weight": "bold"
862
+ },
863
+ "markdown": false
864
+ },
865
+ {
866
+ "value": 0.18680494459051836,
867
+ "description": "min=0.187, mean=0.187, max=0.187, sum=0.187 (1)",
868
+ "style": {
869
+ "font-weight": "bold"
870
+ },
871
+ "markdown": false
872
+ },
873
+ {
874
+ "value": 95.0,
875
+ "description": "min=95, mean=95, max=95, sum=95 (1)",
876
+ "style": {},
877
+ "markdown": false
878
+ },
879
+ {
880
+ "value": 5.0,
881
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
882
+ "style": {},
883
+ "markdown": false
884
+ },
885
+ {
886
+ "value": 0.0,
887
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
888
+ "style": {},
889
+ "markdown": false
890
+ },
891
+ {
892
+ "value": 333.62105263157895,
893
+ "description": "min=333.621, mean=333.621, max=333.621, sum=333.621 (1)",
894
+ "style": {},
895
+ "markdown": false
896
+ },
897
+ {
898
+ "value": 1.0,
899
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
900
+ "style": {},
901
+ "markdown": false
902
+ }
903
+ ]
904
+ ],
905
+ "links": [
906
+ {
907
+ "text": "LaTeX",
908
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/legalbench_legalbench_subset:proa.tex"
909
+ },
910
+ {
911
+ "text": "JSON",
912
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/legalbench_legalbench_subset:proa.json"
913
+ }
914
+ ],
915
+ "name": "legalbench_subset:proa"
916
+ }
917
+ ]
lite_pythia-2.8b-step5000/groups/mmlu.json ADDED
@@ -0,0 +1,917 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "title": "MMLU (Massive Multitask Language Understanding)",
4
+ "header": [
5
+ {
6
+ "value": "Model",
7
+ "markdown": false,
8
+ "metadata": {}
9
+ },
10
+ {
11
+ "value": "EM",
12
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
13
+ "markdown": false,
14
+ "lower_is_better": false,
15
+ "metadata": {
16
+ "metric": "EM",
17
+ "run_group": "MMLU"
18
+ }
19
+ },
20
+ {
21
+ "value": "Observed inference time (s)",
22
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
23
+ "markdown": false,
24
+ "lower_is_better": true,
25
+ "metadata": {
26
+ "metric": "Observed inference time (s)",
27
+ "run_group": "MMLU"
28
+ }
29
+ },
30
+ {
31
+ "value": "# eval",
32
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
33
+ "markdown": false,
34
+ "metadata": {
35
+ "metric": "# eval",
36
+ "run_group": "MMLU"
37
+ }
38
+ },
39
+ {
40
+ "value": "# train",
41
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
42
+ "markdown": false,
43
+ "metadata": {
44
+ "metric": "# train",
45
+ "run_group": "MMLU"
46
+ }
47
+ },
48
+ {
49
+ "value": "truncated",
50
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
51
+ "markdown": false,
52
+ "metadata": {
53
+ "metric": "truncated",
54
+ "run_group": "MMLU"
55
+ }
56
+ },
57
+ {
58
+ "value": "# prompt tokens",
59
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
60
+ "markdown": false,
61
+ "metadata": {
62
+ "metric": "# prompt tokens",
63
+ "run_group": "MMLU"
64
+ }
65
+ },
66
+ {
67
+ "value": "# output tokens",
68
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
69
+ "markdown": false,
70
+ "metadata": {
71
+ "metric": "# output tokens",
72
+ "run_group": "MMLU"
73
+ }
74
+ }
75
+ ],
76
+ "rows": [
77
+ [
78
+ {
79
+ "value": "EleutherAI/pythia-2.8b",
80
+ "description": "",
81
+ "markdown": false
82
+ },
83
+ {
84
+ "value": 0.27487719298245616,
85
+ "description": "min=0.25, mean=0.275, max=0.32, sum=1.374 (5)",
86
+ "style": {
87
+ "font-weight": "bold"
88
+ },
89
+ "markdown": false,
90
+ "run_spec_names": [
91
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
92
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
93
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
94
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
95
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
96
+ ]
97
+ },
98
+ {
99
+ "value": 0.22061018314696193,
100
+ "description": "min=0.173, mean=0.221, max=0.282, sum=1.103 (5)",
101
+ "style": {
102
+ "font-weight": "bold"
103
+ },
104
+ "markdown": false,
105
+ "run_spec_names": [
106
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
107
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
108
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
109
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
110
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
111
+ ]
112
+ },
113
+ {
114
+ "value": 102.8,
115
+ "description": "min=100, mean=102.8, max=114, sum=514 (5)",
116
+ "style": {},
117
+ "markdown": false,
118
+ "run_spec_names": [
119
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
120
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
121
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
122
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
123
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
124
+ ]
125
+ },
126
+ {
127
+ "value": 5.0,
128
+ "description": "min=5, mean=5, max=5, sum=25 (5)",
129
+ "style": {},
130
+ "markdown": false,
131
+ "run_spec_names": [
132
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
133
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
134
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
135
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
136
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
137
+ ]
138
+ },
139
+ {
140
+ "value": 0.0,
141
+ "description": "min=0, mean=0, max=0, sum=0 (5)",
142
+ "style": {},
143
+ "markdown": false,
144
+ "run_spec_names": [
145
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
146
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
147
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
148
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
149
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
150
+ ]
151
+ },
152
+ {
153
+ "value": 467.935649122807,
154
+ "description": "min=358.76, mean=467.936, max=612.798, sum=2339.678 (5)",
155
+ "style": {},
156
+ "markdown": false,
157
+ "run_spec_names": [
158
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
159
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
160
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
161
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
162
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
163
+ ]
164
+ },
165
+ {
166
+ "value": 1.0,
167
+ "description": "min=1, mean=1, max=1, sum=5 (5)",
168
+ "style": {},
169
+ "markdown": false,
170
+ "run_spec_names": [
171
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
172
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
173
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
174
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b",
175
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
176
+ ]
177
+ }
178
+ ]
179
+ ],
180
+ "links": [
181
+ {
182
+ "text": "LaTeX",
183
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu.tex"
184
+ },
185
+ {
186
+ "text": "JSON",
187
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu.json"
188
+ }
189
+ ],
190
+ "name": "mmlu"
191
+ },
192
+ {
193
+ "title": "subject: abstract_algebra",
194
+ "header": [
195
+ {
196
+ "value": "Model",
197
+ "markdown": false,
198
+ "metadata": {}
199
+ },
200
+ {
201
+ "value": "EM",
202
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
203
+ "markdown": false,
204
+ "lower_is_better": false,
205
+ "metadata": {
206
+ "metric": "EM",
207
+ "run_group": "MMLU"
208
+ }
209
+ },
210
+ {
211
+ "value": "Observed inference time (s)",
212
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
213
+ "markdown": false,
214
+ "lower_is_better": true,
215
+ "metadata": {
216
+ "metric": "Observed inference time (s)",
217
+ "run_group": "MMLU"
218
+ }
219
+ },
220
+ {
221
+ "value": "# eval",
222
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
223
+ "markdown": false,
224
+ "metadata": {
225
+ "metric": "# eval",
226
+ "run_group": "MMLU"
227
+ }
228
+ },
229
+ {
230
+ "value": "# train",
231
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
232
+ "markdown": false,
233
+ "metadata": {
234
+ "metric": "# train",
235
+ "run_group": "MMLU"
236
+ }
237
+ },
238
+ {
239
+ "value": "truncated",
240
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
241
+ "markdown": false,
242
+ "metadata": {
243
+ "metric": "truncated",
244
+ "run_group": "MMLU"
245
+ }
246
+ },
247
+ {
248
+ "value": "# prompt tokens",
249
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
250
+ "markdown": false,
251
+ "metadata": {
252
+ "metric": "# prompt tokens",
253
+ "run_group": "MMLU"
254
+ }
255
+ },
256
+ {
257
+ "value": "# output tokens",
258
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
259
+ "markdown": false,
260
+ "metadata": {
261
+ "metric": "# output tokens",
262
+ "run_group": "MMLU"
263
+ }
264
+ }
265
+ ],
266
+ "rows": [
267
+ [
268
+ {
269
+ "value": "EleutherAI/pythia-2.8b",
270
+ "description": "",
271
+ "href": "?group=mmlu&subgroup=subject%3A%20abstract_algebra&runSpecs=%5B%22mmlu%3Asubject%3Dabstract_algebra%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
272
+ "markdown": false,
273
+ "run_spec_names": [
274
+ "mmlu:subject=abstract_algebra,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
275
+ ]
276
+ },
277
+ {
278
+ "value": 0.27,
279
+ "description": "min=0.27, mean=0.27, max=0.27, sum=0.27 (1)",
280
+ "style": {
281
+ "font-weight": "bold"
282
+ },
283
+ "markdown": false
284
+ },
285
+ {
286
+ "value": 0.17254346132278442,
287
+ "description": "min=0.173, mean=0.173, max=0.173, sum=0.173 (1)",
288
+ "style": {
289
+ "font-weight": "bold"
290
+ },
291
+ "markdown": false
292
+ },
293
+ {
294
+ "value": 100.0,
295
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
296
+ "style": {},
297
+ "markdown": false
298
+ },
299
+ {
300
+ "value": 5.0,
301
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
302
+ "style": {},
303
+ "markdown": false
304
+ },
305
+ {
306
+ "value": 0.0,
307
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
308
+ "style": {},
309
+ "markdown": false
310
+ },
311
+ {
312
+ "value": 358.76,
313
+ "description": "min=358.76, mean=358.76, max=358.76, sum=358.76 (1)",
314
+ "style": {},
315
+ "markdown": false
316
+ },
317
+ {
318
+ "value": 1.0,
319
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
320
+ "style": {},
321
+ "markdown": false
322
+ }
323
+ ]
324
+ ],
325
+ "links": [
326
+ {
327
+ "text": "LaTeX",
328
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:abstract_algebra.tex"
329
+ },
330
+ {
331
+ "text": "JSON",
332
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:abstract_algebra.json"
333
+ }
334
+ ],
335
+ "name": "mmlu_subject:abstract_algebra"
336
+ },
337
+ {
338
+ "title": "subject: college_chemistry",
339
+ "header": [
340
+ {
341
+ "value": "Model",
342
+ "markdown": false,
343
+ "metadata": {}
344
+ },
345
+ {
346
+ "value": "EM",
347
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
348
+ "markdown": false,
349
+ "lower_is_better": false,
350
+ "metadata": {
351
+ "metric": "EM",
352
+ "run_group": "MMLU"
353
+ }
354
+ },
355
+ {
356
+ "value": "Observed inference time (s)",
357
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
358
+ "markdown": false,
359
+ "lower_is_better": true,
360
+ "metadata": {
361
+ "metric": "Observed inference time (s)",
362
+ "run_group": "MMLU"
363
+ }
364
+ },
365
+ {
366
+ "value": "# eval",
367
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
368
+ "markdown": false,
369
+ "metadata": {
370
+ "metric": "# eval",
371
+ "run_group": "MMLU"
372
+ }
373
+ },
374
+ {
375
+ "value": "# train",
376
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
377
+ "markdown": false,
378
+ "metadata": {
379
+ "metric": "# train",
380
+ "run_group": "MMLU"
381
+ }
382
+ },
383
+ {
384
+ "value": "truncated",
385
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
386
+ "markdown": false,
387
+ "metadata": {
388
+ "metric": "truncated",
389
+ "run_group": "MMLU"
390
+ }
391
+ },
392
+ {
393
+ "value": "# prompt tokens",
394
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
395
+ "markdown": false,
396
+ "metadata": {
397
+ "metric": "# prompt tokens",
398
+ "run_group": "MMLU"
399
+ }
400
+ },
401
+ {
402
+ "value": "# output tokens",
403
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
404
+ "markdown": false,
405
+ "metadata": {
406
+ "metric": "# output tokens",
407
+ "run_group": "MMLU"
408
+ }
409
+ }
410
+ ],
411
+ "rows": [
412
+ [
413
+ {
414
+ "value": "EleutherAI/pythia-2.8b",
415
+ "description": "",
416
+ "href": "?group=mmlu&subgroup=subject%3A%20college_chemistry&runSpecs=%5B%22mmlu%3Asubject%3Dcollege_chemistry%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
417
+ "markdown": false,
418
+ "run_spec_names": [
419
+ "mmlu:subject=college_chemistry,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
420
+ ]
421
+ },
422
+ {
423
+ "value": 0.32,
424
+ "description": "min=0.32, mean=0.32, max=0.32, sum=0.32 (1)",
425
+ "style": {
426
+ "font-weight": "bold"
427
+ },
428
+ "markdown": false
429
+ },
430
+ {
431
+ "value": 0.2506118679046631,
432
+ "description": "min=0.251, mean=0.251, max=0.251, sum=0.251 (1)",
433
+ "style": {
434
+ "font-weight": "bold"
435
+ },
436
+ "markdown": false
437
+ },
438
+ {
439
+ "value": 100.0,
440
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
441
+ "style": {},
442
+ "markdown": false
443
+ },
444
+ {
445
+ "value": 5.0,
446
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
447
+ "style": {},
448
+ "markdown": false
449
+ },
450
+ {
451
+ "value": 0.0,
452
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
453
+ "style": {},
454
+ "markdown": false
455
+ },
456
+ {
457
+ "value": 535.85,
458
+ "description": "min=535.85, mean=535.85, max=535.85, sum=535.85 (1)",
459
+ "style": {},
460
+ "markdown": false
461
+ },
462
+ {
463
+ "value": 1.0,
464
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
465
+ "style": {},
466
+ "markdown": false
467
+ }
468
+ ]
469
+ ],
470
+ "links": [
471
+ {
472
+ "text": "LaTeX",
473
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:college_chemistry.tex"
474
+ },
475
+ {
476
+ "text": "JSON",
477
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:college_chemistry.json"
478
+ }
479
+ ],
480
+ "name": "mmlu_subject:college_chemistry"
481
+ },
482
+ {
483
+ "title": "subject: computer_security",
484
+ "header": [
485
+ {
486
+ "value": "Model",
487
+ "markdown": false,
488
+ "metadata": {}
489
+ },
490
+ {
491
+ "value": "EM",
492
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
493
+ "markdown": false,
494
+ "lower_is_better": false,
495
+ "metadata": {
496
+ "metric": "EM",
497
+ "run_group": "MMLU"
498
+ }
499
+ },
500
+ {
501
+ "value": "Observed inference time (s)",
502
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
503
+ "markdown": false,
504
+ "lower_is_better": true,
505
+ "metadata": {
506
+ "metric": "Observed inference time (s)",
507
+ "run_group": "MMLU"
508
+ }
509
+ },
510
+ {
511
+ "value": "# eval",
512
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
513
+ "markdown": false,
514
+ "metadata": {
515
+ "metric": "# eval",
516
+ "run_group": "MMLU"
517
+ }
518
+ },
519
+ {
520
+ "value": "# train",
521
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
522
+ "markdown": false,
523
+ "metadata": {
524
+ "metric": "# train",
525
+ "run_group": "MMLU"
526
+ }
527
+ },
528
+ {
529
+ "value": "truncated",
530
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
531
+ "markdown": false,
532
+ "metadata": {
533
+ "metric": "truncated",
534
+ "run_group": "MMLU"
535
+ }
536
+ },
537
+ {
538
+ "value": "# prompt tokens",
539
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
540
+ "markdown": false,
541
+ "metadata": {
542
+ "metric": "# prompt tokens",
543
+ "run_group": "MMLU"
544
+ }
545
+ },
546
+ {
547
+ "value": "# output tokens",
548
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
549
+ "markdown": false,
550
+ "metadata": {
551
+ "metric": "# output tokens",
552
+ "run_group": "MMLU"
553
+ }
554
+ }
555
+ ],
556
+ "rows": [
557
+ [
558
+ {
559
+ "value": "EleutherAI/pythia-2.8b",
560
+ "description": "",
561
+ "href": "?group=mmlu&subgroup=subject%3A%20computer_security&runSpecs=%5B%22mmlu%3Asubject%3Dcomputer_security%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
562
+ "markdown": false,
563
+ "run_spec_names": [
564
+ "mmlu:subject=computer_security,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
565
+ ]
566
+ },
567
+ {
568
+ "value": 0.25,
569
+ "description": "min=0.25, mean=0.25, max=0.25, sum=0.25 (1)",
570
+ "style": {
571
+ "font-weight": "bold"
572
+ },
573
+ "markdown": false
574
+ },
575
+ {
576
+ "value": 0.18395774602890014,
577
+ "description": "min=0.184, mean=0.184, max=0.184, sum=0.184 (1)",
578
+ "style": {
579
+ "font-weight": "bold"
580
+ },
581
+ "markdown": false
582
+ },
583
+ {
584
+ "value": 100.0,
585
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
586
+ "style": {},
587
+ "markdown": false
588
+ },
589
+ {
590
+ "value": 5.0,
591
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
592
+ "style": {},
593
+ "markdown": false
594
+ },
595
+ {
596
+ "value": 0.0,
597
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
598
+ "style": {},
599
+ "markdown": false
600
+ },
601
+ {
602
+ "value": 388.19,
603
+ "description": "min=388.19, mean=388.19, max=388.19, sum=388.19 (1)",
604
+ "style": {},
605
+ "markdown": false
606
+ },
607
+ {
608
+ "value": 1.0,
609
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
610
+ "style": {},
611
+ "markdown": false
612
+ }
613
+ ]
614
+ ],
615
+ "links": [
616
+ {
617
+ "text": "LaTeX",
618
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:computer_security.tex"
619
+ },
620
+ {
621
+ "text": "JSON",
622
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:computer_security.json"
623
+ }
624
+ ],
625
+ "name": "mmlu_subject:computer_security"
626
+ },
627
+ {
628
+ "title": "subject: econometrics",
629
+ "header": [
630
+ {
631
+ "value": "Model",
632
+ "markdown": false,
633
+ "metadata": {}
634
+ },
635
+ {
636
+ "value": "EM",
637
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
638
+ "markdown": false,
639
+ "lower_is_better": false,
640
+ "metadata": {
641
+ "metric": "EM",
642
+ "run_group": "MMLU"
643
+ }
644
+ },
645
+ {
646
+ "value": "Observed inference time (s)",
647
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
648
+ "markdown": false,
649
+ "lower_is_better": true,
650
+ "metadata": {
651
+ "metric": "Observed inference time (s)",
652
+ "run_group": "MMLU"
653
+ }
654
+ },
655
+ {
656
+ "value": "# eval",
657
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
658
+ "markdown": false,
659
+ "metadata": {
660
+ "metric": "# eval",
661
+ "run_group": "MMLU"
662
+ }
663
+ },
664
+ {
665
+ "value": "# train",
666
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
667
+ "markdown": false,
668
+ "metadata": {
669
+ "metric": "# train",
670
+ "run_group": "MMLU"
671
+ }
672
+ },
673
+ {
674
+ "value": "truncated",
675
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
676
+ "markdown": false,
677
+ "metadata": {
678
+ "metric": "truncated",
679
+ "run_group": "MMLU"
680
+ }
681
+ },
682
+ {
683
+ "value": "# prompt tokens",
684
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
685
+ "markdown": false,
686
+ "metadata": {
687
+ "metric": "# prompt tokens",
688
+ "run_group": "MMLU"
689
+ }
690
+ },
691
+ {
692
+ "value": "# output tokens",
693
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
694
+ "markdown": false,
695
+ "metadata": {
696
+ "metric": "# output tokens",
697
+ "run_group": "MMLU"
698
+ }
699
+ }
700
+ ],
701
+ "rows": [
702
+ [
703
+ {
704
+ "value": "EleutherAI/pythia-2.8b",
705
+ "description": "",
706
+ "href": "?group=mmlu&subgroup=subject%3A%20econometrics&runSpecs=%5B%22mmlu%3Asubject%3Deconometrics%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
707
+ "markdown": false,
708
+ "run_spec_names": [
709
+ "mmlu:subject=econometrics,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
710
+ ]
711
+ },
712
+ {
713
+ "value": 0.2543859649122807,
714
+ "description": "min=0.254, mean=0.254, max=0.254, sum=0.254 (1)",
715
+ "style": {
716
+ "font-weight": "bold"
717
+ },
718
+ "markdown": false
719
+ },
720
+ {
721
+ "value": 0.28234320356134784,
722
+ "description": "min=0.282, mean=0.282, max=0.282, sum=0.282 (1)",
723
+ "style": {
724
+ "font-weight": "bold"
725
+ },
726
+ "markdown": false
727
+ },
728
+ {
729
+ "value": 114.0,
730
+ "description": "min=114, mean=114, max=114, sum=114 (1)",
731
+ "style": {},
732
+ "markdown": false
733
+ },
734
+ {
735
+ "value": 5.0,
736
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
737
+ "style": {},
738
+ "markdown": false
739
+ },
740
+ {
741
+ "value": 0.0,
742
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
743
+ "style": {},
744
+ "markdown": false
745
+ },
746
+ {
747
+ "value": 612.7982456140351,
748
+ "description": "min=612.798, mean=612.798, max=612.798, sum=612.798 (1)",
749
+ "style": {},
750
+ "markdown": false
751
+ },
752
+ {
753
+ "value": 1.0,
754
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
755
+ "style": {},
756
+ "markdown": false
757
+ }
758
+ ]
759
+ ],
760
+ "links": [
761
+ {
762
+ "text": "LaTeX",
763
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:econometrics.tex"
764
+ },
765
+ {
766
+ "text": "JSON",
767
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:econometrics.json"
768
+ }
769
+ ],
770
+ "name": "mmlu_subject:econometrics"
771
+ },
772
+ {
773
+ "title": "subject: us_foreign_policy",
774
+ "header": [
775
+ {
776
+ "value": "Model",
777
+ "markdown": false,
778
+ "metadata": {}
779
+ },
780
+ {
781
+ "value": "EM",
782
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
783
+ "markdown": false,
784
+ "lower_is_better": false,
785
+ "metadata": {
786
+ "metric": "EM",
787
+ "run_group": "MMLU"
788
+ }
789
+ },
790
+ {
791
+ "value": "Observed inference time (s)",
792
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
793
+ "markdown": false,
794
+ "lower_is_better": true,
795
+ "metadata": {
796
+ "metric": "Observed inference time (s)",
797
+ "run_group": "MMLU"
798
+ }
799
+ },
800
+ {
801
+ "value": "# eval",
802
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# eval: Number of evaluation instances.",
803
+ "markdown": false,
804
+ "metadata": {
805
+ "metric": "# eval",
806
+ "run_group": "MMLU"
807
+ }
808
+ },
809
+ {
810
+ "value": "# train",
811
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# train: Number of training instances (e.g., in-context examples).",
812
+ "markdown": false,
813
+ "metadata": {
814
+ "metric": "# train",
815
+ "run_group": "MMLU"
816
+ }
817
+ },
818
+ {
819
+ "value": "truncated",
820
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
821
+ "markdown": false,
822
+ "metadata": {
823
+ "metric": "truncated",
824
+ "run_group": "MMLU"
825
+ }
826
+ },
827
+ {
828
+ "value": "# prompt tokens",
829
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# prompt tokens: Number of tokens in the prompt.",
830
+ "markdown": false,
831
+ "metadata": {
832
+ "metric": "# prompt tokens",
833
+ "run_group": "MMLU"
834
+ }
835
+ },
836
+ {
837
+ "value": "# output tokens",
838
+ "description": "The Massive Multitask Language Understanding (MMLU) benchmark for knowledge-intensive question answering across 57 domains [(Hendrycks et al., 2021)](https://arxiv.org/pdf/2009.03300.pdf).\n\n# output tokens: Actual number of output tokens.",
839
+ "markdown": false,
840
+ "metadata": {
841
+ "metric": "# output tokens",
842
+ "run_group": "MMLU"
843
+ }
844
+ }
845
+ ],
846
+ "rows": [
847
+ [
848
+ {
849
+ "value": "EleutherAI/pythia-2.8b",
850
+ "description": "",
851
+ "href": "?group=mmlu&subgroup=subject%3A%20us_foreign_policy&runSpecs=%5B%22mmlu%3Asubject%3Dus_foreign_policy%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
852
+ "markdown": false,
853
+ "run_spec_names": [
854
+ "mmlu:subject=us_foreign_policy,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
855
+ ]
856
+ },
857
+ {
858
+ "value": 0.28,
859
+ "description": "min=0.28, mean=0.28, max=0.28, sum=0.28 (1)",
860
+ "style": {
861
+ "font-weight": "bold"
862
+ },
863
+ "markdown": false
864
+ },
865
+ {
866
+ "value": 0.21359463691711425,
867
+ "description": "min=0.214, mean=0.214, max=0.214, sum=0.214 (1)",
868
+ "style": {
869
+ "font-weight": "bold"
870
+ },
871
+ "markdown": false
872
+ },
873
+ {
874
+ "value": 100.0,
875
+ "description": "min=100, mean=100, max=100, sum=100 (1)",
876
+ "style": {},
877
+ "markdown": false
878
+ },
879
+ {
880
+ "value": 5.0,
881
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
882
+ "style": {},
883
+ "markdown": false
884
+ },
885
+ {
886
+ "value": 0.0,
887
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
888
+ "style": {},
889
+ "markdown": false
890
+ },
891
+ {
892
+ "value": 444.08,
893
+ "description": "min=444.08, mean=444.08, max=444.08, sum=444.08 (1)",
894
+ "style": {},
895
+ "markdown": false
896
+ },
897
+ {
898
+ "value": 1.0,
899
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
900
+ "style": {},
901
+ "markdown": false
902
+ }
903
+ ]
904
+ ],
905
+ "links": [
906
+ {
907
+ "text": "LaTeX",
908
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/mmlu_mmlu_subject:us_foreign_policy.tex"
909
+ },
910
+ {
911
+ "text": "JSON",
912
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/mmlu_mmlu_subject:us_foreign_policy.json"
913
+ }
914
+ ],
915
+ "name": "mmlu_subject:us_foreign_policy"
916
+ }
917
+ ]
lite_pythia-2.8b-step5000/groups/openbookqa.json ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "title": "",
4
+ "header": [
5
+ {
6
+ "value": "Model",
7
+ "markdown": false,
8
+ "metadata": {}
9
+ },
10
+ {
11
+ "value": "EM",
12
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\nExact match: Fraction of instances that the predicted output matches a correct reference exactly.",
13
+ "markdown": false,
14
+ "lower_is_better": false,
15
+ "metadata": {
16
+ "metric": "EM",
17
+ "run_group": "OpenbookQA"
18
+ }
19
+ },
20
+ {
21
+ "value": "Observed inference time (s)",
22
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\nObserved inference runtime (s): Average observed time to process a request to the model (via an API, and thus depends on particular deployment).",
23
+ "markdown": false,
24
+ "lower_is_better": true,
25
+ "metadata": {
26
+ "metric": "Observed inference time (s)",
27
+ "run_group": "OpenbookQA"
28
+ }
29
+ },
30
+ {
31
+ "value": "# eval",
32
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\n# eval: Number of evaluation instances.",
33
+ "markdown": false,
34
+ "metadata": {
35
+ "metric": "# eval",
36
+ "run_group": "OpenbookQA"
37
+ }
38
+ },
39
+ {
40
+ "value": "# train",
41
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\n# train: Number of training instances (e.g., in-context examples).",
42
+ "markdown": false,
43
+ "metadata": {
44
+ "metric": "# train",
45
+ "run_group": "OpenbookQA"
46
+ }
47
+ },
48
+ {
49
+ "value": "truncated",
50
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).",
51
+ "markdown": false,
52
+ "metadata": {
53
+ "metric": "truncated",
54
+ "run_group": "OpenbookQA"
55
+ }
56
+ },
57
+ {
58
+ "value": "# prompt tokens",
59
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\n# prompt tokens: Number of tokens in the prompt.",
60
+ "markdown": false,
61
+ "metadata": {
62
+ "metric": "# prompt tokens",
63
+ "run_group": "OpenbookQA"
64
+ }
65
+ },
66
+ {
67
+ "value": "# output tokens",
68
+ "description": "The OpenbookQA benchmark for commonsense-intensive open book question answering [(Mihaylov et al., 2018)](https://aclanthology.org/D18-1260/).\n\n# output tokens: Actual number of output tokens.",
69
+ "markdown": false,
70
+ "metadata": {
71
+ "metric": "# output tokens",
72
+ "run_group": "OpenbookQA"
73
+ }
74
+ }
75
+ ],
76
+ "rows": [
77
+ [
78
+ {
79
+ "value": "EleutherAI/pythia-2.8b",
80
+ "description": "",
81
+ "href": "?group=openbookqa&subgroup=&runSpecs=%5B%22commonsense%3Adataset%3Dopenbookqa%2Cmethod%3Dmultiple_choice_joint%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D",
82
+ "markdown": false,
83
+ "run_spec_names": [
84
+ "commonsense:dataset=openbookqa,method=multiple_choice_joint,model=EleutherAI_pythia-2.8b"
85
+ ]
86
+ },
87
+ {
88
+ "value": 0.22,
89
+ "description": "min=0.22, mean=0.22, max=0.22, sum=0.22 (1)",
90
+ "style": {
91
+ "font-weight": "bold"
92
+ },
93
+ "markdown": false
94
+ },
95
+ {
96
+ "value": 0.12883714818954467,
97
+ "description": "min=0.129, mean=0.129, max=0.129, sum=0.129 (1)",
98
+ "style": {
99
+ "font-weight": "bold"
100
+ },
101
+ "markdown": false
102
+ },
103
+ {
104
+ "value": 500.0,
105
+ "description": "min=500, mean=500, max=500, sum=500 (1)",
106
+ "style": {},
107
+ "markdown": false
108
+ },
109
+ {
110
+ "value": 5.0,
111
+ "description": "min=5, mean=5, max=5, sum=5 (1)",
112
+ "style": {},
113
+ "markdown": false
114
+ },
115
+ {
116
+ "value": 0.0,
117
+ "description": "min=0, mean=0, max=0, sum=0 (1)",
118
+ "style": {},
119
+ "markdown": false
120
+ },
121
+ {
122
+ "value": 251.556,
123
+ "description": "min=251.556, mean=251.556, max=251.556, sum=251.556 (1)",
124
+ "style": {},
125
+ "markdown": false
126
+ },
127
+ {
128
+ "value": 1.0,
129
+ "description": "min=1, mean=1, max=1, sum=1 (1)",
130
+ "style": {},
131
+ "markdown": false
132
+ }
133
+ ]
134
+ ],
135
+ "links": [
136
+ {
137
+ "text": "LaTeX",
138
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/latex/openbookqa_openbookqa_.tex"
139
+ },
140
+ {
141
+ "text": "JSON",
142
+ "href": "benchmark_output/runs/lite_pythia-2.8b-step5000/groups/json/openbookqa_openbookqa_.json"
143
+ }
144
+ ],
145
+ "name": "openbookqa_"
146
+ }
147
+ ]