[ { "title": "WikiFact", "header": [ { "value": "Model", "markdown": false, "metadata": {} }, { "value": "EM", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.", "markdown": false, "lower_is_better": false, "metadata": { "metric": "EM", "run_group": "WikiFact" } }, { "value": "Denoised inference time (s)", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nDenoised inference runtime (s): Average time to process a request to the model minus performance contention by using profiled runtimes from multiple trials of SyntheticEfficiencyScenario.", "markdown": false, "lower_is_better": true, "metadata": { "metric": "Denoised inference time (s)", "run_group": "WikiFact" } }, { "value": "# eval", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# eval: Number of evaluation instances.", "markdown": false, "metadata": { "metric": "# eval", "run_group": "WikiFact" } }, { "value": "# train", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# train: Number of training instances (e.g., in-context examples).", "markdown": false, "metadata": { "metric": "# train", "run_group": "WikiFact" } }, { "value": "truncated", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).", "markdown": false, "metadata": { "metric": "truncated", "run_group": "WikiFact" } }, { "value": "# prompt tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# prompt tokens: Number of tokens in the prompt.", "markdown": false, "metadata": { "metric": "# prompt tokens", "run_group": "WikiFact" } }, { "value": "# output tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# output tokens: Actual number of output tokens.", "markdown": false, "metadata": { "metric": "# output tokens", "run_group": "WikiFact" } }, { "value": "# trials", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# trials: Number of trials, where in each trial we choose an independent, random set of training instances.", "markdown": false, "metadata": { "metric": "# trials", "run_group": "WikiFact" } } ], "rows": [ [ { "value": "EleutherAI/pythia-2.8b", "description": "", "markdown": false }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (10)", "style": { "font-weight": "bold" }, "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=author,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=currency,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=discoverer_or_inventor,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=instance_of,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=medical_condition_treated,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=part_of,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=place_of_birth,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=plaintiff,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=position_held,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=symptoms_and_signs,model=EleutherAI_pythia-2.8b" ] }, { "description": "10 matching runs, but no matching metrics", "markdown": false }, { "value": 746.2, "description": "min=96, mean=746.2, max=850, sum=7462 (10)", "style": {}, "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=author,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=currency,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=discoverer_or_inventor,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=instance_of,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=medical_condition_treated,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=part_of,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=place_of_birth,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=plaintiff,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=position_held,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=symptoms_and_signs,model=EleutherAI_pythia-2.8b" ] }, { "value": 5.0, "description": "min=5, mean=5, max=5, sum=50 (10)", "style": {}, "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=author,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=currency,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=discoverer_or_inventor,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=instance_of,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=medical_condition_treated,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=part_of,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=place_of_birth,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=plaintiff,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=position_held,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=symptoms_and_signs,model=EleutherAI_pythia-2.8b" ] }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (10)", "style": {}, "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=author,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=currency,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=discoverer_or_inventor,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=instance_of,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=medical_condition_treated,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=part_of,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=place_of_birth,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=plaintiff,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=position_held,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=symptoms_and_signs,model=EleutherAI_pythia-2.8b" ] }, { "value": 83.9422429501836, "description": "min=65.855, mean=83.942, max=128.292, sum=839.422 (10)", "style": {}, "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=author,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=currency,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=discoverer_or_inventor,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=instance_of,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=medical_condition_treated,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=part_of,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=place_of_birth,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=plaintiff,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=position_held,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=symptoms_and_signs,model=EleutherAI_pythia-2.8b" ] }, { "value": 7.866394763735883, "description": "min=7.823, mean=7.866, max=7.915, sum=78.664 (10)", "style": {}, "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=author,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=currency,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=discoverer_or_inventor,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=instance_of,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=medical_condition_treated,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=part_of,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=place_of_birth,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=plaintiff,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=position_held,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=symptoms_and_signs,model=EleutherAI_pythia-2.8b" ] }, { "value": 1.0, "description": "min=1, mean=1, max=1, sum=10 (10)", "style": {}, "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=author,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=currency,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=discoverer_or_inventor,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=instance_of,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=medical_condition_treated,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=part_of,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=place_of_birth,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=plaintiff,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=position_held,model=EleutherAI_pythia-2.8b", "wikifact:k=5,subject=symptoms_and_signs,model=EleutherAI_pythia-2.8b" ] } ] ], "links": [ { "text": "LaTeX", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/latex/wikifact_wikifact.tex" }, { "text": "JSON", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/json/wikifact_wikifact.json" } ], "name": "wikifact" }, { "title": "subject: author", "header": [ { "value": "Model", "markdown": false, "metadata": {} }, { "value": "EM", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.", "markdown": false, "lower_is_better": false, "metadata": { "metric": "EM", "run_group": "WikiFact" } }, { "value": "Denoised inference time (s)", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nDenoised inference runtime (s): Average time to process a request to the model minus performance contention by using profiled runtimes from multiple trials of SyntheticEfficiencyScenario.", "markdown": false, "lower_is_better": true, "metadata": { "metric": "Denoised inference time (s)", "run_group": "WikiFact" } }, { "value": "# eval", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# eval: Number of evaluation instances.", "markdown": false, "metadata": { "metric": "# eval", "run_group": "WikiFact" } }, { "value": "# train", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# train: Number of training instances (e.g., in-context examples).", "markdown": false, "metadata": { "metric": "# train", "run_group": "WikiFact" } }, { "value": "truncated", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).", "markdown": false, "metadata": { "metric": "truncated", "run_group": "WikiFact" } }, { "value": "# prompt tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# prompt tokens: Number of tokens in the prompt.", "markdown": false, "metadata": { "metric": "# prompt tokens", "run_group": "WikiFact" } }, { "value": "# output tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# output tokens: Actual number of output tokens.", "markdown": false, "metadata": { "metric": "# output tokens", "run_group": "WikiFact" } }, { "value": "# trials", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# trials: Number of trials, where in each trial we choose an independent, random set of training instances.", "markdown": false, "metadata": { "metric": "# trials", "run_group": "WikiFact" } } ], "rows": [ [ { "value": "EleutherAI/pythia-2.8b", "description": "", "href": "?group=wikifact&subgroup=subject%3A%20author&runSpecs=%5B%22wikifact%3Ak%3D5%2Csubject%3Dauthor%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D", "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=author,model=EleutherAI_pythia-2.8b" ] }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": { "font-weight": "bold" }, "markdown": false }, { "description": "1 matching runs, but no matching metrics", "markdown": false }, { "value": 850.0, "description": "min=850, mean=850, max=850, sum=850 (1)", "style": {}, "markdown": false }, { "value": 5.0, "description": "min=5, mean=5, max=5, sum=5 (1)", "style": {}, "markdown": false }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": {}, "markdown": false }, { "value": 77.12235294117647, "description": "min=77.122, mean=77.122, max=77.122, sum=77.122 (1)", "style": {}, "markdown": false }, { "value": 7.83764705882353, "description": "min=7.838, mean=7.838, max=7.838, sum=7.838 (1)", "style": {}, "markdown": false }, { "value": 1.0, "description": "min=1, mean=1, max=1, sum=1 (1)", "style": {}, "markdown": false } ] ], "links": [ { "text": "LaTeX", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/latex/wikifact_wikifact_subject:author.tex" }, { "text": "JSON", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/json/wikifact_wikifact_subject:author.json" } ], "name": "wikifact_subject:author" }, { "title": "subject: currency", "header": [ { "value": "Model", "markdown": false, "metadata": {} }, { "value": "EM", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.", "markdown": false, "lower_is_better": false, "metadata": { "metric": "EM", "run_group": "WikiFact" } }, { "value": "Denoised inference time (s)", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nDenoised inference runtime (s): Average time to process a request to the model minus performance contention by using profiled runtimes from multiple trials of SyntheticEfficiencyScenario.", "markdown": false, "lower_is_better": true, "metadata": { "metric": "Denoised inference time (s)", "run_group": "WikiFact" } }, { "value": "# eval", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# eval: Number of evaluation instances.", "markdown": false, "metadata": { "metric": "# eval", "run_group": "WikiFact" } }, { "value": "# train", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# train: Number of training instances (e.g., in-context examples).", "markdown": false, "metadata": { "metric": "# train", "run_group": "WikiFact" } }, { "value": "truncated", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).", "markdown": false, "metadata": { "metric": "truncated", "run_group": "WikiFact" } }, { "value": "# prompt tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# prompt tokens: Number of tokens in the prompt.", "markdown": false, "metadata": { "metric": "# prompt tokens", "run_group": "WikiFact" } }, { "value": "# output tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# output tokens: Actual number of output tokens.", "markdown": false, "metadata": { "metric": "# output tokens", "run_group": "WikiFact" } }, { "value": "# trials", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# trials: Number of trials, where in each trial we choose an independent, random set of training instances.", "markdown": false, "metadata": { "metric": "# trials", "run_group": "WikiFact" } } ], "rows": [ [ { "value": "EleutherAI/pythia-2.8b", "description": "", "href": "?group=wikifact&subgroup=subject%3A%20currency&runSpecs=%5B%22wikifact%3Ak%3D5%2Csubject%3Dcurrency%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D", "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=currency,model=EleutherAI_pythia-2.8b" ] }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": { "font-weight": "bold" }, "markdown": false }, { "description": "1 matching runs, but no matching metrics", "markdown": false }, { "value": 850.0, "description": "min=850, mean=850, max=850, sum=850 (1)", "style": {}, "markdown": false }, { "value": 5.0, "description": "min=5, mean=5, max=5, sum=5 (1)", "style": {}, "markdown": false }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": {}, "markdown": false }, { "value": 67.89294117647059, "description": "min=67.893, mean=67.893, max=67.893, sum=67.893 (1)", "style": {}, "markdown": false }, { "value": 7.857647058823529, "description": "min=7.858, mean=7.858, max=7.858, sum=7.858 (1)", "style": {}, "markdown": false }, { "value": 1.0, "description": "min=1, mean=1, max=1, sum=1 (1)", "style": {}, "markdown": false } ] ], "links": [ { "text": "LaTeX", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/latex/wikifact_wikifact_subject:currency.tex" }, { "text": "JSON", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/json/wikifact_wikifact_subject:currency.json" } ], "name": "wikifact_subject:currency" }, { "title": "subject: discoverer_or_inventor", "header": [ { "value": "Model", "markdown": false, "metadata": {} }, { "value": "EM", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.", "markdown": false, "lower_is_better": false, "metadata": { "metric": "EM", "run_group": "WikiFact" } }, { "value": "Denoised inference time (s)", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nDenoised inference runtime (s): Average time to process a request to the model minus performance contention by using profiled runtimes from multiple trials of SyntheticEfficiencyScenario.", "markdown": false, "lower_is_better": true, "metadata": { "metric": "Denoised inference time (s)", "run_group": "WikiFact" } }, { "value": "# eval", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# eval: Number of evaluation instances.", "markdown": false, "metadata": { "metric": "# eval", "run_group": "WikiFact" } }, { "value": "# train", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# train: Number of training instances (e.g., in-context examples).", "markdown": false, "metadata": { "metric": "# train", "run_group": "WikiFact" } }, { "value": "truncated", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).", "markdown": false, "metadata": { "metric": "truncated", "run_group": "WikiFact" } }, { "value": "# prompt tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# prompt tokens: Number of tokens in the prompt.", "markdown": false, "metadata": { "metric": "# prompt tokens", "run_group": "WikiFact" } }, { "value": "# output tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# output tokens: Actual number of output tokens.", "markdown": false, "metadata": { "metric": "# output tokens", "run_group": "WikiFact" } }, { "value": "# trials", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# trials: Number of trials, where in each trial we choose an independent, random set of training instances.", "markdown": false, "metadata": { "metric": "# trials", "run_group": "WikiFact" } } ], "rows": [ [ { "value": "EleutherAI/pythia-2.8b", "description": "", "href": "?group=wikifact&subgroup=subject%3A%20discoverer_or_inventor&runSpecs=%5B%22wikifact%3Ak%3D5%2Csubject%3Ddiscoverer_or_inventor%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D", "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=discoverer_or_inventor,model=EleutherAI_pythia-2.8b" ] }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": { "font-weight": "bold" }, "markdown": false }, { "description": "1 matching runs, but no matching metrics", "markdown": false }, { "value": 850.0, "description": "min=850, mean=850, max=850, sum=850 (1)", "style": {}, "markdown": false }, { "value": 5.0, "description": "min=5, mean=5, max=5, sum=5 (1)", "style": {}, "markdown": false }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": {}, "markdown": false }, { "value": 87.16117647058823, "description": "min=87.161, mean=87.161, max=87.161, sum=87.161 (1)", "style": {}, "markdown": false }, { "value": 7.905882352941177, "description": "min=7.906, mean=7.906, max=7.906, sum=7.906 (1)", "style": {}, "markdown": false }, { "value": 1.0, "description": "min=1, mean=1, max=1, sum=1 (1)", "style": {}, "markdown": false } ] ], "links": [ { "text": "LaTeX", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/latex/wikifact_wikifact_subject:discoverer_or_inventor.tex" }, { "text": "JSON", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/json/wikifact_wikifact_subject:discoverer_or_inventor.json" } ], "name": "wikifact_subject:discoverer_or_inventor" }, { "title": "subject: instance_of", "header": [ { "value": "Model", "markdown": false, "metadata": {} }, { "value": "EM", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.", "markdown": false, "lower_is_better": false, "metadata": { "metric": "EM", "run_group": "WikiFact" } }, { "value": "Denoised inference time (s)", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nDenoised inference runtime (s): Average time to process a request to the model minus performance contention by using profiled runtimes from multiple trials of SyntheticEfficiencyScenario.", "markdown": false, "lower_is_better": true, "metadata": { "metric": "Denoised inference time (s)", "run_group": "WikiFact" } }, { "value": "# eval", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# eval: Number of evaluation instances.", "markdown": false, "metadata": { "metric": "# eval", "run_group": "WikiFact" } }, { "value": "# train", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# train: Number of training instances (e.g., in-context examples).", "markdown": false, "metadata": { "metric": "# train", "run_group": "WikiFact" } }, { "value": "truncated", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).", "markdown": false, "metadata": { "metric": "truncated", "run_group": "WikiFact" } }, { "value": "# prompt tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# prompt tokens: Number of tokens in the prompt.", "markdown": false, "metadata": { "metric": "# prompt tokens", "run_group": "WikiFact" } }, { "value": "# output tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# output tokens: Actual number of output tokens.", "markdown": false, "metadata": { "metric": "# output tokens", "run_group": "WikiFact" } }, { "value": "# trials", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# trials: Number of trials, where in each trial we choose an independent, random set of training instances.", "markdown": false, "metadata": { "metric": "# trials", "run_group": "WikiFact" } } ], "rows": [ [ { "value": "EleutherAI/pythia-2.8b", "description": "", "href": "?group=wikifact&subgroup=subject%3A%20instance_of&runSpecs=%5B%22wikifact%3Ak%3D5%2Csubject%3Dinstance_of%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D", "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=instance_of,model=EleutherAI_pythia-2.8b" ] }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": { "font-weight": "bold" }, "markdown": false }, { "description": "1 matching runs, but no matching metrics", "markdown": false }, { "value": 850.0, "description": "min=850, mean=850, max=850, sum=850 (1)", "style": {}, "markdown": false }, { "value": 5.0, "description": "min=5, mean=5, max=5, sum=5 (1)", "style": {}, "markdown": false }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": {}, "markdown": false }, { "value": 72.06117647058824, "description": "min=72.061, mean=72.061, max=72.061, sum=72.061 (1)", "style": {}, "markdown": false }, { "value": 7.867058823529412, "description": "min=7.867, mean=7.867, max=7.867, sum=7.867 (1)", "style": {}, "markdown": false }, { "value": 1.0, "description": "min=1, mean=1, max=1, sum=1 (1)", "style": {}, "markdown": false } ] ], "links": [ { "text": "LaTeX", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/latex/wikifact_wikifact_subject:instance_of.tex" }, { "text": "JSON", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/json/wikifact_wikifact_subject:instance_of.json" } ], "name": "wikifact_subject:instance_of" }, { "title": "subject: medical_condition_treated", "header": [ { "value": "Model", "markdown": false, "metadata": {} }, { "value": "EM", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.", "markdown": false, "lower_is_better": false, "metadata": { "metric": "EM", "run_group": "WikiFact" } }, { "value": "Denoised inference time (s)", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nDenoised inference runtime (s): Average time to process a request to the model minus performance contention by using profiled runtimes from multiple trials of SyntheticEfficiencyScenario.", "markdown": false, "lower_is_better": true, "metadata": { "metric": "Denoised inference time (s)", "run_group": "WikiFact" } }, { "value": "# eval", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# eval: Number of evaluation instances.", "markdown": false, "metadata": { "metric": "# eval", "run_group": "WikiFact" } }, { "value": "# train", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# train: Number of training instances (e.g., in-context examples).", "markdown": false, "metadata": { "metric": "# train", "run_group": "WikiFact" } }, { "value": "truncated", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).", "markdown": false, "metadata": { "metric": "truncated", "run_group": "WikiFact" } }, { "value": "# prompt tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# prompt tokens: Number of tokens in the prompt.", "markdown": false, "metadata": { "metric": "# prompt tokens", "run_group": "WikiFact" } }, { "value": "# output tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# output tokens: Actual number of output tokens.", "markdown": false, "metadata": { "metric": "# output tokens", "run_group": "WikiFact" } }, { "value": "# trials", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# trials: Number of trials, where in each trial we choose an independent, random set of training instances.", "markdown": false, "metadata": { "metric": "# trials", "run_group": "WikiFact" } } ], "rows": [ [ { "value": "EleutherAI/pythia-2.8b", "description": "", "href": "?group=wikifact&subgroup=subject%3A%20medical_condition_treated&runSpecs=%5B%22wikifact%3Ak%3D5%2Csubject%3Dmedical_condition_treated%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D", "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=medical_condition_treated,model=EleutherAI_pythia-2.8b" ] }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": { "font-weight": "bold" }, "markdown": false }, { "description": "1 matching runs, but no matching metrics", "markdown": false }, { "value": 850.0, "description": "min=850, mean=850, max=850, sum=850 (1)", "style": {}, "markdown": false }, { "value": 5.0, "description": "min=5, mean=5, max=5, sum=5 (1)", "style": {}, "markdown": false }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": {}, "markdown": false }, { "value": 85.21058823529411, "description": "min=85.211, mean=85.211, max=85.211, sum=85.211 (1)", "style": {}, "markdown": false }, { "value": 7.915294117647059, "description": "min=7.915, mean=7.915, max=7.915, sum=7.915 (1)", "style": {}, "markdown": false }, { "value": 1.0, "description": "min=1, mean=1, max=1, sum=1 (1)", "style": {}, "markdown": false } ] ], "links": [ { "text": "LaTeX", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/latex/wikifact_wikifact_subject:medical_condition_treated.tex" }, { "text": "JSON", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/json/wikifact_wikifact_subject:medical_condition_treated.json" } ], "name": "wikifact_subject:medical_condition_treated" }, { "title": "subject: part_of", "header": [ { "value": "Model", "markdown": false, "metadata": {} }, { "value": "EM", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.", "markdown": false, "lower_is_better": false, "metadata": { "metric": "EM", "run_group": "WikiFact" } }, { "value": "Denoised inference time (s)", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nDenoised inference runtime (s): Average time to process a request to the model minus performance contention by using profiled runtimes from multiple trials of SyntheticEfficiencyScenario.", "markdown": false, "lower_is_better": true, "metadata": { "metric": "Denoised inference time (s)", "run_group": "WikiFact" } }, { "value": "# eval", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# eval: Number of evaluation instances.", "markdown": false, "metadata": { "metric": "# eval", "run_group": "WikiFact" } }, { "value": "# train", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# train: Number of training instances (e.g., in-context examples).", "markdown": false, "metadata": { "metric": "# train", "run_group": "WikiFact" } }, { "value": "truncated", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).", "markdown": false, "metadata": { "metric": "truncated", "run_group": "WikiFact" } }, { "value": "# prompt tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# prompt tokens: Number of tokens in the prompt.", "markdown": false, "metadata": { "metric": "# prompt tokens", "run_group": "WikiFact" } }, { "value": "# output tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# output tokens: Actual number of output tokens.", "markdown": false, "metadata": { "metric": "# output tokens", "run_group": "WikiFact" } }, { "value": "# trials", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# trials: Number of trials, where in each trial we choose an independent, random set of training instances.", "markdown": false, "metadata": { "metric": "# trials", "run_group": "WikiFact" } } ], "rows": [ [ { "value": "EleutherAI/pythia-2.8b", "description": "", "href": "?group=wikifact&subgroup=subject%3A%20part_of&runSpecs=%5B%22wikifact%3Ak%3D5%2Csubject%3Dpart_of%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D", "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=part_of,model=EleutherAI_pythia-2.8b" ] }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": { "font-weight": "bold" }, "markdown": false }, { "description": "1 matching runs, but no matching metrics", "markdown": false }, { "value": 850.0, "description": "min=850, mean=850, max=850, sum=850 (1)", "style": {}, "markdown": false }, { "value": 5.0, "description": "min=5, mean=5, max=5, sum=5 (1)", "style": {}, "markdown": false }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": {}, "markdown": false }, { "value": 96.1329411764706, "description": "min=96.133, mean=96.133, max=96.133, sum=96.133 (1)", "style": {}, "markdown": false }, { "value": 7.856470588235294, "description": "min=7.856, mean=7.856, max=7.856, sum=7.856 (1)", "style": {}, "markdown": false }, { "value": 1.0, "description": "min=1, mean=1, max=1, sum=1 (1)", "style": {}, "markdown": false } ] ], "links": [ { "text": "LaTeX", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/latex/wikifact_wikifact_subject:part_of.tex" }, { "text": "JSON", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/json/wikifact_wikifact_subject:part_of.json" } ], "name": "wikifact_subject:part_of" }, { "title": "subject: place_of_birth", "header": [ { "value": "Model", "markdown": false, "metadata": {} }, { "value": "EM", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.", "markdown": false, "lower_is_better": false, "metadata": { "metric": "EM", "run_group": "WikiFact" } }, { "value": "Denoised inference time (s)", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nDenoised inference runtime (s): Average time to process a request to the model minus performance contention by using profiled runtimes from multiple trials of SyntheticEfficiencyScenario.", "markdown": false, "lower_is_better": true, "metadata": { "metric": "Denoised inference time (s)", "run_group": "WikiFact" } }, { "value": "# eval", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# eval: Number of evaluation instances.", "markdown": false, "metadata": { "metric": "# eval", "run_group": "WikiFact" } }, { "value": "# train", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# train: Number of training instances (e.g., in-context examples).", "markdown": false, "metadata": { "metric": "# train", "run_group": "WikiFact" } }, { "value": "truncated", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).", "markdown": false, "metadata": { "metric": "truncated", "run_group": "WikiFact" } }, { "value": "# prompt tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# prompt tokens: Number of tokens in the prompt.", "markdown": false, "metadata": { "metric": "# prompt tokens", "run_group": "WikiFact" } }, { "value": "# output tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# output tokens: Actual number of output tokens.", "markdown": false, "metadata": { "metric": "# output tokens", "run_group": "WikiFact" } }, { "value": "# trials", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# trials: Number of trials, where in each trial we choose an independent, random set of training instances.", "markdown": false, "metadata": { "metric": "# trials", "run_group": "WikiFact" } } ], "rows": [ [ { "value": "EleutherAI/pythia-2.8b", "description": "", "href": "?group=wikifact&subgroup=subject%3A%20place_of_birth&runSpecs=%5B%22wikifact%3Ak%3D5%2Csubject%3Dplace_of_birth%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D", "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=place_of_birth,model=EleutherAI_pythia-2.8b" ] }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": { "font-weight": "bold" }, "markdown": false }, { "description": "1 matching runs, but no matching metrics", "markdown": false }, { "value": 850.0, "description": "min=850, mean=850, max=850, sum=850 (1)", "style": {}, "markdown": false }, { "value": 5.0, "description": "min=5, mean=5, max=5, sum=5 (1)", "style": {}, "markdown": false }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": {}, "markdown": false }, { "value": 65.85529411764706, "description": "min=65.855, mean=65.855, max=65.855, sum=65.855 (1)", "style": {}, "markdown": false }, { "value": 7.881176470588235, "description": "min=7.881, mean=7.881, max=7.881, sum=7.881 (1)", "style": {}, "markdown": false }, { "value": 1.0, "description": "min=1, mean=1, max=1, sum=1 (1)", "style": {}, "markdown": false } ] ], "links": [ { "text": "LaTeX", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/latex/wikifact_wikifact_subject:place_of_birth.tex" }, { "text": "JSON", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/json/wikifact_wikifact_subject:place_of_birth.json" } ], "name": "wikifact_subject:place_of_birth" }, { "title": "subject: plaintiff", "header": [ { "value": "Model", "markdown": false, "metadata": {} }, { "value": "EM", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.", "markdown": false, "lower_is_better": false, "metadata": { "metric": "EM", "run_group": "WikiFact" } }, { "value": "Denoised inference time (s)", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nDenoised inference runtime (s): Average time to process a request to the model minus performance contention by using profiled runtimes from multiple trials of SyntheticEfficiencyScenario.", "markdown": false, "lower_is_better": true, "metadata": { "metric": "Denoised inference time (s)", "run_group": "WikiFact" } }, { "value": "# eval", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# eval: Number of evaluation instances.", "markdown": false, "metadata": { "metric": "# eval", "run_group": "WikiFact" } }, { "value": "# train", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# train: Number of training instances (e.g., in-context examples).", "markdown": false, "metadata": { "metric": "# train", "run_group": "WikiFact" } }, { "value": "truncated", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).", "markdown": false, "metadata": { "metric": "truncated", "run_group": "WikiFact" } }, { "value": "# prompt tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# prompt tokens: Number of tokens in the prompt.", "markdown": false, "metadata": { "metric": "# prompt tokens", "run_group": "WikiFact" } }, { "value": "# output tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# output tokens: Actual number of output tokens.", "markdown": false, "metadata": { "metric": "# output tokens", "run_group": "WikiFact" } }, { "value": "# trials", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# trials: Number of trials, where in each trial we choose an independent, random set of training instances.", "markdown": false, "metadata": { "metric": "# trials", "run_group": "WikiFact" } } ], "rows": [ [ { "value": "EleutherAI/pythia-2.8b", "description": "", "href": "?group=wikifact&subgroup=subject%3A%20plaintiff&runSpecs=%5B%22wikifact%3Ak%3D5%2Csubject%3Dplaintiff%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D", "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=plaintiff,model=EleutherAI_pythia-2.8b" ] }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": { "font-weight": "bold" }, "markdown": false }, { "description": "1 matching runs, but no matching metrics", "markdown": false }, { "value": 96.0, "description": "min=96, mean=96, max=96, sum=96 (1)", "style": {}, "markdown": false }, { "value": 5.0, "description": "min=5, mean=5, max=5, sum=5 (1)", "style": {}, "markdown": false }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": {}, "markdown": false }, { "value": 128.29166666666666, "description": "min=128.292, mean=128.292, max=128.292, sum=128.292 (1)", "style": {}, "markdown": false }, { "value": 7.822916666666667, "description": "min=7.823, mean=7.823, max=7.823, sum=7.823 (1)", "style": {}, "markdown": false }, { "value": 1.0, "description": "min=1, mean=1, max=1, sum=1 (1)", "style": {}, "markdown": false } ] ], "links": [ { "text": "LaTeX", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/latex/wikifact_wikifact_subject:plaintiff.tex" }, { "text": "JSON", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/json/wikifact_wikifact_subject:plaintiff.json" } ], "name": "wikifact_subject:plaintiff" }, { "title": "subject: position_held", "header": [ { "value": "Model", "markdown": false, "metadata": {} }, { "value": "EM", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.", "markdown": false, "lower_is_better": false, "metadata": { "metric": "EM", "run_group": "WikiFact" } }, { "value": "Denoised inference time (s)", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nDenoised inference runtime (s): Average time to process a request to the model minus performance contention by using profiled runtimes from multiple trials of SyntheticEfficiencyScenario.", "markdown": false, "lower_is_better": true, "metadata": { "metric": "Denoised inference time (s)", "run_group": "WikiFact" } }, { "value": "# eval", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# eval: Number of evaluation instances.", "markdown": false, "metadata": { "metric": "# eval", "run_group": "WikiFact" } }, { "value": "# train", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# train: Number of training instances (e.g., in-context examples).", "markdown": false, "metadata": { "metric": "# train", "run_group": "WikiFact" } }, { "value": "truncated", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).", "markdown": false, "metadata": { "metric": "truncated", "run_group": "WikiFact" } }, { "value": "# prompt tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# prompt tokens: Number of tokens in the prompt.", "markdown": false, "metadata": { "metric": "# prompt tokens", "run_group": "WikiFact" } }, { "value": "# output tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# output tokens: Actual number of output tokens.", "markdown": false, "metadata": { "metric": "# output tokens", "run_group": "WikiFact" } }, { "value": "# trials", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# trials: Number of trials, where in each trial we choose an independent, random set of training instances.", "markdown": false, "metadata": { "metric": "# trials", "run_group": "WikiFact" } } ], "rows": [ [ { "value": "EleutherAI/pythia-2.8b", "description": "", "href": "?group=wikifact&subgroup=subject%3A%20position_held&runSpecs=%5B%22wikifact%3Ak%3D5%2Csubject%3Dposition_held%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D", "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=position_held,model=EleutherAI_pythia-2.8b" ] }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": { "font-weight": "bold" }, "markdown": false }, { "description": "1 matching runs, but no matching metrics", "markdown": false }, { "value": 850.0, "description": "min=850, mean=850, max=850, sum=850 (1)", "style": {}, "markdown": false }, { "value": 5.0, "description": "min=5, mean=5, max=5, sum=5 (1)", "style": {}, "markdown": false }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": {}, "markdown": false }, { "value": 89.04588235294118, "description": "min=89.046, mean=89.046, max=89.046, sum=89.046 (1)", "style": {}, "markdown": false }, { "value": 7.843529411764706, "description": "min=7.844, mean=7.844, max=7.844, sum=7.844 (1)", "style": {}, "markdown": false }, { "value": 1.0, "description": "min=1, mean=1, max=1, sum=1 (1)", "style": {}, "markdown": false } ] ], "links": [ { "text": "LaTeX", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/latex/wikifact_wikifact_subject:position_held.tex" }, { "text": "JSON", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/json/wikifact_wikifact_subject:position_held.json" } ], "name": "wikifact_subject:position_held" }, { "title": "subject: symptoms_and_signs", "header": [ { "value": "Model", "markdown": false, "metadata": {} }, { "value": "EM", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nQuasi-exact match: Fraction of instances that the predicted output matches a correct reference up to light processing.", "markdown": false, "lower_is_better": false, "metadata": { "metric": "EM", "run_group": "WikiFact" } }, { "value": "Denoised inference time (s)", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\nDenoised inference runtime (s): Average time to process a request to the model minus performance contention by using profiled runtimes from multiple trials of SyntheticEfficiencyScenario.", "markdown": false, "lower_is_better": true, "metadata": { "metric": "Denoised inference time (s)", "run_group": "WikiFact" } }, { "value": "# eval", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# eval: Number of evaluation instances.", "markdown": false, "metadata": { "metric": "# eval", "run_group": "WikiFact" } }, { "value": "# train", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# train: Number of training instances (e.g., in-context examples).", "markdown": false, "metadata": { "metric": "# train", "run_group": "WikiFact" } }, { "value": "truncated", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\ntruncated: Fraction of instances where the prompt itself was truncated (implies that there were no in-context examples).", "markdown": false, "metadata": { "metric": "truncated", "run_group": "WikiFact" } }, { "value": "# prompt tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# prompt tokens: Number of tokens in the prompt.", "markdown": false, "metadata": { "metric": "# prompt tokens", "run_group": "WikiFact" } }, { "value": "# output tokens", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# output tokens: Actual number of output tokens.", "markdown": false, "metadata": { "metric": "# output tokens", "run_group": "WikiFact" } }, { "value": "# trials", "description": "Scenario introduced in this work, inspired by [Petroni et al. (2019)](https://aclanthology.org/D19-1250/), to more extensively test factual knowledge.\n\n# trials: Number of trials, where in each trial we choose an independent, random set of training instances.", "markdown": false, "metadata": { "metric": "# trials", "run_group": "WikiFact" } } ], "rows": [ [ { "value": "EleutherAI/pythia-2.8b", "description": "", "href": "?group=wikifact&subgroup=subject%3A%20symptoms_and_signs&runSpecs=%5B%22wikifact%3Ak%3D5%2Csubject%3Dsymptoms_and_signs%2Cmodel%3DEleutherAI_pythia-2.8b%22%5D", "markdown": false, "run_spec_names": [ "wikifact:k=5,subject=symptoms_and_signs,model=EleutherAI_pythia-2.8b" ] }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": { "font-weight": "bold" }, "markdown": false }, { "description": "1 matching runs, but no matching metrics", "markdown": false }, { "value": 566.0, "description": "min=566, mean=566, max=566, sum=566 (1)", "style": {}, "markdown": false }, { "value": 5.0, "description": "min=5, mean=5, max=5, sum=5 (1)", "style": {}, "markdown": false }, { "value": 0.0, "description": "min=0, mean=0, max=0, sum=0 (1)", "style": {}, "markdown": false }, { "value": 70.64840989399293, "description": "min=70.648, mean=70.648, max=70.648, sum=70.648 (1)", "style": {}, "markdown": false }, { "value": 7.876325088339223, "description": "min=7.876, mean=7.876, max=7.876, sum=7.876 (1)", "style": {}, "markdown": false }, { "value": 1.0, "description": "min=1, mean=1, max=1, sum=1 (1)", "style": {}, "markdown": false } ] ], "links": [ { "text": "LaTeX", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/latex/wikifact_wikifact_subject:symptoms_and_signs.tex" }, { "text": "JSON", "href": "benchmark_output/runs/classic_pythia-2.8b-step2/groups/json/wikifact_wikifact_subject:symptoms_and_signs.json" } ], "name": "wikifact_subject:symptoms_and_signs" } ]