id
stringlengths
2
115
lastModified
stringlengths
24
24
tags
list
author
stringlengths
2
42
description
stringlengths
0
68.7k
citation
stringlengths
0
10.7k
cardData
null
likes
int64
0
3.55k
downloads
int64
0
10.1M
card
stringlengths
0
1.01M
autoevaluate/autoeval-eval-dair-ai__emotion-split-240c41-63739145509
2023-10-04T17:06:26.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-374bef-63814145510
2023-10-04T17:06:37.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-amazon_reviews_multi-en-24024a-63483145496
2023-10-04T17:06:49.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-57ee79-63845145511
2023-10-04T17:06:53.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-717623-63909145512
2023-10-04T17:07:01.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-4107a2-63488145497
2023-10-04T17:07:12.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-cnn_dailymail-3.0.0-66cf5e-63956145513
2023-10-04T18:49:07.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - cnn_dailymail eval_info: task: summarization model: 0x70DA/pegasus-cnn_dailymail metrics: [] dataset_name: cnn_dailymail dataset_config: 3.0.0 dataset_split: test col_mapping: text: article target: highlights --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: 0x70DA/pegasus-cnn_dailymail * Dataset: cnn_dailymail * Config: 3.0.0 * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@gauiru1998](https://huggingface.co/gauiru1998) for evaluating this model.
newsmediabias/debiased-data-synthetic
2023-10-04T17:07:56.000Z
[ "region:us" ]
newsmediabias
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-d0226f-63704145506
2023-10-04T17:07:26.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-Jean-Baptiste__wikiner_fr-Jean-Baptiste__wikiner_fr-0fef71-64035145514
2023-10-04T17:07:28.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-glue-cola-469170-64108145515
2023-10-04T17:07:32.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-glue-cola-9913ac-64114145516
2023-10-04T17:07:44.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-123a2c-64159145517
2023-10-04T17:07:54.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-f8eab9-64187145518
2023-10-04T17:08:05.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-25f0ea-64248145519
2023-10-04T17:08:15.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-e2aeb6-64266145520
2023-10-04T17:08:27.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-dair-ai__emotion-split-37576f-64278145521
2023-10-04T17:08:35.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-fabiochiu__medium-articles-fabiochiu__medium-articles-5207ae-64303145522
2023-10-05T00:20:22.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - fabiochiu/medium-articles eval_info: task: summarization model: SamAct/PromptGeneration-base metrics: ['f1', 'accuracy'] dataset_name: fabiochiu/medium-articles dataset_config: fabiochiu--medium-articles dataset_split: train col_mapping: text: text target: title --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: SamAct/PromptGeneration-base * Dataset: fabiochiu/medium-articles * Config: fabiochiu--medium-articles * Split: train To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@SamAct](https://huggingface.co/SamAct) for evaluating this model.
autoevaluate/autoeval-eval-adversarial_qa-adversarialQA-efa810-64484145523
2023-10-04T17:09:43.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - adversarial_qa eval_info: task: extractive_question_answering model: saikatkumardey/my_awesome_qa_model metrics: ['accuracy'] dataset_name: adversarial_qa dataset_config: adversarialQA dataset_split: validation col_mapping: context: context question: question answers-text: answers.text answers-answer_start: answers.answer_start --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Question Answering * Model: saikatkumardey/my_awesome_qa_model * Dataset: adversarial_qa * Config: adversarialQA * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@Jyostna](https://huggingface.co/Jyostna) for evaluating this model.
autoevaluate/autoeval-eval-acronym_identification-default-e9e4cd-64568145524
2023-10-04T17:09:04.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-da15a4-64620145525
2023-10-04T17:09:17.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-squad-plain_text-7290ef-64745145527
2023-10-04T17:09:24.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-ef21a1-64788145528
2023-10-04T17:09:34.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-ag_news-default-5b1609-64790145529
2023-10-04T17:09:45.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-ade_corpus_v2-Ade_corpus_v2_classification-12dd94-64935145530
2023-10-04T17:09:55.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-cnn_dailymail-3.0.0-e26065-64936145531
2023-10-04T18:48:46.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - cnn_dailymail eval_info: task: summarization model: google/roberta2roberta_L-24_cnn_daily_mail metrics: [] dataset_name: cnn_dailymail dataset_config: 3.0.0 dataset_split: test col_mapping: text: article target: highlights --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: google/roberta2roberta_L-24_cnn_daily_mail * Dataset: cnn_dailymail * Config: 3.0.0 * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@https://huggingface.co/Sini](https://huggingface.co/https://huggingface.co/Sini) for evaluating this model.
autoevaluate/autoeval-eval-cnn_dailymail-3.0.0-e26065-64936145532
2023-10-04T17:13:21.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - cnn_dailymail eval_info: task: summarization model: minhtoan/t5-finetune-cnndaily-news metrics: [] dataset_name: cnn_dailymail dataset_config: 3.0.0 dataset_split: test col_mapping: text: article target: highlights --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: minhtoan/t5-finetune-cnndaily-news * Dataset: cnn_dailymail * Config: 3.0.0 * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@https://huggingface.co/Sini](https://huggingface.co/https://huggingface.co/Sini) for evaluating this model.
autoevaluate/autoeval-eval-cnn_dailymail-3.0.0-e26065-64936145533
2023-10-04T17:19:34.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - cnn_dailymail eval_info: task: summarization model: amagzari/bart-base-finetuned-cnn_dailymail metrics: [] dataset_name: cnn_dailymail dataset_config: 3.0.0 dataset_split: test col_mapping: text: article target: highlights --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: amagzari/bart-base-finetuned-cnn_dailymail * Dataset: cnn_dailymail * Config: 3.0.0 * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@https://huggingface.co/Sini](https://huggingface.co/https://huggingface.co/Sini) for evaluating this model.
autoevaluate/autoeval-eval-cnn_dailymail-3.0.0-e26065-64936145534
2023-10-04T17:10:38.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-cnn_dailymail-3.0.0-e26065-64936145535
2023-10-04T18:48:08.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - cnn_dailymail eval_info: task: summarization model: 0x70DA/pegasus-cnn_dailymail metrics: [] dataset_name: cnn_dailymail dataset_config: 3.0.0 dataset_split: test col_mapping: text: article target: highlights --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: 0x70DA/pegasus-cnn_dailymail * Dataset: cnn_dailymail * Config: 3.0.0 * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@https://huggingface.co/Sini](https://huggingface.co/https://huggingface.co/Sini) for evaluating this model.
songlab/gpn-msa-sapiens-dataset
2023-10-08T12:35:17.000Z
[ "license:mit", "region:us" ]
songlab
null
null
null
0
0
--- license: mit --- See https://github.com/songlab-cal/gpn for more details. Path in Snakemake: `results/dataset/multiz100way/89/128/64/True/defined.phastCons.percentile-75_0.05_0.001`
autoevaluate/autoeval-eval-cnn_dailymail-3.0.0-e26065-64936145536
2023-10-04T17:11:30.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-4c9bfb-64991145537
2023-10-04T17:11:49.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-allocine-allocine-3ab564-65116145540
2023-10-04T17:12:28.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-b2440b-65136145541
2023-10-04T17:13:10.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-glue-cola-7df5e3-65200145544
2023-10-04T17:13:16.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-de41e4-65224145545
2023-10-04T17:13:50.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-glue-cola-7eda1f-65227145546
2023-10-04T17:13:56.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-super_glue-rte-77d5f0-65249145548
2023-10-04T17:14:00.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-allocine-allocine-5883bf-65148145542
2023-10-04T17:14:28.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-ceb797-65268145549
2023-10-04T17:14:33.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-51e861-65294145550
2023-10-04T17:14:39.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-a6ef92-65325145551
2023-10-04T17:14:41.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-5036c3-65326145552
2023-10-04T17:15:11.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-cdb155-65337145553
2023-10-04T17:15:16.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-451087-65338145554
2023-10-04T17:15:22.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-94ea0e-65176145543
2023-10-04T17:15:22.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-482467-65395145555
2023-10-04T17:15:22.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-conll2003-conll2003-0a1842-65397145556
2023-10-04T17:15:49.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-conll2003-conll2003-0a1842-65397145557
2023-10-04T17:17:19.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - conll2003 eval_info: task: entity_extraction model: Abelll/bert-finetuned-ner metrics: [] dataset_name: conll2003 dataset_config: conll2003 dataset_split: test col_mapping: tokens: tokens tags: ner_tags --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Token Classification * Model: Abelll/bert-finetuned-ner * Dataset: conll2003 * Config: conll2003 * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@chnlyi](https://huggingface.co/chnlyi) for evaluating this model.
autoevaluate/autoeval-eval-conll2003-conll2003-0a1842-65397145558
2023-10-04T17:15:59.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-d775e5-65410145559
2023-10-04T17:16:07.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-1136f4-65080145538
2023-10-04T17:16:06.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-ade_corpus_v2-Ade_corpus_v2_classification-bea911-65081145539
2023-10-04T17:16:31.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-199908-65247145547
2023-10-04T17:16:44.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-imdb-plain_text-c9e01d-65528145562
2023-10-04T17:16:48.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-squad-plain_text-c5ffde-65417145560
2023-10-04T17:16:49.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-7b7734-65497145561
2023-10-04T17:17:10.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-fe48cb-65541145563
2023-10-04T17:17:22.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
open-llm-leaderboard/details_upstage__SOLAR-0-70b-16bit
2023-10-04T17:18:22.000Z
[ "region:us" ]
open-llm-leaderboard
null
null
null
0
0
--- pretty_name: Evaluation run of upstage/SOLAR-0-70b-16bit dataset_summary: "Dataset automatically created during the evaluation run of model\ \ [upstage/SOLAR-0-70b-16bit](https://huggingface.co/upstage/SOLAR-0-70b-16bit)\ \ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\ \nThe dataset is composed of 61 configuration, each one coresponding to one of the\ \ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\ \ found as a specific split in each configuration, the split being named using the\ \ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\ \nAn additional configuration \"results\" store all the aggregated results of the\ \ run (and is used to compute and display the agregated metrics on the [Open LLM\ \ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\ \nTo load the details from a run, you can for instance do the following:\n```python\n\ from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_upstage__SOLAR-0-70b-16bit\"\ ,\n\t\"harness_truthfulqa_mc_0\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\ \nThese are the [latest results from run 2023-10-04T17:16:57.736703](https://huggingface.co/datasets/open-llm-leaderboard/details_upstage__SOLAR-0-70b-16bit/blob/main/results_2023-10-04T17-16-57.736703.json)(note\ \ that their might be results for other tasks in the repos if successive evals didn't\ \ cover the same tasks. You find each in the results and the \"latest\" split for\ \ each eval):\n\n```python\n{\n \"all\": {\n \"acc\": 0.7050740464217434,\n\ \ \"acc_stderr\": 0.03085018588043536,\n \"acc_norm\": 0.7087855823993987,\n\ \ \"acc_norm_stderr\": 0.03081992944181276,\n \"mc1\": 0.44430844553243576,\n\ \ \"mc1_stderr\": 0.017394586250743173,\n \"mc2\": 0.6224972679005382,\n\ \ \"mc2_stderr\": 0.014880875055625352\n },\n \"harness|arc:challenge|25\"\ : {\n \"acc\": 0.6732081911262798,\n \"acc_stderr\": 0.013706665975587333,\n\ \ \"acc_norm\": 0.7107508532423208,\n \"acc_norm_stderr\": 0.013250012579393441\n\ \ },\n \"harness|hellaswag|10\": {\n \"acc\": 0.6974706233817964,\n\ \ \"acc_stderr\": 0.00458414401465495,\n \"acc_norm\": 0.8789085839474209,\n\ \ \"acc_norm_stderr\": 0.0032556675321152857\n },\n \"harness|hendrycksTest-abstract_algebra|5\"\ : {\n \"acc\": 0.33,\n \"acc_stderr\": 0.04725815626252605,\n \ \ \"acc_norm\": 0.33,\n \"acc_norm_stderr\": 0.04725815626252605\n \ \ },\n \"harness|hendrycksTest-anatomy|5\": {\n \"acc\": 0.6518518518518519,\n\ \ \"acc_stderr\": 0.041153246103369526,\n \"acc_norm\": 0.6518518518518519,\n\ \ \"acc_norm_stderr\": 0.041153246103369526\n },\n \"harness|hendrycksTest-astronomy|5\"\ : {\n \"acc\": 0.8421052631578947,\n \"acc_stderr\": 0.029674167520101453,\n\ \ \"acc_norm\": 0.8421052631578947,\n \"acc_norm_stderr\": 0.029674167520101453\n\ \ },\n \"harness|hendrycksTest-business_ethics|5\": {\n \"acc\": 0.74,\n\ \ \"acc_stderr\": 0.044084400227680794,\n \"acc_norm\": 0.74,\n \ \ \"acc_norm_stderr\": 0.044084400227680794\n },\n \"harness|hendrycksTest-clinical_knowledge|5\"\ : {\n \"acc\": 0.7320754716981132,\n \"acc_stderr\": 0.027257260322494845,\n\ \ \"acc_norm\": 0.7320754716981132,\n \"acc_norm_stderr\": 0.027257260322494845\n\ \ },\n \"harness|hendrycksTest-college_biology|5\": {\n \"acc\": 0.8402777777777778,\n\ \ \"acc_stderr\": 0.030635578972093274,\n \"acc_norm\": 0.8402777777777778,\n\ \ \"acc_norm_stderr\": 0.030635578972093274\n },\n \"harness|hendrycksTest-college_chemistry|5\"\ : {\n \"acc\": 0.49,\n \"acc_stderr\": 0.05024183937956912,\n \ \ \"acc_norm\": 0.49,\n \"acc_norm_stderr\": 0.05024183937956912\n \ \ },\n \"harness|hendrycksTest-college_computer_science|5\": {\n \"acc\"\ : 0.6,\n \"acc_stderr\": 0.04923659639173309,\n \"acc_norm\": 0.6,\n\ \ \"acc_norm_stderr\": 0.04923659639173309\n },\n \"harness|hendrycksTest-college_mathematics|5\"\ : {\n \"acc\": 0.42,\n \"acc_stderr\": 0.049604496374885836,\n \ \ \"acc_norm\": 0.42,\n \"acc_norm_stderr\": 0.049604496374885836\n \ \ },\n \"harness|hendrycksTest-college_medicine|5\": {\n \"acc\": 0.6589595375722543,\n\ \ \"acc_stderr\": 0.036146654241808254,\n \"acc_norm\": 0.6589595375722543,\n\ \ \"acc_norm_stderr\": 0.036146654241808254\n },\n \"harness|hendrycksTest-college_physics|5\"\ : {\n \"acc\": 0.4215686274509804,\n \"acc_stderr\": 0.04913595201274498,\n\ \ \"acc_norm\": 0.4215686274509804,\n \"acc_norm_stderr\": 0.04913595201274498\n\ \ },\n \"harness|hendrycksTest-computer_security|5\": {\n \"acc\":\ \ 0.77,\n \"acc_stderr\": 0.04229525846816507,\n \"acc_norm\": 0.77,\n\ \ \"acc_norm_stderr\": 0.04229525846816507\n },\n \"harness|hendrycksTest-conceptual_physics|5\"\ : {\n \"acc\": 0.7063829787234043,\n \"acc_stderr\": 0.029771642712491227,\n\ \ \"acc_norm\": 0.7063829787234043,\n \"acc_norm_stderr\": 0.029771642712491227\n\ \ },\n \"harness|hendrycksTest-econometrics|5\": {\n \"acc\": 0.4649122807017544,\n\ \ \"acc_stderr\": 0.04692008381368909,\n \"acc_norm\": 0.4649122807017544,\n\ \ \"acc_norm_stderr\": 0.04692008381368909\n },\n \"harness|hendrycksTest-electrical_engineering|5\"\ : {\n \"acc\": 0.6482758620689655,\n \"acc_stderr\": 0.0397923663749741,\n\ \ \"acc_norm\": 0.6482758620689655,\n \"acc_norm_stderr\": 0.0397923663749741\n\ \ },\n \"harness|hendrycksTest-elementary_mathematics|5\": {\n \"acc\"\ : 0.46825396825396826,\n \"acc_stderr\": 0.025699352832131792,\n \"\ acc_norm\": 0.46825396825396826,\n \"acc_norm_stderr\": 0.025699352832131792\n\ \ },\n \"harness|hendrycksTest-formal_logic|5\": {\n \"acc\": 0.46825396825396826,\n\ \ \"acc_stderr\": 0.04463112720677173,\n \"acc_norm\": 0.46825396825396826,\n\ \ \"acc_norm_stderr\": 0.04463112720677173\n },\n \"harness|hendrycksTest-global_facts|5\"\ : {\n \"acc\": 0.47,\n \"acc_stderr\": 0.05016135580465919,\n \ \ \"acc_norm\": 0.47,\n \"acc_norm_stderr\": 0.05016135580465919\n \ \ },\n \"harness|hendrycksTest-high_school_biology|5\": {\n \"acc\": 0.8096774193548387,\n\ \ \"acc_stderr\": 0.02233170761182307,\n \"acc_norm\": 0.8096774193548387,\n\ \ \"acc_norm_stderr\": 0.02233170761182307\n },\n \"harness|hendrycksTest-high_school_chemistry|5\"\ : {\n \"acc\": 0.5615763546798029,\n \"acc_stderr\": 0.03491207857486519,\n\ \ \"acc_norm\": 0.5615763546798029,\n \"acc_norm_stderr\": 0.03491207857486519\n\ \ },\n \"harness|hendrycksTest-high_school_computer_science|5\": {\n \ \ \"acc\": 0.79,\n \"acc_stderr\": 0.040936018074033256,\n \"acc_norm\"\ : 0.79,\n \"acc_norm_stderr\": 0.040936018074033256\n },\n \"harness|hendrycksTest-high_school_european_history|5\"\ : {\n \"acc\": 0.8424242424242424,\n \"acc_stderr\": 0.02845038880528436,\n\ \ \"acc_norm\": 0.8424242424242424,\n \"acc_norm_stderr\": 0.02845038880528436\n\ \ },\n \"harness|hendrycksTest-high_school_geography|5\": {\n \"acc\"\ : 0.8737373737373737,\n \"acc_stderr\": 0.023664359402880242,\n \"\ acc_norm\": 0.8737373737373737,\n \"acc_norm_stderr\": 0.023664359402880242\n\ \ },\n \"harness|hendrycksTest-high_school_government_and_politics|5\": {\n\ \ \"acc\": 0.9378238341968912,\n \"acc_stderr\": 0.017426974154240528,\n\ \ \"acc_norm\": 0.9378238341968912,\n \"acc_norm_stderr\": 0.017426974154240528\n\ \ },\n \"harness|hendrycksTest-high_school_macroeconomics|5\": {\n \ \ \"acc\": 0.7102564102564103,\n \"acc_stderr\": 0.023000628243687968,\n\ \ \"acc_norm\": 0.7102564102564103,\n \"acc_norm_stderr\": 0.023000628243687968\n\ \ },\n \"harness|hendrycksTest-high_school_mathematics|5\": {\n \"\ acc\": 0.31851851851851853,\n \"acc_stderr\": 0.028406533090608463,\n \ \ \"acc_norm\": 0.31851851851851853,\n \"acc_norm_stderr\": 0.028406533090608463\n\ \ },\n \"harness|hendrycksTest-high_school_microeconomics|5\": {\n \ \ \"acc\": 0.7647058823529411,\n \"acc_stderr\": 0.02755361446786381,\n \ \ \"acc_norm\": 0.7647058823529411,\n \"acc_norm_stderr\": 0.02755361446786381\n\ \ },\n \"harness|hendrycksTest-high_school_physics|5\": {\n \"acc\"\ : 0.47019867549668876,\n \"acc_stderr\": 0.04075224992216979,\n \"\ acc_norm\": 0.47019867549668876,\n \"acc_norm_stderr\": 0.04075224992216979\n\ \ },\n \"harness|hendrycksTest-high_school_psychology|5\": {\n \"acc\"\ : 0.9027522935779817,\n \"acc_stderr\": 0.012703533408540366,\n \"\ acc_norm\": 0.9027522935779817,\n \"acc_norm_stderr\": 0.012703533408540366\n\ \ },\n \"harness|hendrycksTest-high_school_statistics|5\": {\n \"acc\"\ : 0.6018518518518519,\n \"acc_stderr\": 0.033384734032074016,\n \"\ acc_norm\": 0.6018518518518519,\n \"acc_norm_stderr\": 0.033384734032074016\n\ \ },\n \"harness|hendrycksTest-high_school_us_history|5\": {\n \"acc\"\ : 0.9264705882352942,\n \"acc_stderr\": 0.01831885585008968,\n \"\ acc_norm\": 0.9264705882352942,\n \"acc_norm_stderr\": 0.01831885585008968\n\ \ },\n \"harness|hendrycksTest-high_school_world_history|5\": {\n \"\ acc\": 0.8945147679324894,\n \"acc_stderr\": 0.01999556072375854,\n \ \ \"acc_norm\": 0.8945147679324894,\n \"acc_norm_stderr\": 0.01999556072375854\n\ \ },\n \"harness|hendrycksTest-human_aging|5\": {\n \"acc\": 0.7937219730941704,\n\ \ \"acc_stderr\": 0.02715715047956382,\n \"acc_norm\": 0.7937219730941704,\n\ \ \"acc_norm_stderr\": 0.02715715047956382\n },\n \"harness|hendrycksTest-human_sexuality|5\"\ : {\n \"acc\": 0.8625954198473282,\n \"acc_stderr\": 0.030194823996804475,\n\ \ \"acc_norm\": 0.8625954198473282,\n \"acc_norm_stderr\": 0.030194823996804475\n\ \ },\n \"harness|hendrycksTest-international_law|5\": {\n \"acc\":\ \ 0.859504132231405,\n \"acc_stderr\": 0.03172233426002157,\n \"acc_norm\"\ : 0.859504132231405,\n \"acc_norm_stderr\": 0.03172233426002157\n },\n\ \ \"harness|hendrycksTest-jurisprudence|5\": {\n \"acc\": 0.8240740740740741,\n\ \ \"acc_stderr\": 0.036809181416738807,\n \"acc_norm\": 0.8240740740740741,\n\ \ \"acc_norm_stderr\": 0.036809181416738807\n },\n \"harness|hendrycksTest-logical_fallacies|5\"\ : {\n \"acc\": 0.803680981595092,\n \"acc_stderr\": 0.031207970394709218,\n\ \ \"acc_norm\": 0.803680981595092,\n \"acc_norm_stderr\": 0.031207970394709218\n\ \ },\n \"harness|hendrycksTest-machine_learning|5\": {\n \"acc\": 0.5089285714285714,\n\ \ \"acc_stderr\": 0.04745033255489122,\n \"acc_norm\": 0.5089285714285714,\n\ \ \"acc_norm_stderr\": 0.04745033255489122\n },\n \"harness|hendrycksTest-management|5\"\ : {\n \"acc\": 0.8252427184466019,\n \"acc_stderr\": 0.0376017800602662,\n\ \ \"acc_norm\": 0.8252427184466019,\n \"acc_norm_stderr\": 0.0376017800602662\n\ \ },\n \"harness|hendrycksTest-marketing|5\": {\n \"acc\": 0.9017094017094017,\n\ \ \"acc_stderr\": 0.019503444900757567,\n \"acc_norm\": 0.9017094017094017,\n\ \ \"acc_norm_stderr\": 0.019503444900757567\n },\n \"harness|hendrycksTest-medical_genetics|5\"\ : {\n \"acc\": 0.71,\n \"acc_stderr\": 0.04560480215720684,\n \ \ \"acc_norm\": 0.71,\n \"acc_norm_stderr\": 0.04560480215720684\n \ \ },\n \"harness|hendrycksTest-miscellaneous|5\": {\n \"acc\": 0.8684546615581098,\n\ \ \"acc_stderr\": 0.01208670521425043,\n \"acc_norm\": 0.8684546615581098,\n\ \ \"acc_norm_stderr\": 0.01208670521425043\n },\n \"harness|hendrycksTest-moral_disputes|5\"\ : {\n \"acc\": 0.7803468208092486,\n \"acc_stderr\": 0.022289638852617893,\n\ \ \"acc_norm\": 0.7803468208092486,\n \"acc_norm_stderr\": 0.022289638852617893\n\ \ },\n \"harness|hendrycksTest-moral_scenarios|5\": {\n \"acc\": 0.6044692737430167,\n\ \ \"acc_stderr\": 0.01635341541007577,\n \"acc_norm\": 0.6044692737430167,\n\ \ \"acc_norm_stderr\": 0.01635341541007577\n },\n \"harness|hendrycksTest-nutrition|5\"\ : {\n \"acc\": 0.7679738562091504,\n \"acc_stderr\": 0.024170840879340873,\n\ \ \"acc_norm\": 0.7679738562091504,\n \"acc_norm_stderr\": 0.024170840879340873\n\ \ },\n \"harness|hendrycksTest-philosophy|5\": {\n \"acc\": 0.7781350482315113,\n\ \ \"acc_stderr\": 0.02359885829286305,\n \"acc_norm\": 0.7781350482315113,\n\ \ \"acc_norm_stderr\": 0.02359885829286305\n },\n \"harness|hendrycksTest-prehistory|5\"\ : {\n \"acc\": 0.8333333333333334,\n \"acc_stderr\": 0.020736358408060006,\n\ \ \"acc_norm\": 0.8333333333333334,\n \"acc_norm_stderr\": 0.020736358408060006\n\ \ },\n \"harness|hendrycksTest-professional_accounting|5\": {\n \"\ acc\": 0.574468085106383,\n \"acc_stderr\": 0.029494827600144366,\n \ \ \"acc_norm\": 0.574468085106383,\n \"acc_norm_stderr\": 0.029494827600144366\n\ \ },\n \"harness|hendrycksTest-professional_law|5\": {\n \"acc\": 0.5521512385919165,\n\ \ \"acc_stderr\": 0.012700582404768235,\n \"acc_norm\": 0.5521512385919165,\n\ \ \"acc_norm_stderr\": 0.012700582404768235\n },\n \"harness|hendrycksTest-professional_medicine|5\"\ : {\n \"acc\": 0.7389705882352942,\n \"acc_stderr\": 0.02667925227010314,\n\ \ \"acc_norm\": 0.7389705882352942,\n \"acc_norm_stderr\": 0.02667925227010314\n\ \ },\n \"harness|hendrycksTest-professional_psychology|5\": {\n \"\ acc\": 0.7647058823529411,\n \"acc_stderr\": 0.01716058723504635,\n \ \ \"acc_norm\": 0.7647058823529411,\n \"acc_norm_stderr\": 0.01716058723504635\n\ \ },\n \"harness|hendrycksTest-public_relations|5\": {\n \"acc\": 0.7454545454545455,\n\ \ \"acc_stderr\": 0.041723430387053825,\n \"acc_norm\": 0.7454545454545455,\n\ \ \"acc_norm_stderr\": 0.041723430387053825\n },\n \"harness|hendrycksTest-security_studies|5\"\ : {\n \"acc\": 0.8204081632653061,\n \"acc_stderr\": 0.024573293589585637,\n\ \ \"acc_norm\": 0.8204081632653061,\n \"acc_norm_stderr\": 0.024573293589585637\n\ \ },\n \"harness|hendrycksTest-sociology|5\": {\n \"acc\": 0.8756218905472637,\n\ \ \"acc_stderr\": 0.023335401790166327,\n \"acc_norm\": 0.8756218905472637,\n\ \ \"acc_norm_stderr\": 0.023335401790166327\n },\n \"harness|hendrycksTest-us_foreign_policy|5\"\ : {\n \"acc\": 0.9,\n \"acc_stderr\": 0.030151134457776334,\n \ \ \"acc_norm\": 0.9,\n \"acc_norm_stderr\": 0.030151134457776334\n \ \ },\n \"harness|hendrycksTest-virology|5\": {\n \"acc\": 0.5301204819277109,\n\ \ \"acc_stderr\": 0.03885425420866767,\n \"acc_norm\": 0.5301204819277109,\n\ \ \"acc_norm_stderr\": 0.03885425420866767\n },\n \"harness|hendrycksTest-world_religions|5\"\ : {\n \"acc\": 0.8771929824561403,\n \"acc_stderr\": 0.02517298435015575,\n\ \ \"acc_norm\": 0.8771929824561403,\n \"acc_norm_stderr\": 0.02517298435015575\n\ \ },\n \"harness|truthfulqa:mc|0\": {\n \"mc1\": 0.44430844553243576,\n\ \ \"mc1_stderr\": 0.017394586250743173,\n \"mc2\": 0.6224972679005382,\n\ \ \"mc2_stderr\": 0.014880875055625352\n }\n}\n```" repo_url: https://huggingface.co/upstage/SOLAR-0-70b-16bit leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard point_of_contact: clementine@hf.co configs: - config_name: harness_arc_challenge_25 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|arc:challenge|25_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|arc:challenge|25_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hellaswag_10 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hellaswag|10_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hellaswag|10_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-anatomy|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-astronomy|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-business_ethics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-college_biology|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-college_chemistry|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-college_computer_science|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-college_mathematics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-college_medicine|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-college_physics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-computer_security|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-conceptual_physics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-econometrics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-electrical_engineering|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-formal_logic|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-global_facts|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_biology|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_european_history|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_geography|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_physics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_psychology|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_statistics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_us_history|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_world_history|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-human_aging|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-human_sexuality|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-international_law|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-jurisprudence|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-logical_fallacies|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-machine_learning|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-management|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-marketing|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-medical_genetics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-miscellaneous|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-moral_disputes|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-moral_scenarios|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-nutrition|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-philosophy|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-prehistory|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-professional_accounting|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-professional_law|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-professional_medicine|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-professional_psychology|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-public_relations|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-security_studies|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-sociology|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-virology|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-world_religions|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-anatomy|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-astronomy|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-business_ethics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-college_biology|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-college_chemistry|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-college_computer_science|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-college_mathematics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-college_medicine|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-college_physics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-computer_security|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-conceptual_physics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-econometrics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-electrical_engineering|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-formal_logic|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-global_facts|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_biology|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_european_history|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_geography|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_physics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_psychology|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_statistics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_us_history|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-high_school_world_history|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-human_aging|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-human_sexuality|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-international_law|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-jurisprudence|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-logical_fallacies|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-machine_learning|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-management|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-marketing|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-medical_genetics|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-miscellaneous|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-moral_disputes|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-moral_scenarios|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-nutrition|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-philosophy|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-prehistory|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-professional_accounting|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-professional_law|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-professional_medicine|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-professional_psychology|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-public_relations|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-security_studies|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-sociology|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-virology|5_2023-10-04T17-16-57.736703.parquet' - '**/details_harness|hendrycksTest-world_religions|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_abstract_algebra_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_anatomy_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-anatomy|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-anatomy|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_astronomy_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-astronomy|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-astronomy|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_business_ethics_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-business_ethics|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-business_ethics|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_clinical_knowledge_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_college_biology_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-college_biology|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_biology|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_college_chemistry_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-college_chemistry|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_chemistry|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_college_computer_science_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-college_computer_science|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_computer_science|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_college_mathematics_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-college_mathematics|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_mathematics|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_college_medicine_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-college_medicine|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_medicine|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_college_physics_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-college_physics|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_physics|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_computer_security_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-computer_security|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-computer_security|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_conceptual_physics_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-conceptual_physics|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-conceptual_physics|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_econometrics_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-econometrics|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-econometrics|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_electrical_engineering_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-electrical_engineering|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-electrical_engineering|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_elementary_mathematics_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_formal_logic_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-formal_logic|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-formal_logic|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_global_facts_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-global_facts|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-global_facts|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_high_school_biology_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-high_school_biology|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_biology|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_high_school_chemistry_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_high_school_computer_science_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_high_school_european_history_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-high_school_european_history|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_european_history|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_high_school_geography_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-high_school_geography|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_geography|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_high_school_government_and_politics_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_high_school_macroeconomics_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_high_school_mathematics_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_high_school_microeconomics_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_high_school_physics_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-high_school_physics|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_physics|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_high_school_psychology_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-high_school_psychology|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_psychology|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_high_school_statistics_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-high_school_statistics|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_statistics|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_high_school_us_history_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-high_school_us_history|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_us_history|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_high_school_world_history_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-high_school_world_history|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_world_history|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_human_aging_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-human_aging|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-human_aging|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_human_sexuality_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-human_sexuality|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-human_sexuality|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_international_law_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-international_law|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-international_law|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_jurisprudence_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-jurisprudence|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-jurisprudence|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_logical_fallacies_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-logical_fallacies|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-logical_fallacies|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_machine_learning_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-machine_learning|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-machine_learning|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_management_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-management|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-management|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_marketing_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-marketing|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-marketing|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_medical_genetics_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-medical_genetics|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-medical_genetics|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_miscellaneous_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-miscellaneous|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-miscellaneous|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_moral_disputes_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-moral_disputes|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-moral_disputes|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_moral_scenarios_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-moral_scenarios|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-moral_scenarios|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_nutrition_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-nutrition|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-nutrition|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_philosophy_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-philosophy|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-philosophy|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_prehistory_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-prehistory|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-prehistory|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_professional_accounting_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-professional_accounting|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-professional_accounting|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_professional_law_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-professional_law|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-professional_law|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_professional_medicine_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-professional_medicine|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-professional_medicine|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_professional_psychology_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-professional_psychology|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-professional_psychology|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_public_relations_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-public_relations|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-public_relations|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_security_studies_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-security_studies|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-security_studies|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_sociology_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-sociology|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-sociology|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_us_foreign_policy_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_virology_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-virology|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-virology|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_hendrycksTest_world_religions_5 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|hendrycksTest-world_religions|5_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|hendrycksTest-world_religions|5_2023-10-04T17-16-57.736703.parquet' - config_name: harness_truthfulqa_mc_0 data_files: - split: 2023_10_04T17_16_57.736703 path: - '**/details_harness|truthfulqa:mc|0_2023-10-04T17-16-57.736703.parquet' - split: latest path: - '**/details_harness|truthfulqa:mc|0_2023-10-04T17-16-57.736703.parquet' - config_name: results data_files: - split: 2023_10_04T17_16_57.736703 path: - results_2023-10-04T17-16-57.736703.parquet - split: latest path: - results_2023-10-04T17-16-57.736703.parquet --- # Dataset Card for Evaluation run of upstage/SOLAR-0-70b-16bit ## Dataset Description - **Homepage:** - **Repository:** https://huggingface.co/upstage/SOLAR-0-70b-16bit - **Paper:** - **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard - **Point of Contact:** clementine@hf.co ### Dataset Summary Dataset automatically created during the evaluation run of model [upstage/SOLAR-0-70b-16bit](https://huggingface.co/upstage/SOLAR-0-70b-16bit) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). The dataset is composed of 61 configuration, each one coresponding to one of the evaluated task. The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results. An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)). To load the details from a run, you can for instance do the following: ```python from datasets import load_dataset data = load_dataset("open-llm-leaderboard/details_upstage__SOLAR-0-70b-16bit", "harness_truthfulqa_mc_0", split="train") ``` ## Latest results These are the [latest results from run 2023-10-04T17:16:57.736703](https://huggingface.co/datasets/open-llm-leaderboard/details_upstage__SOLAR-0-70b-16bit/blob/main/results_2023-10-04T17-16-57.736703.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval): ```python { "all": { "acc": 0.7050740464217434, "acc_stderr": 0.03085018588043536, "acc_norm": 0.7087855823993987, "acc_norm_stderr": 0.03081992944181276, "mc1": 0.44430844553243576, "mc1_stderr": 0.017394586250743173, "mc2": 0.6224972679005382, "mc2_stderr": 0.014880875055625352 }, "harness|arc:challenge|25": { "acc": 0.6732081911262798, "acc_stderr": 0.013706665975587333, "acc_norm": 0.7107508532423208, "acc_norm_stderr": 0.013250012579393441 }, "harness|hellaswag|10": { "acc": 0.6974706233817964, "acc_stderr": 0.00458414401465495, "acc_norm": 0.8789085839474209, "acc_norm_stderr": 0.0032556675321152857 }, "harness|hendrycksTest-abstract_algebra|5": { "acc": 0.33, "acc_stderr": 0.04725815626252605, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252605 }, "harness|hendrycksTest-anatomy|5": { "acc": 0.6518518518518519, "acc_stderr": 0.041153246103369526, "acc_norm": 0.6518518518518519, "acc_norm_stderr": 0.041153246103369526 }, "harness|hendrycksTest-astronomy|5": { "acc": 0.8421052631578947, "acc_stderr": 0.029674167520101453, "acc_norm": 0.8421052631578947, "acc_norm_stderr": 0.029674167520101453 }, "harness|hendrycksTest-business_ethics|5": { "acc": 0.74, "acc_stderr": 0.044084400227680794, "acc_norm": 0.74, "acc_norm_stderr": 0.044084400227680794 }, "harness|hendrycksTest-clinical_knowledge|5": { "acc": 0.7320754716981132, "acc_stderr": 0.027257260322494845, "acc_norm": 0.7320754716981132, "acc_norm_stderr": 0.027257260322494845 }, "harness|hendrycksTest-college_biology|5": { "acc": 0.8402777777777778, "acc_stderr": 0.030635578972093274, "acc_norm": 0.8402777777777778, "acc_norm_stderr": 0.030635578972093274 }, "harness|hendrycksTest-college_chemistry|5": { "acc": 0.49, "acc_stderr": 0.05024183937956912, "acc_norm": 0.49, "acc_norm_stderr": 0.05024183937956912 }, "harness|hendrycksTest-college_computer_science|5": { "acc": 0.6, "acc_stderr": 0.04923659639173309, "acc_norm": 0.6, "acc_norm_stderr": 0.04923659639173309 }, "harness|hendrycksTest-college_mathematics|5": { "acc": 0.42, "acc_stderr": 0.049604496374885836, "acc_norm": 0.42, "acc_norm_stderr": 0.049604496374885836 }, "harness|hendrycksTest-college_medicine|5": { "acc": 0.6589595375722543, "acc_stderr": 0.036146654241808254, "acc_norm": 0.6589595375722543, "acc_norm_stderr": 0.036146654241808254 }, "harness|hendrycksTest-college_physics|5": { "acc": 0.4215686274509804, "acc_stderr": 0.04913595201274498, "acc_norm": 0.4215686274509804, "acc_norm_stderr": 0.04913595201274498 }, "harness|hendrycksTest-computer_security|5": { "acc": 0.77, "acc_stderr": 0.04229525846816507, "acc_norm": 0.77, "acc_norm_stderr": 0.04229525846816507 }, "harness|hendrycksTest-conceptual_physics|5": { "acc": 0.7063829787234043, "acc_stderr": 0.029771642712491227, "acc_norm": 0.7063829787234043, "acc_norm_stderr": 0.029771642712491227 }, "harness|hendrycksTest-econometrics|5": { "acc": 0.4649122807017544, "acc_stderr": 0.04692008381368909, "acc_norm": 0.4649122807017544, "acc_norm_stderr": 0.04692008381368909 }, "harness|hendrycksTest-electrical_engineering|5": { "acc": 0.6482758620689655, "acc_stderr": 0.0397923663749741, "acc_norm": 0.6482758620689655, "acc_norm_stderr": 0.0397923663749741 }, "harness|hendrycksTest-elementary_mathematics|5": { "acc": 0.46825396825396826, "acc_stderr": 0.025699352832131792, "acc_norm": 0.46825396825396826, "acc_norm_stderr": 0.025699352832131792 }, "harness|hendrycksTest-formal_logic|5": { "acc": 0.46825396825396826, "acc_stderr": 0.04463112720677173, "acc_norm": 0.46825396825396826, "acc_norm_stderr": 0.04463112720677173 }, "harness|hendrycksTest-global_facts|5": { "acc": 0.47, "acc_stderr": 0.05016135580465919, "acc_norm": 0.47, "acc_norm_stderr": 0.05016135580465919 }, "harness|hendrycksTest-high_school_biology|5": { "acc": 0.8096774193548387, "acc_stderr": 0.02233170761182307, "acc_norm": 0.8096774193548387, "acc_norm_stderr": 0.02233170761182307 }, "harness|hendrycksTest-high_school_chemistry|5": { "acc": 0.5615763546798029, "acc_stderr": 0.03491207857486519, "acc_norm": 0.5615763546798029, "acc_norm_stderr": 0.03491207857486519 }, "harness|hendrycksTest-high_school_computer_science|5": { "acc": 0.79, "acc_stderr": 0.040936018074033256, "acc_norm": 0.79, "acc_norm_stderr": 0.040936018074033256 }, "harness|hendrycksTest-high_school_european_history|5": { "acc": 0.8424242424242424, "acc_stderr": 0.02845038880528436, "acc_norm": 0.8424242424242424, "acc_norm_stderr": 0.02845038880528436 }, "harness|hendrycksTest-high_school_geography|5": { "acc": 0.8737373737373737, "acc_stderr": 0.023664359402880242, "acc_norm": 0.8737373737373737, "acc_norm_stderr": 0.023664359402880242 }, "harness|hendrycksTest-high_school_government_and_politics|5": { "acc": 0.9378238341968912, "acc_stderr": 0.017426974154240528, "acc_norm": 0.9378238341968912, "acc_norm_stderr": 0.017426974154240528 }, "harness|hendrycksTest-high_school_macroeconomics|5": { "acc": 0.7102564102564103, "acc_stderr": 0.023000628243687968, "acc_norm": 0.7102564102564103, "acc_norm_stderr": 0.023000628243687968 }, "harness|hendrycksTest-high_school_mathematics|5": { "acc": 0.31851851851851853, "acc_stderr": 0.028406533090608463, "acc_norm": 0.31851851851851853, "acc_norm_stderr": 0.028406533090608463 }, "harness|hendrycksTest-high_school_microeconomics|5": { "acc": 0.7647058823529411, "acc_stderr": 0.02755361446786381, "acc_norm": 0.7647058823529411, "acc_norm_stderr": 0.02755361446786381 }, "harness|hendrycksTest-high_school_physics|5": { "acc": 0.47019867549668876, "acc_stderr": 0.04075224992216979, "acc_norm": 0.47019867549668876, "acc_norm_stderr": 0.04075224992216979 }, "harness|hendrycksTest-high_school_psychology|5": { "acc": 0.9027522935779817, "acc_stderr": 0.012703533408540366, "acc_norm": 0.9027522935779817, "acc_norm_stderr": 0.012703533408540366 }, "harness|hendrycksTest-high_school_statistics|5": { "acc": 0.6018518518518519, "acc_stderr": 0.033384734032074016, "acc_norm": 0.6018518518518519, "acc_norm_stderr": 0.033384734032074016 }, "harness|hendrycksTest-high_school_us_history|5": { "acc": 0.9264705882352942, "acc_stderr": 0.01831885585008968, "acc_norm": 0.9264705882352942, "acc_norm_stderr": 0.01831885585008968 }, "harness|hendrycksTest-high_school_world_history|5": { "acc": 0.8945147679324894, "acc_stderr": 0.01999556072375854, "acc_norm": 0.8945147679324894, "acc_norm_stderr": 0.01999556072375854 }, "harness|hendrycksTest-human_aging|5": { "acc": 0.7937219730941704, "acc_stderr": 0.02715715047956382, "acc_norm": 0.7937219730941704, "acc_norm_stderr": 0.02715715047956382 }, "harness|hendrycksTest-human_sexuality|5": { "acc": 0.8625954198473282, "acc_stderr": 0.030194823996804475, "acc_norm": 0.8625954198473282, "acc_norm_stderr": 0.030194823996804475 }, "harness|hendrycksTest-international_law|5": { "acc": 0.859504132231405, "acc_stderr": 0.03172233426002157, "acc_norm": 0.859504132231405, "acc_norm_stderr": 0.03172233426002157 }, "harness|hendrycksTest-jurisprudence|5": { "acc": 0.8240740740740741, "acc_stderr": 0.036809181416738807, "acc_norm": 0.8240740740740741, "acc_norm_stderr": 0.036809181416738807 }, "harness|hendrycksTest-logical_fallacies|5": { "acc": 0.803680981595092, "acc_stderr": 0.031207970394709218, "acc_norm": 0.803680981595092, "acc_norm_stderr": 0.031207970394709218 }, "harness|hendrycksTest-machine_learning|5": { "acc": 0.5089285714285714, "acc_stderr": 0.04745033255489122, "acc_norm": 0.5089285714285714, "acc_norm_stderr": 0.04745033255489122 }, "harness|hendrycksTest-management|5": { "acc": 0.8252427184466019, "acc_stderr": 0.0376017800602662, "acc_norm": 0.8252427184466019, "acc_norm_stderr": 0.0376017800602662 }, "harness|hendrycksTest-marketing|5": { "acc": 0.9017094017094017, "acc_stderr": 0.019503444900757567, "acc_norm": 0.9017094017094017, "acc_norm_stderr": 0.019503444900757567 }, "harness|hendrycksTest-medical_genetics|5": { "acc": 0.71, "acc_stderr": 0.04560480215720684, "acc_norm": 0.71, "acc_norm_stderr": 0.04560480215720684 }, "harness|hendrycksTest-miscellaneous|5": { "acc": 0.8684546615581098, "acc_stderr": 0.01208670521425043, "acc_norm": 0.8684546615581098, "acc_norm_stderr": 0.01208670521425043 }, "harness|hendrycksTest-moral_disputes|5": { "acc": 0.7803468208092486, "acc_stderr": 0.022289638852617893, "acc_norm": 0.7803468208092486, "acc_norm_stderr": 0.022289638852617893 }, "harness|hendrycksTest-moral_scenarios|5": { "acc": 0.6044692737430167, "acc_stderr": 0.01635341541007577, "acc_norm": 0.6044692737430167, "acc_norm_stderr": 0.01635341541007577 }, "harness|hendrycksTest-nutrition|5": { "acc": 0.7679738562091504, "acc_stderr": 0.024170840879340873, "acc_norm": 0.7679738562091504, "acc_norm_stderr": 0.024170840879340873 }, "harness|hendrycksTest-philosophy|5": { "acc": 0.7781350482315113, "acc_stderr": 0.02359885829286305, "acc_norm": 0.7781350482315113, "acc_norm_stderr": 0.02359885829286305 }, "harness|hendrycksTest-prehistory|5": { "acc": 0.8333333333333334, "acc_stderr": 0.020736358408060006, "acc_norm": 0.8333333333333334, "acc_norm_stderr": 0.020736358408060006 }, "harness|hendrycksTest-professional_accounting|5": { "acc": 0.574468085106383, "acc_stderr": 0.029494827600144366, "acc_norm": 0.574468085106383, "acc_norm_stderr": 0.029494827600144366 }, "harness|hendrycksTest-professional_law|5": { "acc": 0.5521512385919165, "acc_stderr": 0.012700582404768235, "acc_norm": 0.5521512385919165, "acc_norm_stderr": 0.012700582404768235 }, "harness|hendrycksTest-professional_medicine|5": { "acc": 0.7389705882352942, "acc_stderr": 0.02667925227010314, "acc_norm": 0.7389705882352942, "acc_norm_stderr": 0.02667925227010314 }, "harness|hendrycksTest-professional_psychology|5": { "acc": 0.7647058823529411, "acc_stderr": 0.01716058723504635, "acc_norm": 0.7647058823529411, "acc_norm_stderr": 0.01716058723504635 }, "harness|hendrycksTest-public_relations|5": { "acc": 0.7454545454545455, "acc_stderr": 0.041723430387053825, "acc_norm": 0.7454545454545455, "acc_norm_stderr": 0.041723430387053825 }, "harness|hendrycksTest-security_studies|5": { "acc": 0.8204081632653061, "acc_stderr": 0.024573293589585637, "acc_norm": 0.8204081632653061, "acc_norm_stderr": 0.024573293589585637 }, "harness|hendrycksTest-sociology|5": { "acc": 0.8756218905472637, "acc_stderr": 0.023335401790166327, "acc_norm": 0.8756218905472637, "acc_norm_stderr": 0.023335401790166327 }, "harness|hendrycksTest-us_foreign_policy|5": { "acc": 0.9, "acc_stderr": 0.030151134457776334, "acc_norm": 0.9, "acc_norm_stderr": 0.030151134457776334 }, "harness|hendrycksTest-virology|5": { "acc": 0.5301204819277109, "acc_stderr": 0.03885425420866767, "acc_norm": 0.5301204819277109, "acc_norm_stderr": 0.03885425420866767 }, "harness|hendrycksTest-world_religions|5": { "acc": 0.8771929824561403, "acc_stderr": 0.02517298435015575, "acc_norm": 0.8771929824561403, "acc_norm_stderr": 0.02517298435015575 }, "harness|truthfulqa:mc|0": { "mc1": 0.44430844553243576, "mc1_stderr": 0.017394586250743173, "mc2": 0.6224972679005382, "mc2_stderr": 0.014880875055625352 } } ``` ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions [More Information Needed]
autoevaluate/autoeval-eval-squad-plain_text-15c7c2-65587145564
2023-10-04T17:17:28.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-5dd0bd-65620145565
2023-10-04T17:17:49.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-amazon_polarity-amazon_polarity-56b5a5-65818145567
2023-10-04T17:17:56.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-8c99dc-65834145568
2023-10-04T17:18:01.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-adversarial_qa-adversarialQA-0cf9bf-65912145569
2023-10-04T17:18:57.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - adversarial_qa eval_info: task: extractive_question_answering model: Chetna19/my_awesome_qa_model metrics: ['perplexity', 'accuracy', 'bleu'] dataset_name: adversarial_qa dataset_config: adversarialQA dataset_split: validation col_mapping: context: context question: question answers-text: answers.text answers-answer_start: answers.answer_start --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Question Answering * Model: Chetna19/my_awesome_qa_model * Dataset: adversarial_qa * Config: adversarialQA * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@Hizafa](https://huggingface.co/Hizafa) for evaluating this model.
autoevaluate/autoeval-eval-ag_news-default-63bee1-65942145570
2023-10-04T17:18:26.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-6908fe-65995145571
2023-10-04T17:18:39.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-f521b7-65708145566
2023-10-04T17:18:52.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-09876d-66054145572
2023-10-04T17:19:06.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-684f5b-66055145573
2023-10-04T17:19:17.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-b1d1bf-66058145574
2023-10-04T17:19:30.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-adversarial_qa-adversarialQA-6a5d0b-66069145577
2023-10-04T17:20:31.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - adversarial_qa eval_info: task: extractive_question_answering model: jgammack/SAE-roberta-base-squad metrics: [] dataset_name: adversarial_qa dataset_config: adversarialQA dataset_split: validation col_mapping: context: context question: question answers-text: answers.text answers-answer_start: answers.answer_start --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Question Answering * Model: jgammack/SAE-roberta-base-squad * Dataset: adversarial_qa * Config: adversarialQA * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@jtatman](https://huggingface.co/jtatman) for evaluating this model.
autoevaluate/autoeval-eval-glue-cola-46ed06-66108145578
2023-10-04T17:19:38.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-multi_news-default-1231f6-66166145579
2023-10-04T17:44:36.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - multi_news eval_info: task: summarization model: facebook/bart-large-cnn metrics: [] dataset_name: multi_news dataset_config: default dataset_split: test col_mapping: text: document target: summary --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: facebook/bart-large-cnn * Dataset: multi_news * Config: default * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@bart-multi-news](https://huggingface.co/bart-multi-news) for evaluating this model.
autoevaluate/autoeval-eval-adversarial_qa-adversarialQA-6a5d0b-66069145576
2023-10-04T17:20:39.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - adversarial_qa eval_info: task: extractive_question_answering model: Laurie/QA-distilbert metrics: [] dataset_name: adversarial_qa dataset_config: adversarialQA dataset_split: validation col_mapping: context: context question: question answers-text: answers.text answers-answer_start: answers.answer_start --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Question Answering * Model: Laurie/QA-distilbert * Dataset: adversarial_qa * Config: adversarialQA * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@jtatman](https://huggingface.co/jtatman) for evaluating this model.
autoevaluate/autoeval-eval-tweet_eval-emotion-dbaa98-66233145580
2023-10-04T17:20:23.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - tweet_eval eval_info: task: multi_class_classification model: 095ey11/bert-emotion metrics: [] dataset_name: tweet_eval dataset_config: emotion dataset_split: train col_mapping: text: text target: label --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Multi-class Text Classification * Model: 095ey11/bert-emotion * Dataset: tweet_eval * Config: emotion * Split: train To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@Ayushkm2799](https://huggingface.co/Ayushkm2799) for evaluating this model.
autoevaluate/autoeval-eval-tweet_eval-emotion-dbaa98-66233145581
2023-10-04T17:20:32.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - tweet_eval eval_info: task: multi_class_classification model: FelixHonikker/bert-emotion metrics: [] dataset_name: tweet_eval dataset_config: emotion dataset_split: train col_mapping: text: text target: label --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Multi-class Text Classification * Model: FelixHonikker/bert-emotion * Dataset: tweet_eval * Config: emotion * Split: train To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@Ayushkm2799](https://huggingface.co/Ayushkm2799) for evaluating this model.
autoevaluate/autoeval-eval-acronym_identification-default-912d39-66059145575
2023-10-04T17:20:09.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-4b31ed-66272145582
2023-10-04T17:20:17.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-tweet_eval-emotion-11ffbd-66309145583
2023-10-04T17:20:45.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - tweet_eval eval_info: task: multi_class_classification model: ShoneRan/bert-emotion metrics: [] dataset_name: tweet_eval dataset_config: emotion dataset_split: test col_mapping: text: text target: label --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Multi-class Text Classification * Model: ShoneRan/bert-emotion * Dataset: tweet_eval * Config: emotion * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lenague](https://huggingface.co/lenague) for evaluating this model.
autoevaluate/autoeval-eval-tweet_eval-emotion-cb9f8a-66323145584
2023-10-04T17:21:04.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - tweet_eval eval_info: task: multi_class_classification model: ShoneRan/bert-emotion metrics: [] dataset_name: tweet_eval dataset_config: emotion dataset_split: test col_mapping: text: text target: label --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Multi-class Text Classification * Model: ShoneRan/bert-emotion * Dataset: tweet_eval * Config: emotion * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@Lenague](https://huggingface.co/Lenague) for evaluating this model.
autoevaluate/autoeval-eval-acronym_identification-default-30917a-66451145585
2023-10-04T17:20:56.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-c77ec0-66995145586
2023-10-04T17:20:58.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-super_glue-axg-09004e-67010145587
2023-10-04T17:21:09.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-super_glue-axg-09004e-67010145588
2023-10-04T17:21:18.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-super_glue-axg-09004e-67010145589
2023-10-04T17:21:28.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-imdb-plain_text-5b4ba5-67068145590
2023-10-04T17:21:40.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-imdb-plain_text-fdc5b9-67091145591
2023-10-04T17:26:01.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - imdb eval_info: task: summarization model: t5-small metrics: [] dataset_name: imdb dataset_config: plain_text dataset_split: test col_mapping: text: text target: label --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: t5-small * Dataset: imdb * Config: plain_text * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@michaeldesmond](https://huggingface.co/michaeldesmond) for evaluating this model.
autoevaluate/autoeval-eval-imdb-plain_text-87fbde-67095145592
2023-10-04T17:26:09.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - imdb eval_info: task: summarization model: t5-small metrics: [] dataset_name: imdb dataset_config: plain_text dataset_split: test col_mapping: text: text target: label --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: t5-small * Dataset: imdb * Config: plain_text * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@michaeldesmond](https://huggingface.co/michaeldesmond) for evaluating this model.
autoevaluate/autoeval-eval-xsum-default-5381b8-67099145593
2023-10-04T17:24:35.000Z
[ "autotrain", "evaluation", "region:us" ]
autoevaluate
null
null
null
0
0
--- type: predictions tags: - autotrain - evaluation datasets: - xsum eval_info: task: summarization model: t5-small metrics: [] dataset_name: xsum dataset_config: default dataset_split: test col_mapping: text: document target: summary --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: t5-small * Dataset: xsum * Config: default * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@michaeldesmond](https://huggingface.co/michaeldesmond) for evaluating this model.
niffelheim/DatasetVozAbby
2023-10-04T18:33:36.000Z
[ "license:openrail", "region:us" ]
niffelheim
null
null
null
0
0
--- license: openrail ---
autoevaluate/autoeval-eval-acronym_identification-default-66c824-67126145594
2023-10-04T17:22:21.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-86b43b-67127145595
2023-10-04T17:22:31.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-glue-cola-6c1861-67190145596
2023-10-04T17:22:39.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-b5d5f1-67267145597
2023-10-04T17:22:50.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-acronym_identification-default-53b15c-67271145598
2023-10-04T17:22:59.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
devjirawat8/Air_Vehicle
2023-10-04T17:23:03.000Z
[ "license:mit", "region:us" ]
devjirawat8
null
null
null
0
0
--- license: mit ---
autoevaluate/autoeval-eval-financial_phrasebank-sentences_50agree-10e19c-67276145599
2023-10-04T17:23:10.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-amazon_polarity-amazon_polarity-eccfe6-67532145600
2023-10-04T17:23:21.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found
autoevaluate/autoeval-eval-squad_v2-squad_v2-f30cdf-67536145601
2023-10-04T17:23:31.000Z
[ "region:us" ]
autoevaluate
null
null
null
0
0
Entry not found