datasetId stringlengths 2 117 | card stringlengths 19 1.01M |
|---|---|
korotkov/glaive-function-calling-v2-ru-parsed | ---
task_categories:
- text-generation
language:
- ru
pretty_name: Glaive Function Calling V2 Parsed
size_categories:
- 100K<n<1M
---
# Glaive Function Calling V2 Parsed (RU)
This dataset is a fork of an original [mizinovmv/glaive-function-calling-v2-ru](https://huggingface.co/datasets/mizinovmv/glaive-function-calling-v2-ru).
## Structure
* train (102860 rows)
- messages
- functions
* test (9681 rows)
- messages
- functions
Messages field contains JSON string with list of messages. Each message has role (one of: system, user, assistant, function_call, function_response) and content (message string) fields.
Functions field contains JSON string with list of functions for current chat (their number may vary from 0 to 2).
Functions in train and test sets do not intersect.
## How to use
Example of converting messages list into one string
```python
import json
from datasets import load_dataset
# Download dataset from huggingface
dataset = load_dataset("korotkov/glaive-function-calling-v2-ru-parsed")
# Convert messages list into string
messages_string = []
for message in json.loads(dataset["train"]["messages"][1]):
messages_string.append(f"{message['role'].upper()}: {message['content']}")
messages_string = "\n\n".join(messages_string)
print(messages_string)
```
Output:
```json
SYSTEM: Ты полезный помощник с доступом к следующим функциям. Используй их при необходимости:
{
"name": "calculate_square",
"description": "Вычислить квадрат числа",
"parameters": {
"type": "object",
"properties": {
"number": {
"type": "number",
"description": "Число для вычисления квадрата"
}
},
"required": [
"number"
]
}
},
{
"name": "schedule_appointment",
"description": "Запланировать встречу",
"parameters": {
"type": "object",
"properties": {
"date": {
"type": "string",
"description": "Дата назначения"
},
"time": {
"type": "string",
"description": "Время встречи"
},
"description": {
"type": "string",
"description": "Описание встречи"
}
},
"required": [
"date",
"time"
]
}
}
USER: Привет, мне нужна помощь с задачей по математике.
ASSISTANT: Конечно! Я был бы рад помочь. В чем проблема?
USER: Мне нужно найти квадрат 7.
FUNCTION_CALL: {"name": "calculate_square", "arguments": {"number": 7}}
FUNCTION_RESPONSE: {"result": 49}
ASSISTANT: Квадрат 7 равен 49.
USER: Большое спасибо! А теперь можешь помочь мне заказать пиццу?
ASSISTANT: Извините, но я, как ИИ, не имею возможности выполнять внешние задачи, например, заказывать пиццу. Я могу оказывать помощь только в соответствии с запрограммированными мной функциями. В настоящее время я могу помочь с вычислением квадрата числа.
``` |
SINAI/share | ---
license: cc-by-nc-sa-4.0
language:
- es
tags:
- hate speech
- insults
- spanish
- offensive expressions
- offensive language
pretty_name: SHARE
size_categories:
- 10K<n<100K
---
### Dataset Description
**Paper**: [SHARE: A Lexicon of Harmful Expressions by Spanish Speakers](https://aclanthology.org/2022.lrec-1.139.pdf)
**Point of Contact**: flor.plaza@unibocconi.it, maite@ujaen.es
### Source Data
Telegram
### Licensing Information
SHARE is released under the [Apache-2.0 License](http://www.apache.org/licenses/LICENSE-2.0).
### Citation Information
```
@inproceedings{plaza-del-arco-etal-2022-share,
title = "{SHARE}: A Lexicon of Harmful Expressions by {S}panish Speakers",
author = "{Plaza-del-Arco}, Flor Miriam and
Parras Portillo, Ana Belén and
López-Úbeda, Pilar and
Gil, Beatriz and
Martín-Valdivia, María-Teresa",
booktitle = "Proceedings of the Thirteenth Language Resources and Evaluation Conference",
month = jun,
year = "2022",
address = "Marseille, France",
publisher = "European Language Resources Association",
url = "https://aclanthology.org/2022.lrec-1.139",
pages = "1307--1316"}
``` |
renumics/spotlight-mnist-enrichment | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: image.embedding
sequence: float32
length: 2
splits:
- name: train
num_bytes: 480000
num_examples: 60000
- name: test
num_bytes: 80000
num_examples: 10000
download_size: 778435
dataset_size: 560000
---
# Dataset Card for "spotlight-mnist-enrichment"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
AdapterOcean/med_alpaca_standardized_cluster_52_alpaca | ---
dataset_info:
features:
- name: input
dtype: string
- name: output
dtype: string
splits:
- name: train
num_bytes: 29809228
num_examples: 15142
download_size: 14860344
dataset_size: 29809228
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "med_alpaca_standardized_cluster_52_alpaca"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
joey234/mmlu-philosophy-original-neg | ---
dataset_info:
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: test
num_bytes: 11326.958199356914
num_examples: 44
download_size: 14783
dataset_size: 11326.958199356914
---
# Dataset Card for "mmlu-philosophy-original-neg"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
mask-distilled-onesec-cv12-each-chunk-uniq/chunk_204 | ---
dataset_info:
features:
- name: logits
sequence: float32
- name: mfcc
sequence:
sequence: float64
splits:
- name: train
num_bytes: 1123478512.0
num_examples: 220636
download_size: 1147792515
dataset_size: 1123478512.0
---
# Dataset Card for "chunk_204"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
lumijek/deeplense-diffusion | ---
language:
- en
size_categories:
- 10K<n<100K
--- |
SadeghJafari/CVC | ---
license: apache-2.0
---
|
maitri-vv/UN16_Peace-Justice | ---
task_categories:
- question-answering
tags:
- legal
size_categories:
- n<1K
---
# Dataset Card for Dataset Name
## Dataset Description
- **Homepage:**
- **Repository:**
- **Paper:**
- **Leaderboard:**
- **Point of Contact:**
### Dataset Summary
This dataset card aims to be a base template for new datasets. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/datasetcard_template.md?plain=1).
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] |
open-llm-leaderboard/details_Minami-su__Qwen1.5-0.5B-Chat_mistral | ---
pretty_name: Evaluation run of Minami-su/Qwen1.5-0.5B-Chat_mistral
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [Minami-su/Qwen1.5-0.5B-Chat_mistral](https://huggingface.co/Minami-su/Qwen1.5-0.5B-Chat_mistral)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 63 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the aggregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_Minami-su__Qwen1.5-0.5B-Chat_mistral\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2024-03-25T09:24:56.116599](https://huggingface.co/datasets/open-llm-leaderboard/details_Minami-su__Qwen1.5-0.5B-Chat_mistral/blob/main/results_2024-03-25T09-24-56.116599.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"acc\": 0.23146707204075107,\n\
\ \"acc_stderr\": 0.029930144939974264,\n \"acc_norm\": 0.23180919521674745,\n\
\ \"acc_norm_stderr\": 0.03071922270148664,\n \"mc1\": 0.23011015911872704,\n\
\ \"mc1_stderr\": 0.014734557959807756,\n \"mc2\": 0.49059783076199204,\n\
\ \"mc2_stderr\": 0.01679727487548572\n },\n \"harness|arc:challenge|25\"\
: {\n \"acc\": 0.2158703071672355,\n \"acc_stderr\": 0.012022975360030679,\n\
\ \"acc_norm\": 0.2551194539249147,\n \"acc_norm_stderr\": 0.012739038695202102\n\
\ },\n \"harness|hellaswag|10\": {\n \"acc\": 0.2591117307309301,\n\
\ \"acc_stderr\": 0.0043725160601647455,\n \"acc_norm\": 0.2640908185620394,\n\
\ \"acc_norm_stderr\": 0.004399469168272512\n },\n \"harness|hendrycksTest-abstract_algebra|5\"\
: {\n \"acc\": 0.22,\n \"acc_stderr\": 0.04163331998932268,\n \
\ \"acc_norm\": 0.22,\n \"acc_norm_stderr\": 0.04163331998932268\n \
\ },\n \"harness|hendrycksTest-anatomy|5\": {\n \"acc\": 0.18518518518518517,\n\
\ \"acc_stderr\": 0.03355677216313142,\n \"acc_norm\": 0.18518518518518517,\n\
\ \"acc_norm_stderr\": 0.03355677216313142\n },\n \"harness|hendrycksTest-astronomy|5\"\
: {\n \"acc\": 0.17763157894736842,\n \"acc_stderr\": 0.031103182383123398,\n\
\ \"acc_norm\": 0.17763157894736842,\n \"acc_norm_stderr\": 0.031103182383123398\n\
\ },\n \"harness|hendrycksTest-business_ethics|5\": {\n \"acc\": 0.3,\n\
\ \"acc_stderr\": 0.046056618647183814,\n \"acc_norm\": 0.3,\n \
\ \"acc_norm_stderr\": 0.046056618647183814\n },\n \"harness|hendrycksTest-clinical_knowledge|5\"\
: {\n \"acc\": 0.21509433962264152,\n \"acc_stderr\": 0.02528839450289137,\n\
\ \"acc_norm\": 0.21509433962264152,\n \"acc_norm_stderr\": 0.02528839450289137\n\
\ },\n \"harness|hendrycksTest-college_biology|5\": {\n \"acc\": 0.2638888888888889,\n\
\ \"acc_stderr\": 0.03685651095897532,\n \"acc_norm\": 0.2638888888888889,\n\
\ \"acc_norm_stderr\": 0.03685651095897532\n },\n \"harness|hendrycksTest-college_chemistry|5\"\
: {\n \"acc\": 0.2,\n \"acc_stderr\": 0.04020151261036845,\n \
\ \"acc_norm\": 0.2,\n \"acc_norm_stderr\": 0.04020151261036845\n },\n\
\ \"harness|hendrycksTest-college_computer_science|5\": {\n \"acc\": 0.26,\n\
\ \"acc_stderr\": 0.0440844002276808,\n \"acc_norm\": 0.26,\n \
\ \"acc_norm_stderr\": 0.0440844002276808\n },\n \"harness|hendrycksTest-college_mathematics|5\"\
: {\n \"acc\": 0.21,\n \"acc_stderr\": 0.040936018074033256,\n \
\ \"acc_norm\": 0.21,\n \"acc_norm_stderr\": 0.040936018074033256\n \
\ },\n \"harness|hendrycksTest-college_medicine|5\": {\n \"acc\": 0.21965317919075145,\n\
\ \"acc_stderr\": 0.031568093627031744,\n \"acc_norm\": 0.21965317919075145,\n\
\ \"acc_norm_stderr\": 0.031568093627031744\n },\n \"harness|hendrycksTest-college_physics|5\"\
: {\n \"acc\": 0.21568627450980393,\n \"acc_stderr\": 0.04092563958237654,\n\
\ \"acc_norm\": 0.21568627450980393,\n \"acc_norm_stderr\": 0.04092563958237654\n\
\ },\n \"harness|hendrycksTest-computer_security|5\": {\n \"acc\":\
\ 0.28,\n \"acc_stderr\": 0.045126085985421276,\n \"acc_norm\": 0.28,\n\
\ \"acc_norm_stderr\": 0.045126085985421276\n },\n \"harness|hendrycksTest-conceptual_physics|5\"\
: {\n \"acc\": 0.26382978723404255,\n \"acc_stderr\": 0.028809989854102973,\n\
\ \"acc_norm\": 0.26382978723404255,\n \"acc_norm_stderr\": 0.028809989854102973\n\
\ },\n \"harness|hendrycksTest-econometrics|5\": {\n \"acc\": 0.23684210526315788,\n\
\ \"acc_stderr\": 0.039994238792813365,\n \"acc_norm\": 0.23684210526315788,\n\
\ \"acc_norm_stderr\": 0.039994238792813365\n },\n \"harness|hendrycksTest-electrical_engineering|5\"\
: {\n \"acc\": 0.2413793103448276,\n \"acc_stderr\": 0.03565998174135302,\n\
\ \"acc_norm\": 0.2413793103448276,\n \"acc_norm_stderr\": 0.03565998174135302\n\
\ },\n \"harness|hendrycksTest-elementary_mathematics|5\": {\n \"acc\"\
: 0.20899470899470898,\n \"acc_stderr\": 0.02094048156533486,\n \"\
acc_norm\": 0.20899470899470898,\n \"acc_norm_stderr\": 0.02094048156533486\n\
\ },\n \"harness|hendrycksTest-formal_logic|5\": {\n \"acc\": 0.2857142857142857,\n\
\ \"acc_stderr\": 0.04040610178208841,\n \"acc_norm\": 0.2857142857142857,\n\
\ \"acc_norm_stderr\": 0.04040610178208841\n },\n \"harness|hendrycksTest-global_facts|5\"\
: {\n \"acc\": 0.18,\n \"acc_stderr\": 0.038612291966536934,\n \
\ \"acc_norm\": 0.18,\n \"acc_norm_stderr\": 0.038612291966536934\n \
\ },\n \"harness|hendrycksTest-high_school_biology|5\": {\n \"acc\"\
: 0.1774193548387097,\n \"acc_stderr\": 0.02173254068932927,\n \"\
acc_norm\": 0.1774193548387097,\n \"acc_norm_stderr\": 0.02173254068932927\n\
\ },\n \"harness|hendrycksTest-high_school_chemistry|5\": {\n \"acc\"\
: 0.15270935960591134,\n \"acc_stderr\": 0.02530890453938063,\n \"\
acc_norm\": 0.15270935960591134,\n \"acc_norm_stderr\": 0.02530890453938063\n\
\ },\n \"harness|hendrycksTest-high_school_computer_science|5\": {\n \
\ \"acc\": 0.25,\n \"acc_stderr\": 0.04351941398892446,\n \"acc_norm\"\
: 0.25,\n \"acc_norm_stderr\": 0.04351941398892446\n },\n \"harness|hendrycksTest-high_school_european_history|5\"\
: {\n \"acc\": 0.21818181818181817,\n \"acc_stderr\": 0.03225078108306289,\n\
\ \"acc_norm\": 0.21818181818181817,\n \"acc_norm_stderr\": 0.03225078108306289\n\
\ },\n \"harness|hendrycksTest-high_school_geography|5\": {\n \"acc\"\
: 0.17676767676767677,\n \"acc_stderr\": 0.027178752639044915,\n \"\
acc_norm\": 0.17676767676767677,\n \"acc_norm_stderr\": 0.027178752639044915\n\
\ },\n \"harness|hendrycksTest-high_school_government_and_politics|5\": {\n\
\ \"acc\": 0.19689119170984457,\n \"acc_stderr\": 0.028697873971860664,\n\
\ \"acc_norm\": 0.19689119170984457,\n \"acc_norm_stderr\": 0.028697873971860664\n\
\ },\n \"harness|hendrycksTest-high_school_macroeconomics|5\": {\n \
\ \"acc\": 0.20256410256410257,\n \"acc_stderr\": 0.020377660970371372,\n\
\ \"acc_norm\": 0.20256410256410257,\n \"acc_norm_stderr\": 0.020377660970371372\n\
\ },\n \"harness|hendrycksTest-high_school_mathematics|5\": {\n \"\
acc\": 0.2111111111111111,\n \"acc_stderr\": 0.024882116857655075,\n \
\ \"acc_norm\": 0.2111111111111111,\n \"acc_norm_stderr\": 0.024882116857655075\n\
\ },\n \"harness|hendrycksTest-high_school_microeconomics|5\": {\n \
\ \"acc\": 0.21008403361344538,\n \"acc_stderr\": 0.026461398717471874,\n\
\ \"acc_norm\": 0.21008403361344538,\n \"acc_norm_stderr\": 0.026461398717471874\n\
\ },\n \"harness|hendrycksTest-high_school_physics|5\": {\n \"acc\"\
: 0.1986754966887417,\n \"acc_stderr\": 0.03257847384436776,\n \"\
acc_norm\": 0.1986754966887417,\n \"acc_norm_stderr\": 0.03257847384436776\n\
\ },\n \"harness|hendrycksTest-high_school_psychology|5\": {\n \"acc\"\
: 0.1926605504587156,\n \"acc_stderr\": 0.016909276884936094,\n \"\
acc_norm\": 0.1926605504587156,\n \"acc_norm_stderr\": 0.016909276884936094\n\
\ },\n \"harness|hendrycksTest-high_school_statistics|5\": {\n \"acc\"\
: 0.1527777777777778,\n \"acc_stderr\": 0.024536326026134224,\n \"\
acc_norm\": 0.1527777777777778,\n \"acc_norm_stderr\": 0.024536326026134224\n\
\ },\n \"harness|hendrycksTest-high_school_us_history|5\": {\n \"acc\"\
: 0.24509803921568626,\n \"acc_stderr\": 0.030190282453501954,\n \"\
acc_norm\": 0.24509803921568626,\n \"acc_norm_stderr\": 0.030190282453501954\n\
\ },\n \"harness|hendrycksTest-high_school_world_history|5\": {\n \"\
acc\": 0.270042194092827,\n \"acc_stderr\": 0.028900721906293426,\n \
\ \"acc_norm\": 0.270042194092827,\n \"acc_norm_stderr\": 0.028900721906293426\n\
\ },\n \"harness|hendrycksTest-human_aging|5\": {\n \"acc\": 0.31390134529147984,\n\
\ \"acc_stderr\": 0.031146796482972465,\n \"acc_norm\": 0.31390134529147984,\n\
\ \"acc_norm_stderr\": 0.031146796482972465\n },\n \"harness|hendrycksTest-human_sexuality|5\"\
: {\n \"acc\": 0.2595419847328244,\n \"acc_stderr\": 0.03844876139785271,\n\
\ \"acc_norm\": 0.2595419847328244,\n \"acc_norm_stderr\": 0.03844876139785271\n\
\ },\n \"harness|hendrycksTest-international_law|5\": {\n \"acc\":\
\ 0.2396694214876033,\n \"acc_stderr\": 0.03896878985070417,\n \"\
acc_norm\": 0.2396694214876033,\n \"acc_norm_stderr\": 0.03896878985070417\n\
\ },\n \"harness|hendrycksTest-jurisprudence|5\": {\n \"acc\": 0.25925925925925924,\n\
\ \"acc_stderr\": 0.042365112580946336,\n \"acc_norm\": 0.25925925925925924,\n\
\ \"acc_norm_stderr\": 0.042365112580946336\n },\n \"harness|hendrycksTest-logical_fallacies|5\"\
: {\n \"acc\": 0.22085889570552147,\n \"acc_stderr\": 0.032591773927421776,\n\
\ \"acc_norm\": 0.22085889570552147,\n \"acc_norm_stderr\": 0.032591773927421776\n\
\ },\n \"harness|hendrycksTest-machine_learning|5\": {\n \"acc\": 0.3125,\n\
\ \"acc_stderr\": 0.043994650575715215,\n \"acc_norm\": 0.3125,\n\
\ \"acc_norm_stderr\": 0.043994650575715215\n },\n \"harness|hendrycksTest-management|5\"\
: {\n \"acc\": 0.17475728155339806,\n \"acc_stderr\": 0.037601780060266224,\n\
\ \"acc_norm\": 0.17475728155339806,\n \"acc_norm_stderr\": 0.037601780060266224\n\
\ },\n \"harness|hendrycksTest-marketing|5\": {\n \"acc\": 0.2905982905982906,\n\
\ \"acc_stderr\": 0.02974504857267404,\n \"acc_norm\": 0.2905982905982906,\n\
\ \"acc_norm_stderr\": 0.02974504857267404\n },\n \"harness|hendrycksTest-medical_genetics|5\"\
: {\n \"acc\": 0.3,\n \"acc_stderr\": 0.046056618647183814,\n \
\ \"acc_norm\": 0.3,\n \"acc_norm_stderr\": 0.046056618647183814\n \
\ },\n \"harness|hendrycksTest-miscellaneous|5\": {\n \"acc\": 0.23754789272030652,\n\
\ \"acc_stderr\": 0.015218733046150193,\n \"acc_norm\": 0.23754789272030652,\n\
\ \"acc_norm_stderr\": 0.015218733046150193\n },\n \"harness|hendrycksTest-moral_disputes|5\"\
: {\n \"acc\": 0.24855491329479767,\n \"acc_stderr\": 0.023267528432100174,\n\
\ \"acc_norm\": 0.24855491329479767,\n \"acc_norm_stderr\": 0.023267528432100174\n\
\ },\n \"harness|hendrycksTest-moral_scenarios|5\": {\n \"acc\": 0.23798882681564246,\n\
\ \"acc_stderr\": 0.014242630070574915,\n \"acc_norm\": 0.23798882681564246,\n\
\ \"acc_norm_stderr\": 0.014242630070574915\n },\n \"harness|hendrycksTest-nutrition|5\"\
: {\n \"acc\": 0.22549019607843138,\n \"acc_stderr\": 0.023929155517351284,\n\
\ \"acc_norm\": 0.22549019607843138,\n \"acc_norm_stderr\": 0.023929155517351284\n\
\ },\n \"harness|hendrycksTest-philosophy|5\": {\n \"acc\": 0.1864951768488746,\n\
\ \"acc_stderr\": 0.022122439772480792,\n \"acc_norm\": 0.1864951768488746,\n\
\ \"acc_norm_stderr\": 0.022122439772480792\n },\n \"harness|hendrycksTest-prehistory|5\"\
: {\n \"acc\": 0.21604938271604937,\n \"acc_stderr\": 0.022899162918445806,\n\
\ \"acc_norm\": 0.21604938271604937,\n \"acc_norm_stderr\": 0.022899162918445806\n\
\ },\n \"harness|hendrycksTest-professional_accounting|5\": {\n \"\
acc\": 0.23404255319148937,\n \"acc_stderr\": 0.025257861359432417,\n \
\ \"acc_norm\": 0.23404255319148937,\n \"acc_norm_stderr\": 0.025257861359432417\n\
\ },\n \"harness|hendrycksTest-professional_law|5\": {\n \"acc\": 0.2457627118644068,\n\
\ \"acc_stderr\": 0.010996156635142692,\n \"acc_norm\": 0.2457627118644068,\n\
\ \"acc_norm_stderr\": 0.010996156635142692\n },\n \"harness|hendrycksTest-professional_medicine|5\"\
: {\n \"acc\": 0.18382352941176472,\n \"acc_stderr\": 0.023529242185193106,\n\
\ \"acc_norm\": 0.18382352941176472,\n \"acc_norm_stderr\": 0.023529242185193106\n\
\ },\n \"harness|hendrycksTest-professional_psychology|5\": {\n \"\
acc\": 0.2173202614379085,\n \"acc_stderr\": 0.01668482092914861,\n \
\ \"acc_norm\": 0.2173202614379085,\n \"acc_norm_stderr\": 0.01668482092914861\n\
\ },\n \"harness|hendrycksTest-public_relations|5\": {\n \"acc\": 0.21818181818181817,\n\
\ \"acc_stderr\": 0.03955932861795833,\n \"acc_norm\": 0.21818181818181817,\n\
\ \"acc_norm_stderr\": 0.03955932861795833\n },\n \"harness|hendrycksTest-security_studies|5\"\
: {\n \"acc\": 0.18775510204081633,\n \"acc_stderr\": 0.02500025603954621,\n\
\ \"acc_norm\": 0.18775510204081633,\n \"acc_norm_stderr\": 0.02500025603954621\n\
\ },\n \"harness|hendrycksTest-sociology|5\": {\n \"acc\": 0.24378109452736318,\n\
\ \"acc_stderr\": 0.03036049015401465,\n \"acc_norm\": 0.24378109452736318,\n\
\ \"acc_norm_stderr\": 0.03036049015401465\n },\n \"harness|hendrycksTest-us_foreign_policy|5\"\
: {\n \"acc\": 0.28,\n \"acc_stderr\": 0.04512608598542128,\n \
\ \"acc_norm\": 0.28,\n \"acc_norm_stderr\": 0.04512608598542128\n \
\ },\n \"harness|hendrycksTest-virology|5\": {\n \"acc\": 0.28313253012048195,\n\
\ \"acc_stderr\": 0.03507295431370518,\n \"acc_norm\": 0.28313253012048195,\n\
\ \"acc_norm_stderr\": 0.03507295431370518\n },\n \"harness|hendrycksTest-world_religions|5\"\
: {\n \"acc\": 0.3216374269005848,\n \"acc_stderr\": 0.03582529442573122,\n\
\ \"acc_norm\": 0.3216374269005848,\n \"acc_norm_stderr\": 0.03582529442573122\n\
\ },\n \"harness|truthfulqa:mc|0\": {\n \"mc1\": 0.23011015911872704,\n\
\ \"mc1_stderr\": 0.014734557959807756,\n \"mc2\": 0.49059783076199204,\n\
\ \"mc2_stderr\": 0.01679727487548572\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.48697711128650356,\n \"acc_stderr\": 0.014047718393997667\n\
\ },\n \"harness|gsm8k|5\": {\n \"acc\": 0.0,\n \"acc_stderr\"\
: 0.0\n }\n}\n```"
repo_url: https://huggingface.co/Minami-su/Qwen1.5-0.5B-Chat_mistral
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|arc:challenge|25_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|gsm8k|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hellaswag|10_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-management|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-virology|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-management|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-virology|5_2024-03-25T09-24-56.116599.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-anatomy|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-astronomy|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-college_biology|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-college_physics|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-computer_security|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-econometrics|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-global_facts|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-human_aging|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-international_law|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-management|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-marketing|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-nutrition|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-philosophy|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-prehistory|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-professional_law|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-public_relations|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-security_studies|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-sociology|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-virology|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|hendrycksTest-world_religions|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|truthfulqa:mc|0_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2024-03-25T09-24-56.116599.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- '**/details_harness|winogrande|5_2024-03-25T09-24-56.116599.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2024-03-25T09-24-56.116599.parquet'
- config_name: results
data_files:
- split: 2024_03_25T09_24_56.116599
path:
- results_2024-03-25T09-24-56.116599.parquet
- split: latest
path:
- results_2024-03-25T09-24-56.116599.parquet
---
# Dataset Card for Evaluation run of Minami-su/Qwen1.5-0.5B-Chat_mistral
<!-- Provide a quick summary of the dataset. -->
Dataset automatically created during the evaluation run of model [Minami-su/Qwen1.5-0.5B-Chat_mistral](https://huggingface.co/Minami-su/Qwen1.5-0.5B-Chat_mistral) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 63 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_Minami-su__Qwen1.5-0.5B-Chat_mistral",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2024-03-25T09:24:56.116599](https://huggingface.co/datasets/open-llm-leaderboard/details_Minami-su__Qwen1.5-0.5B-Chat_mistral/blob/main/results_2024-03-25T09-24-56.116599.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"acc": 0.23146707204075107,
"acc_stderr": 0.029930144939974264,
"acc_norm": 0.23180919521674745,
"acc_norm_stderr": 0.03071922270148664,
"mc1": 0.23011015911872704,
"mc1_stderr": 0.014734557959807756,
"mc2": 0.49059783076199204,
"mc2_stderr": 0.01679727487548572
},
"harness|arc:challenge|25": {
"acc": 0.2158703071672355,
"acc_stderr": 0.012022975360030679,
"acc_norm": 0.2551194539249147,
"acc_norm_stderr": 0.012739038695202102
},
"harness|hellaswag|10": {
"acc": 0.2591117307309301,
"acc_stderr": 0.0043725160601647455,
"acc_norm": 0.2640908185620394,
"acc_norm_stderr": 0.004399469168272512
},
"harness|hendrycksTest-abstract_algebra|5": {
"acc": 0.22,
"acc_stderr": 0.04163331998932268,
"acc_norm": 0.22,
"acc_norm_stderr": 0.04163331998932268
},
"harness|hendrycksTest-anatomy|5": {
"acc": 0.18518518518518517,
"acc_stderr": 0.03355677216313142,
"acc_norm": 0.18518518518518517,
"acc_norm_stderr": 0.03355677216313142
},
"harness|hendrycksTest-astronomy|5": {
"acc": 0.17763157894736842,
"acc_stderr": 0.031103182383123398,
"acc_norm": 0.17763157894736842,
"acc_norm_stderr": 0.031103182383123398
},
"harness|hendrycksTest-business_ethics|5": {
"acc": 0.3,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.3,
"acc_norm_stderr": 0.046056618647183814
},
"harness|hendrycksTest-clinical_knowledge|5": {
"acc": 0.21509433962264152,
"acc_stderr": 0.02528839450289137,
"acc_norm": 0.21509433962264152,
"acc_norm_stderr": 0.02528839450289137
},
"harness|hendrycksTest-college_biology|5": {
"acc": 0.2638888888888889,
"acc_stderr": 0.03685651095897532,
"acc_norm": 0.2638888888888889,
"acc_norm_stderr": 0.03685651095897532
},
"harness|hendrycksTest-college_chemistry|5": {
"acc": 0.2,
"acc_stderr": 0.04020151261036845,
"acc_norm": 0.2,
"acc_norm_stderr": 0.04020151261036845
},
"harness|hendrycksTest-college_computer_science|5": {
"acc": 0.26,
"acc_stderr": 0.0440844002276808,
"acc_norm": 0.26,
"acc_norm_stderr": 0.0440844002276808
},
"harness|hendrycksTest-college_mathematics|5": {
"acc": 0.21,
"acc_stderr": 0.040936018074033256,
"acc_norm": 0.21,
"acc_norm_stderr": 0.040936018074033256
},
"harness|hendrycksTest-college_medicine|5": {
"acc": 0.21965317919075145,
"acc_stderr": 0.031568093627031744,
"acc_norm": 0.21965317919075145,
"acc_norm_stderr": 0.031568093627031744
},
"harness|hendrycksTest-college_physics|5": {
"acc": 0.21568627450980393,
"acc_stderr": 0.04092563958237654,
"acc_norm": 0.21568627450980393,
"acc_norm_stderr": 0.04092563958237654
},
"harness|hendrycksTest-computer_security|5": {
"acc": 0.28,
"acc_stderr": 0.045126085985421276,
"acc_norm": 0.28,
"acc_norm_stderr": 0.045126085985421276
},
"harness|hendrycksTest-conceptual_physics|5": {
"acc": 0.26382978723404255,
"acc_stderr": 0.028809989854102973,
"acc_norm": 0.26382978723404255,
"acc_norm_stderr": 0.028809989854102973
},
"harness|hendrycksTest-econometrics|5": {
"acc": 0.23684210526315788,
"acc_stderr": 0.039994238792813365,
"acc_norm": 0.23684210526315788,
"acc_norm_stderr": 0.039994238792813365
},
"harness|hendrycksTest-electrical_engineering|5": {
"acc": 0.2413793103448276,
"acc_stderr": 0.03565998174135302,
"acc_norm": 0.2413793103448276,
"acc_norm_stderr": 0.03565998174135302
},
"harness|hendrycksTest-elementary_mathematics|5": {
"acc": 0.20899470899470898,
"acc_stderr": 0.02094048156533486,
"acc_norm": 0.20899470899470898,
"acc_norm_stderr": 0.02094048156533486
},
"harness|hendrycksTest-formal_logic|5": {
"acc": 0.2857142857142857,
"acc_stderr": 0.04040610178208841,
"acc_norm": 0.2857142857142857,
"acc_norm_stderr": 0.04040610178208841
},
"harness|hendrycksTest-global_facts|5": {
"acc": 0.18,
"acc_stderr": 0.038612291966536934,
"acc_norm": 0.18,
"acc_norm_stderr": 0.038612291966536934
},
"harness|hendrycksTest-high_school_biology|5": {
"acc": 0.1774193548387097,
"acc_stderr": 0.02173254068932927,
"acc_norm": 0.1774193548387097,
"acc_norm_stderr": 0.02173254068932927
},
"harness|hendrycksTest-high_school_chemistry|5": {
"acc": 0.15270935960591134,
"acc_stderr": 0.02530890453938063,
"acc_norm": 0.15270935960591134,
"acc_norm_stderr": 0.02530890453938063
},
"harness|hendrycksTest-high_school_computer_science|5": {
"acc": 0.25,
"acc_stderr": 0.04351941398892446,
"acc_norm": 0.25,
"acc_norm_stderr": 0.04351941398892446
},
"harness|hendrycksTest-high_school_european_history|5": {
"acc": 0.21818181818181817,
"acc_stderr": 0.03225078108306289,
"acc_norm": 0.21818181818181817,
"acc_norm_stderr": 0.03225078108306289
},
"harness|hendrycksTest-high_school_geography|5": {
"acc": 0.17676767676767677,
"acc_stderr": 0.027178752639044915,
"acc_norm": 0.17676767676767677,
"acc_norm_stderr": 0.027178752639044915
},
"harness|hendrycksTest-high_school_government_and_politics|5": {
"acc": 0.19689119170984457,
"acc_stderr": 0.028697873971860664,
"acc_norm": 0.19689119170984457,
"acc_norm_stderr": 0.028697873971860664
},
"harness|hendrycksTest-high_school_macroeconomics|5": {
"acc": 0.20256410256410257,
"acc_stderr": 0.020377660970371372,
"acc_norm": 0.20256410256410257,
"acc_norm_stderr": 0.020377660970371372
},
"harness|hendrycksTest-high_school_mathematics|5": {
"acc": 0.2111111111111111,
"acc_stderr": 0.024882116857655075,
"acc_norm": 0.2111111111111111,
"acc_norm_stderr": 0.024882116857655075
},
"harness|hendrycksTest-high_school_microeconomics|5": {
"acc": 0.21008403361344538,
"acc_stderr": 0.026461398717471874,
"acc_norm": 0.21008403361344538,
"acc_norm_stderr": 0.026461398717471874
},
"harness|hendrycksTest-high_school_physics|5": {
"acc": 0.1986754966887417,
"acc_stderr": 0.03257847384436776,
"acc_norm": 0.1986754966887417,
"acc_norm_stderr": 0.03257847384436776
},
"harness|hendrycksTest-high_school_psychology|5": {
"acc": 0.1926605504587156,
"acc_stderr": 0.016909276884936094,
"acc_norm": 0.1926605504587156,
"acc_norm_stderr": 0.016909276884936094
},
"harness|hendrycksTest-high_school_statistics|5": {
"acc": 0.1527777777777778,
"acc_stderr": 0.024536326026134224,
"acc_norm": 0.1527777777777778,
"acc_norm_stderr": 0.024536326026134224
},
"harness|hendrycksTest-high_school_us_history|5": {
"acc": 0.24509803921568626,
"acc_stderr": 0.030190282453501954,
"acc_norm": 0.24509803921568626,
"acc_norm_stderr": 0.030190282453501954
},
"harness|hendrycksTest-high_school_world_history|5": {
"acc": 0.270042194092827,
"acc_stderr": 0.028900721906293426,
"acc_norm": 0.270042194092827,
"acc_norm_stderr": 0.028900721906293426
},
"harness|hendrycksTest-human_aging|5": {
"acc": 0.31390134529147984,
"acc_stderr": 0.031146796482972465,
"acc_norm": 0.31390134529147984,
"acc_norm_stderr": 0.031146796482972465
},
"harness|hendrycksTest-human_sexuality|5": {
"acc": 0.2595419847328244,
"acc_stderr": 0.03844876139785271,
"acc_norm": 0.2595419847328244,
"acc_norm_stderr": 0.03844876139785271
},
"harness|hendrycksTest-international_law|5": {
"acc": 0.2396694214876033,
"acc_stderr": 0.03896878985070417,
"acc_norm": 0.2396694214876033,
"acc_norm_stderr": 0.03896878985070417
},
"harness|hendrycksTest-jurisprudence|5": {
"acc": 0.25925925925925924,
"acc_stderr": 0.042365112580946336,
"acc_norm": 0.25925925925925924,
"acc_norm_stderr": 0.042365112580946336
},
"harness|hendrycksTest-logical_fallacies|5": {
"acc": 0.22085889570552147,
"acc_stderr": 0.032591773927421776,
"acc_norm": 0.22085889570552147,
"acc_norm_stderr": 0.032591773927421776
},
"harness|hendrycksTest-machine_learning|5": {
"acc": 0.3125,
"acc_stderr": 0.043994650575715215,
"acc_norm": 0.3125,
"acc_norm_stderr": 0.043994650575715215
},
"harness|hendrycksTest-management|5": {
"acc": 0.17475728155339806,
"acc_stderr": 0.037601780060266224,
"acc_norm": 0.17475728155339806,
"acc_norm_stderr": 0.037601780060266224
},
"harness|hendrycksTest-marketing|5": {
"acc": 0.2905982905982906,
"acc_stderr": 0.02974504857267404,
"acc_norm": 0.2905982905982906,
"acc_norm_stderr": 0.02974504857267404
},
"harness|hendrycksTest-medical_genetics|5": {
"acc": 0.3,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.3,
"acc_norm_stderr": 0.046056618647183814
},
"harness|hendrycksTest-miscellaneous|5": {
"acc": 0.23754789272030652,
"acc_stderr": 0.015218733046150193,
"acc_norm": 0.23754789272030652,
"acc_norm_stderr": 0.015218733046150193
},
"harness|hendrycksTest-moral_disputes|5": {
"acc": 0.24855491329479767,
"acc_stderr": 0.023267528432100174,
"acc_norm": 0.24855491329479767,
"acc_norm_stderr": 0.023267528432100174
},
"harness|hendrycksTest-moral_scenarios|5": {
"acc": 0.23798882681564246,
"acc_stderr": 0.014242630070574915,
"acc_norm": 0.23798882681564246,
"acc_norm_stderr": 0.014242630070574915
},
"harness|hendrycksTest-nutrition|5": {
"acc": 0.22549019607843138,
"acc_stderr": 0.023929155517351284,
"acc_norm": 0.22549019607843138,
"acc_norm_stderr": 0.023929155517351284
},
"harness|hendrycksTest-philosophy|5": {
"acc": 0.1864951768488746,
"acc_stderr": 0.022122439772480792,
"acc_norm": 0.1864951768488746,
"acc_norm_stderr": 0.022122439772480792
},
"harness|hendrycksTest-prehistory|5": {
"acc": 0.21604938271604937,
"acc_stderr": 0.022899162918445806,
"acc_norm": 0.21604938271604937,
"acc_norm_stderr": 0.022899162918445806
},
"harness|hendrycksTest-professional_accounting|5": {
"acc": 0.23404255319148937,
"acc_stderr": 0.025257861359432417,
"acc_norm": 0.23404255319148937,
"acc_norm_stderr": 0.025257861359432417
},
"harness|hendrycksTest-professional_law|5": {
"acc": 0.2457627118644068,
"acc_stderr": 0.010996156635142692,
"acc_norm": 0.2457627118644068,
"acc_norm_stderr": 0.010996156635142692
},
"harness|hendrycksTest-professional_medicine|5": {
"acc": 0.18382352941176472,
"acc_stderr": 0.023529242185193106,
"acc_norm": 0.18382352941176472,
"acc_norm_stderr": 0.023529242185193106
},
"harness|hendrycksTest-professional_psychology|5": {
"acc": 0.2173202614379085,
"acc_stderr": 0.01668482092914861,
"acc_norm": 0.2173202614379085,
"acc_norm_stderr": 0.01668482092914861
},
"harness|hendrycksTest-public_relations|5": {
"acc": 0.21818181818181817,
"acc_stderr": 0.03955932861795833,
"acc_norm": 0.21818181818181817,
"acc_norm_stderr": 0.03955932861795833
},
"harness|hendrycksTest-security_studies|5": {
"acc": 0.18775510204081633,
"acc_stderr": 0.02500025603954621,
"acc_norm": 0.18775510204081633,
"acc_norm_stderr": 0.02500025603954621
},
"harness|hendrycksTest-sociology|5": {
"acc": 0.24378109452736318,
"acc_stderr": 0.03036049015401465,
"acc_norm": 0.24378109452736318,
"acc_norm_stderr": 0.03036049015401465
},
"harness|hendrycksTest-us_foreign_policy|5": {
"acc": 0.28,
"acc_stderr": 0.04512608598542128,
"acc_norm": 0.28,
"acc_norm_stderr": 0.04512608598542128
},
"harness|hendrycksTest-virology|5": {
"acc": 0.28313253012048195,
"acc_stderr": 0.03507295431370518,
"acc_norm": 0.28313253012048195,
"acc_norm_stderr": 0.03507295431370518
},
"harness|hendrycksTest-world_religions|5": {
"acc": 0.3216374269005848,
"acc_stderr": 0.03582529442573122,
"acc_norm": 0.3216374269005848,
"acc_norm_stderr": 0.03582529442573122
},
"harness|truthfulqa:mc|0": {
"mc1": 0.23011015911872704,
"mc1_stderr": 0.014734557959807756,
"mc2": 0.49059783076199204,
"mc2_stderr": 0.01679727487548572
},
"harness|winogrande|5": {
"acc": 0.48697711128650356,
"acc_stderr": 0.014047718393997667
},
"harness|gsm8k|5": {
"acc": 0.0,
"acc_stderr": 0.0
}
}
```
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] |
distilled-one-sec-cv12-each-chunk-uniq/chunk_234 | ---
dataset_info:
features:
- name: logits
sequence: float32
- name: mfcc
sequence:
sequence: float64
splits:
- name: train
num_bytes: 1145344364.0
num_examples: 223177
download_size: 1174640337
dataset_size: 1145344364.0
---
# Dataset Card for "chunk_234"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
kpriyanshu256/MultiTabQA-multitable_pretraining-Salesforce-codet5-base_train-markdown-61000 | ---
dataset_info:
features:
- name: input_ids
sequence:
sequence: int32
- name: attention_mask
sequence:
sequence: int8
- name: labels
sequence:
sequence: int64
splits:
- name: train
num_bytes: 13336000
num_examples: 1000
download_size: 1095907
dataset_size: 13336000
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
|
iamnguyen/cdnc_law | ---
dataset_info:
features:
- name: content
dtype: string
- name: citation
dtype: string
- name: meta
struct:
- name: effective_date
dtype: string
- name: issuing_agency
dtype: string
- name: promulgation_date
dtype: string
- name: sign_number
dtype: string
- name: signer
dtype: string
- name: type
dtype: string
- name: url
dtype: string
- name: text
dtype: string
- name: law
sequence: string
splits:
- name: luat
num_bytes: 12032705.477916697
num_examples: 4310
download_size: 3185785
dataset_size: 12032705.477916697
configs:
- config_name: default
data_files:
- split: luat
path: data/luat-*
---
|
open-llm-leaderboard/details_RatanRohith__NeuralPizza-7B-V0.2 | ---
pretty_name: Evaluation run of RatanRohith/NeuralPizza-7B-V0.2
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [RatanRohith/NeuralPizza-7B-V0.2](https://huggingface.co/RatanRohith/NeuralPizza-7B-V0.2)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 63 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the aggregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_RatanRohith__NeuralPizza-7B-V0.2\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2024-01-21T21:04:32.551333](https://huggingface.co/datasets/open-llm-leaderboard/details_RatanRohith__NeuralPizza-7B-V0.2/blob/main/results_2024-01-21T21-04-32.551333.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"acc\": 0.6473430412019332,\n\
\ \"acc_stderr\": 0.03217057797429981,\n \"acc_norm\": 0.64765430784867,\n\
\ \"acc_norm_stderr\": 0.032828743060941856,\n \"mc1\": 0.4565483476132191,\n\
\ \"mc1_stderr\": 0.017437280953183695,\n \"mc2\": 0.6137860051903924,\n\
\ \"mc2_stderr\": 0.015431238638634374\n },\n \"harness|arc:challenge|25\"\
: {\n \"acc\": 0.6535836177474402,\n \"acc_stderr\": 0.013905011180063239,\n\
\ \"acc_norm\": 0.6877133105802048,\n \"acc_norm_stderr\": 0.013542598541688065\n\
\ },\n \"harness|hellaswag|10\": {\n \"acc\": 0.6819358693487353,\n\
\ \"acc_stderr\": 0.004647727222445386,\n \"acc_norm\": 0.8610834495120494,\n\
\ \"acc_norm_stderr\": 0.003451525868724678\n },\n \"harness|hendrycksTest-abstract_algebra|5\"\
: {\n \"acc\": 0.3,\n \"acc_stderr\": 0.046056618647183814,\n \
\ \"acc_norm\": 0.3,\n \"acc_norm_stderr\": 0.046056618647183814\n \
\ },\n \"harness|hendrycksTest-anatomy|5\": {\n \"acc\": 0.6370370370370371,\n\
\ \"acc_stderr\": 0.04153948404742398,\n \"acc_norm\": 0.6370370370370371,\n\
\ \"acc_norm_stderr\": 0.04153948404742398\n },\n \"harness|hendrycksTest-astronomy|5\"\
: {\n \"acc\": 0.6842105263157895,\n \"acc_stderr\": 0.0378272898086547,\n\
\ \"acc_norm\": 0.6842105263157895,\n \"acc_norm_stderr\": 0.0378272898086547\n\
\ },\n \"harness|hendrycksTest-business_ethics|5\": {\n \"acc\": 0.63,\n\
\ \"acc_stderr\": 0.04852365870939099,\n \"acc_norm\": 0.63,\n \
\ \"acc_norm_stderr\": 0.04852365870939099\n },\n \"harness|hendrycksTest-clinical_knowledge|5\"\
: {\n \"acc\": 0.7056603773584905,\n \"acc_stderr\": 0.02804918631569525,\n\
\ \"acc_norm\": 0.7056603773584905,\n \"acc_norm_stderr\": 0.02804918631569525\n\
\ },\n \"harness|hendrycksTest-college_biology|5\": {\n \"acc\": 0.7708333333333334,\n\
\ \"acc_stderr\": 0.03514697467862388,\n \"acc_norm\": 0.7708333333333334,\n\
\ \"acc_norm_stderr\": 0.03514697467862388\n },\n \"harness|hendrycksTest-college_chemistry|5\"\
: {\n \"acc\": 0.5,\n \"acc_stderr\": 0.050251890762960605,\n \
\ \"acc_norm\": 0.5,\n \"acc_norm_stderr\": 0.050251890762960605\n \
\ },\n \"harness|hendrycksTest-college_computer_science|5\": {\n \"acc\"\
: 0.58,\n \"acc_stderr\": 0.049604496374885836,\n \"acc_norm\": 0.58,\n\
\ \"acc_norm_stderr\": 0.049604496374885836\n },\n \"harness|hendrycksTest-college_mathematics|5\"\
: {\n \"acc\": 0.32,\n \"acc_stderr\": 0.046882617226215034,\n \
\ \"acc_norm\": 0.32,\n \"acc_norm_stderr\": 0.046882617226215034\n \
\ },\n \"harness|hendrycksTest-college_medicine|5\": {\n \"acc\": 0.6878612716763006,\n\
\ \"acc_stderr\": 0.035331333893236574,\n \"acc_norm\": 0.6878612716763006,\n\
\ \"acc_norm_stderr\": 0.035331333893236574\n },\n \"harness|hendrycksTest-college_physics|5\"\
: {\n \"acc\": 0.45098039215686275,\n \"acc_stderr\": 0.049512182523962625,\n\
\ \"acc_norm\": 0.45098039215686275,\n \"acc_norm_stderr\": 0.049512182523962625\n\
\ },\n \"harness|hendrycksTest-computer_security|5\": {\n \"acc\":\
\ 0.76,\n \"acc_stderr\": 0.04292346959909283,\n \"acc_norm\": 0.76,\n\
\ \"acc_norm_stderr\": 0.04292346959909283\n },\n \"harness|hendrycksTest-conceptual_physics|5\"\
: {\n \"acc\": 0.5914893617021276,\n \"acc_stderr\": 0.032134180267015755,\n\
\ \"acc_norm\": 0.5914893617021276,\n \"acc_norm_stderr\": 0.032134180267015755\n\
\ },\n \"harness|hendrycksTest-econometrics|5\": {\n \"acc\": 0.4824561403508772,\n\
\ \"acc_stderr\": 0.04700708033551038,\n \"acc_norm\": 0.4824561403508772,\n\
\ \"acc_norm_stderr\": 0.04700708033551038\n },\n \"harness|hendrycksTest-electrical_engineering|5\"\
: {\n \"acc\": 0.5586206896551724,\n \"acc_stderr\": 0.04137931034482757,\n\
\ \"acc_norm\": 0.5586206896551724,\n \"acc_norm_stderr\": 0.04137931034482757\n\
\ },\n \"harness|hendrycksTest-elementary_mathematics|5\": {\n \"acc\"\
: 0.4126984126984127,\n \"acc_stderr\": 0.025355741263055256,\n \"\
acc_norm\": 0.4126984126984127,\n \"acc_norm_stderr\": 0.025355741263055256\n\
\ },\n \"harness|hendrycksTest-formal_logic|5\": {\n \"acc\": 0.4523809523809524,\n\
\ \"acc_stderr\": 0.044518079590553275,\n \"acc_norm\": 0.4523809523809524,\n\
\ \"acc_norm_stderr\": 0.044518079590553275\n },\n \"harness|hendrycksTest-global_facts|5\"\
: {\n \"acc\": 0.33,\n \"acc_stderr\": 0.047258156262526045,\n \
\ \"acc_norm\": 0.33,\n \"acc_norm_stderr\": 0.047258156262526045\n \
\ },\n \"harness|hendrycksTest-high_school_biology|5\": {\n \"acc\"\
: 0.7548387096774194,\n \"acc_stderr\": 0.024472243840895518,\n \"\
acc_norm\": 0.7548387096774194,\n \"acc_norm_stderr\": 0.024472243840895518\n\
\ },\n \"harness|hendrycksTest-high_school_chemistry|5\": {\n \"acc\"\
: 0.4827586206896552,\n \"acc_stderr\": 0.035158955511656986,\n \"\
acc_norm\": 0.4827586206896552,\n \"acc_norm_stderr\": 0.035158955511656986\n\
\ },\n \"harness|hendrycksTest-high_school_computer_science|5\": {\n \
\ \"acc\": 0.71,\n \"acc_stderr\": 0.045604802157206845,\n \"acc_norm\"\
: 0.71,\n \"acc_norm_stderr\": 0.045604802157206845\n },\n \"harness|hendrycksTest-high_school_european_history|5\"\
: {\n \"acc\": 0.7757575757575758,\n \"acc_stderr\": 0.032568666616811015,\n\
\ \"acc_norm\": 0.7757575757575758,\n \"acc_norm_stderr\": 0.032568666616811015\n\
\ },\n \"harness|hendrycksTest-high_school_geography|5\": {\n \"acc\"\
: 0.7929292929292929,\n \"acc_stderr\": 0.02886977846026705,\n \"\
acc_norm\": 0.7929292929292929,\n \"acc_norm_stderr\": 0.02886977846026705\n\
\ },\n \"harness|hendrycksTest-high_school_government_and_politics|5\": {\n\
\ \"acc\": 0.8808290155440415,\n \"acc_stderr\": 0.02338193534812143,\n\
\ \"acc_norm\": 0.8808290155440415,\n \"acc_norm_stderr\": 0.02338193534812143\n\
\ },\n \"harness|hendrycksTest-high_school_macroeconomics|5\": {\n \
\ \"acc\": 0.6666666666666666,\n \"acc_stderr\": 0.023901157979402534,\n\
\ \"acc_norm\": 0.6666666666666666,\n \"acc_norm_stderr\": 0.023901157979402534\n\
\ },\n \"harness|hendrycksTest-high_school_mathematics|5\": {\n \"\
acc\": 0.34074074074074073,\n \"acc_stderr\": 0.028897748741131143,\n \
\ \"acc_norm\": 0.34074074074074073,\n \"acc_norm_stderr\": 0.028897748741131143\n\
\ },\n \"harness|hendrycksTest-high_school_microeconomics|5\": {\n \
\ \"acc\": 0.6890756302521008,\n \"acc_stderr\": 0.03006676158297793,\n \
\ \"acc_norm\": 0.6890756302521008,\n \"acc_norm_stderr\": 0.03006676158297793\n\
\ },\n \"harness|hendrycksTest-high_school_physics|5\": {\n \"acc\"\
: 0.31788079470198677,\n \"acc_stderr\": 0.038020397601079024,\n \"\
acc_norm\": 0.31788079470198677,\n \"acc_norm_stderr\": 0.038020397601079024\n\
\ },\n \"harness|hendrycksTest-high_school_psychology|5\": {\n \"acc\"\
: 0.8495412844036697,\n \"acc_stderr\": 0.015328563932669237,\n \"\
acc_norm\": 0.8495412844036697,\n \"acc_norm_stderr\": 0.015328563932669237\n\
\ },\n \"harness|hendrycksTest-high_school_statistics|5\": {\n \"acc\"\
: 0.5509259259259259,\n \"acc_stderr\": 0.03392238405321617,\n \"\
acc_norm\": 0.5509259259259259,\n \"acc_norm_stderr\": 0.03392238405321617\n\
\ },\n \"harness|hendrycksTest-high_school_us_history|5\": {\n \"acc\"\
: 0.8088235294117647,\n \"acc_stderr\": 0.027599174300640766,\n \"\
acc_norm\": 0.8088235294117647,\n \"acc_norm_stderr\": 0.027599174300640766\n\
\ },\n \"harness|hendrycksTest-high_school_world_history|5\": {\n \"\
acc\": 0.7974683544303798,\n \"acc_stderr\": 0.026160568246601443,\n \
\ \"acc_norm\": 0.7974683544303798,\n \"acc_norm_stderr\": 0.026160568246601443\n\
\ },\n \"harness|hendrycksTest-human_aging|5\": {\n \"acc\": 0.6816143497757847,\n\
\ \"acc_stderr\": 0.03126580522513713,\n \"acc_norm\": 0.6816143497757847,\n\
\ \"acc_norm_stderr\": 0.03126580522513713\n },\n \"harness|hendrycksTest-human_sexuality|5\"\
: {\n \"acc\": 0.7786259541984732,\n \"acc_stderr\": 0.03641297081313729,\n\
\ \"acc_norm\": 0.7786259541984732,\n \"acc_norm_stderr\": 0.03641297081313729\n\
\ },\n \"harness|hendrycksTest-international_law|5\": {\n \"acc\":\
\ 0.7768595041322314,\n \"acc_stderr\": 0.03800754475228733,\n \"\
acc_norm\": 0.7768595041322314,\n \"acc_norm_stderr\": 0.03800754475228733\n\
\ },\n \"harness|hendrycksTest-jurisprudence|5\": {\n \"acc\": 0.7685185185185185,\n\
\ \"acc_stderr\": 0.04077494709252626,\n \"acc_norm\": 0.7685185185185185,\n\
\ \"acc_norm_stderr\": 0.04077494709252626\n },\n \"harness|hendrycksTest-logical_fallacies|5\"\
: {\n \"acc\": 0.7607361963190185,\n \"acc_stderr\": 0.033519538795212696,\n\
\ \"acc_norm\": 0.7607361963190185,\n \"acc_norm_stderr\": 0.033519538795212696\n\
\ },\n \"harness|hendrycksTest-machine_learning|5\": {\n \"acc\": 0.4375,\n\
\ \"acc_stderr\": 0.04708567521880525,\n \"acc_norm\": 0.4375,\n \
\ \"acc_norm_stderr\": 0.04708567521880525\n },\n \"harness|hendrycksTest-management|5\"\
: {\n \"acc\": 0.8058252427184466,\n \"acc_stderr\": 0.039166677628225836,\n\
\ \"acc_norm\": 0.8058252427184466,\n \"acc_norm_stderr\": 0.039166677628225836\n\
\ },\n \"harness|hendrycksTest-marketing|5\": {\n \"acc\": 0.8717948717948718,\n\
\ \"acc_stderr\": 0.02190190511507333,\n \"acc_norm\": 0.8717948717948718,\n\
\ \"acc_norm_stderr\": 0.02190190511507333\n },\n \"harness|hendrycksTest-medical_genetics|5\"\
: {\n \"acc\": 0.72,\n \"acc_stderr\": 0.045126085985421276,\n \
\ \"acc_norm\": 0.72,\n \"acc_norm_stderr\": 0.045126085985421276\n \
\ },\n \"harness|hendrycksTest-miscellaneous|5\": {\n \"acc\": 0.8263090676883781,\n\
\ \"acc_stderr\": 0.01354741565866226,\n \"acc_norm\": 0.8263090676883781,\n\
\ \"acc_norm_stderr\": 0.01354741565866226\n },\n \"harness|hendrycksTest-moral_disputes|5\"\
: {\n \"acc\": 0.7427745664739884,\n \"acc_stderr\": 0.023532925431044287,\n\
\ \"acc_norm\": 0.7427745664739884,\n \"acc_norm_stderr\": 0.023532925431044287\n\
\ },\n \"harness|hendrycksTest-moral_scenarios|5\": {\n \"acc\": 0.3877094972067039,\n\
\ \"acc_stderr\": 0.01629533232815581,\n \"acc_norm\": 0.3877094972067039,\n\
\ \"acc_norm_stderr\": 0.01629533232815581\n },\n \"harness|hendrycksTest-nutrition|5\"\
: {\n \"acc\": 0.7222222222222222,\n \"acc_stderr\": 0.025646863097137897,\n\
\ \"acc_norm\": 0.7222222222222222,\n \"acc_norm_stderr\": 0.025646863097137897\n\
\ },\n \"harness|hendrycksTest-philosophy|5\": {\n \"acc\": 0.7170418006430869,\n\
\ \"acc_stderr\": 0.02558306248998481,\n \"acc_norm\": 0.7170418006430869,\n\
\ \"acc_norm_stderr\": 0.02558306248998481\n },\n \"harness|hendrycksTest-prehistory|5\"\
: {\n \"acc\": 0.7469135802469136,\n \"acc_stderr\": 0.024191808600712992,\n\
\ \"acc_norm\": 0.7469135802469136,\n \"acc_norm_stderr\": 0.024191808600712992\n\
\ },\n \"harness|hendrycksTest-professional_accounting|5\": {\n \"\
acc\": 0.4397163120567376,\n \"acc_stderr\": 0.029609912075594106,\n \
\ \"acc_norm\": 0.4397163120567376,\n \"acc_norm_stderr\": 0.029609912075594106\n\
\ },\n \"harness|hendrycksTest-professional_law|5\": {\n \"acc\": 0.455019556714472,\n\
\ \"acc_stderr\": 0.012718456618701766,\n \"acc_norm\": 0.455019556714472,\n\
\ \"acc_norm_stderr\": 0.012718456618701766\n },\n \"harness|hendrycksTest-professional_medicine|5\"\
: {\n \"acc\": 0.6654411764705882,\n \"acc_stderr\": 0.028661996202335307,\n\
\ \"acc_norm\": 0.6654411764705882,\n \"acc_norm_stderr\": 0.028661996202335307\n\
\ },\n \"harness|hendrycksTest-professional_psychology|5\": {\n \"\
acc\": 0.6699346405228758,\n \"acc_stderr\": 0.019023726160724556,\n \
\ \"acc_norm\": 0.6699346405228758,\n \"acc_norm_stderr\": 0.019023726160724556\n\
\ },\n \"harness|hendrycksTest-public_relations|5\": {\n \"acc\": 0.6636363636363637,\n\
\ \"acc_stderr\": 0.04525393596302506,\n \"acc_norm\": 0.6636363636363637,\n\
\ \"acc_norm_stderr\": 0.04525393596302506\n },\n \"harness|hendrycksTest-security_studies|5\"\
: {\n \"acc\": 0.7183673469387755,\n \"acc_stderr\": 0.028795185574291296,\n\
\ \"acc_norm\": 0.7183673469387755,\n \"acc_norm_stderr\": 0.028795185574291296\n\
\ },\n \"harness|hendrycksTest-sociology|5\": {\n \"acc\": 0.8407960199004975,\n\
\ \"acc_stderr\": 0.025870646766169143,\n \"acc_norm\": 0.8407960199004975,\n\
\ \"acc_norm_stderr\": 0.025870646766169143\n },\n \"harness|hendrycksTest-us_foreign_policy|5\"\
: {\n \"acc\": 0.85,\n \"acc_stderr\": 0.03588702812826371,\n \
\ \"acc_norm\": 0.85,\n \"acc_norm_stderr\": 0.03588702812826371\n \
\ },\n \"harness|hendrycksTest-virology|5\": {\n \"acc\": 0.5301204819277109,\n\
\ \"acc_stderr\": 0.03885425420866767,\n \"acc_norm\": 0.5301204819277109,\n\
\ \"acc_norm_stderr\": 0.03885425420866767\n },\n \"harness|hendrycksTest-world_religions|5\"\
: {\n \"acc\": 0.8128654970760234,\n \"acc_stderr\": 0.029913127232368036,\n\
\ \"acc_norm\": 0.8128654970760234,\n \"acc_norm_stderr\": 0.029913127232368036\n\
\ },\n \"harness|truthfulqa:mc|0\": {\n \"mc1\": 0.4565483476132191,\n\
\ \"mc1_stderr\": 0.017437280953183695,\n \"mc2\": 0.6137860051903924,\n\
\ \"mc2_stderr\": 0.015431238638634374\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.8034727703235991,\n \"acc_stderr\": 0.011168120593569563\n\
\ },\n \"harness|gsm8k|5\": {\n \"acc\": 0.686125852918878,\n \
\ \"acc_stderr\": 0.012782681251053201\n }\n}\n```"
repo_url: https://huggingface.co/RatanRohith/NeuralPizza-7B-V0.2
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|arc:challenge|25_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|gsm8k|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hellaswag|10_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-management|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-virology|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-management|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-virology|5_2024-01-21T21-04-32.551333.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-anatomy|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-astronomy|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-college_biology|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-college_physics|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-computer_security|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-econometrics|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-global_facts|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-human_aging|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-international_law|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-management|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-marketing|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-nutrition|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-philosophy|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-prehistory|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-professional_law|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-public_relations|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-security_studies|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-sociology|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-virology|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|hendrycksTest-world_religions|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|truthfulqa:mc|0_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2024-01-21T21-04-32.551333.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- '**/details_harness|winogrande|5_2024-01-21T21-04-32.551333.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2024-01-21T21-04-32.551333.parquet'
- config_name: results
data_files:
- split: 2024_01_21T21_04_32.551333
path:
- results_2024-01-21T21-04-32.551333.parquet
- split: latest
path:
- results_2024-01-21T21-04-32.551333.parquet
---
# Dataset Card for Evaluation run of RatanRohith/NeuralPizza-7B-V0.2
<!-- Provide a quick summary of the dataset. -->
Dataset automatically created during the evaluation run of model [RatanRohith/NeuralPizza-7B-V0.2](https://huggingface.co/RatanRohith/NeuralPizza-7B-V0.2) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 63 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_RatanRohith__NeuralPizza-7B-V0.2",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2024-01-21T21:04:32.551333](https://huggingface.co/datasets/open-llm-leaderboard/details_RatanRohith__NeuralPizza-7B-V0.2/blob/main/results_2024-01-21T21-04-32.551333.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"acc": 0.6473430412019332,
"acc_stderr": 0.03217057797429981,
"acc_norm": 0.64765430784867,
"acc_norm_stderr": 0.032828743060941856,
"mc1": 0.4565483476132191,
"mc1_stderr": 0.017437280953183695,
"mc2": 0.6137860051903924,
"mc2_stderr": 0.015431238638634374
},
"harness|arc:challenge|25": {
"acc": 0.6535836177474402,
"acc_stderr": 0.013905011180063239,
"acc_norm": 0.6877133105802048,
"acc_norm_stderr": 0.013542598541688065
},
"harness|hellaswag|10": {
"acc": 0.6819358693487353,
"acc_stderr": 0.004647727222445386,
"acc_norm": 0.8610834495120494,
"acc_norm_stderr": 0.003451525868724678
},
"harness|hendrycksTest-abstract_algebra|5": {
"acc": 0.3,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.3,
"acc_norm_stderr": 0.046056618647183814
},
"harness|hendrycksTest-anatomy|5": {
"acc": 0.6370370370370371,
"acc_stderr": 0.04153948404742398,
"acc_norm": 0.6370370370370371,
"acc_norm_stderr": 0.04153948404742398
},
"harness|hendrycksTest-astronomy|5": {
"acc": 0.6842105263157895,
"acc_stderr": 0.0378272898086547,
"acc_norm": 0.6842105263157895,
"acc_norm_stderr": 0.0378272898086547
},
"harness|hendrycksTest-business_ethics|5": {
"acc": 0.63,
"acc_stderr": 0.04852365870939099,
"acc_norm": 0.63,
"acc_norm_stderr": 0.04852365870939099
},
"harness|hendrycksTest-clinical_knowledge|5": {
"acc": 0.7056603773584905,
"acc_stderr": 0.02804918631569525,
"acc_norm": 0.7056603773584905,
"acc_norm_stderr": 0.02804918631569525
},
"harness|hendrycksTest-college_biology|5": {
"acc": 0.7708333333333334,
"acc_stderr": 0.03514697467862388,
"acc_norm": 0.7708333333333334,
"acc_norm_stderr": 0.03514697467862388
},
"harness|hendrycksTest-college_chemistry|5": {
"acc": 0.5,
"acc_stderr": 0.050251890762960605,
"acc_norm": 0.5,
"acc_norm_stderr": 0.050251890762960605
},
"harness|hendrycksTest-college_computer_science|5": {
"acc": 0.58,
"acc_stderr": 0.049604496374885836,
"acc_norm": 0.58,
"acc_norm_stderr": 0.049604496374885836
},
"harness|hendrycksTest-college_mathematics|5": {
"acc": 0.32,
"acc_stderr": 0.046882617226215034,
"acc_norm": 0.32,
"acc_norm_stderr": 0.046882617226215034
},
"harness|hendrycksTest-college_medicine|5": {
"acc": 0.6878612716763006,
"acc_stderr": 0.035331333893236574,
"acc_norm": 0.6878612716763006,
"acc_norm_stderr": 0.035331333893236574
},
"harness|hendrycksTest-college_physics|5": {
"acc": 0.45098039215686275,
"acc_stderr": 0.049512182523962625,
"acc_norm": 0.45098039215686275,
"acc_norm_stderr": 0.049512182523962625
},
"harness|hendrycksTest-computer_security|5": {
"acc": 0.76,
"acc_stderr": 0.04292346959909283,
"acc_norm": 0.76,
"acc_norm_stderr": 0.04292346959909283
},
"harness|hendrycksTest-conceptual_physics|5": {
"acc": 0.5914893617021276,
"acc_stderr": 0.032134180267015755,
"acc_norm": 0.5914893617021276,
"acc_norm_stderr": 0.032134180267015755
},
"harness|hendrycksTest-econometrics|5": {
"acc": 0.4824561403508772,
"acc_stderr": 0.04700708033551038,
"acc_norm": 0.4824561403508772,
"acc_norm_stderr": 0.04700708033551038
},
"harness|hendrycksTest-electrical_engineering|5": {
"acc": 0.5586206896551724,
"acc_stderr": 0.04137931034482757,
"acc_norm": 0.5586206896551724,
"acc_norm_stderr": 0.04137931034482757
},
"harness|hendrycksTest-elementary_mathematics|5": {
"acc": 0.4126984126984127,
"acc_stderr": 0.025355741263055256,
"acc_norm": 0.4126984126984127,
"acc_norm_stderr": 0.025355741263055256
},
"harness|hendrycksTest-formal_logic|5": {
"acc": 0.4523809523809524,
"acc_stderr": 0.044518079590553275,
"acc_norm": 0.4523809523809524,
"acc_norm_stderr": 0.044518079590553275
},
"harness|hendrycksTest-global_facts|5": {
"acc": 0.33,
"acc_stderr": 0.047258156262526045,
"acc_norm": 0.33,
"acc_norm_stderr": 0.047258156262526045
},
"harness|hendrycksTest-high_school_biology|5": {
"acc": 0.7548387096774194,
"acc_stderr": 0.024472243840895518,
"acc_norm": 0.7548387096774194,
"acc_norm_stderr": 0.024472243840895518
},
"harness|hendrycksTest-high_school_chemistry|5": {
"acc": 0.4827586206896552,
"acc_stderr": 0.035158955511656986,
"acc_norm": 0.4827586206896552,
"acc_norm_stderr": 0.035158955511656986
},
"harness|hendrycksTest-high_school_computer_science|5": {
"acc": 0.71,
"acc_stderr": 0.045604802157206845,
"acc_norm": 0.71,
"acc_norm_stderr": 0.045604802157206845
},
"harness|hendrycksTest-high_school_european_history|5": {
"acc": 0.7757575757575758,
"acc_stderr": 0.032568666616811015,
"acc_norm": 0.7757575757575758,
"acc_norm_stderr": 0.032568666616811015
},
"harness|hendrycksTest-high_school_geography|5": {
"acc": 0.7929292929292929,
"acc_stderr": 0.02886977846026705,
"acc_norm": 0.7929292929292929,
"acc_norm_stderr": 0.02886977846026705
},
"harness|hendrycksTest-high_school_government_and_politics|5": {
"acc": 0.8808290155440415,
"acc_stderr": 0.02338193534812143,
"acc_norm": 0.8808290155440415,
"acc_norm_stderr": 0.02338193534812143
},
"harness|hendrycksTest-high_school_macroeconomics|5": {
"acc": 0.6666666666666666,
"acc_stderr": 0.023901157979402534,
"acc_norm": 0.6666666666666666,
"acc_norm_stderr": 0.023901157979402534
},
"harness|hendrycksTest-high_school_mathematics|5": {
"acc": 0.34074074074074073,
"acc_stderr": 0.028897748741131143,
"acc_norm": 0.34074074074074073,
"acc_norm_stderr": 0.028897748741131143
},
"harness|hendrycksTest-high_school_microeconomics|5": {
"acc": 0.6890756302521008,
"acc_stderr": 0.03006676158297793,
"acc_norm": 0.6890756302521008,
"acc_norm_stderr": 0.03006676158297793
},
"harness|hendrycksTest-high_school_physics|5": {
"acc": 0.31788079470198677,
"acc_stderr": 0.038020397601079024,
"acc_norm": 0.31788079470198677,
"acc_norm_stderr": 0.038020397601079024
},
"harness|hendrycksTest-high_school_psychology|5": {
"acc": 0.8495412844036697,
"acc_stderr": 0.015328563932669237,
"acc_norm": 0.8495412844036697,
"acc_norm_stderr": 0.015328563932669237
},
"harness|hendrycksTest-high_school_statistics|5": {
"acc": 0.5509259259259259,
"acc_stderr": 0.03392238405321617,
"acc_norm": 0.5509259259259259,
"acc_norm_stderr": 0.03392238405321617
},
"harness|hendrycksTest-high_school_us_history|5": {
"acc": 0.8088235294117647,
"acc_stderr": 0.027599174300640766,
"acc_norm": 0.8088235294117647,
"acc_norm_stderr": 0.027599174300640766
},
"harness|hendrycksTest-high_school_world_history|5": {
"acc": 0.7974683544303798,
"acc_stderr": 0.026160568246601443,
"acc_norm": 0.7974683544303798,
"acc_norm_stderr": 0.026160568246601443
},
"harness|hendrycksTest-human_aging|5": {
"acc": 0.6816143497757847,
"acc_stderr": 0.03126580522513713,
"acc_norm": 0.6816143497757847,
"acc_norm_stderr": 0.03126580522513713
},
"harness|hendrycksTest-human_sexuality|5": {
"acc": 0.7786259541984732,
"acc_stderr": 0.03641297081313729,
"acc_norm": 0.7786259541984732,
"acc_norm_stderr": 0.03641297081313729
},
"harness|hendrycksTest-international_law|5": {
"acc": 0.7768595041322314,
"acc_stderr": 0.03800754475228733,
"acc_norm": 0.7768595041322314,
"acc_norm_stderr": 0.03800754475228733
},
"harness|hendrycksTest-jurisprudence|5": {
"acc": 0.7685185185185185,
"acc_stderr": 0.04077494709252626,
"acc_norm": 0.7685185185185185,
"acc_norm_stderr": 0.04077494709252626
},
"harness|hendrycksTest-logical_fallacies|5": {
"acc": 0.7607361963190185,
"acc_stderr": 0.033519538795212696,
"acc_norm": 0.7607361963190185,
"acc_norm_stderr": 0.033519538795212696
},
"harness|hendrycksTest-machine_learning|5": {
"acc": 0.4375,
"acc_stderr": 0.04708567521880525,
"acc_norm": 0.4375,
"acc_norm_stderr": 0.04708567521880525
},
"harness|hendrycksTest-management|5": {
"acc": 0.8058252427184466,
"acc_stderr": 0.039166677628225836,
"acc_norm": 0.8058252427184466,
"acc_norm_stderr": 0.039166677628225836
},
"harness|hendrycksTest-marketing|5": {
"acc": 0.8717948717948718,
"acc_stderr": 0.02190190511507333,
"acc_norm": 0.8717948717948718,
"acc_norm_stderr": 0.02190190511507333
},
"harness|hendrycksTest-medical_genetics|5": {
"acc": 0.72,
"acc_stderr": 0.045126085985421276,
"acc_norm": 0.72,
"acc_norm_stderr": 0.045126085985421276
},
"harness|hendrycksTest-miscellaneous|5": {
"acc": 0.8263090676883781,
"acc_stderr": 0.01354741565866226,
"acc_norm": 0.8263090676883781,
"acc_norm_stderr": 0.01354741565866226
},
"harness|hendrycksTest-moral_disputes|5": {
"acc": 0.7427745664739884,
"acc_stderr": 0.023532925431044287,
"acc_norm": 0.7427745664739884,
"acc_norm_stderr": 0.023532925431044287
},
"harness|hendrycksTest-moral_scenarios|5": {
"acc": 0.3877094972067039,
"acc_stderr": 0.01629533232815581,
"acc_norm": 0.3877094972067039,
"acc_norm_stderr": 0.01629533232815581
},
"harness|hendrycksTest-nutrition|5": {
"acc": 0.7222222222222222,
"acc_stderr": 0.025646863097137897,
"acc_norm": 0.7222222222222222,
"acc_norm_stderr": 0.025646863097137897
},
"harness|hendrycksTest-philosophy|5": {
"acc": 0.7170418006430869,
"acc_stderr": 0.02558306248998481,
"acc_norm": 0.7170418006430869,
"acc_norm_stderr": 0.02558306248998481
},
"harness|hendrycksTest-prehistory|5": {
"acc": 0.7469135802469136,
"acc_stderr": 0.024191808600712992,
"acc_norm": 0.7469135802469136,
"acc_norm_stderr": 0.024191808600712992
},
"harness|hendrycksTest-professional_accounting|5": {
"acc": 0.4397163120567376,
"acc_stderr": 0.029609912075594106,
"acc_norm": 0.4397163120567376,
"acc_norm_stderr": 0.029609912075594106
},
"harness|hendrycksTest-professional_law|5": {
"acc": 0.455019556714472,
"acc_stderr": 0.012718456618701766,
"acc_norm": 0.455019556714472,
"acc_norm_stderr": 0.012718456618701766
},
"harness|hendrycksTest-professional_medicine|5": {
"acc": 0.6654411764705882,
"acc_stderr": 0.028661996202335307,
"acc_norm": 0.6654411764705882,
"acc_norm_stderr": 0.028661996202335307
},
"harness|hendrycksTest-professional_psychology|5": {
"acc": 0.6699346405228758,
"acc_stderr": 0.019023726160724556,
"acc_norm": 0.6699346405228758,
"acc_norm_stderr": 0.019023726160724556
},
"harness|hendrycksTest-public_relations|5": {
"acc": 0.6636363636363637,
"acc_stderr": 0.04525393596302506,
"acc_norm": 0.6636363636363637,
"acc_norm_stderr": 0.04525393596302506
},
"harness|hendrycksTest-security_studies|5": {
"acc": 0.7183673469387755,
"acc_stderr": 0.028795185574291296,
"acc_norm": 0.7183673469387755,
"acc_norm_stderr": 0.028795185574291296
},
"harness|hendrycksTest-sociology|5": {
"acc": 0.8407960199004975,
"acc_stderr": 0.025870646766169143,
"acc_norm": 0.8407960199004975,
"acc_norm_stderr": 0.025870646766169143
},
"harness|hendrycksTest-us_foreign_policy|5": {
"acc": 0.85,
"acc_stderr": 0.03588702812826371,
"acc_norm": 0.85,
"acc_norm_stderr": 0.03588702812826371
},
"harness|hendrycksTest-virology|5": {
"acc": 0.5301204819277109,
"acc_stderr": 0.03885425420866767,
"acc_norm": 0.5301204819277109,
"acc_norm_stderr": 0.03885425420866767
},
"harness|hendrycksTest-world_religions|5": {
"acc": 0.8128654970760234,
"acc_stderr": 0.029913127232368036,
"acc_norm": 0.8128654970760234,
"acc_norm_stderr": 0.029913127232368036
},
"harness|truthfulqa:mc|0": {
"mc1": 0.4565483476132191,
"mc1_stderr": 0.017437280953183695,
"mc2": 0.6137860051903924,
"mc2_stderr": 0.015431238638634374
},
"harness|winogrande|5": {
"acc": 0.8034727703235991,
"acc_stderr": 0.011168120593569563
},
"harness|gsm8k|5": {
"acc": 0.686125852918878,
"acc_stderr": 0.012782681251053201
}
}
```
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] |
juraj-juraj/doc_gen | ---
language:
- en
license: mit
task_categories:
- text-generation
pretty_name: py_code_doc
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
- split: test
path: data/test-*
dataset_info:
features:
- name: docstring
dtype: string
- name: function
dtype: string
- name: __index_level_0__
dtype: int64
splits:
- name: train
num_bytes: 270465917
num_examples: 313366
- name: validation
num_bytes: 763140
num_examples: 1000
- name: test
num_bytes: 878385
num_examples: 1000
download_size: 107450380
dataset_size: 272107442
---
# Code documentation dataset
This dataset aims leverage usage of lm to automatically generate documenation to undocumented python code. Dataset consists of pairs code and its documenation
Content of dataset is created from CodeSearchNet dataset. |
Taegyuu/KoAlpaca-v1.1a | ---
dataset_info:
features:
- name: instruction
dtype: string
- name: output
dtype: string
- name: url
dtype: string
splits:
- name: train
num_bytes: 23371027
num_examples: 21155
download_size: 12856014
dataset_size: 23371027
task_categories:
- text-generation
language:
- ko
tags:
- KoAlpaca
pretty_name: KoAlpaca-v1.1a
---
# Dataset Card for "KoAlpaca-v1.1a"
## Project Repo
- Github Repo: [Beomi/KoAlpaca](https://github.com/Beomi/KoAlpaca)
## How to use
```python
>>> from datasets import load_dataset
>>> ds = load_dataset("beomi/KoAlpaca-v1.1a", split="train")
>>> ds
Dataset({
features: ['instruction', 'input', 'output'],
num_rows: 21155
})
```
```python
>>> ds[0]
{'instruction': '양파는 어떤 식물 부위인가요? 그리고 고구마는 뿌리인가요?',
'output': '양파는 잎이 아닌 식물의 줄기 부분입니다. 고구마는 식물의 뿌리 부분입니다. \n\n식물의 부위의 구분에 대해 궁금해하는 분이라면 분명 이 질문에 대한 답을 찾고 있을 것입니다. 양파는 잎이 아닌 줄기 부분입니다. 고구마는 다른 질문과 답변에서 언급된 것과 같이 뿌리 부분입니다. 따라서, 양파는 식물의 줄기 부분이 되고, 고구마는 식물의 뿌리 부분입니다.\n\n 덧붙이는 답변: 고구마 줄기도 볶아먹을 수 있나요? \n\n고구마 줄기도 식용으로 볶아먹을 수 있습니다. 하지만 줄기 뿐만 아니라, 잎, 씨, 뿌리까지 모든 부위가 식용으로 활용되기도 합니다. 다만, 한국에서는 일반적으로 뿌리 부분인 고구마를 주로 먹습니다.',
'url': 'https://kin.naver.com/qna/detail.naver?d1id=11&dirId=1116&docId=55320268'} |
clips/20Q | ---
annotations_creators: []
language:
- en
language_creators: []
license: []
multilinguality:
- monolingual
pretty_name: 20Q - World Knowledge Benchmark
size_categories:
- 1K<n<10K
source_datasets: []
tags:
- 20Q
- Twenty Questions
- 20 Questions
task_categories:
- question-answering
task_ids: []
---
# Dataset Card for 20Q
|
ntkctksnb1/mini_platypus | ---
dataset_info:
config_name: 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
features:
- name: instruction
dtype: string
- name: output
dtype: string
splits:
- name: train
num_bytes: 4201526
num_examples: 1000
download_size: 2247084
dataset_size: 4201526
configs:
- config_name: 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
data_files:
- split: train
path: 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/train-*
---
|
pawkanarek/spraix_1024_9frames | ---
license: gpl-3.0
tags:
- art
pretty_name: 96 images to train sprite animations
size_categories:
- n<1K
---
This dataset inclueds labeled sprite animations with 9 frames. More info https://github.com/PawKanarek/spraix |
Aashi/All_About_Apple_Devices | ---
license: mit
task_categories:
- question-answering
- conversational
language:
- en
tags:
- tech
pretty_name: A3D 🚀
size_categories:
- n<1K
---
This dataset contains Q&A based on what kind of questions do people ask online with their respective answers. This can be used as an LLM project where we finetune an LLM and create a chatbot where one can ask the chatbot any question related to Apple products and get respective answers.
## To use this dataset try running the following code:
- dataset = load_dataset("Aashi/All_About_Apple_Devices", data_files={"train": "QandA.csv"})
- df_train = dataset['train'].to_pandas() |
Lienid/chat_perplexity_scored | ---
dataset_info:
features:
- name: question
dtype: string
- name: response
dtype: string
- name: perplexity
dtype: float32
splits:
- name: train
num_bytes: 370318195.41288
num_examples: 173496
download_size: 185184757
dataset_size: 370318195.41288
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
|
ArmelRandy/most_precious_4 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: prompt
dtype: string
splits:
- name: train
num_bytes: 8031615.129808251
num_examples: 7744
- name: test
num_bytes: 892977.870191749
num_examples: 861
download_size: 5178596
dataset_size: 8924593.0
---
# Dataset Card for "most_precious_4"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
autoevaluate/autoeval-eval-lener_br-lener_br-b36dee-1776161642 | ---
type: predictions
tags:
- autotrain
- evaluation
datasets:
- lener_br
eval_info:
task: entity_extraction
model: Luciano/xlm-roberta-base-finetuned-lener-br
metrics: []
dataset_name: lener_br
dataset_config: lener_br
dataset_split: validation
col_mapping:
tokens: tokens
tags: ner_tags
---
# Dataset Card for AutoTrain Evaluator
This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset:
* Task: Token Classification
* Model: Luciano/xlm-roberta-base-finetuned-lener-br
* Dataset: lener_br
* Config: lener_br
* Split: validation
To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator).
## Contributions
Thanks to [@Luciano](https://huggingface.co/Luciano) for evaluating this model. |
autoevaluate/autoeval-staging-eval-project-squad-ef91144d-11985603 | ---
type: predictions
tags:
- autotrain
- evaluation
datasets:
- squad
eval_info:
task: extractive_question_answering
model: nlpconnect/roberta-base-squad2-nq
metrics: []
dataset_name: squad
dataset_config: plain_text
dataset_split: validation
col_mapping:
context: context
question: question
answers-text: answers.text
answers-answer_start: answers.answer_start
---
# Dataset Card for AutoTrain Evaluator
This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset:
* Task: Question Answering
* Model: nlpconnect/roberta-base-squad2-nq
* Dataset: squad
* Config: plain_text
* Split: validation
To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator).
## Contributions
Thanks to [@ankur310794](https://huggingface.co/ankur310794) for evaluating this model. |
idning/ffhq64-caption | ---
license: mit
dataset_info:
features:
- name: image
dtype: image
- name: text
dtype: string
splits:
- name: train
num_bytes: 586217154.0
num_examples: 70000
download_size: 584117488
dataset_size: 586217154.0
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
|
autoevaluate/autoeval-staging-eval-project-bc0462a6-7584895 | ---
type: predictions
tags:
- autotrain
- evaluation
datasets:
- xtreme
eval_info:
task: entity_extraction
model: moghis/xlm-roberta-base-finetuned-panx-fr
metrics: []
dataset_name: xtreme
dataset_config: PAN-X.fr
dataset_split: test
col_mapping:
tokens: tokens
tags: ner_tags
---
# Dataset Card for AutoTrain Evaluator
This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset:
* Task: Token Classification
* Model: moghis/xlm-roberta-base-finetuned-panx-fr
* Dataset: xtreme
To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator).
## Contributions
Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model. |
BerMaker/beans | ---
license: apache-2.0
---
|
euclaise/mathqa_programs | ---
dataset_info:
features:
- name: options
dtype: string
- name: correct
dtype: string
- name: annotated_formula
dtype: string
- name: problem
dtype: string
- name: rationale
dtype: string
- name: program
dtype: string
splits:
- name: train
num_bytes: 17017833
num_examples: 28851
download_size: 8877888
dataset_size: 17017833
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
license: apache-2.0
---
math_qa converted to Python snippets |
liuyanchen1015/MULTI_VALUE_mnli_flat_adj_for_adv | ---
dataset_info:
features:
- name: premise
dtype: string
- name: hypothesis
dtype: string
- name: label
dtype: int64
- name: idx
dtype: int64
- name: score
dtype: int64
splits:
- name: dev_matched
num_bytes: 52060
num_examples: 221
- name: dev_mismatched
num_bytes: 51025
num_examples: 207
- name: test_matched
num_bytes: 45543
num_examples: 194
- name: test_mismatched
num_bytes: 63860
num_examples: 243
- name: train
num_bytes: 1870439
num_examples: 8415
download_size: 1258577
dataset_size: 2082927
---
# Dataset Card for "MULTI_VALUE_mnli_flat_adj_for_adv"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
juliaturc/captioned-cartoons | ---
dataset_info:
features:
- name: image
dtype: image
- name: text
dtype: string
splits:
- name: train
num_bytes: 22981331.0
num_examples: 100
download_size: 22873699
dataset_size: 22981331.0
---
# Dataset Card for "captioned-cartoons"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
open-llm-leaderboard/details_Weyaxi__MetaMath-NeuralHermes-2.5-Mistral-7B-Linear | ---
pretty_name: Evaluation run of Weyaxi/MetaMath-NeuralHermes-2.5-Mistral-7B-Linear
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [Weyaxi/MetaMath-NeuralHermes-2.5-Mistral-7B-Linear](https://huggingface.co/Weyaxi/MetaMath-NeuralHermes-2.5-Mistral-7B-Linear)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 63 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the aggregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_Weyaxi__MetaMath-NeuralHermes-2.5-Mistral-7B-Linear\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-12-29T12:58:30.448615](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__MetaMath-NeuralHermes-2.5-Mistral-7B-Linear/blob/main/results_2023-12-29T12-58-30.448615.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"acc\": 0.6372944871407636,\n\
\ \"acc_stderr\": 0.0321517342901721,\n \"acc_norm\": 0.6376839473607636,\n\
\ \"acc_norm_stderr\": 0.03280413968071383,\n \"mc1\": 0.3329253365973072,\n\
\ \"mc1_stderr\": 0.016497402382012052,\n \"mc2\": 0.48572469546439073,\n\
\ \"mc2_stderr\": 0.01547664094918918\n },\n \"harness|arc:challenge|25\"\
: {\n \"acc\": 0.5964163822525598,\n \"acc_stderr\": 0.014337158914268438,\n\
\ \"acc_norm\": 0.6279863481228669,\n \"acc_norm_stderr\": 0.014124597881844461\n\
\ },\n \"harness|hellaswag|10\": {\n \"acc\": 0.6590320653256323,\n\
\ \"acc_stderr\": 0.004730658073041556,\n \"acc_norm\": 0.8420633339972117,\n\
\ \"acc_norm_stderr\": 0.003639363021784421\n },\n \"harness|hendrycksTest-abstract_algebra|5\"\
: {\n \"acc\": 0.28,\n \"acc_stderr\": 0.04512608598542128,\n \
\ \"acc_norm\": 0.28,\n \"acc_norm_stderr\": 0.04512608598542128\n \
\ },\n \"harness|hendrycksTest-anatomy|5\": {\n \"acc\": 0.5925925925925926,\n\
\ \"acc_stderr\": 0.04244633238353227,\n \"acc_norm\": 0.5925925925925926,\n\
\ \"acc_norm_stderr\": 0.04244633238353227\n },\n \"harness|hendrycksTest-astronomy|5\"\
: {\n \"acc\": 0.6973684210526315,\n \"acc_stderr\": 0.03738520676119667,\n\
\ \"acc_norm\": 0.6973684210526315,\n \"acc_norm_stderr\": 0.03738520676119667\n\
\ },\n \"harness|hendrycksTest-business_ethics|5\": {\n \"acc\": 0.58,\n\
\ \"acc_stderr\": 0.049604496374885836,\n \"acc_norm\": 0.58,\n \
\ \"acc_norm_stderr\": 0.049604496374885836\n },\n \"harness|hendrycksTest-clinical_knowledge|5\"\
: {\n \"acc\": 0.6792452830188679,\n \"acc_stderr\": 0.028727502957880267,\n\
\ \"acc_norm\": 0.6792452830188679,\n \"acc_norm_stderr\": 0.028727502957880267\n\
\ },\n \"harness|hendrycksTest-college_biology|5\": {\n \"acc\": 0.7430555555555556,\n\
\ \"acc_stderr\": 0.03653946969442099,\n \"acc_norm\": 0.7430555555555556,\n\
\ \"acc_norm_stderr\": 0.03653946969442099\n },\n \"harness|hendrycksTest-college_chemistry|5\"\
: {\n \"acc\": 0.44,\n \"acc_stderr\": 0.04988876515698589,\n \
\ \"acc_norm\": 0.44,\n \"acc_norm_stderr\": 0.04988876515698589\n \
\ },\n \"harness|hendrycksTest-college_computer_science|5\": {\n \"acc\"\
: 0.45,\n \"acc_stderr\": 0.05,\n \"acc_norm\": 0.45,\n \"\
acc_norm_stderr\": 0.05\n },\n \"harness|hendrycksTest-college_mathematics|5\"\
: {\n \"acc\": 0.3,\n \"acc_stderr\": 0.046056618647183814,\n \
\ \"acc_norm\": 0.3,\n \"acc_norm_stderr\": 0.046056618647183814\n \
\ },\n \"harness|hendrycksTest-college_medicine|5\": {\n \"acc\": 0.6069364161849711,\n\
\ \"acc_stderr\": 0.03724249595817731,\n \"acc_norm\": 0.6069364161849711,\n\
\ \"acc_norm_stderr\": 0.03724249595817731\n },\n \"harness|hendrycksTest-college_physics|5\"\
: {\n \"acc\": 0.35294117647058826,\n \"acc_stderr\": 0.04755129616062946,\n\
\ \"acc_norm\": 0.35294117647058826,\n \"acc_norm_stderr\": 0.04755129616062946\n\
\ },\n \"harness|hendrycksTest-computer_security|5\": {\n \"acc\":\
\ 0.79,\n \"acc_stderr\": 0.040936018074033256,\n \"acc_norm\": 0.79,\n\
\ \"acc_norm_stderr\": 0.040936018074033256\n },\n \"harness|hendrycksTest-conceptual_physics|5\"\
: {\n \"acc\": 0.5702127659574469,\n \"acc_stderr\": 0.03236214467715564,\n\
\ \"acc_norm\": 0.5702127659574469,\n \"acc_norm_stderr\": 0.03236214467715564\n\
\ },\n \"harness|hendrycksTest-econometrics|5\": {\n \"acc\": 0.5,\n\
\ \"acc_stderr\": 0.047036043419179864,\n \"acc_norm\": 0.5,\n \
\ \"acc_norm_stderr\": 0.047036043419179864\n },\n \"harness|hendrycksTest-electrical_engineering|5\"\
: {\n \"acc\": 0.5448275862068965,\n \"acc_stderr\": 0.04149886942192117,\n\
\ \"acc_norm\": 0.5448275862068965,\n \"acc_norm_stderr\": 0.04149886942192117\n\
\ },\n \"harness|hendrycksTest-elementary_mathematics|5\": {\n \"acc\"\
: 0.41798941798941797,\n \"acc_stderr\": 0.02540255550326091,\n \"\
acc_norm\": 0.41798941798941797,\n \"acc_norm_stderr\": 0.02540255550326091\n\
\ },\n \"harness|hendrycksTest-formal_logic|5\": {\n \"acc\": 0.3968253968253968,\n\
\ \"acc_stderr\": 0.043758884927270605,\n \"acc_norm\": 0.3968253968253968,\n\
\ \"acc_norm_stderr\": 0.043758884927270605\n },\n \"harness|hendrycksTest-global_facts|5\"\
: {\n \"acc\": 0.36,\n \"acc_stderr\": 0.048241815132442176,\n \
\ \"acc_norm\": 0.36,\n \"acc_norm_stderr\": 0.048241815132442176\n \
\ },\n \"harness|hendrycksTest-high_school_biology|5\": {\n \"acc\"\
: 0.7677419354838709,\n \"acc_stderr\": 0.024022256130308235,\n \"\
acc_norm\": 0.7677419354838709,\n \"acc_norm_stderr\": 0.024022256130308235\n\
\ },\n \"harness|hendrycksTest-high_school_chemistry|5\": {\n \"acc\"\
: 0.5270935960591133,\n \"acc_stderr\": 0.03512819077876106,\n \"\
acc_norm\": 0.5270935960591133,\n \"acc_norm_stderr\": 0.03512819077876106\n\
\ },\n \"harness|hendrycksTest-high_school_computer_science|5\": {\n \
\ \"acc\": 0.68,\n \"acc_stderr\": 0.04688261722621505,\n \"acc_norm\"\
: 0.68,\n \"acc_norm_stderr\": 0.04688261722621505\n },\n \"harness|hendrycksTest-high_school_european_history|5\"\
: {\n \"acc\": 0.7575757575757576,\n \"acc_stderr\": 0.03346409881055953,\n\
\ \"acc_norm\": 0.7575757575757576,\n \"acc_norm_stderr\": 0.03346409881055953\n\
\ },\n \"harness|hendrycksTest-high_school_geography|5\": {\n \"acc\"\
: 0.7727272727272727,\n \"acc_stderr\": 0.029857515673386414,\n \"\
acc_norm\": 0.7727272727272727,\n \"acc_norm_stderr\": 0.029857515673386414\n\
\ },\n \"harness|hendrycksTest-high_school_government_and_politics|5\": {\n\
\ \"acc\": 0.8911917098445595,\n \"acc_stderr\": 0.022473253332768776,\n\
\ \"acc_norm\": 0.8911917098445595,\n \"acc_norm_stderr\": 0.022473253332768776\n\
\ },\n \"harness|hendrycksTest-high_school_macroeconomics|5\": {\n \
\ \"acc\": 0.6282051282051282,\n \"acc_stderr\": 0.024503472557110936,\n\
\ \"acc_norm\": 0.6282051282051282,\n \"acc_norm_stderr\": 0.024503472557110936\n\
\ },\n \"harness|hendrycksTest-high_school_mathematics|5\": {\n \"\
acc\": 0.34814814814814815,\n \"acc_stderr\": 0.029045600290616258,\n \
\ \"acc_norm\": 0.34814814814814815,\n \"acc_norm_stderr\": 0.029045600290616258\n\
\ },\n \"harness|hendrycksTest-high_school_microeconomics|5\": {\n \
\ \"acc\": 0.6512605042016807,\n \"acc_stderr\": 0.030956636328566545,\n\
\ \"acc_norm\": 0.6512605042016807,\n \"acc_norm_stderr\": 0.030956636328566545\n\
\ },\n \"harness|hendrycksTest-high_school_physics|5\": {\n \"acc\"\
: 0.304635761589404,\n \"acc_stderr\": 0.03757949922943343,\n \"acc_norm\"\
: 0.304635761589404,\n \"acc_norm_stderr\": 0.03757949922943343\n },\n\
\ \"harness|hendrycksTest-high_school_psychology|5\": {\n \"acc\": 0.8330275229357799,\n\
\ \"acc_stderr\": 0.015990154885073393,\n \"acc_norm\": 0.8330275229357799,\n\
\ \"acc_norm_stderr\": 0.015990154885073393\n },\n \"harness|hendrycksTest-high_school_statistics|5\"\
: {\n \"acc\": 0.49074074074074076,\n \"acc_stderr\": 0.034093869469927006,\n\
\ \"acc_norm\": 0.49074074074074076,\n \"acc_norm_stderr\": 0.034093869469927006\n\
\ },\n \"harness|hendrycksTest-high_school_us_history|5\": {\n \"acc\"\
: 0.803921568627451,\n \"acc_stderr\": 0.027865942286639318,\n \"\
acc_norm\": 0.803921568627451,\n \"acc_norm_stderr\": 0.027865942286639318\n\
\ },\n \"harness|hendrycksTest-high_school_world_history|5\": {\n \"\
acc\": 0.7805907172995781,\n \"acc_stderr\": 0.026939106581553945,\n \
\ \"acc_norm\": 0.7805907172995781,\n \"acc_norm_stderr\": 0.026939106581553945\n\
\ },\n \"harness|hendrycksTest-human_aging|5\": {\n \"acc\": 0.7085201793721974,\n\
\ \"acc_stderr\": 0.03050028317654585,\n \"acc_norm\": 0.7085201793721974,\n\
\ \"acc_norm_stderr\": 0.03050028317654585\n },\n \"harness|hendrycksTest-human_sexuality|5\"\
: {\n \"acc\": 0.7862595419847328,\n \"acc_stderr\": 0.0359546161177469,\n\
\ \"acc_norm\": 0.7862595419847328,\n \"acc_norm_stderr\": 0.0359546161177469\n\
\ },\n \"harness|hendrycksTest-international_law|5\": {\n \"acc\":\
\ 0.768595041322314,\n \"acc_stderr\": 0.03849856098794088,\n \"acc_norm\"\
: 0.768595041322314,\n \"acc_norm_stderr\": 0.03849856098794088\n },\n\
\ \"harness|hendrycksTest-jurisprudence|5\": {\n \"acc\": 0.7870370370370371,\n\
\ \"acc_stderr\": 0.0395783547198098,\n \"acc_norm\": 0.7870370370370371,\n\
\ \"acc_norm_stderr\": 0.0395783547198098\n },\n \"harness|hendrycksTest-logical_fallacies|5\"\
: {\n \"acc\": 0.7914110429447853,\n \"acc_stderr\": 0.031921934489347235,\n\
\ \"acc_norm\": 0.7914110429447853,\n \"acc_norm_stderr\": 0.031921934489347235\n\
\ },\n \"harness|hendrycksTest-machine_learning|5\": {\n \"acc\": 0.5089285714285714,\n\
\ \"acc_stderr\": 0.04745033255489123,\n \"acc_norm\": 0.5089285714285714,\n\
\ \"acc_norm_stderr\": 0.04745033255489123\n },\n \"harness|hendrycksTest-management|5\"\
: {\n \"acc\": 0.8155339805825242,\n \"acc_stderr\": 0.03840423627288276,\n\
\ \"acc_norm\": 0.8155339805825242,\n \"acc_norm_stderr\": 0.03840423627288276\n\
\ },\n \"harness|hendrycksTest-marketing|5\": {\n \"acc\": 0.8760683760683761,\n\
\ \"acc_stderr\": 0.021586494001281382,\n \"acc_norm\": 0.8760683760683761,\n\
\ \"acc_norm_stderr\": 0.021586494001281382\n },\n \"harness|hendrycksTest-medical_genetics|5\"\
: {\n \"acc\": 0.69,\n \"acc_stderr\": 0.04648231987117316,\n \
\ \"acc_norm\": 0.69,\n \"acc_norm_stderr\": 0.04648231987117316\n \
\ },\n \"harness|hendrycksTest-miscellaneous|5\": {\n \"acc\": 0.8160919540229885,\n\
\ \"acc_stderr\": 0.013853724170922526,\n \"acc_norm\": 0.8160919540229885,\n\
\ \"acc_norm_stderr\": 0.013853724170922526\n },\n \"harness|hendrycksTest-moral_disputes|5\"\
: {\n \"acc\": 0.7398843930635838,\n \"acc_stderr\": 0.023618678310069363,\n\
\ \"acc_norm\": 0.7398843930635838,\n \"acc_norm_stderr\": 0.023618678310069363\n\
\ },\n \"harness|hendrycksTest-moral_scenarios|5\": {\n \"acc\": 0.35307262569832404,\n\
\ \"acc_stderr\": 0.015984204545268565,\n \"acc_norm\": 0.35307262569832404,\n\
\ \"acc_norm_stderr\": 0.015984204545268565\n },\n \"harness|hendrycksTest-nutrition|5\"\
: {\n \"acc\": 0.7777777777777778,\n \"acc_stderr\": 0.023805186524888135,\n\
\ \"acc_norm\": 0.7777777777777778,\n \"acc_norm_stderr\": 0.023805186524888135\n\
\ },\n \"harness|hendrycksTest-philosophy|5\": {\n \"acc\": 0.6945337620578779,\n\
\ \"acc_stderr\": 0.02616058445014045,\n \"acc_norm\": 0.6945337620578779,\n\
\ \"acc_norm_stderr\": 0.02616058445014045\n },\n \"harness|hendrycksTest-prehistory|5\"\
: {\n \"acc\": 0.75,\n \"acc_stderr\": 0.02409347123262133,\n \
\ \"acc_norm\": 0.75,\n \"acc_norm_stderr\": 0.02409347123262133\n \
\ },\n \"harness|hendrycksTest-professional_accounting|5\": {\n \"acc\"\
: 0.4929078014184397,\n \"acc_stderr\": 0.02982449855912901,\n \"\
acc_norm\": 0.4929078014184397,\n \"acc_norm_stderr\": 0.02982449855912901\n\
\ },\n \"harness|hendrycksTest-professional_law|5\": {\n \"acc\": 0.4602346805736636,\n\
\ \"acc_stderr\": 0.012729785386598564,\n \"acc_norm\": 0.4602346805736636,\n\
\ \"acc_norm_stderr\": 0.012729785386598564\n },\n \"harness|hendrycksTest-professional_medicine|5\"\
: {\n \"acc\": 0.6470588235294118,\n \"acc_stderr\": 0.029029422815681393,\n\
\ \"acc_norm\": 0.6470588235294118,\n \"acc_norm_stderr\": 0.029029422815681393\n\
\ },\n \"harness|hendrycksTest-professional_psychology|5\": {\n \"\
acc\": 0.6584967320261438,\n \"acc_stderr\": 0.019184639328092487,\n \
\ \"acc_norm\": 0.6584967320261438,\n \"acc_norm_stderr\": 0.019184639328092487\n\
\ },\n \"harness|hendrycksTest-public_relations|5\": {\n \"acc\": 0.6636363636363637,\n\
\ \"acc_stderr\": 0.04525393596302506,\n \"acc_norm\": 0.6636363636363637,\n\
\ \"acc_norm_stderr\": 0.04525393596302506\n },\n \"harness|hendrycksTest-security_studies|5\"\
: {\n \"acc\": 0.7387755102040816,\n \"acc_stderr\": 0.028123429335142783,\n\
\ \"acc_norm\": 0.7387755102040816,\n \"acc_norm_stderr\": 0.028123429335142783\n\
\ },\n \"harness|hendrycksTest-sociology|5\": {\n \"acc\": 0.8308457711442786,\n\
\ \"acc_stderr\": 0.026508590656233268,\n \"acc_norm\": 0.8308457711442786,\n\
\ \"acc_norm_stderr\": 0.026508590656233268\n },\n \"harness|hendrycksTest-us_foreign_policy|5\"\
: {\n \"acc\": 0.88,\n \"acc_stderr\": 0.032659863237109066,\n \
\ \"acc_norm\": 0.88,\n \"acc_norm_stderr\": 0.032659863237109066\n \
\ },\n \"harness|hendrycksTest-virology|5\": {\n \"acc\": 0.5542168674698795,\n\
\ \"acc_stderr\": 0.03869543323472101,\n \"acc_norm\": 0.5542168674698795,\n\
\ \"acc_norm_stderr\": 0.03869543323472101\n },\n \"harness|hendrycksTest-world_religions|5\"\
: {\n \"acc\": 0.8245614035087719,\n \"acc_stderr\": 0.02917088550072767,\n\
\ \"acc_norm\": 0.8245614035087719,\n \"acc_norm_stderr\": 0.02917088550072767\n\
\ },\n \"harness|truthfulqa:mc|0\": {\n \"mc1\": 0.3329253365973072,\n\
\ \"mc1_stderr\": 0.016497402382012052,\n \"mc2\": 0.48572469546439073,\n\
\ \"mc2_stderr\": 0.01547664094918918\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.7679558011049724,\n \"acc_stderr\": 0.011864149691827936\n\
\ },\n \"harness|gsm8k|5\": {\n \"acc\": 0.6982562547384382,\n \
\ \"acc_stderr\": 0.012643544762873358\n }\n}\n```"
repo_url: https://huggingface.co/Weyaxi/MetaMath-NeuralHermes-2.5-Mistral-7B-Linear
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|arc:challenge|25_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|gsm8k|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hellaswag|10_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-12-29T12-58-30.448615.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-management|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-virology|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|truthfulqa:mc|0_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2023-12-29T12-58-30.448615.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- '**/details_harness|winogrande|5_2023-12-29T12-58-30.448615.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-12-29T12-58-30.448615.parquet'
- config_name: results
data_files:
- split: 2023_12_29T12_58_30.448615
path:
- results_2023-12-29T12-58-30.448615.parquet
- split: latest
path:
- results_2023-12-29T12-58-30.448615.parquet
---
# Dataset Card for Evaluation run of Weyaxi/MetaMath-NeuralHermes-2.5-Mistral-7B-Linear
<!-- Provide a quick summary of the dataset. -->
Dataset automatically created during the evaluation run of model [Weyaxi/MetaMath-NeuralHermes-2.5-Mistral-7B-Linear](https://huggingface.co/Weyaxi/MetaMath-NeuralHermes-2.5-Mistral-7B-Linear) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 63 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_Weyaxi__MetaMath-NeuralHermes-2.5-Mistral-7B-Linear",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-12-29T12:58:30.448615](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__MetaMath-NeuralHermes-2.5-Mistral-7B-Linear/blob/main/results_2023-12-29T12-58-30.448615.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"acc": 0.6372944871407636,
"acc_stderr": 0.0321517342901721,
"acc_norm": 0.6376839473607636,
"acc_norm_stderr": 0.03280413968071383,
"mc1": 0.3329253365973072,
"mc1_stderr": 0.016497402382012052,
"mc2": 0.48572469546439073,
"mc2_stderr": 0.01547664094918918
},
"harness|arc:challenge|25": {
"acc": 0.5964163822525598,
"acc_stderr": 0.014337158914268438,
"acc_norm": 0.6279863481228669,
"acc_norm_stderr": 0.014124597881844461
},
"harness|hellaswag|10": {
"acc": 0.6590320653256323,
"acc_stderr": 0.004730658073041556,
"acc_norm": 0.8420633339972117,
"acc_norm_stderr": 0.003639363021784421
},
"harness|hendrycksTest-abstract_algebra|5": {
"acc": 0.28,
"acc_stderr": 0.04512608598542128,
"acc_norm": 0.28,
"acc_norm_stderr": 0.04512608598542128
},
"harness|hendrycksTest-anatomy|5": {
"acc": 0.5925925925925926,
"acc_stderr": 0.04244633238353227,
"acc_norm": 0.5925925925925926,
"acc_norm_stderr": 0.04244633238353227
},
"harness|hendrycksTest-astronomy|5": {
"acc": 0.6973684210526315,
"acc_stderr": 0.03738520676119667,
"acc_norm": 0.6973684210526315,
"acc_norm_stderr": 0.03738520676119667
},
"harness|hendrycksTest-business_ethics|5": {
"acc": 0.58,
"acc_stderr": 0.049604496374885836,
"acc_norm": 0.58,
"acc_norm_stderr": 0.049604496374885836
},
"harness|hendrycksTest-clinical_knowledge|5": {
"acc": 0.6792452830188679,
"acc_stderr": 0.028727502957880267,
"acc_norm": 0.6792452830188679,
"acc_norm_stderr": 0.028727502957880267
},
"harness|hendrycksTest-college_biology|5": {
"acc": 0.7430555555555556,
"acc_stderr": 0.03653946969442099,
"acc_norm": 0.7430555555555556,
"acc_norm_stderr": 0.03653946969442099
},
"harness|hendrycksTest-college_chemistry|5": {
"acc": 0.44,
"acc_stderr": 0.04988876515698589,
"acc_norm": 0.44,
"acc_norm_stderr": 0.04988876515698589
},
"harness|hendrycksTest-college_computer_science|5": {
"acc": 0.45,
"acc_stderr": 0.05,
"acc_norm": 0.45,
"acc_norm_stderr": 0.05
},
"harness|hendrycksTest-college_mathematics|5": {
"acc": 0.3,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.3,
"acc_norm_stderr": 0.046056618647183814
},
"harness|hendrycksTest-college_medicine|5": {
"acc": 0.6069364161849711,
"acc_stderr": 0.03724249595817731,
"acc_norm": 0.6069364161849711,
"acc_norm_stderr": 0.03724249595817731
},
"harness|hendrycksTest-college_physics|5": {
"acc": 0.35294117647058826,
"acc_stderr": 0.04755129616062946,
"acc_norm": 0.35294117647058826,
"acc_norm_stderr": 0.04755129616062946
},
"harness|hendrycksTest-computer_security|5": {
"acc": 0.79,
"acc_stderr": 0.040936018074033256,
"acc_norm": 0.79,
"acc_norm_stderr": 0.040936018074033256
},
"harness|hendrycksTest-conceptual_physics|5": {
"acc": 0.5702127659574469,
"acc_stderr": 0.03236214467715564,
"acc_norm": 0.5702127659574469,
"acc_norm_stderr": 0.03236214467715564
},
"harness|hendrycksTest-econometrics|5": {
"acc": 0.5,
"acc_stderr": 0.047036043419179864,
"acc_norm": 0.5,
"acc_norm_stderr": 0.047036043419179864
},
"harness|hendrycksTest-electrical_engineering|5": {
"acc": 0.5448275862068965,
"acc_stderr": 0.04149886942192117,
"acc_norm": 0.5448275862068965,
"acc_norm_stderr": 0.04149886942192117
},
"harness|hendrycksTest-elementary_mathematics|5": {
"acc": 0.41798941798941797,
"acc_stderr": 0.02540255550326091,
"acc_norm": 0.41798941798941797,
"acc_norm_stderr": 0.02540255550326091
},
"harness|hendrycksTest-formal_logic|5": {
"acc": 0.3968253968253968,
"acc_stderr": 0.043758884927270605,
"acc_norm": 0.3968253968253968,
"acc_norm_stderr": 0.043758884927270605
},
"harness|hendrycksTest-global_facts|5": {
"acc": 0.36,
"acc_stderr": 0.048241815132442176,
"acc_norm": 0.36,
"acc_norm_stderr": 0.048241815132442176
},
"harness|hendrycksTest-high_school_biology|5": {
"acc": 0.7677419354838709,
"acc_stderr": 0.024022256130308235,
"acc_norm": 0.7677419354838709,
"acc_norm_stderr": 0.024022256130308235
},
"harness|hendrycksTest-high_school_chemistry|5": {
"acc": 0.5270935960591133,
"acc_stderr": 0.03512819077876106,
"acc_norm": 0.5270935960591133,
"acc_norm_stderr": 0.03512819077876106
},
"harness|hendrycksTest-high_school_computer_science|5": {
"acc": 0.68,
"acc_stderr": 0.04688261722621505,
"acc_norm": 0.68,
"acc_norm_stderr": 0.04688261722621505
},
"harness|hendrycksTest-high_school_european_history|5": {
"acc": 0.7575757575757576,
"acc_stderr": 0.03346409881055953,
"acc_norm": 0.7575757575757576,
"acc_norm_stderr": 0.03346409881055953
},
"harness|hendrycksTest-high_school_geography|5": {
"acc": 0.7727272727272727,
"acc_stderr": 0.029857515673386414,
"acc_norm": 0.7727272727272727,
"acc_norm_stderr": 0.029857515673386414
},
"harness|hendrycksTest-high_school_government_and_politics|5": {
"acc": 0.8911917098445595,
"acc_stderr": 0.022473253332768776,
"acc_norm": 0.8911917098445595,
"acc_norm_stderr": 0.022473253332768776
},
"harness|hendrycksTest-high_school_macroeconomics|5": {
"acc": 0.6282051282051282,
"acc_stderr": 0.024503472557110936,
"acc_norm": 0.6282051282051282,
"acc_norm_stderr": 0.024503472557110936
},
"harness|hendrycksTest-high_school_mathematics|5": {
"acc": 0.34814814814814815,
"acc_stderr": 0.029045600290616258,
"acc_norm": 0.34814814814814815,
"acc_norm_stderr": 0.029045600290616258
},
"harness|hendrycksTest-high_school_microeconomics|5": {
"acc": 0.6512605042016807,
"acc_stderr": 0.030956636328566545,
"acc_norm": 0.6512605042016807,
"acc_norm_stderr": 0.030956636328566545
},
"harness|hendrycksTest-high_school_physics|5": {
"acc": 0.304635761589404,
"acc_stderr": 0.03757949922943343,
"acc_norm": 0.304635761589404,
"acc_norm_stderr": 0.03757949922943343
},
"harness|hendrycksTest-high_school_psychology|5": {
"acc": 0.8330275229357799,
"acc_stderr": 0.015990154885073393,
"acc_norm": 0.8330275229357799,
"acc_norm_stderr": 0.015990154885073393
},
"harness|hendrycksTest-high_school_statistics|5": {
"acc": 0.49074074074074076,
"acc_stderr": 0.034093869469927006,
"acc_norm": 0.49074074074074076,
"acc_norm_stderr": 0.034093869469927006
},
"harness|hendrycksTest-high_school_us_history|5": {
"acc": 0.803921568627451,
"acc_stderr": 0.027865942286639318,
"acc_norm": 0.803921568627451,
"acc_norm_stderr": 0.027865942286639318
},
"harness|hendrycksTest-high_school_world_history|5": {
"acc": 0.7805907172995781,
"acc_stderr": 0.026939106581553945,
"acc_norm": 0.7805907172995781,
"acc_norm_stderr": 0.026939106581553945
},
"harness|hendrycksTest-human_aging|5": {
"acc": 0.7085201793721974,
"acc_stderr": 0.03050028317654585,
"acc_norm": 0.7085201793721974,
"acc_norm_stderr": 0.03050028317654585
},
"harness|hendrycksTest-human_sexuality|5": {
"acc": 0.7862595419847328,
"acc_stderr": 0.0359546161177469,
"acc_norm": 0.7862595419847328,
"acc_norm_stderr": 0.0359546161177469
},
"harness|hendrycksTest-international_law|5": {
"acc": 0.768595041322314,
"acc_stderr": 0.03849856098794088,
"acc_norm": 0.768595041322314,
"acc_norm_stderr": 0.03849856098794088
},
"harness|hendrycksTest-jurisprudence|5": {
"acc": 0.7870370370370371,
"acc_stderr": 0.0395783547198098,
"acc_norm": 0.7870370370370371,
"acc_norm_stderr": 0.0395783547198098
},
"harness|hendrycksTest-logical_fallacies|5": {
"acc": 0.7914110429447853,
"acc_stderr": 0.031921934489347235,
"acc_norm": 0.7914110429447853,
"acc_norm_stderr": 0.031921934489347235
},
"harness|hendrycksTest-machine_learning|5": {
"acc": 0.5089285714285714,
"acc_stderr": 0.04745033255489123,
"acc_norm": 0.5089285714285714,
"acc_norm_stderr": 0.04745033255489123
},
"harness|hendrycksTest-management|5": {
"acc": 0.8155339805825242,
"acc_stderr": 0.03840423627288276,
"acc_norm": 0.8155339805825242,
"acc_norm_stderr": 0.03840423627288276
},
"harness|hendrycksTest-marketing|5": {
"acc": 0.8760683760683761,
"acc_stderr": 0.021586494001281382,
"acc_norm": 0.8760683760683761,
"acc_norm_stderr": 0.021586494001281382
},
"harness|hendrycksTest-medical_genetics|5": {
"acc": 0.69,
"acc_stderr": 0.04648231987117316,
"acc_norm": 0.69,
"acc_norm_stderr": 0.04648231987117316
},
"harness|hendrycksTest-miscellaneous|5": {
"acc": 0.8160919540229885,
"acc_stderr": 0.013853724170922526,
"acc_norm": 0.8160919540229885,
"acc_norm_stderr": 0.013853724170922526
},
"harness|hendrycksTest-moral_disputes|5": {
"acc": 0.7398843930635838,
"acc_stderr": 0.023618678310069363,
"acc_norm": 0.7398843930635838,
"acc_norm_stderr": 0.023618678310069363
},
"harness|hendrycksTest-moral_scenarios|5": {
"acc": 0.35307262569832404,
"acc_stderr": 0.015984204545268565,
"acc_norm": 0.35307262569832404,
"acc_norm_stderr": 0.015984204545268565
},
"harness|hendrycksTest-nutrition|5": {
"acc": 0.7777777777777778,
"acc_stderr": 0.023805186524888135,
"acc_norm": 0.7777777777777778,
"acc_norm_stderr": 0.023805186524888135
},
"harness|hendrycksTest-philosophy|5": {
"acc": 0.6945337620578779,
"acc_stderr": 0.02616058445014045,
"acc_norm": 0.6945337620578779,
"acc_norm_stderr": 0.02616058445014045
},
"harness|hendrycksTest-prehistory|5": {
"acc": 0.75,
"acc_stderr": 0.02409347123262133,
"acc_norm": 0.75,
"acc_norm_stderr": 0.02409347123262133
},
"harness|hendrycksTest-professional_accounting|5": {
"acc": 0.4929078014184397,
"acc_stderr": 0.02982449855912901,
"acc_norm": 0.4929078014184397,
"acc_norm_stderr": 0.02982449855912901
},
"harness|hendrycksTest-professional_law|5": {
"acc": 0.4602346805736636,
"acc_stderr": 0.012729785386598564,
"acc_norm": 0.4602346805736636,
"acc_norm_stderr": 0.012729785386598564
},
"harness|hendrycksTest-professional_medicine|5": {
"acc": 0.6470588235294118,
"acc_stderr": 0.029029422815681393,
"acc_norm": 0.6470588235294118,
"acc_norm_stderr": 0.029029422815681393
},
"harness|hendrycksTest-professional_psychology|5": {
"acc": 0.6584967320261438,
"acc_stderr": 0.019184639328092487,
"acc_norm": 0.6584967320261438,
"acc_norm_stderr": 0.019184639328092487
},
"harness|hendrycksTest-public_relations|5": {
"acc": 0.6636363636363637,
"acc_stderr": 0.04525393596302506,
"acc_norm": 0.6636363636363637,
"acc_norm_stderr": 0.04525393596302506
},
"harness|hendrycksTest-security_studies|5": {
"acc": 0.7387755102040816,
"acc_stderr": 0.028123429335142783,
"acc_norm": 0.7387755102040816,
"acc_norm_stderr": 0.028123429335142783
},
"harness|hendrycksTest-sociology|5": {
"acc": 0.8308457711442786,
"acc_stderr": 0.026508590656233268,
"acc_norm": 0.8308457711442786,
"acc_norm_stderr": 0.026508590656233268
},
"harness|hendrycksTest-us_foreign_policy|5": {
"acc": 0.88,
"acc_stderr": 0.032659863237109066,
"acc_norm": 0.88,
"acc_norm_stderr": 0.032659863237109066
},
"harness|hendrycksTest-virology|5": {
"acc": 0.5542168674698795,
"acc_stderr": 0.03869543323472101,
"acc_norm": 0.5542168674698795,
"acc_norm_stderr": 0.03869543323472101
},
"harness|hendrycksTest-world_religions|5": {
"acc": 0.8245614035087719,
"acc_stderr": 0.02917088550072767,
"acc_norm": 0.8245614035087719,
"acc_norm_stderr": 0.02917088550072767
},
"harness|truthfulqa:mc|0": {
"mc1": 0.3329253365973072,
"mc1_stderr": 0.016497402382012052,
"mc2": 0.48572469546439073,
"mc2_stderr": 0.01547664094918918
},
"harness|winogrande|5": {
"acc": 0.7679558011049724,
"acc_stderr": 0.011864149691827936
},
"harness|gsm8k|5": {
"acc": 0.6982562547384382,
"acc_stderr": 0.012643544762873358
}
}
```
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] |
CyberHarem/provence_arknights | ---
license: mit
task_categories:
- text-to-image
tags:
- art
- not-for-all-audiences
size_categories:
- n<1K
---
# Dataset of provence/プロヴァンス/普罗旺斯 (Arknights)
This is the dataset of provence/プロヴァンス/普罗旺斯 (Arknights), containing 153 images and their tags.
The core tags of this character are `animal_ears, wolf_ears, tail, long_hair, purple_hair, wolf_tail, yellow_eyes, braid, hair_between_eyes, wolf_girl, breasts, large_tail, single_braid, medium_breasts`, which are pruned in this dataset.
Images are crawled from many sites (e.g. danbooru, pixiv, zerochan ...), the auto-crawling system is powered by [DeepGHS Team](https://github.com/deepghs)([huggingface organization](https://huggingface.co/deepghs)).
## List of Packages
| Name | Images | Size | Download | Type | Description |
|:-----------------|---------:|:-----------|:--------------------------------------------------------------------------------------------------------------------|:-----------|:---------------------------------------------------------------------|
| raw | 153 | 252.16 MiB | [Download](https://huggingface.co/datasets/CyberHarem/provence_arknights/resolve/main/dataset-raw.zip) | Waifuc-Raw | Raw data with meta information (min edge aligned to 1400 if larger). |
| 1200 | 153 | 207.20 MiB | [Download](https://huggingface.co/datasets/CyberHarem/provence_arknights/resolve/main/dataset-1200.zip) | IMG+TXT | dataset with the shorter side not exceeding 1200 pixels. |
| stage3-p480-1200 | 398 | 409.73 MiB | [Download](https://huggingface.co/datasets/CyberHarem/provence_arknights/resolve/main/dataset-stage3-p480-1200.zip) | IMG+TXT | 3-stage cropped dataset with the area not less than 480x480 pixels. |
### Load Raw Dataset with Waifuc
We provide raw dataset (including tagged images) for [waifuc](https://deepghs.github.io/waifuc/main/tutorials/installation/index.html) loading. If you need this, just run the following code
```python
import os
import zipfile
from huggingface_hub import hf_hub_download
from waifuc.source import LocalSource
# download raw archive file
zip_file = hf_hub_download(
repo_id='CyberHarem/provence_arknights',
repo_type='dataset',
filename='dataset-raw.zip',
)
# extract files to your directory
dataset_dir = 'dataset_dir'
os.makedirs(dataset_dir, exist_ok=True)
with zipfile.ZipFile(zip_file, 'r') as zf:
zf.extractall(dataset_dir)
# load the dataset with waifuc
source = LocalSource(dataset_dir)
for item in source:
print(item.image, item.meta['filename'], item.meta['tags'])
```
## List of Clusters
List of tag clustering result, maybe some outfits can be mined here.
### Raw Text Version
| # | Samples | Img-1 | Img-2 | Img-3 | Img-4 | Img-5 | Tags |
|----:|----------:|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| 0 | 17 |  |  |  |  |  | 1girl, looking_at_viewer, solo, simple_background, white_background, black_gloves, upper_body, jacket, :d, open_mouth, hood, puffy_short_sleeves, purple_shirt |
| 1 | 7 |  |  |  |  |  | 1girl, bare_shoulders, black_shorts, cleavage, hip_vent, looking_at_viewer, navel, official_alternate_costume, short_shorts, solo, stomach, black_bikini, cowboy_shot, midriff, short_sleeves, thighs, hand_up, open_mouth, shirt, single_glove, :d, blue_gloves, crop_top, fingerless_gloves, holding_food, popsicle, standing, thigh_strap, bikini_under_clothes, clothing_cutout, groin, off_shoulder, simple_background, white_background |
| 2 | 7 |  |  |  |  |  | 1girl, blush, hetero, open_mouth, solo_focus, 1boy, large_breasts, navel, nipples, penis, sex, vaginal, mosaic_censoring, extra_ears, looking_at_viewer, spread_legs, completely_nude, cum_in_pussy, cum_overflow, lying, tongue_out |
### Table Version
| # | Samples | Img-1 | Img-2 | Img-3 | Img-4 | Img-5 | 1girl | looking_at_viewer | solo | simple_background | white_background | black_gloves | upper_body | jacket | :d | open_mouth | hood | puffy_short_sleeves | purple_shirt | bare_shoulders | black_shorts | cleavage | hip_vent | navel | official_alternate_costume | short_shorts | stomach | black_bikini | cowboy_shot | midriff | short_sleeves | thighs | hand_up | shirt | single_glove | blue_gloves | crop_top | fingerless_gloves | holding_food | popsicle | standing | thigh_strap | bikini_under_clothes | clothing_cutout | groin | off_shoulder | blush | hetero | solo_focus | 1boy | large_breasts | nipples | penis | sex | vaginal | mosaic_censoring | extra_ears | spread_legs | completely_nude | cum_in_pussy | cum_overflow | lying | tongue_out |
|----:|----------:|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------|:--------|:--------------------|:-------|:--------------------|:-------------------|:---------------|:-------------|:---------|:-----|:-------------|:-------|:----------------------|:---------------|:-----------------|:---------------|:-----------|:-----------|:--------|:-----------------------------|:---------------|:----------|:---------------|:--------------|:----------|:----------------|:---------|:----------|:--------|:---------------|:--------------|:-----------|:--------------------|:---------------|:-----------|:-----------|:--------------|:-----------------------|:------------------|:--------|:---------------|:--------|:---------|:-------------|:-------|:----------------|:----------|:--------|:------|:----------|:-------------------|:-------------|:--------------|:------------------|:---------------|:---------------|:--------|:-------------|
| 0 | 17 |  |  |  |  |  | X | X | X | X | X | X | X | X | X | X | X | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | |
| 1 | 7 |  |  |  |  |  | X | X | X | X | X | | | | X | X | | | | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | | | | | | | | | | | | | | | | | |
| 2 | 7 |  |  |  |  |  | X | X | | | | | | | | X | | | | | | | | X | | | | | | | | | | | | | | | | | | | | | | | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X |
|
mazkobot/1_digit_mask_ensemble_distilled_from_cv12_balanced_mfcc | ---
dataset_info:
features:
- name: logits
sequence: float32
- name: mfcc
sequence:
sequence: float64
splits:
- name: train
num_bytes: 25771854448.0
num_examples: 5061244
download_size: 26308842420
dataset_size: 25771854448.0
---
# Dataset Card for "1_digit_mask_ensemble_distilled_from_cv12_balanced_mfcc"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
kpriyanshu256/MultiTabQA-multitable_pretraining-Salesforce-codet5-base_train-latex-23000 | ---
dataset_info:
features:
- name: input_ids
sequence:
sequence: int32
- name: attention_mask
sequence:
sequence: int8
- name: labels
sequence:
sequence: int64
splits:
- name: train
num_bytes: 13336000
num_examples: 1000
download_size: 1006931
dataset_size: 13336000
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
|
zhangshuoming/math_23k_train_numeric | ---
dataset_info:
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 20552640.641394995
num_examples: 21094
download_size: 2752930
dataset_size: 20552640.641394995
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "math_23k_train_numeric"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
ickma/sec_shots | ---
dataset_info:
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 142
num_examples: 3
download_size: 1090
dataset_size: 142
---
# Dataset Card for "sec_shots"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
TigerResearch/pretrain_en | ---
dataset_info:
features:
- name: content
dtype: string
splits:
- name: train
num_bytes: 48490123196
num_examples: 22690306
download_size: 5070161762
dataset_size: 48490123196
license: apache-2.0
task_categories:
- text-generation
language:
- en
size_categories:
- 10M<n<100M
---
# Dataset Card for "pretrain_en"
[Tigerbot](https://github.com/TigerResearch/TigerBot) pretrain数据的英文部分。
## Usage
```python
import datasets
ds_sft = datasets.load_dataset('TigerResearch/pretrain_en')
``` |
urdu_sentiment_corpus | ---
annotations_creators:
- expert-generated
language_creators:
- crowdsourced
language:
- ur
license:
- unknown
multilinguality:
- monolingual
size_categories:
- 1K<n<10K
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- sentiment-classification
paperswithcode_id: urdu-sentiment-corpus
pretty_name: Urdu Sentiment Corpus (USC)
dataset_info:
features:
- name: sentence
dtype: string
- name: sentiment
dtype:
class_label:
names:
'0': P
'1': N
'2': O
splits:
- name: train
num_bytes: 161190
num_examples: 1000
download_size: 51583
dataset_size: 161190
---
# Dataset Card for Urdu Sentiment Corpus (USC)
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [Github](https://github.com/MuhammadYaseenKhan/Urdu-Sentiment-Corpus)
- **Repository:** [Github](https://github.com/MuhammadYaseenKhan/Urdu-Sentiment-Corpus)
- **Paper:** [IEEE](https://ieeexplore.ieee.org/abstract/document/9080043)
- **Leaderboard:**
- **Point of Contact:** [Muhammad Yaseen Khan](https://github.com/MuhammadYaseenKhan)
### Dataset Summary
[More Information Needed]
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
- sentences: The Urdu tweet
- sentiment: The sentiment that was exhibited in the tweet, which can be Positive(P) or Negative(N) or Objective(O).
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
Thanks to [@chaitnayabasava](https://github.com/chaitnayabasava) for adding this dataset. |
shrinath-suresh/so_5k_with_short_answer | ---
license: apache-2.0
---
|
Falah/cabo_da_roca_light_conditions | ---
dataset_info:
features:
- name: prompts
dtype: string
splits:
- name: train
num_bytes: 707338
num_examples: 10000
download_size: 11725
dataset_size: 707338
---
# Dataset Card for "cabo_da_roca_light_conditions"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
Lollitor/FineTuneDataset1024 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
dataset_info:
features:
- name: sequence
dtype: string
- name: label
dtype: float64
- name: __index_level_0__
dtype: int64
splits:
- name: train
num_bytes: 15148943
num_examples: 15389
- name: validation
num_bytes: 1683634
num_examples: 1710
download_size: 7973519
dataset_size: 16832577
---
# Dataset Card for "FineTuneDataset1024"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
promptora11/dataset_train | ---
dataset_info:
features:
- name: Query
dtype: string
- name: Response
dtype: string
- name: __index_level_0__
dtype: int64
splits:
- name: train
num_bytes: 6750
num_examples: 32
download_size: 6692
dataset_size: 6750
---
# Dataset Card for "dataset_train"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
cahya/wikipedia_id_20230101 | ---
dataset_info:
features:
- name: id
dtype: string
- name: url
dtype: string
- name: title
dtype: string
- name: text
dtype: string
splits:
- name: train
num_bytes: 1051737365
num_examples: 634559
download_size: 544132473
dataset_size: 1051737365
---
# Dataset Card for "wikipedia_id_20230101"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
Vrushali/my-image-captioning-dataset | ---
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': blipdataset
- name: text
dtype: string
splits:
- name: train
num_bytes: 456995431.0
num_examples: 29048
download_size: 461550035
dataset_size: 456995431.0
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "my-image-captioning-dataset"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
reuters21578 | ---
language:
- en
license: other
paperswithcode_id: reuters-21578
pretty_name: Reuters-21578 Text Categorization Collection
dataset_info:
- config_name: ModApte
features:
- name: text
dtype: string
- name: text_type
dtype: string
- name: topics
sequence: string
- name: lewis_split
dtype: string
- name: cgis_split
dtype: string
- name: old_id
dtype: string
- name: new_id
dtype: string
- name: places
sequence: string
- name: people
sequence: string
- name: orgs
sequence: string
- name: exchanges
sequence: string
- name: date
dtype: string
- name: title
dtype: string
splits:
- name: test
num_bytes: 2971653
num_examples: 3299
- name: train
num_bytes: 9161179
num_examples: 9603
- name: unused
num_bytes: 948244
num_examples: 722
download_size: 8150596
dataset_size: 13081076
- config_name: ModHayes
features:
- name: text
dtype: string
- name: text_type
dtype: string
- name: topics
sequence: string
- name: lewis_split
dtype: string
- name: cgis_split
dtype: string
- name: old_id
dtype: string
- name: new_id
dtype: string
- name: places
sequence: string
- name: people
sequence: string
- name: orgs
sequence: string
- name: exchanges
sequence: string
- name: date
dtype: string
- name: title
dtype: string
splits:
- name: test
num_bytes: 948244
num_examples: 722
- name: train
num_bytes: 19071106
num_examples: 20856
download_size: 8150596
dataset_size: 20019350
- config_name: ModLewis
features:
- name: text
dtype: string
- name: text_type
dtype: string
- name: topics
sequence: string
- name: lewis_split
dtype: string
- name: cgis_split
dtype: string
- name: old_id
dtype: string
- name: new_id
dtype: string
- name: places
sequence: string
- name: people
sequence: string
- name: orgs
sequence: string
- name: exchanges
sequence: string
- name: date
dtype: string
- name: title
dtype: string
splits:
- name: test
num_bytes: 5400506
num_examples: 6188
- name: train
num_bytes: 12994591
num_examples: 13625
- name: unused
num_bytes: 948244
num_examples: 722
download_size: 8150596
dataset_size: 19343341
---
# Dataset Card for "reuters21578"
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** https://archive.ics.uci.edu/dataset/137/reuters+21578+text+categorization+collection
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 24.45 MB
- **Size of the generated dataset:** 52.22 MB
- **Total amount of disk used:** 76.67 MB
### Dataset Summary
The Reuters-21578 dataset is one of the most widely used data collections for text
categorization research. It is collected from the Reuters financial newswire service in 1987.
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Dataset Structure
### Data Instances
#### ModApte
- **Size of downloaded dataset files:** 8.15 MB
- **Size of the generated dataset:** 13.05 MB
- **Total amount of disk used:** 21.21 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"cgis_split": "\"TRAINING-SET\"",
"date": "19-MAR-1987 06:17:22.36",
"exchanges": [],
"lewis_split": "\"TRAIN\"",
"new_id": "\"7001\"",
"old_id": "\"11914\"",
"orgs": [],
"people": [],
"places": ["australia"],
"text": "\"Media group John Fairfax Ltd <FFXA.S>\\nsaid that its flat first half net profit partly reflected the\\nimpact of changes in t...",
"title": "FAIRFAX SAYS HIGHER TAX HITS FIRST HALF EARNINGS",
"topics": ["earn"]
}
```
#### ModHayes
- **Size of downloaded dataset files:** 8.15 MB
- **Size of the generated dataset:** 19.79 MB
- **Total amount of disk used:** 27.93 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"cgis_split": "\"TRAINING-SET\"",
"date": "19-OCT-1987 23:49:31.45",
"exchanges": [],
"lewis_split": "\"TEST\"",
"new_id": "\"20001\"",
"old_id": "\"20596\"",
"orgs": [],
"people": [],
"places": ["japan", "usa"],
"text": "\"If the dollar goes the way of Wall Street,\\nJapanese will finally move out of dollar investments in a\\nserious way, Japan inves...",
"title": "IF DOLLAR FOLLOWS WALL STREET JAPANESE WILL DIVEST",
"topics": ["money-fx"]
}
```
#### ModLewis
- **Size of downloaded dataset files:** 8.15 MB
- **Size of the generated dataset:** 19.38 MB
- **Total amount of disk used:** 27.54 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"cgis_split": "\"TRAINING-SET\"",
"date": "19-MAR-1987 06:17:22.36",
"exchanges": [],
"lewis_split": "\"TRAIN\"",
"new_id": "\"7001\"",
"old_id": "\"11914\"",
"orgs": [],
"people": [],
"places": ["australia"],
"text": "\"Media group John Fairfax Ltd <FFXA.S>\\nsaid that its flat first half net profit partly reflected the\\nimpact of changes in t...",
"title": "FAIRFAX SAYS HIGHER TAX HITS FIRST HALF EARNINGS",
"topics": ["earn"]
}
```
### Data Fields
The data fields are the same among all splits.
#### ModApte
- `text`: a `string` feature.
- `topics`: a `list` of `string` features.
- `lewis_split`: a `string` feature.
- `cgis_split`: a `string` feature.
- `old_id`: a `string` feature.
- `new_id`: a `string` feature.
- `places`: a `list` of `string` features.
- `people`: a `list` of `string` features.
- `orgs`: a `list` of `string` features.
- `exchanges`: a `list` of `string` features.
- `date`: a `string` feature.
- `title`: a `string` feature.
#### ModHayes
- `text`: a `string` feature.
- `topics`: a `list` of `string` features.
- `lewis_split`: a `string` feature.
- `cgis_split`: a `string` feature.
- `old_id`: a `string` feature.
- `new_id`: a `string` feature.
- `places`: a `list` of `string` features.
- `people`: a `list` of `string` features.
- `orgs`: a `list` of `string` features.
- `exchanges`: a `list` of `string` features.
- `date`: a `string` feature.
- `title`: a `string` feature.
#### ModLewis
- `text`: a `string` feature.
- `topics`: a `list` of `string` features.
- `lewis_split`: a `string` feature.
- `cgis_split`: a `string` feature.
- `old_id`: a `string` feature.
- `new_id`: a `string` feature.
- `places`: a `list` of `string` features.
- `people`: a `list` of `string` features.
- `orgs`: a `list` of `string` features.
- `exchanges`: a `list` of `string` features.
- `date`: a `string` feature.
- `title`: a `string` feature.
### Data Splits
#### ModApte
| |train|unused|test|
|-------|----:|-----:|---:|
|ModApte| 8762| 720|3009|
#### ModHayes
| |train|test|
|--------|----:|---:|
|ModHayes|18323| 720|
#### ModLewis
| |train|unused|test|
|--------|----:|-----:|---:|
|ModLewis|12449| 720|5458|
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
According to the dataset website (https://archive.ics.uci.edu/dataset/137/reuters+21578+text+categorization+collection),
this dataset is licensed under [Creative Commons Attribution 4.0 International](https://creativecommons.org/licenses/by/4.0/legalcode)
(C BY 4.0) license.
However, the source data file contains a `README.txt` file with the following information under the
**Copyright & Notification** section:
> The copyright for the text of newswire articles and Reuters
annotations in the Reuters-21578 collection resides with Reuters Ltd.
Reuters Ltd. and Carnegie Group, Inc. have agreed to allow the free
distribution of this data *for research purposes only*.
> If you publish results based on this data set, please acknowledge
its use, refer to the data set by the name "Reuters-21578,
Distribution 1.0", and inform your readers of the current location of
the data set (see "Availability & Questions").
### Citation Information
```
@article{APTE94,
author = {Chidanand Apt{'{e}} and Fred Damerau and Sholom M. Weiss},
title = {Automated Learning of Decision Rules for Text Categorization},
journal = {ACM Transactions on Information Systems},
year = {1994},
note = {To appear.}
}
@inproceedings{APTE94b,
author = {Chidanand Apt{'{e}} and Fred Damerau and Sholom M. Weiss},
title = {Toward Language Independent Automated Learning of Text Categorization Models},
booktitle = {sigir94},
year = {1994},
note = {To appear.}
}
@inproceedings{HAYES8},
author = {Philip J. Hayes and Peggy M. Anderson and Irene B. Nirenburg and
Linda M. Schmandt},
title = {{TCS}: A Shell for Content-Based Text Categorization},
booktitle = {IEEE Conference on Artificial Intelligence Applications},
year = {1990}
}
@inproceedings{HAYES90b,
author = {Philip J. Hayes and Steven P. Weinstein},
title = {{CONSTRUE/TIS:} A System for Content-Based Indexing of a
Database of News Stories},
booktitle = {Second Annual Conference on Innovative Applications of
Artificial Intelligence},
year = {1990}
}
@incollection{HAYES92 ,
author = {Philip J. Hayes},
title = {Intelligent High-Volume Text Processing using Shallow,
Domain-Specific Techniques},
booktitle = {Text-Based Intelligent Systems},
publisher = {Lawrence Erlbaum},
address = {Hillsdale, NJ},
year = {1992},
editor = {Paul S. Jacobs}
}
@inproceedings{LEWIS91c ,
author = {David D. Lewis},
title = {Evaluating Text Categorization},
booktitle = {Proceedings of Speech and Natural Language Workshop},
year = {1991},
month = {feb},
organization = {Defense Advanced Research Projects Agency},
publisher = {Morgan Kaufmann},
pages = {312--318}
}
@phdthesis{LEWIS91d,
author = {David Dolan Lewis},
title = {Representation and Learning in Information Retrieval},
school = {Computer Science Dept.; Univ. of Massachusetts; Amherst, MA 01003},
year = 1992},
note = {Technical Report 91--93.}
}
@inproceedings{LEWIS91e,
author = {David D. Lewis},
title = {Data Extraction as Text Categorization: An Experiment with
the {MUC-3} Corpus},
booktitle = {Proceedings of the Third Message Understanding Evaluation
and Conference},
year = {1991},
month = {may},
organization = {Defense Advanced Research Projects Agency},
publisher = {Morgan Kaufmann},
address = {Los Altos, CA}
}
@inproceedings{LEWIS92b,
author = {David D. Lewis},
title = {An Evaluation of Phrasal and Clustered Representations on a Text
Categorization Task},
booktitle = {Fifteenth Annual International ACM SIGIR Conference on
Research and Development in Information Retrieval},
year = {1992},
pages = {37--50}
}
@inproceedings{LEWIS92d ,
author = {David D. Lewis and Richard M. Tong},
title = {Text Filtering in {MUC-3} and {MUC-4}},
booktitle = {Proceedings of the Fourth Message Understanding Conference ({MUC-4})},
year = {1992},
month = {jun},
organization = {Defense Advanced Research Projects Agency},
publisher = {Morgan Kaufmann},
address = {Los Altos, CA}
}
@inproceedings{LEWIS92e,
author = {David D. Lewis},
title = {Feature Selection and Feature Extraction for Text Categorization},
booktitle = {Proceedings of Speech and Natural Language Workshop},
year = {1992},
month = {feb} ,
organization = {Defense Advanced Research Projects Agency},
publisher = {Morgan Kaufmann},
pages = {212--217}
}
@inproceedings{LEWIS94b,
author = {David D. Lewis and Marc Ringuette},
title = {A Comparison of Two Learning Algorithms for Text Categorization},
booktitle = {Symposium on Document Analysis and Information Retrieval},
year = {1994},
organization = {ISRI; Univ. of Nevada, Las Vegas},
address = {Las Vegas, NV},
month = {apr},
pages = {81--93}
}
@article{LEWIS94d,
author = {David D. Lewis and Philip J. Hayes},
title = {Guest Editorial},
journal = {ACM Transactions on Information Systems},
year = {1994},
volume = {12},
number = {3},
pages = {231},
month = {jul}
}
@article{SPARCKJONES76,
author = {K. {Sparck Jones} and C. J. {van Rijsbergen}},
title = {Information Retrieval Test Collections},
journal = {Journal of Documentation},
year = {1976},
volume = {32},
number = {1},
pages = {59--75}
}
@book{WEISS91,
author = {Sholom M. Weiss and Casimir A. Kulikowski},
title = {Computer Systems That Learn},
publisher = {Morgan Kaufmann},
year = {1991},
address = {San Mateo, CA}
}
```
### Contributions
Thanks to [@jplu](https://github.com/jplu), [@jbragg](https://github.com/jbragg), [@thomwolf](https://github.com/thomwolf), [@mariamabarham](https://github.com/mariamabarham), [@lhoestq](https://github.com/lhoestq) for adding this dataset. |
coroperv/oxford-flowers | ---
license: unknown
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': '1'
'1': '10'
'2': '100'
'3': '102'
'4': '12'
'5': '13'
'6': '15'
'7': '16'
'8': '17'
'9': '19'
'10': '22'
'11': '26'
'12': '28'
'13': '29'
'14': '31'
'15': '33'
'16': '34'
'17': '36'
'18': '38'
'19': '40'
'20': '41'
'21': '42'
'22': '45'
'23': '47'
'24': '48'
'25': '49'
'26': '5'
'27': '50'
'28': '51'
'29': '52'
'30': '54'
'31': '55'
'32': '56'
'33': '57'
'34': '59'
'35': '62'
'36': '63'
'37': '64'
'38': '65'
'39': '66'
'40': '68'
'41': '69'
'42': '7'
'43': '70'
'44': '71'
'45': '73'
'46': '74'
'47': '75'
'48': '76'
'49': '77'
'50': '78'
'51': '81'
'52': '82'
'53': '83'
'54': '85'
'55': '86'
'56': '89'
'57': '91'
'58': '96'
'59': '98'
'60': '99'
splits:
- name: train
num_bytes: 103936242.384
num_examples: 2452
download_size: 99259189
dataset_size: 103936242.384
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
|
AdapterOcean/med_alpaca_standardized_cluster_0_alpaca | ---
dataset_info:
features:
- name: input
dtype: string
- name: output
dtype: string
splits:
- name: train
num_bytes: 18113077
num_examples: 10247
download_size: 10005035
dataset_size: 18113077
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "med_alpaca_standardized_cluster_0_alpaca"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
open-llm-leaderboard/details_BreadAi__gpt-YA-1-1_160M | ---
pretty_name: Evaluation run of BreadAi/gpt-YA-1-1_160M
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [BreadAi/gpt-YA-1-1_160M](https://huggingface.co/BreadAi/gpt-YA-1-1_160M) on the\
\ [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 64 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 2 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the agregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_BreadAi__gpt-YA-1-1_160M\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-09-16T23:15:39.203631](https://huggingface.co/datasets/open-llm-leaderboard/details_BreadAi__gpt-YA-1-1_160M/blob/main/results_2023-09-16T23-15-39.203631.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"em\": 0.0037751677852348995,\n\
\ \"em_stderr\": 0.0006280387809484644,\n \"f1\": 0.02323720637583893,\n\
\ \"f1_stderr\": 0.0010591010187142242,\n \"acc\": 0.2533543804262036,\n\
\ \"acc_stderr\": 0.007025610346165173\n },\n \"harness|drop|3\": {\n\
\ \"em\": 0.0037751677852348995,\n \"em_stderr\": 0.0006280387809484644,\n\
\ \"f1\": 0.02323720637583893,\n \"f1_stderr\": 0.0010591010187142242\n\
\ },\n \"harness|gsm8k|5\": {\n \"acc\": 0.0,\n \"acc_stderr\"\
: 0.0\n },\n \"harness|winogrande|5\": {\n \"acc\": 0.5067087608524072,\n\
\ \"acc_stderr\": 0.014051220692330346\n }\n}\n```"
repo_url: https://huggingface.co/BreadAi/gpt-YA-1-1_160M
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|arc:challenge|25_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_drop_3
data_files:
- split: 2023_09_16T23_15_39.203631
path:
- '**/details_harness|drop|3_2023-09-16T23-15-39.203631.parquet'
- split: latest
path:
- '**/details_harness|drop|3_2023-09-16T23-15-39.203631.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_09_16T23_15_39.203631
path:
- '**/details_harness|gsm8k|5_2023-09-16T23-15-39.203631.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-09-16T23-15-39.203631.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hellaswag|10_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-07-19T14:02:10.207194.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-management|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-virology|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- '**/details_harness|truthfulqa:mc|0_2023-07-19T14:02:10.207194.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2023-07-19T14:02:10.207194.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_09_16T23_15_39.203631
path:
- '**/details_harness|winogrande|5_2023-09-16T23-15-39.203631.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-09-16T23-15-39.203631.parquet'
- config_name: results
data_files:
- split: 2023_07_19T14_02_10.207194
path:
- results_2023-07-19T14:02:10.207194.parquet
- split: 2023_09_16T23_15_39.203631
path:
- results_2023-09-16T23-15-39.203631.parquet
- split: latest
path:
- results_2023-09-16T23-15-39.203631.parquet
---
# Dataset Card for Evaluation run of BreadAi/gpt-YA-1-1_160M
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/BreadAi/gpt-YA-1-1_160M
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [BreadAi/gpt-YA-1-1_160M](https://huggingface.co/BreadAi/gpt-YA-1-1_160M) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 64 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 2 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_BreadAi__gpt-YA-1-1_160M",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-09-16T23:15:39.203631](https://huggingface.co/datasets/open-llm-leaderboard/details_BreadAi__gpt-YA-1-1_160M/blob/main/results_2023-09-16T23-15-39.203631.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"em": 0.0037751677852348995,
"em_stderr": 0.0006280387809484644,
"f1": 0.02323720637583893,
"f1_stderr": 0.0010591010187142242,
"acc": 0.2533543804262036,
"acc_stderr": 0.007025610346165173
},
"harness|drop|3": {
"em": 0.0037751677852348995,
"em_stderr": 0.0006280387809484644,
"f1": 0.02323720637583893,
"f1_stderr": 0.0010591010187142242
},
"harness|gsm8k|5": {
"acc": 0.0,
"acc_stderr": 0.0
},
"harness|winogrande|5": {
"acc": 0.5067087608524072,
"acc_stderr": 0.014051220692330346
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] |
open-llm-leaderboard/details_allknowingroger__NeuralCeptrix-7B-slerp | ---
pretty_name: Evaluation run of allknowingroger/NeuralCeptrix-7B-slerp
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [allknowingroger/NeuralCeptrix-7B-slerp](https://huggingface.co/allknowingroger/NeuralCeptrix-7B-slerp)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 63 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the aggregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_allknowingroger__NeuralCeptrix-7B-slerp\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2024-04-10T21:14:16.828642](https://huggingface.co/datasets/open-llm-leaderboard/details_allknowingroger__NeuralCeptrix-7B-slerp/blob/main/results_2024-04-10T21-14-16.828642.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"acc\": 0.6495434097360565,\n\
\ \"acc_stderr\": 0.032279899940585416,\n \"acc_norm\": 0.6487090853177374,\n\
\ \"acc_norm_stderr\": 0.03296063025774512,\n \"mc1\": 0.6523867809057528,\n\
\ \"mc1_stderr\": 0.01667076918889731,\n \"mc2\": 0.7984722132506933,\n\
\ \"mc2_stderr\": 0.013281143496846105\n },\n \"harness|arc:challenge|25\"\
: {\n \"acc\": 0.7022184300341296,\n \"acc_stderr\": 0.013363080107244482,\n\
\ \"acc_norm\": 0.7209897610921502,\n \"acc_norm_stderr\": 0.013106784883601334\n\
\ },\n \"harness|hellaswag|10\": {\n \"acc\": 0.7197769368651663,\n\
\ \"acc_stderr\": 0.004481902637505651,\n \"acc_norm\": 0.8919537940649273,\n\
\ \"acc_norm_stderr\": 0.0030980431017758477\n },\n \"harness|hendrycksTest-abstract_algebra|5\"\
: {\n \"acc\": 0.36,\n \"acc_stderr\": 0.048241815132442176,\n \
\ \"acc_norm\": 0.36,\n \"acc_norm_stderr\": 0.048241815132442176\n \
\ },\n \"harness|hendrycksTest-anatomy|5\": {\n \"acc\": 0.6222222222222222,\n\
\ \"acc_stderr\": 0.04188307537595853,\n \"acc_norm\": 0.6222222222222222,\n\
\ \"acc_norm_stderr\": 0.04188307537595853\n },\n \"harness|hendrycksTest-astronomy|5\"\
: {\n \"acc\": 0.6973684210526315,\n \"acc_stderr\": 0.03738520676119669,\n\
\ \"acc_norm\": 0.6973684210526315,\n \"acc_norm_stderr\": 0.03738520676119669\n\
\ },\n \"harness|hendrycksTest-business_ethics|5\": {\n \"acc\": 0.62,\n\
\ \"acc_stderr\": 0.048783173121456316,\n \"acc_norm\": 0.62,\n \
\ \"acc_norm_stderr\": 0.048783173121456316\n },\n \"harness|hendrycksTest-clinical_knowledge|5\"\
: {\n \"acc\": 0.6792452830188679,\n \"acc_stderr\": 0.028727502957880267,\n\
\ \"acc_norm\": 0.6792452830188679,\n \"acc_norm_stderr\": 0.028727502957880267\n\
\ },\n \"harness|hendrycksTest-college_biology|5\": {\n \"acc\": 0.7777777777777778,\n\
\ \"acc_stderr\": 0.03476590104304134,\n \"acc_norm\": 0.7777777777777778,\n\
\ \"acc_norm_stderr\": 0.03476590104304134\n },\n \"harness|hendrycksTest-college_chemistry|5\"\
: {\n \"acc\": 0.52,\n \"acc_stderr\": 0.050211673156867795,\n \
\ \"acc_norm\": 0.52,\n \"acc_norm_stderr\": 0.050211673156867795\n \
\ },\n \"harness|hendrycksTest-college_computer_science|5\": {\n \"\
acc\": 0.57,\n \"acc_stderr\": 0.04975698519562428,\n \"acc_norm\"\
: 0.57,\n \"acc_norm_stderr\": 0.04975698519562428\n },\n \"harness|hendrycksTest-college_mathematics|5\"\
: {\n \"acc\": 0.35,\n \"acc_stderr\": 0.04793724854411019,\n \
\ \"acc_norm\": 0.35,\n \"acc_norm_stderr\": 0.04793724854411019\n \
\ },\n \"harness|hendrycksTest-college_medicine|5\": {\n \"acc\": 0.6358381502890174,\n\
\ \"acc_stderr\": 0.03669072477416907,\n \"acc_norm\": 0.6358381502890174,\n\
\ \"acc_norm_stderr\": 0.03669072477416907\n },\n \"harness|hendrycksTest-college_physics|5\"\
: {\n \"acc\": 0.38235294117647056,\n \"acc_stderr\": 0.04835503696107224,\n\
\ \"acc_norm\": 0.38235294117647056,\n \"acc_norm_stderr\": 0.04835503696107224\n\
\ },\n \"harness|hendrycksTest-computer_security|5\": {\n \"acc\":\
\ 0.73,\n \"acc_stderr\": 0.044619604333847394,\n \"acc_norm\": 0.73,\n\
\ \"acc_norm_stderr\": 0.044619604333847394\n },\n \"harness|hendrycksTest-conceptual_physics|5\"\
: {\n \"acc\": 0.5617021276595745,\n \"acc_stderr\": 0.03243618636108102,\n\
\ \"acc_norm\": 0.5617021276595745,\n \"acc_norm_stderr\": 0.03243618636108102\n\
\ },\n \"harness|hendrycksTest-econometrics|5\": {\n \"acc\": 0.47368421052631576,\n\
\ \"acc_stderr\": 0.04697085136647863,\n \"acc_norm\": 0.47368421052631576,\n\
\ \"acc_norm_stderr\": 0.04697085136647863\n },\n \"harness|hendrycksTest-electrical_engineering|5\"\
: {\n \"acc\": 0.5586206896551724,\n \"acc_stderr\": 0.04137931034482757,\n\
\ \"acc_norm\": 0.5586206896551724,\n \"acc_norm_stderr\": 0.04137931034482757\n\
\ },\n \"harness|hendrycksTest-elementary_mathematics|5\": {\n \"acc\"\
: 0.41798941798941797,\n \"acc_stderr\": 0.025402555503260912,\n \"\
acc_norm\": 0.41798941798941797,\n \"acc_norm_stderr\": 0.025402555503260912\n\
\ },\n \"harness|hendrycksTest-formal_logic|5\": {\n \"acc\": 0.47619047619047616,\n\
\ \"acc_stderr\": 0.04467062628403273,\n \"acc_norm\": 0.47619047619047616,\n\
\ \"acc_norm_stderr\": 0.04467062628403273\n },\n \"harness|hendrycksTest-global_facts|5\"\
: {\n \"acc\": 0.31,\n \"acc_stderr\": 0.04648231987117316,\n \
\ \"acc_norm\": 0.31,\n \"acc_norm_stderr\": 0.04648231987117316\n \
\ },\n \"harness|hendrycksTest-high_school_biology|5\": {\n \"acc\": 0.7709677419354839,\n\
\ \"acc_stderr\": 0.02390491431178265,\n \"acc_norm\": 0.7709677419354839,\n\
\ \"acc_norm_stderr\": 0.02390491431178265\n },\n \"harness|hendrycksTest-high_school_chemistry|5\"\
: {\n \"acc\": 0.5073891625615764,\n \"acc_stderr\": 0.035176035403610105,\n\
\ \"acc_norm\": 0.5073891625615764,\n \"acc_norm_stderr\": 0.035176035403610105\n\
\ },\n \"harness|hendrycksTest-high_school_computer_science|5\": {\n \
\ \"acc\": 0.7,\n \"acc_stderr\": 0.046056618647183814,\n \"acc_norm\"\
: 0.7,\n \"acc_norm_stderr\": 0.046056618647183814\n },\n \"harness|hendrycksTest-high_school_european_history|5\"\
: {\n \"acc\": 0.7636363636363637,\n \"acc_stderr\": 0.03317505930009181,\n\
\ \"acc_norm\": 0.7636363636363637,\n \"acc_norm_stderr\": 0.03317505930009181\n\
\ },\n \"harness|hendrycksTest-high_school_geography|5\": {\n \"acc\"\
: 0.8080808080808081,\n \"acc_stderr\": 0.028057791672989017,\n \"\
acc_norm\": 0.8080808080808081,\n \"acc_norm_stderr\": 0.028057791672989017\n\
\ },\n \"harness|hendrycksTest-high_school_government_and_politics|5\": {\n\
\ \"acc\": 0.8911917098445595,\n \"acc_stderr\": 0.022473253332768763,\n\
\ \"acc_norm\": 0.8911917098445595,\n \"acc_norm_stderr\": 0.022473253332768763\n\
\ },\n \"harness|hendrycksTest-high_school_macroeconomics|5\": {\n \
\ \"acc\": 0.658974358974359,\n \"acc_stderr\": 0.02403548967633508,\n \
\ \"acc_norm\": 0.658974358974359,\n \"acc_norm_stderr\": 0.02403548967633508\n\
\ },\n \"harness|hendrycksTest-high_school_mathematics|5\": {\n \"\
acc\": 0.34444444444444444,\n \"acc_stderr\": 0.028972648884844267,\n \
\ \"acc_norm\": 0.34444444444444444,\n \"acc_norm_stderr\": 0.028972648884844267\n\
\ },\n \"harness|hendrycksTest-high_school_microeconomics|5\": {\n \
\ \"acc\": 0.6764705882352942,\n \"acc_stderr\": 0.030388353551886793,\n\
\ \"acc_norm\": 0.6764705882352942,\n \"acc_norm_stderr\": 0.030388353551886793\n\
\ },\n \"harness|hendrycksTest-high_school_physics|5\": {\n \"acc\"\
: 0.3708609271523179,\n \"acc_stderr\": 0.03943966699183629,\n \"\
acc_norm\": 0.3708609271523179,\n \"acc_norm_stderr\": 0.03943966699183629\n\
\ },\n \"harness|hendrycksTest-high_school_psychology|5\": {\n \"acc\"\
: 0.8403669724770643,\n \"acc_stderr\": 0.015703498348461766,\n \"\
acc_norm\": 0.8403669724770643,\n \"acc_norm_stderr\": 0.015703498348461766\n\
\ },\n \"harness|hendrycksTest-high_school_statistics|5\": {\n \"acc\"\
: 0.5277777777777778,\n \"acc_stderr\": 0.0340470532865388,\n \"acc_norm\"\
: 0.5277777777777778,\n \"acc_norm_stderr\": 0.0340470532865388\n },\n\
\ \"harness|hendrycksTest-high_school_us_history|5\": {\n \"acc\": 0.8431372549019608,\n\
\ \"acc_stderr\": 0.025524722324553353,\n \"acc_norm\": 0.8431372549019608,\n\
\ \"acc_norm_stderr\": 0.025524722324553353\n },\n \"harness|hendrycksTest-high_school_world_history|5\"\
: {\n \"acc\": 0.8016877637130801,\n \"acc_stderr\": 0.02595502084162113,\n\
\ \"acc_norm\": 0.8016877637130801,\n \"acc_norm_stderr\": 0.02595502084162113\n\
\ },\n \"harness|hendrycksTest-human_aging|5\": {\n \"acc\": 0.6681614349775785,\n\
\ \"acc_stderr\": 0.03160295143776679,\n \"acc_norm\": 0.6681614349775785,\n\
\ \"acc_norm_stderr\": 0.03160295143776679\n },\n \"harness|hendrycksTest-human_sexuality|5\"\
: {\n \"acc\": 0.7862595419847328,\n \"acc_stderr\": 0.0359546161177469,\n\
\ \"acc_norm\": 0.7862595419847328,\n \"acc_norm_stderr\": 0.0359546161177469\n\
\ },\n \"harness|hendrycksTest-international_law|5\": {\n \"acc\":\
\ 0.743801652892562,\n \"acc_stderr\": 0.03984979653302872,\n \"acc_norm\"\
: 0.743801652892562,\n \"acc_norm_stderr\": 0.03984979653302872\n },\n\
\ \"harness|hendrycksTest-jurisprudence|5\": {\n \"acc\": 0.7777777777777778,\n\
\ \"acc_stderr\": 0.0401910747255735,\n \"acc_norm\": 0.7777777777777778,\n\
\ \"acc_norm_stderr\": 0.0401910747255735\n },\n \"harness|hendrycksTest-logical_fallacies|5\"\
: {\n \"acc\": 0.7668711656441718,\n \"acc_stderr\": 0.0332201579577674,\n\
\ \"acc_norm\": 0.7668711656441718,\n \"acc_norm_stderr\": 0.0332201579577674\n\
\ },\n \"harness|hendrycksTest-machine_learning|5\": {\n \"acc\": 0.42857142857142855,\n\
\ \"acc_stderr\": 0.04697113923010212,\n \"acc_norm\": 0.42857142857142855,\n\
\ \"acc_norm_stderr\": 0.04697113923010212\n },\n \"harness|hendrycksTest-management|5\"\
: {\n \"acc\": 0.7669902912621359,\n \"acc_stderr\": 0.04185832598928315,\n\
\ \"acc_norm\": 0.7669902912621359,\n \"acc_norm_stderr\": 0.04185832598928315\n\
\ },\n \"harness|hendrycksTest-marketing|5\": {\n \"acc\": 0.8760683760683761,\n\
\ \"acc_stderr\": 0.021586494001281365,\n \"acc_norm\": 0.8760683760683761,\n\
\ \"acc_norm_stderr\": 0.021586494001281365\n },\n \"harness|hendrycksTest-medical_genetics|5\"\
: {\n \"acc\": 0.7,\n \"acc_stderr\": 0.046056618647183814,\n \
\ \"acc_norm\": 0.7,\n \"acc_norm_stderr\": 0.046056618647183814\n \
\ },\n \"harness|hendrycksTest-miscellaneous|5\": {\n \"acc\": 0.8160919540229885,\n\
\ \"acc_stderr\": 0.01385372417092253,\n \"acc_norm\": 0.8160919540229885,\n\
\ \"acc_norm_stderr\": 0.01385372417092253\n },\n \"harness|hendrycksTest-moral_disputes|5\"\
: {\n \"acc\": 0.7225433526011561,\n \"acc_stderr\": 0.024105712607754307,\n\
\ \"acc_norm\": 0.7225433526011561,\n \"acc_norm_stderr\": 0.024105712607754307\n\
\ },\n \"harness|hendrycksTest-moral_scenarios|5\": {\n \"acc\": 0.4245810055865922,\n\
\ \"acc_stderr\": 0.016531170993278888,\n \"acc_norm\": 0.4245810055865922,\n\
\ \"acc_norm_stderr\": 0.016531170993278888\n },\n \"harness|hendrycksTest-nutrition|5\"\
: {\n \"acc\": 0.7156862745098039,\n \"acc_stderr\": 0.02582916327275748,\n\
\ \"acc_norm\": 0.7156862745098039,\n \"acc_norm_stderr\": 0.02582916327275748\n\
\ },\n \"harness|hendrycksTest-philosophy|5\": {\n \"acc\": 0.6945337620578779,\n\
\ \"acc_stderr\": 0.02616058445014045,\n \"acc_norm\": 0.6945337620578779,\n\
\ \"acc_norm_stderr\": 0.02616058445014045\n },\n \"harness|hendrycksTest-prehistory|5\"\
: {\n \"acc\": 0.7314814814814815,\n \"acc_stderr\": 0.024659685185967284,\n\
\ \"acc_norm\": 0.7314814814814815,\n \"acc_norm_stderr\": 0.024659685185967284\n\
\ },\n \"harness|hendrycksTest-professional_accounting|5\": {\n \"\
acc\": 0.48936170212765956,\n \"acc_stderr\": 0.02982074719142248,\n \
\ \"acc_norm\": 0.48936170212765956,\n \"acc_norm_stderr\": 0.02982074719142248\n\
\ },\n \"harness|hendrycksTest-professional_law|5\": {\n \"acc\": 0.47131681877444587,\n\
\ \"acc_stderr\": 0.012749206007657476,\n \"acc_norm\": 0.47131681877444587,\n\
\ \"acc_norm_stderr\": 0.012749206007657476\n },\n \"harness|hendrycksTest-professional_medicine|5\"\
: {\n \"acc\": 0.6838235294117647,\n \"acc_stderr\": 0.028245687391462923,\n\
\ \"acc_norm\": 0.6838235294117647,\n \"acc_norm_stderr\": 0.028245687391462923\n\
\ },\n \"harness|hendrycksTest-professional_psychology|5\": {\n \"\
acc\": 0.6797385620915033,\n \"acc_stderr\": 0.018875682938069443,\n \
\ \"acc_norm\": 0.6797385620915033,\n \"acc_norm_stderr\": 0.018875682938069443\n\
\ },\n \"harness|hendrycksTest-public_relations|5\": {\n \"acc\": 0.6727272727272727,\n\
\ \"acc_stderr\": 0.0449429086625209,\n \"acc_norm\": 0.6727272727272727,\n\
\ \"acc_norm_stderr\": 0.0449429086625209\n },\n \"harness|hendrycksTest-security_studies|5\"\
: {\n \"acc\": 0.7387755102040816,\n \"acc_stderr\": 0.02812342933514278,\n\
\ \"acc_norm\": 0.7387755102040816,\n \"acc_norm_stderr\": 0.02812342933514278\n\
\ },\n \"harness|hendrycksTest-sociology|5\": {\n \"acc\": 0.8308457711442786,\n\
\ \"acc_stderr\": 0.02650859065623327,\n \"acc_norm\": 0.8308457711442786,\n\
\ \"acc_norm_stderr\": 0.02650859065623327\n },\n \"harness|hendrycksTest-us_foreign_policy|5\"\
: {\n \"acc\": 0.86,\n \"acc_stderr\": 0.03487350880197771,\n \
\ \"acc_norm\": 0.86,\n \"acc_norm_stderr\": 0.03487350880197771\n \
\ },\n \"harness|hendrycksTest-virology|5\": {\n \"acc\": 0.5481927710843374,\n\
\ \"acc_stderr\": 0.03874371556587953,\n \"acc_norm\": 0.5481927710843374,\n\
\ \"acc_norm_stderr\": 0.03874371556587953\n },\n \"harness|hendrycksTest-world_religions|5\"\
: {\n \"acc\": 0.8187134502923976,\n \"acc_stderr\": 0.029547741687640044,\n\
\ \"acc_norm\": 0.8187134502923976,\n \"acc_norm_stderr\": 0.029547741687640044\n\
\ },\n \"harness|truthfulqa:mc|0\": {\n \"mc1\": 0.6523867809057528,\n\
\ \"mc1_stderr\": 0.01667076918889731,\n \"mc2\": 0.7984722132506933,\n\
\ \"mc2_stderr\": 0.013281143496846105\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.8516179952644041,\n \"acc_stderr\": 0.009990706005184136\n\
\ },\n \"harness|gsm8k|5\": {\n \"acc\": 0.6876421531463229,\n \
\ \"acc_stderr\": 0.012765850404191407\n }\n}\n```"
repo_url: https://huggingface.co/allknowingroger/NeuralCeptrix-7B-slerp
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|arc:challenge|25_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|gsm8k|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hellaswag|10_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-management|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-virology|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-management|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-virology|5_2024-04-10T21-14-16.828642.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-anatomy|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-astronomy|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-college_biology|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-college_physics|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-computer_security|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-econometrics|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-global_facts|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-human_aging|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-international_law|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-management|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-marketing|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-nutrition|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-philosophy|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-prehistory|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-professional_law|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-public_relations|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-security_studies|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-sociology|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-virology|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|hendrycksTest-world_religions|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|truthfulqa:mc|0_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2024-04-10T21-14-16.828642.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- '**/details_harness|winogrande|5_2024-04-10T21-14-16.828642.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2024-04-10T21-14-16.828642.parquet'
- config_name: results
data_files:
- split: 2024_04_10T21_14_16.828642
path:
- results_2024-04-10T21-14-16.828642.parquet
- split: latest
path:
- results_2024-04-10T21-14-16.828642.parquet
---
# Dataset Card for Evaluation run of allknowingroger/NeuralCeptrix-7B-slerp
<!-- Provide a quick summary of the dataset. -->
Dataset automatically created during the evaluation run of model [allknowingroger/NeuralCeptrix-7B-slerp](https://huggingface.co/allknowingroger/NeuralCeptrix-7B-slerp) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 63 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_allknowingroger__NeuralCeptrix-7B-slerp",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2024-04-10T21:14:16.828642](https://huggingface.co/datasets/open-llm-leaderboard/details_allknowingroger__NeuralCeptrix-7B-slerp/blob/main/results_2024-04-10T21-14-16.828642.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"acc": 0.6495434097360565,
"acc_stderr": 0.032279899940585416,
"acc_norm": 0.6487090853177374,
"acc_norm_stderr": 0.03296063025774512,
"mc1": 0.6523867809057528,
"mc1_stderr": 0.01667076918889731,
"mc2": 0.7984722132506933,
"mc2_stderr": 0.013281143496846105
},
"harness|arc:challenge|25": {
"acc": 0.7022184300341296,
"acc_stderr": 0.013363080107244482,
"acc_norm": 0.7209897610921502,
"acc_norm_stderr": 0.013106784883601334
},
"harness|hellaswag|10": {
"acc": 0.7197769368651663,
"acc_stderr": 0.004481902637505651,
"acc_norm": 0.8919537940649273,
"acc_norm_stderr": 0.0030980431017758477
},
"harness|hendrycksTest-abstract_algebra|5": {
"acc": 0.36,
"acc_stderr": 0.048241815132442176,
"acc_norm": 0.36,
"acc_norm_stderr": 0.048241815132442176
},
"harness|hendrycksTest-anatomy|5": {
"acc": 0.6222222222222222,
"acc_stderr": 0.04188307537595853,
"acc_norm": 0.6222222222222222,
"acc_norm_stderr": 0.04188307537595853
},
"harness|hendrycksTest-astronomy|5": {
"acc": 0.6973684210526315,
"acc_stderr": 0.03738520676119669,
"acc_norm": 0.6973684210526315,
"acc_norm_stderr": 0.03738520676119669
},
"harness|hendrycksTest-business_ethics|5": {
"acc": 0.62,
"acc_stderr": 0.048783173121456316,
"acc_norm": 0.62,
"acc_norm_stderr": 0.048783173121456316
},
"harness|hendrycksTest-clinical_knowledge|5": {
"acc": 0.6792452830188679,
"acc_stderr": 0.028727502957880267,
"acc_norm": 0.6792452830188679,
"acc_norm_stderr": 0.028727502957880267
},
"harness|hendrycksTest-college_biology|5": {
"acc": 0.7777777777777778,
"acc_stderr": 0.03476590104304134,
"acc_norm": 0.7777777777777778,
"acc_norm_stderr": 0.03476590104304134
},
"harness|hendrycksTest-college_chemistry|5": {
"acc": 0.52,
"acc_stderr": 0.050211673156867795,
"acc_norm": 0.52,
"acc_norm_stderr": 0.050211673156867795
},
"harness|hendrycksTest-college_computer_science|5": {
"acc": 0.57,
"acc_stderr": 0.04975698519562428,
"acc_norm": 0.57,
"acc_norm_stderr": 0.04975698519562428
},
"harness|hendrycksTest-college_mathematics|5": {
"acc": 0.35,
"acc_stderr": 0.04793724854411019,
"acc_norm": 0.35,
"acc_norm_stderr": 0.04793724854411019
},
"harness|hendrycksTest-college_medicine|5": {
"acc": 0.6358381502890174,
"acc_stderr": 0.03669072477416907,
"acc_norm": 0.6358381502890174,
"acc_norm_stderr": 0.03669072477416907
},
"harness|hendrycksTest-college_physics|5": {
"acc": 0.38235294117647056,
"acc_stderr": 0.04835503696107224,
"acc_norm": 0.38235294117647056,
"acc_norm_stderr": 0.04835503696107224
},
"harness|hendrycksTest-computer_security|5": {
"acc": 0.73,
"acc_stderr": 0.044619604333847394,
"acc_norm": 0.73,
"acc_norm_stderr": 0.044619604333847394
},
"harness|hendrycksTest-conceptual_physics|5": {
"acc": 0.5617021276595745,
"acc_stderr": 0.03243618636108102,
"acc_norm": 0.5617021276595745,
"acc_norm_stderr": 0.03243618636108102
},
"harness|hendrycksTest-econometrics|5": {
"acc": 0.47368421052631576,
"acc_stderr": 0.04697085136647863,
"acc_norm": 0.47368421052631576,
"acc_norm_stderr": 0.04697085136647863
},
"harness|hendrycksTest-electrical_engineering|5": {
"acc": 0.5586206896551724,
"acc_stderr": 0.04137931034482757,
"acc_norm": 0.5586206896551724,
"acc_norm_stderr": 0.04137931034482757
},
"harness|hendrycksTest-elementary_mathematics|5": {
"acc": 0.41798941798941797,
"acc_stderr": 0.025402555503260912,
"acc_norm": 0.41798941798941797,
"acc_norm_stderr": 0.025402555503260912
},
"harness|hendrycksTest-formal_logic|5": {
"acc": 0.47619047619047616,
"acc_stderr": 0.04467062628403273,
"acc_norm": 0.47619047619047616,
"acc_norm_stderr": 0.04467062628403273
},
"harness|hendrycksTest-global_facts|5": {
"acc": 0.31,
"acc_stderr": 0.04648231987117316,
"acc_norm": 0.31,
"acc_norm_stderr": 0.04648231987117316
},
"harness|hendrycksTest-high_school_biology|5": {
"acc": 0.7709677419354839,
"acc_stderr": 0.02390491431178265,
"acc_norm": 0.7709677419354839,
"acc_norm_stderr": 0.02390491431178265
},
"harness|hendrycksTest-high_school_chemistry|5": {
"acc": 0.5073891625615764,
"acc_stderr": 0.035176035403610105,
"acc_norm": 0.5073891625615764,
"acc_norm_stderr": 0.035176035403610105
},
"harness|hendrycksTest-high_school_computer_science|5": {
"acc": 0.7,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.7,
"acc_norm_stderr": 0.046056618647183814
},
"harness|hendrycksTest-high_school_european_history|5": {
"acc": 0.7636363636363637,
"acc_stderr": 0.03317505930009181,
"acc_norm": 0.7636363636363637,
"acc_norm_stderr": 0.03317505930009181
},
"harness|hendrycksTest-high_school_geography|5": {
"acc": 0.8080808080808081,
"acc_stderr": 0.028057791672989017,
"acc_norm": 0.8080808080808081,
"acc_norm_stderr": 0.028057791672989017
},
"harness|hendrycksTest-high_school_government_and_politics|5": {
"acc": 0.8911917098445595,
"acc_stderr": 0.022473253332768763,
"acc_norm": 0.8911917098445595,
"acc_norm_stderr": 0.022473253332768763
},
"harness|hendrycksTest-high_school_macroeconomics|5": {
"acc": 0.658974358974359,
"acc_stderr": 0.02403548967633508,
"acc_norm": 0.658974358974359,
"acc_norm_stderr": 0.02403548967633508
},
"harness|hendrycksTest-high_school_mathematics|5": {
"acc": 0.34444444444444444,
"acc_stderr": 0.028972648884844267,
"acc_norm": 0.34444444444444444,
"acc_norm_stderr": 0.028972648884844267
},
"harness|hendrycksTest-high_school_microeconomics|5": {
"acc": 0.6764705882352942,
"acc_stderr": 0.030388353551886793,
"acc_norm": 0.6764705882352942,
"acc_norm_stderr": 0.030388353551886793
},
"harness|hendrycksTest-high_school_physics|5": {
"acc": 0.3708609271523179,
"acc_stderr": 0.03943966699183629,
"acc_norm": 0.3708609271523179,
"acc_norm_stderr": 0.03943966699183629
},
"harness|hendrycksTest-high_school_psychology|5": {
"acc": 0.8403669724770643,
"acc_stderr": 0.015703498348461766,
"acc_norm": 0.8403669724770643,
"acc_norm_stderr": 0.015703498348461766
},
"harness|hendrycksTest-high_school_statistics|5": {
"acc": 0.5277777777777778,
"acc_stderr": 0.0340470532865388,
"acc_norm": 0.5277777777777778,
"acc_norm_stderr": 0.0340470532865388
},
"harness|hendrycksTest-high_school_us_history|5": {
"acc": 0.8431372549019608,
"acc_stderr": 0.025524722324553353,
"acc_norm": 0.8431372549019608,
"acc_norm_stderr": 0.025524722324553353
},
"harness|hendrycksTest-high_school_world_history|5": {
"acc": 0.8016877637130801,
"acc_stderr": 0.02595502084162113,
"acc_norm": 0.8016877637130801,
"acc_norm_stderr": 0.02595502084162113
},
"harness|hendrycksTest-human_aging|5": {
"acc": 0.6681614349775785,
"acc_stderr": 0.03160295143776679,
"acc_norm": 0.6681614349775785,
"acc_norm_stderr": 0.03160295143776679
},
"harness|hendrycksTest-human_sexuality|5": {
"acc": 0.7862595419847328,
"acc_stderr": 0.0359546161177469,
"acc_norm": 0.7862595419847328,
"acc_norm_stderr": 0.0359546161177469
},
"harness|hendrycksTest-international_law|5": {
"acc": 0.743801652892562,
"acc_stderr": 0.03984979653302872,
"acc_norm": 0.743801652892562,
"acc_norm_stderr": 0.03984979653302872
},
"harness|hendrycksTest-jurisprudence|5": {
"acc": 0.7777777777777778,
"acc_stderr": 0.0401910747255735,
"acc_norm": 0.7777777777777778,
"acc_norm_stderr": 0.0401910747255735
},
"harness|hendrycksTest-logical_fallacies|5": {
"acc": 0.7668711656441718,
"acc_stderr": 0.0332201579577674,
"acc_norm": 0.7668711656441718,
"acc_norm_stderr": 0.0332201579577674
},
"harness|hendrycksTest-machine_learning|5": {
"acc": 0.42857142857142855,
"acc_stderr": 0.04697113923010212,
"acc_norm": 0.42857142857142855,
"acc_norm_stderr": 0.04697113923010212
},
"harness|hendrycksTest-management|5": {
"acc": 0.7669902912621359,
"acc_stderr": 0.04185832598928315,
"acc_norm": 0.7669902912621359,
"acc_norm_stderr": 0.04185832598928315
},
"harness|hendrycksTest-marketing|5": {
"acc": 0.8760683760683761,
"acc_stderr": 0.021586494001281365,
"acc_norm": 0.8760683760683761,
"acc_norm_stderr": 0.021586494001281365
},
"harness|hendrycksTest-medical_genetics|5": {
"acc": 0.7,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.7,
"acc_norm_stderr": 0.046056618647183814
},
"harness|hendrycksTest-miscellaneous|5": {
"acc": 0.8160919540229885,
"acc_stderr": 0.01385372417092253,
"acc_norm": 0.8160919540229885,
"acc_norm_stderr": 0.01385372417092253
},
"harness|hendrycksTest-moral_disputes|5": {
"acc": 0.7225433526011561,
"acc_stderr": 0.024105712607754307,
"acc_norm": 0.7225433526011561,
"acc_norm_stderr": 0.024105712607754307
},
"harness|hendrycksTest-moral_scenarios|5": {
"acc": 0.4245810055865922,
"acc_stderr": 0.016531170993278888,
"acc_norm": 0.4245810055865922,
"acc_norm_stderr": 0.016531170993278888
},
"harness|hendrycksTest-nutrition|5": {
"acc": 0.7156862745098039,
"acc_stderr": 0.02582916327275748,
"acc_norm": 0.7156862745098039,
"acc_norm_stderr": 0.02582916327275748
},
"harness|hendrycksTest-philosophy|5": {
"acc": 0.6945337620578779,
"acc_stderr": 0.02616058445014045,
"acc_norm": 0.6945337620578779,
"acc_norm_stderr": 0.02616058445014045
},
"harness|hendrycksTest-prehistory|5": {
"acc": 0.7314814814814815,
"acc_stderr": 0.024659685185967284,
"acc_norm": 0.7314814814814815,
"acc_norm_stderr": 0.024659685185967284
},
"harness|hendrycksTest-professional_accounting|5": {
"acc": 0.48936170212765956,
"acc_stderr": 0.02982074719142248,
"acc_norm": 0.48936170212765956,
"acc_norm_stderr": 0.02982074719142248
},
"harness|hendrycksTest-professional_law|5": {
"acc": 0.47131681877444587,
"acc_stderr": 0.012749206007657476,
"acc_norm": 0.47131681877444587,
"acc_norm_stderr": 0.012749206007657476
},
"harness|hendrycksTest-professional_medicine|5": {
"acc": 0.6838235294117647,
"acc_stderr": 0.028245687391462923,
"acc_norm": 0.6838235294117647,
"acc_norm_stderr": 0.028245687391462923
},
"harness|hendrycksTest-professional_psychology|5": {
"acc": 0.6797385620915033,
"acc_stderr": 0.018875682938069443,
"acc_norm": 0.6797385620915033,
"acc_norm_stderr": 0.018875682938069443
},
"harness|hendrycksTest-public_relations|5": {
"acc": 0.6727272727272727,
"acc_stderr": 0.0449429086625209,
"acc_norm": 0.6727272727272727,
"acc_norm_stderr": 0.0449429086625209
},
"harness|hendrycksTest-security_studies|5": {
"acc": 0.7387755102040816,
"acc_stderr": 0.02812342933514278,
"acc_norm": 0.7387755102040816,
"acc_norm_stderr": 0.02812342933514278
},
"harness|hendrycksTest-sociology|5": {
"acc": 0.8308457711442786,
"acc_stderr": 0.02650859065623327,
"acc_norm": 0.8308457711442786,
"acc_norm_stderr": 0.02650859065623327
},
"harness|hendrycksTest-us_foreign_policy|5": {
"acc": 0.86,
"acc_stderr": 0.03487350880197771,
"acc_norm": 0.86,
"acc_norm_stderr": 0.03487350880197771
},
"harness|hendrycksTest-virology|5": {
"acc": 0.5481927710843374,
"acc_stderr": 0.03874371556587953,
"acc_norm": 0.5481927710843374,
"acc_norm_stderr": 0.03874371556587953
},
"harness|hendrycksTest-world_religions|5": {
"acc": 0.8187134502923976,
"acc_stderr": 0.029547741687640044,
"acc_norm": 0.8187134502923976,
"acc_norm_stderr": 0.029547741687640044
},
"harness|truthfulqa:mc|0": {
"mc1": 0.6523867809057528,
"mc1_stderr": 0.01667076918889731,
"mc2": 0.7984722132506933,
"mc2_stderr": 0.013281143496846105
},
"harness|winogrande|5": {
"acc": 0.8516179952644041,
"acc_stderr": 0.009990706005184136
},
"harness|gsm8k|5": {
"acc": 0.6876421531463229,
"acc_stderr": 0.012765850404191407
}
}
```
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] |
owanr/gpt-vs-human-annots | ---
license: apache-2.0
---
|
Vijay06nh/sample_v1 | ---
license: apache-2.0
---
|
rhaymison/orca-math-portuguese-64k | ---
dataset_info:
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: train
num_bytes: 67908462
num_examples: 64081
download_size: 24062641
dataset_size: 67908462
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
license: apache-2.0
task_categories:
- text-generation
- question-answering
language:
- pt
tags:
- math
- matematica
size_categories:
- 10K<n<100K
pretty_name: orca math portuguese
---
translated for:
<!-- Provide the basic links for the dataset. -->
- **Repository:** [microsoft/orca-math-word-problems-200k](https://huggingface.co/datasets/microsoft/orca-math-word-problems-200k)
- **Paper:** [Orca-Math: Unlocking the potential of
SLMs in Grade School Math](https://arxiv.org/pdf/2402.14830.pdf) |
asun17904/no_bank_examples | ---
dataset_info:
features:
- name: id
dtype: string
- name: url
dtype: string
- name: title
dtype: string
- name: text
dtype: string
- name: sentences
sequence: string
splits:
- name: train
num_bytes: 40311360919
num_examples: 6458670
download_size: 23461049304
dataset_size: 40311360919
---
# Dataset Card for "no_bank_examples"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
Shoubhik8/instruct_data_train | ---
dataset_info:
features:
- name: instructions
dtype: string
- name: output
dtype: string
- name: __index_level_0__
dtype: int64
splits:
- name: train
num_bytes: 219317265
num_examples: 220000
download_size: 9631975
dataset_size: 219317265
---
# Dataset Card for "instruct_data_train"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
southlemon/s-test1 | ---
license: unknown
---
|
yRodrigus/_ssalokin | ---
license: openrail
---
|
namanyash/client-3-preprocessed-platypus | ---
dataset_info:
features:
- name: instruction
dtype: string
- name: output
dtype: string
splits:
- name: train
num_bytes: 7821169
num_examples: 6000
download_size: 4319064
dataset_size: 7821169
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
|
liuyanchen1015/MULTI_VALUE_sst2_your_yalls | ---
dataset_info:
features:
- name: sentence
dtype: string
- name: label
dtype: int64
- name: idx
dtype: int64
- name: score
dtype: int64
splits:
- name: dev
num_bytes: 1813
num_examples: 14
- name: test
num_bytes: 5815
num_examples: 39
- name: train
num_bytes: 99647
num_examples: 883
download_size: 45138
dataset_size: 107275
---
# Dataset Card for "MULTI_VALUE_sst2_your_yalls"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
jan27/guanaco-llama2-1k | ---
dataset_info:
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 1654448
num_examples: 1000
download_size: 966693
dataset_size: 1654448
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "guanaco-llama2-1k"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
Yddvcodes/test_dataset | ---
license: wtfpl
task_categories:
- question-answering
language:
- en
pretty_name: testy
--- |
liuyanchen1015/MULTI_VALUE_rte_possessives_for_post | ---
dataset_info:
features:
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype: int64
- name: idx
dtype: int64
- name: value_score
dtype: int64
splits:
- name: test
num_bytes: 617408
num_examples: 1492
- name: train
num_bytes: 554538
num_examples: 1311
download_size: 754396
dataset_size: 1171946
---
# Dataset Card for "MULTI_VALUE_rte_possessives_for_post"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
BPHoops/futurama | ---
license: apache-2.0
---
|
abideen/pretrain_corpus | ---
dataset_info:
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 199296832162.51273
num_examples: 65123248
- name: validation
num_bytes: 35170032625.48728
num_examples: 11492339
download_size: 136513929683
dataset_size: 234466864788.0
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
---
|
Karim-Gamal/SemEval-2018-Task-2-english-emojis | ---
license: apache-2.0
---
|
nikchar/20k_claims_train_final | ---
dataset_info:
features:
- name: claim
dtype: string
- name: text
dtype: string
- name: label
dtype: int64
splits:
- name: train
num_bytes: 30738751.0
num_examples: 19998
download_size: 17098290
dataset_size: 30738751.0
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "20k_claims_train_final"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
medalpaca/medical_meadow_pubmed_causal | ---
task_categories:
- question-answering
language:
- en
---
# Dataset Card for Pubmed Causal
## Dataset Description
- **Paper:** https://aclanthology.org/D19-1473/
### Dataset Summary
This is the dataset used in the paper: Detecting Causal Language Use in Science Findings.
### Citation Information
```
@inproceedings{yu-etal-2019-detecting,
title = "Detecting Causal Language Use in Science Findings",
author = "Yu, Bei and
Li, Yingya and
Wang, Jun",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)",
month = nov,
year = "2019",
address = "Hong Kong, China",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/D19-1473",
doi = "10.18653/v1/D19-1473",
pages = "4664--4674",
}
``` |
edumunozsala/Bactrian-X-instruction-es | ---
dataset_info:
features:
- name: instruction
dtype: string
- name: input
dtype: string
- name: id
dtype: string
- name: output
dtype: string
- name: text
dtype: string
splits:
- name: train
num_bytes: 122397731
num_examples: 67017
download_size: 63536232
dataset_size: 122397731
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
|
kichanj/customchan | ---
dataset_info:
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 4790
num_examples: 29
download_size: 3544
dataset_size: 4790
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
|
StofEzz/data1000_preprocessed | ---
configs:
- config_name: default
data_files:
- split: test
path: data/test-*
- split: train
path: data/train-*
- split: validation
path: data/validation-*
dataset_info:
features:
- name: input_values
sequence: float32
- name: labels
sequence: int64
splits:
- name: test
num_bytes: 33556448
num_examples: 100
- name: train
num_bytes: 287053448
num_examples: 1000
- name: validation
num_bytes: 30626216
num_examples: 100
download_size: 350661232
dataset_size: 351236112
---
# Dataset Card for "data1000_preprocessed"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
NobodyExistsOnTheInternet/10kexamplesample | ---
license: other
---
|
maywell/ELLL_sorted | ---
dataset_info:
features:
- name: question
dtype: string
- name: answer
dtype: string
- name: eng_percentage
dtype: float64
- name: question_token_count
dtype: int64
- name: answer_token_count
dtype: int64
- name: __index_level_0__
dtype: int64
splits:
- name: train
num_bytes: 116961414
num_examples: 61587
download_size: 73312742
dataset_size: 116961414
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "ELLL_sorted"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
svjack/diffusiondb_random_10k_zh_v1 | ---
annotations_creators:
- machine-generated
language:
- en
- zh
language_creators:
- other
multilinguality:
- multilingual
pretty_name: 'Pokémon BLIP captions'
size_categories:
- 10K
dataset_info:
features:
- name: image
dtype: image
- name: prompt
dtype: string
- name: seed
dtype: int64
- name: step
dtype: int64
- name: cfg
dtype: float32
- name: sampler
dtype: string
- name: zh_prompt
dtype: string
splits:
- name: train
num_bytes: 5826763233.4353
num_examples: 9841
download_size: 5829710525
dataset_size: 5826763233.4353
---
# Dataset Card for "diffusiondb_random_10k_zh_v1"
svjack/diffusiondb_random_10k_zh_v1 is a dataset that random sample 10k English samples from [diffusiondb](https://github.com/poloclub/diffusiondb) and use [NMT](https://en.wikipedia.org/wiki/Neural_machine_translation) translate them into Chinese with some corrections.<br/>
it used to train stable diffusion models in <br/> [svjack/Stable-Diffusion-FineTuned-zh-v0](https://huggingface.co/svjack/Stable-Diffusion-FineTuned-zh-v0)<br/>
[svjack/Stable-Diffusion-FineTuned-zh-v1](https://huggingface.co/svjack/Stable-Diffusion-FineTuned-zh-v1)<br/>
[svjack/Stable-Diffusion-FineTuned-zh-v2](https://huggingface.co/svjack/Stable-Diffusion-FineTuned-zh-v2)<br/>
And is the data support of [https://github.com/svjack/Stable-Diffusion-Chinese-Extend](https://github.com/svjack/Stable-Diffusion-Chinese-Extend) which is a fine tune version of Stable Diffusion model on self-translate 10k diffusiondb Chinese Corpus and "extend" it.
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
roborovski/glaive-tool-usage-dpo | ---
dataset_info:
features:
- name: system
dtype: string
- name: question
dtype: string
- name: chosen
dtype: string
- name: rejected
dtype: string
splits:
- name: train
num_bytes: 60460164
num_examples: 42018
download_size: 25338925
dataset_size: 60460164
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
|
nirantk/nli-triplets | ---
dataset_info:
features:
- name: anchor
dtype: string
- name: entailment
dtype: string
- name: contradiction
dtype: string
- name: metadata
struct:
- name: objective
struct:
- name: paired
sequence:
sequence: string
- name: self
sequence: 'null'
- name: triplet
sequence:
sequence: string
- name: topic
dtype: string
splits:
- name: train
num_bytes: 79858702
num_examples: 277122
download_size: 31073286
dataset_size: 79858702
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
|
AdapterOcean/data-standardized_cluster_24 | ---
dataset_info:
features:
- name: text
dtype: string
- name: conversation_id
dtype: int64
- name: embedding
sequence: float64
- name: cluster
dtype: int64
splits:
- name: train
num_bytes: 87435740
num_examples: 8434
download_size: 25087384
dataset_size: 87435740
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "data-standardized_cluster_24"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
mbeaty2/exoplanet-data | ---
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': data_huggingface
- name: text
dtype: string
splits:
- name: train
num_bytes: 2965375.0
num_examples: 127
download_size: 1537208
dataset_size: 2965375.0
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "exoplanet-data"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
bigscience-data/roots_en_royal_society_corpus | ---
language: en
license: cc-by-nc-sa-4.0
extra_gated_prompt: 'By accessing this dataset, you agree to abide by the BigScience
Ethical Charter. The charter can be found at:
https://hf.co/spaces/bigscience/ethical-charter'
extra_gated_fields:
I have read and agree to abide by the BigScience Ethical Charter: checkbox
---
ROOTS Subset: roots_en_royal_society_corpus
# Royal Society Corpus
- Dataset uid: `royal_society_corpus`
### Description
The Royal Society Corpus (RSC) 6.0 Open is based on the first centuries of the Philosophical Transactions of the Royal Society of London from its beginning in 1665 to 1920. It includes all publications of the journal written in English and containing running text. The Philosophical Transactions was the first periodical of scientific writing in England. Founded in 1665 by Henry Oldenburg, the first secretary of the Royal Society, it initially contained excerpts of letters of his scientific correspondence, reviews and summaries of recently-published books, and accounts of observations and experiments.
### Homepage
https://fedora.clarin-d.uni-saarland.de/rsc_v6/index.html
### Licensing
- public domain
- cc0-1.0: Creative Commons Zero v1.0 Universal
### Speaker Locations
- Northern Europe
- United Kingdom
### Sizes
- 0.0334 % of total
- 0.1808 % of en
### BigScience processing steps
#### Filters applied to: en
- dedup_document
- dedup_template_soft
- filter_remove_empty_docs
- filter_small_docs_bytes_1024
|
wanng/wukong100m | ---
annotations_creators:
- crowdsourced
language_creators:
- crowdsourced
language:
- zh
license:
- cc-by-nc-sa-4.0
multilinguality:
- monolingual
pretty_name: Wukong100M
task_categories:
- feature-extraction
---
# wukong100m
## 简介 Brief Introduction
取自Noah-Wukong多语言多模态数据集中的中文部分,一共100M个图文对。
A subset from Noah-Wukong (a multimodal dataset), around 100M image-text pairs (only Chinese).
## 数据集信息 Dataset Information
大约一共100M个中文图文对。大约占用16GB空间(仅仅是url等文本信息,不包含图片)。下载成功率在80%左右。(虽然我没有统计下载之后会占用多少空间,但是,可以说非常非常大)
- Homepage: [Noah-Wukong](https://wukong-dataset.github.io/wukong-dataset/index.html)
## 下载 Download
```bash
mkdir wukong100m && cd wukong100m
for i in {00000..00031}; do wget https://huggingface.co/datasets/wanng/wukong100m/resolve/main/data/train-$i-of-00032.parquet; done
cd ..
```
## Lisence
CC BY-NC-SA 4.0
|
nglaura/scielo-summarization | ---
license: apache-2.0
task_categories:
- summarization
language:
- fr
pretty_name: SciELO
---
# LoRaLay: A Multilingual and Multimodal Dataset for Long Range and Layout-Aware Summarization
A collaboration between [reciTAL](https://recital.ai/en/), [MLIA](https://mlia.lip6.fr/) (ISIR, Sorbonne Université), [Meta AI](https://ai.facebook.com/), and [Università di Trento](https://www.unitn.it/)
## SciELO dataset for summarization
SciELO is a dataset for summarization of research papers written in Spanish and Portuguese, for which layout information is provided.
### Data Fields
- `article_id`: article id
- `article_words`: sequence of words constituting the body of the article
- `article_bboxes`: sequence of corresponding word bounding boxes
- `norm_article_bboxes`: sequence of corresponding normalized word bounding boxes
- `abstract`: a string containing the abstract of the article
- `article_pdf_url`: URL of the article's PDF
### Data Splits
This dataset has 3 splits: _train_, _validation_, and _test_.
| Dataset Split | Number of Instances (ES/PT) |
| ------------- | ----------------------------|
| Train | 20,853 / 19,407 |
| Validation | 1,158 / 1,078 |
| Test | 1,159 / 1,078 |
## Citation
``` latex
@article{nguyen2023loralay,
title={LoRaLay: A Multilingual and Multimodal Dataset for Long Range and Layout-Aware Summarization},
author={Nguyen, Laura and Scialom, Thomas and Piwowarski, Benjamin and Staiano, Jacopo},
journal={arXiv preprint arXiv:2301.11312},
year={2023}
}
``` |
lucasmccabe/logiqa | ---
task_categories:
- question-answering
language:
- en
pretty_name: LogiQA
size_categories:
- 1K<n<10K
paperswithcode_id: logiqa
dataset_info:
features:
- name: context
dtype: string
- name: query
dtype: string
- name: options
sequence:
dtype: string
- name: correct_option
dtype: string
splits:
- name: train
num_examples: 7376
- name: validation
num_examples: 651
- name: test
num_examples: 651
---
# Dataset Card for LogiQA
## Dataset Description
- **Homepage:**
- **Repository:**
- **Paper:**
- **Leaderboard:**
- **Point of Contact:**
### Dataset Summary
LogiQA is constructed from the logical comprehension problems from publically available questions of the National Civil Servants Examination of China, which are designed to test the civil servant candidates’ critical thinking and problem solving. This dataset includes the English versions only; the Chinese versions are available via the homepage/original source.
## Dataset Structure
### Data Instances
An example from `train` looks as follows:
```
{'context': 'Continuous exposure to indoor fluorescent lights is beneficial to the health of hamsters with heart disease. One group of hamsters exposed to continuous exposure to fluorescent lights has an average lifespan that is 2.5% longer than another one of the same species but living in a black wall.',
'query': 'Which of the following questions was the initial motivation for conducting the above experiment?',
'options': ['Can hospital light therapy be proved to promote patient recovery?',
'Which one lives longer, the hamster living under the light or the hamster living in the dark?',
'What kind of illness does the hamster have?',
'Do some hamsters need a period of darkness?'],
'correct_option': 0}
```
### Data Fields
- `context`: a `string` feature.
- `query`: a `string` feature.
- `answers`: a `list` feature containing `string` features.
- `correct_option`: a `string` feature.
### Data Splits
|train|validation|test|
|----:|---------:|---:|
| 7376| 651| 651|
## Additional Information
### Dataset Curators
The original LogiQA was produced by Jian Liu, Leyang Cui , Hanmeng Liu, Dandan Huang, Yile Wang, and Yue Zhang.
### Licensing Information
[More Information Needed]
### Citation Information
```
@article{liu2020logiqa,
title={Logiqa: A challenge dataset for machine reading comprehension with logical reasoning},
author={Liu, Jian and Cui, Leyang and Liu, Hanmeng and Huang, Dandan and Wang, Yile and Zhang, Yue},
journal={arXiv preprint arXiv:2007.08124},
year={2020}
}
```
### Contributions
[@lucasmccabe](https://github.com/lucasmccabe) added this dataset. |
ibranze/araproje_hellaswag_en_conf_mgpt_nearestscore_true_y | ---
dataset_info:
features:
- name: ind
dtype: int32
- name: activity_label
dtype: string
- name: ctx_a
dtype: string
- name: ctx_b
dtype: string
- name: ctx
dtype: string
- name: endings
sequence: string
- name: source_id
dtype: string
- name: split
dtype: string
- name: split_type
dtype: string
- name: label
dtype: string
splits:
- name: validation
num_bytes: 149738.0
num_examples: 250
download_size: 81214
dataset_size: 149738.0
configs:
- config_name: default
data_files:
- split: validation
path: data/validation-*
---
# Dataset Card for "araproje_hellaswag_en_conf_mgpt_nearestscore_true_y"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
jjonhwa/raw3_v1 | ---
dataset_info:
features:
- name: context
dtype: string
- name: is_impossible
dtype: bool
- name: question
dtype: string
- name: answer
dtype: string
- name: answer_start
dtype: int64
splits:
- name: train
num_bytes: 131956852
num_examples: 51888
download_size: 41299212
dataset_size: 131956852
---
# Dataset Card for "raw3_v1"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
Hiltoon/gabriel | ---
license: openrail
---
|
jeremyc/Alpaca-Lora-GPT4-Swedish | ---
language:
- sv
pretty_name: Alpaca-Lora GPT4 Swedish
size_categories:
- 10K<n<100K
---
This dataset is the machine translation of the GPT4 dataset provided on Alpaca-Lora github repository.
We provide two version: The full translation, and a translation of a subset of ~50 000 entries that was cleaned and do not contain instances of "I am an AI language model" or similar.
This work was inspired from the French alpaca lora variant **Vigogne** and the Ukrainian alpaca lora variante **Kruk**. |
Rogashbabu/Training-Bert-Model-Analysis | ---
license: other
task_categories:
- text-classification
language:
- en
--- |
feilongfl/ChineseNewsSummary | ---
license: apache-2.0
---
generate by chatgpt3.5 with prompt:
``` text
你是一个用于整理新闻的AI助手,请根据分割线后爬虫在{{ $json.isoDate }}爬取的新闻,使用简体中文按照包含:title,summary,class,time四个节点的json格式输出结果。
title字段:请为新闻起一个30到60个字的简体中文标题。应当是包含上下文简洁说明的结论性内容的陈述句,词汇简单,信息全面。
class:请分类为:财经、汽车、房产、家居、教育、科技、社会、时政、体育、游戏、娱乐等。
time字段:如果新闻中有提供时间,请使用新闻中的时间;否则使用爬虫时间。时间应当以年/月/日格式输出,例如:2024/1/16。
summary字段:应当用100字以内简单的简体中文陈述已经发生的事实,着重于结论和支撑结论的数据,不要假设和预测,不要重复标题。请过滤原文中可能包含的问题或反问、猜测/情绪化表达、政治口号、联系方式、股票代码或广告词。如果原文有晦涩或不常见的词汇或多重否定,请换用简单的词语进行描述。
---
{{ $json.contentSnippet }}
```
|
mike-ravkine/can-ai-code_junior-dev_v1 | ---
language:
- en
---
First version of the [can-ai-code](https://github.com/the-crypt-keeper/can-ai-code/) junior-dev interview. |
Cheetor1996/Sayoko_Kagami | ---
license: cc-by-2.0
language:
- en
tags:
- art
---
**Sayoko Kagami** from Helter Skelter Hakudaku no Mura
- *Trained with anime (full-final-pruned) model*
- *Works the best with ALL, MIDD, OUTD, and OUTALL LoRA weight block, and with 0.5+ weights.* |
pkr7098/gpt2-small-bookcorpus-wiki-2022030-en-vocab_size-32000 | ---
dataset_info:
config_name: truncate-1024
features:
- name: input_ids
sequence: int32
- name: attention_mask
sequence: int8
- name: special_tokens_mask
sequence: int8
splits:
- name: train
num_bytes: 22599642492
num_examples: 3671157
- name: validation
num_bytes: 397295928
num_examples: 64538
download_size: 0
dataset_size: 22996938420
configs:
- config_name: truncate-1024
data_files:
- split: train
path: truncate-1024/train-*
- split: validation
path: truncate-1024/validation-*
---
# Dataset Card for "gpt2-small-bookcorpus-wiki-2022030-en-vocab_size-32000"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
yuvalkirstain/pexel | ---
dataset_info:
features:
- name: image
dtype: image
- name: text
dtype: string
splits:
- name: train
num_bytes: 653318190.0
num_examples: 1560
- name: validation
num_bytes: 7122908.0
num_examples: 20
download_size: 653521442
dataset_size: 660441098.0
---
# Dataset Card for "pexel"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
Femur19899079/Fat | ---
license: lgpl-3.0
---
|
ashutosh09/samplecoredge | ---
license: mit
---
|
open-llm-leaderboard/details_Sao10K__Stheno-1.8-L2-13B | ---
pretty_name: Evaluation run of Sao10K/Stheno-1.8-L2-13B
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [Sao10K/Stheno-1.8-L2-13B](https://huggingface.co/Sao10K/Stheno-1.8-L2-13B) on\
\ the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 64 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 2 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the agregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_Sao10K__Stheno-1.8-L2-13B\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-10-29T00:06:59.711419](https://huggingface.co/datasets/open-llm-leaderboard/details_Sao10K__Stheno-1.8-L2-13B/blob/main/results_2023-10-29T00-06-59.711419.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"em\": 0.06564597315436242,\n\
\ \"em_stderr\": 0.002536293026031132,\n \"f1\": 0.15330012583892566,\n\
\ \"f1_stderr\": 0.0029061084519907457,\n \"acc\": 0.44834287054661603,\n\
\ \"acc_stderr\": 0.01063887899085892\n },\n \"harness|drop|3\": {\n\
\ \"em\": 0.06564597315436242,\n \"em_stderr\": 0.002536293026031132,\n\
\ \"f1\": 0.15330012583892566,\n \"f1_stderr\": 0.0029061084519907457\n\
\ },\n \"harness|gsm8k|5\": {\n \"acc\": 0.1326762699014405,\n \
\ \"acc_stderr\": 0.009343929131442216\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.7640094711917916,\n \"acc_stderr\": 0.011933828850275625\n\
\ }\n}\n```"
repo_url: https://huggingface.co/Sao10K/Stheno-1.8-L2-13B
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|arc:challenge|25_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_drop_3
data_files:
- split: 2023_10_29T00_06_59.711419
path:
- '**/details_harness|drop|3_2023-10-29T00-06-59.711419.parquet'
- split: latest
path:
- '**/details_harness|drop|3_2023-10-29T00-06-59.711419.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_10_29T00_06_59.711419
path:
- '**/details_harness|gsm8k|5_2023-10-29T00-06-59.711419.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-10-29T00-06-59.711419.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hellaswag|10_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-10-04T00-45-37.800224.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-management|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-virology|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- '**/details_harness|truthfulqa:mc|0_2023-10-04T00-45-37.800224.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2023-10-04T00-45-37.800224.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_10_29T00_06_59.711419
path:
- '**/details_harness|winogrande|5_2023-10-29T00-06-59.711419.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-10-29T00-06-59.711419.parquet'
- config_name: results
data_files:
- split: 2023_10_04T00_45_37.800224
path:
- results_2023-10-04T00-45-37.800224.parquet
- split: 2023_10_29T00_06_59.711419
path:
- results_2023-10-29T00-06-59.711419.parquet
- split: latest
path:
- results_2023-10-29T00-06-59.711419.parquet
---
# Dataset Card for Evaluation run of Sao10K/Stheno-1.8-L2-13B
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/Sao10K/Stheno-1.8-L2-13B
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [Sao10K/Stheno-1.8-L2-13B](https://huggingface.co/Sao10K/Stheno-1.8-L2-13B) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 64 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 2 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_Sao10K__Stheno-1.8-L2-13B",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-10-29T00:06:59.711419](https://huggingface.co/datasets/open-llm-leaderboard/details_Sao10K__Stheno-1.8-L2-13B/blob/main/results_2023-10-29T00-06-59.711419.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"em": 0.06564597315436242,
"em_stderr": 0.002536293026031132,
"f1": 0.15330012583892566,
"f1_stderr": 0.0029061084519907457,
"acc": 0.44834287054661603,
"acc_stderr": 0.01063887899085892
},
"harness|drop|3": {
"em": 0.06564597315436242,
"em_stderr": 0.002536293026031132,
"f1": 0.15330012583892566,
"f1_stderr": 0.0029061084519907457
},
"harness|gsm8k|5": {
"acc": 0.1326762699014405,
"acc_stderr": 0.009343929131442216
},
"harness|winogrande|5": {
"acc": 0.7640094711917916,
"acc_stderr": 0.011933828850275625
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] |
CyberHarem/aihara_yuzu_citrus | ---
license: mit
task_categories:
- text-to-image
tags:
- art
- not-for-all-audiences
size_categories:
- n<1K
---
# Dataset of Aihara Yuzu
This is the dataset of Aihara Yuzu, containing 300 images and their tags.
Images are crawled from many sites (e.g. danbooru, pixiv, zerochan ...), the auto-crawling system is powered by [DeepGHS Team](https://github.com/deepghs)([huggingface organization](https://huggingface.co/deepghs)).
| Name | Images | Download | Description |
|:----------------|---------:|:----------------------------------------|:-----------------------------------------------------------------------------------------|
| raw | 300 | [Download](dataset-raw.zip) | Raw data with meta information. |
| raw-stage3 | 691 | [Download](dataset-raw-stage3.zip) | 3-stage cropped raw data with meta information. |
| raw-stage3-eyes | 891 | [Download](dataset-raw-stage3-eyes.zip) | 3-stage cropped (with eye-focus) raw data with meta information. |
| 384x512 | 300 | [Download](dataset-384x512.zip) | 384x512 aligned dataset. |
| 512x704 | 300 | [Download](dataset-512x704.zip) | 512x704 aligned dataset. |
| 640x880 | 300 | [Download](dataset-640x880.zip) | 640x880 aligned dataset. |
| stage3-640 | 691 | [Download](dataset-stage3-640.zip) | 3-stage cropped dataset with the shorter side not exceeding 640 pixels. |
| stage3-800 | 691 | [Download](dataset-stage3-800.zip) | 3-stage cropped dataset with the shorter side not exceeding 800 pixels. |
| stage3-p512-640 | 599 | [Download](dataset-stage3-p512-640.zip) | 3-stage cropped dataset with the area not less than 512x512 pixels. |
| stage3-eyes-640 | 891 | [Download](dataset-stage3-eyes-640.zip) | 3-stage cropped (with eye-focus) dataset with the shorter side not exceeding 640 pixels. |
| stage3-eyes-800 | 891 | [Download](dataset-stage3-eyes-800.zip) | 3-stage cropped (with eye-focus) dataset with the shorter side not exceeding 800 pixels. |
|
chekahchek/ai6127-ghostbuster | ---
dataset_info:
features:
- name: texts
dtype: string
- name: labels
dtype: string
- name: domains
dtype: string
splits:
- name: train
num_bytes: 13188360
num_examples: 3837
- name: val
num_bytes: 3251702
num_examples: 963
- name: test
num_bytes: 4093967
num_examples: 1200
download_size: 11526069
dataset_size: 20534029
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: val
path: data/val-*
- split: test
path: data/test-*
---
|
Shamus/Medical_Speech_Transcription_and_Intent | ---
language:
- en
size_categories:
- 1K<n<10K
---
This dataset came from Kaggle and was contributed by Paul Mooney.
https://www.kaggle.com/datasets/paultimothymooney/medical-speech-transcription-and-intent/data
Context
8.5 hours of audio utterances paired with text for common medical symptoms.
Content
This data contains thousands of audio utterances for common medical symptoms like “knee pain” or “headache,” totaling more than 8 hours in aggregate. Each utterance was created by individual human contributors based on a given symptom. These audio snippets can be used to train conversational agents in the medical field.
This Figure Eight dataset was created via a multi-job workflow. The first involved contributors writing text phrases to describe symptoms given. For example, for “headache,” a contributor might write “I need help with my migraines.” Subsequent jobs captured audio utterances for accepted text strings.
Note that some of the labels are incorrect and some of the audio files have poor quality. I would recommend cleaning the dataset before training any machine learning models.
This dataset contains both the audio utterances and corresponding transcriptions.
Acknowledgements
This dataset was developed by figure-eight and can be downloaded from https://www.figure-eight.com/dataset/audio-recording-and-transcription-for-medical-scenarios/ along with instructions on how to make similar datasets using the figure-eight platform.
https://www.figure-eight.com/dataset/audio-recording-and-transcription-for-medical-scenarios/
Banner Photo by rawpixel on Unsplash |
japanese-asr/whisper_transcriptions.reazonspeech.all_35 | ---
dataset_info:
config_name: all
features:
- name: name
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 16000
- name: transcription
dtype: string
- name: whisper_transcript
sequence: int64
splits:
- name: train
num_bytes: 30456979591.0
num_examples: 267182
download_size: 30215181472
dataset_size: 30456979591.0
configs:
- config_name: all
data_files:
- split: train
path: all/train-*
---
|
UserNae3/FLIR_aligned | ---
license: other
license_name: other
license_link: https://www.flir.com/oem/adas/adas-dataset-form/
---
# 对齐的 FLIR 数据集
[[中文](README.md)] [[English](README_en.md)]
FLIR 数据集由 FLIR 公司发布,双光谱目标检测数据集一般是使用来自 Zhang等人的修改的版本[Paper](https://arxiv.org/abs/2009.12664), [Dataset](https://drive.google.com/file/d/1xHDMGl6HJZwtarNWkEV3T4O9X4ZQYz2Y/view)
> 这里存储了[[张等人对齐的FLIR数据集](https://arxiv.org/abs/2009.12664)]的备份 和 其 [[转换成COCO标注格式的数据集](https://huggingface.co/datasets/UserNae3/FLIR_aligned/resolve/main/flir_align.7z)]
**下载**
- 【推荐】转换成的COCO标注格式:https://huggingface.co/datasets/UserNae3/FLIR_aligned/resolve/main/flir_align.7z?download=true
- Zhang等人原始发布下载:https://drive.google.com/file/d/1xHDMGl6HJZwtarNWkEV3T4O9X4ZQYz2Y/view
- 对Zhang等人发布的备份:https://huggingface.co/datasets/UserNae3/FLIR_aligned/resolve/main/aligned.zip?download=true
**版权**
版权链接: https://www.flir.com/oem/adas/adas-dataset-form/ |
LCA-PORVID/delexicalized_n_grams | ---
dataset_info:
- config_name: journalistic
features:
- name: text
dtype: string
- name: label
dtype: int64
splits:
- name: train
num_bytes: 201756790
num_examples: 615012
download_size: 123762232
dataset_size: 201756790
- config_name: legal
features:
- name: text
dtype: string
- name: label
dtype: int64
splits:
- name: train
num_bytes: 1215937
num_examples: 7440
download_size: 687763
dataset_size: 1215937
- config_name: literature
features:
- name: text
dtype: string
- name: label
dtype: int64
splits:
- name: train
num_bytes: 8812642
num_examples: 40878
download_size: 5541660
dataset_size: 8812642
- config_name: politics
features:
- name: text
dtype: string
- name: label
dtype: int64
splits:
- name: train
num_bytes: 4082098
num_examples: 5778
download_size: 2340087
dataset_size: 4082098
- config_name: social_media
features:
- name: text
dtype: string
- name: label
dtype: int64
splits:
- name: train
num_bytes: 564540
num_examples: 8246
download_size: 340777
dataset_size: 564540
- config_name: web
features:
- name: text
dtype: string
- name: label
dtype: int64
splits:
- name: train
num_bytes: 85208776
num_examples: 79318
download_size: 49328021
dataset_size: 85208776
configs:
- config_name: journalistic
data_files:
- split: train
path: journalistic/train-*
- config_name: legal
data_files:
- split: train
path: legal/train-*
- config_name: literature
data_files:
- split: train
path: literature/train-*
- config_name: politics
data_files:
- split: train
path: politics/train-*
- config_name: social_media
data_files:
- split: train
path: social_media/train-*
- config_name: web
data_files:
- split: train
path: web/train-*
---
|
varcoder/Semantic_Segmentation_CE | ---
dataset_info:
features:
- name: pixel_values
dtype: image
- name: label
dtype: image
splits:
- name: train
num_bytes: 27784142.0
num_examples: 300
- name: test
num_bytes: 90363667.0
num_examples: 237
download_size: 0
dataset_size: 118147809.0
---
# Dataset Card for "Semantic_Segmentation_CE"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.