id stringlengths 2 115 | author stringlengths 2 42 ⌀ | last_modified timestamp[us, tz=UTC] | downloads int64 0 8.87M | likes int64 0 3.84k | paperswithcode_id stringlengths 2 45 ⌀ | tags list | lastModified timestamp[us, tz=UTC] | createdAt stringlengths 24 24 | key stringclasses 1 value | created timestamp[us] | card stringlengths 1 1.01M | embedding list | library_name stringclasses 21 values | pipeline_tag stringclasses 27 values | mask_token null | card_data null | widget_data null | model_index null | config null | transformers_info null | spaces null | safetensors null | transformersInfo null | modelId stringlengths 5 111 ⌀ | embeddings list |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
ugshanyu/new_data | ugshanyu | 2023-11-09T10:52:44Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T10:52:44Z | 2023-11-09T09:15:58.000Z | 2023-11-09T09:15:58 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
NghiemAbe/translation-vietnamese-english | NghiemAbe | 2023-11-09T09:34:40Z | 0 | 0 | null | [
"task_categories:translation",
"size_categories:100M<n<1B",
"language:vi",
"language:en",
"license:mit",
"region:us"
] | 2023-11-09T09:34:40Z | 2023-11-09T09:22:40.000Z | 2023-11-09T09:22:40 | ---
license: mit
task_categories:
- translation
language:
- vi
- en
size_categories:
- 100M<n<1B
---
Test data: PhoMT
Train data: PhoMT (filter len between 40 to 100) | [
-0.9889453649520874,
-0.4709046483039856,
0.4053713083267212,
0.416940301656723,
-0.8030042052268982,
-0.346955806016922,
-0.24646665155887604,
0.3127727806568146,
0.06776068359613419,
0.7702150344848633,
-0.610602617263794,
-0.30613383650779724,
-0.5105075836181641,
0.062325943261384964,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
tinnghuynh/procd-vie-speech-corpus | tinnghuynh | 2023-11-09T09:32:17Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T09:32:17Z | 2023-11-09T09:30:30.000Z | 2023-11-09T09:30:30 | ---
dataset_info:
features:
- name: input_ids
sequence: int32
- name: labels
sequence:
sequence: float32
- name: speaker_embeddings
sequence: float32
splits:
- name: train
num_bytes: 2679530792
num_examples: 22884
download_size: 2484517887
dataset_size: 2679530792
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "procd-vie-speech-corpus"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.6571577787399292,
-0.49413183331489563,
-0.0034883683547377586,
0.1886088252067566,
-0.13075710833072662,
0.04843626916408539,
-0.002569471951574087,
-0.08733372390270233,
0.9109706282615662,
0.6116182208061218,
-0.6700118184089661,
-0.8109080195426941,
-0.6091322302818298,
-0.410088956... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ameerazam08/paper-exp | ameerazam08 | 2023-11-09T09:46:52Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T09:46:52Z | 2023-11-09T09:39:11.000Z | 2023-11-09T09:39:11 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
razat-ag/test | razat-ag | 2023-11-09T10:11:57Z | 0 | 0 | null | [
"license:apache-2.0",
"region:us"
] | 2023-11-09T10:11:57Z | 2023-11-09T09:52:07.000Z | 2023-11-09T09:52:07 | ---
license: apache-2.0
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Sevll/Momo_Momone | Sevll | 2023-11-09T09:59:55Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T09:59:55Z | 2023-11-09T09:59:55.000Z | 2023-11-09T09:59:55 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
roszcz/giant-midi-sustain-v2 | roszcz | 2023-11-09T17:14:18Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T17:14:18Z | 2023-11-09T10:00:02.000Z | 2023-11-09T10:00:02 | ---
dataset_info:
features:
- name: notes
struct:
- name: duration
sequence: float64
- name: end
sequence: float64
- name: pitch
sequence: int64
- name: start
sequence: float64
- name: velocity
sequence: int64
- name: source
struct:
- name: artist
dtype: string
- name: dataset
dtype: string
- name: title
dtype: string
- name: youtube_id
dtype: string
splits:
- name: train
num_bytes: 1548965873
num_examples: 10853
download_size: 483522534
dataset_size: 1548965873
---
# Dataset Card for "giant-midi-sustain-v2"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.571007251739502,
-0.2129603624343872,
0.30461418628692627,
0.3420429229736328,
-0.2073623687028885,
-0.023701246827840805,
0.1674223691225052,
-0.26964715123176575,
0.9338016510009766,
0.4763435125350952,
-0.8999829292297363,
-0.4731869101524353,
-0.43452075123786926,
-0.427520632743835... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
alvarobartt/ultrafeedback-enable-checkpoint-100 | alvarobartt | 2023-11-09T10:06:20Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T10:06:20Z | 2023-11-09T10:06:16.000Z | 2023-11-09T10:06:16 | ---
dataset_info:
features:
- name: input
dtype: string
- name: generation_prompt
dtype: string
- name: raw_generation_responses
sequence: string
- name: generations
sequence: string
- name: labelling_prompt
list:
- name: content
dtype: string
- name: role
dtype: string
- name: raw_labelling_response
struct:
- name: choices
list:
- name: finish_reason
dtype: string
- name: index
dtype: int64
- name: message
struct:
- name: content
dtype: string
- name: role
dtype: string
- name: created
dtype: int64
- name: id
dtype: string
- name: model
dtype: string
- name: object
dtype: string
- name: usage
struct:
- name: completion_tokens
dtype: int64
- name: prompt_tokens
dtype: int64
- name: total_tokens
dtype: int64
- name: rating
sequence: int64
- name: rationale
sequence: string
splits:
- name: train
num_bytes: 926832
num_examples: 100
download_size: 391509
dataset_size: 926832
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "ultrafeedback-enable-checkpoint-100"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5919630527496338,
-0.16257400810718536,
0.35854625701904297,
0.4011992812156677,
-0.08838657289743423,
0.044310372322797775,
0.5242874622344971,
0.03298192098736763,
0.9150717258453369,
0.6379848718643188,
-1.078258752822876,
-0.5951421856880188,
-0.056303635239601135,
-0.22805766761302... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
open-llm-leaderboard/details_AIDC-ai-business__Marcoroni-70B-v1_public | open-llm-leaderboard | 2023-11-09T10:11:08Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T10:11:08Z | 2023-11-09T10:11:00.000Z | 2023-11-09T10:11:00 | ---
pretty_name: Evaluation run of AIDC-ai-business/Marcoroni-70B-v1
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [AIDC-ai-business/Marcoroni-70B-v1](https://huggingface.co/AIDC-ai-business/Marcoroni-70B-v1)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 3 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the aggregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_AIDC-ai-business__Marcoroni-70B-v1_public\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-11-09T10:10:41.822023](https://huggingface.co/datasets/open-llm-leaderboard/details_AIDC-ai-business__Marcoroni-70B-v1_public/blob/main/results_2023-11-09T10-10-41.822023.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"em\": 0.3132340604026846,\n\
\ \"em_stderr\": 0.004749834913438157,\n \"f1\": 0.456531040268459,\n\
\ \"f1_stderr\": 0.004364621394991152,\n \"acc\": 0.5835410217852969,\n\
\ \"acc_stderr\": 0.01171539602098445\n },\n \"harness|drop|3\": {\n\
\ \"em\": 0.3132340604026846,\n \"em_stderr\": 0.004749834913438157,\n\
\ \"f1\": 0.456531040268459,\n \"f1_stderr\": 0.004364621394991152\n\
\ },\n \"harness|gsm8k|5\": {\n \"acc\": 0.332827899924185,\n \
\ \"acc_stderr\": 0.012979892496598271\n },\n \"harness|winogrande|5\":\
\ {\n \"acc\": 0.8342541436464088,\n \"acc_stderr\": 0.010450899545370628\n\
\ }\n}\n```"
repo_url: https://huggingface.co/AIDC-ai-business/Marcoroni-70B-v1
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_drop_3
data_files:
- split: 2023_11_09T10_10_41.822023
path:
- '**/details_harness|drop|3_2023-11-09T10-10-41.822023.parquet'
- split: latest
path:
- '**/details_harness|drop|3_2023-11-09T10-10-41.822023.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_11_09T10_10_41.822023
path:
- '**/details_harness|gsm8k|5_2023-11-09T10-10-41.822023.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-11-09T10-10-41.822023.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_11_09T10_10_41.822023
path:
- '**/details_harness|winogrande|5_2023-11-09T10-10-41.822023.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-11-09T10-10-41.822023.parquet'
- config_name: results
data_files:
- split: 2023_11_09T10_10_41.822023
path:
- results_2023-11-09T10-10-41.822023.parquet
- split: latest
path:
- results_2023-11-09T10-10-41.822023.parquet
---
# Dataset Card for Evaluation run of AIDC-ai-business/Marcoroni-70B-v1
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/AIDC-ai-business/Marcoroni-70B-v1
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [AIDC-ai-business/Marcoroni-70B-v1](https://huggingface.co/AIDC-ai-business/Marcoroni-70B-v1) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 3 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_AIDC-ai-business__Marcoroni-70B-v1_public",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-11-09T10:10:41.822023](https://huggingface.co/datasets/open-llm-leaderboard/details_AIDC-ai-business__Marcoroni-70B-v1_public/blob/main/results_2023-11-09T10-10-41.822023.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"em": 0.3132340604026846,
"em_stderr": 0.004749834913438157,
"f1": 0.456531040268459,
"f1_stderr": 0.004364621394991152,
"acc": 0.5835410217852969,
"acc_stderr": 0.01171539602098445
},
"harness|drop|3": {
"em": 0.3132340604026846,
"em_stderr": 0.004749834913438157,
"f1": 0.456531040268459,
"f1_stderr": 0.004364621394991152
},
"harness|gsm8k|5": {
"acc": 0.332827899924185,
"acc_stderr": 0.012979892496598271
},
"harness|winogrande|5": {
"acc": 0.8342541436464088,
"acc_stderr": 0.010450899545370628
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] | [
-0.42966392636299133,
-0.6760127544403076,
0.1522652506828308,
0.1846376359462738,
-0.10651286691427231,
0.10921092331409454,
-0.4291481375694275,
-0.23212480545043945,
0.41343000531196594,
0.48301568627357483,
-0.6638513207435608,
-0.9583002924919128,
-0.6207636594772339,
0.29301899671554... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Chandu76749/Data | Chandu76749 | 2023-11-09T10:24:08Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T10:24:08Z | 2023-11-09T10:24:08.000Z | 2023-11-09T10:24:08 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
kpptdll/SMD | kpptdll | 2023-11-17T06:35:40Z | 0 | 0 | null | [
"region:us"
] | 2023-11-17T06:35:40Z | 2023-11-09T10:34:54.000Z | 2023-11-09T10:34:54 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Enderfga/sample | Enderfga | 2023-11-13T20:10:58Z | 0 | 0 | null | [
"region:us"
] | 2023-11-13T20:10:58Z | 2023-11-09T10:38:51.000Z | 2023-11-09T10:38:51 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
MruganKulkarni/restomenuu | MruganKulkarni | 2023-11-09T10:49:08Z | 0 | 0 | null | [
"license:mit",
"region:us"
] | 2023-11-09T10:49:08Z | 2023-11-09T10:49:08.000Z | 2023-11-09T10:49:08 | ---
license: mit
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
zeroman1318/daegu-ai-06 | zeroman1318 | 2023-11-09T11:11:42Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:11:42Z | 2023-11-09T10:56:05.000Z | 2023-11-09T10:56:05 | ---
configs:
- config_name: default
data_files:
- split: train
path: data.csv
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.503662645816803,
-0.5130204558372498,
0.18480271100997925,
0.20869815349578857,
-0.3474426865577698,
-0.05577763170003891,
-0.022632520645856857,
-0.6274707913398743,
0.4583321809768677,
0.810380756855011,
-0.7633895874023438,
-0.9683904647827148,
-0.5347056984901428,
0.1252623945474624... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
yangzhou321/ImageNet1k_Corrupt | yangzhou321 | 2023-11-09T11:02:10Z | 0 | 0 | null | [
"license:mit",
"region:us"
] | 2023-11-09T11:02:10Z | 2023-11-09T11:02:10.000Z | 2023-11-09T11:02:10 | ---
license: mit
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
BoyuanJackchen/APPS_zeroshot_for_code_generation | BoyuanJackchen | 2023-11-09T11:05:52Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:05:52Z | 2023-11-09T11:02:41.000Z | 2023-11-09T11:02:41 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
leonardPKU/llava1.5_data | leonardPKU | 2023-11-09T11:09:05Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:09:05Z | 2023-11-09T11:09:05.000Z | 2023-11-09T11:09:05 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
KevinJung/dataset_repository_name | KevinJung | 2023-11-09T11:10:11Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:10:11Z | 2023-11-09T11:10:10.000Z | 2023-11-09T11:10:10 | ---
configs:
- config_name: default
data_files:
- split: train
path: data.csv
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.503662645816803,
-0.5130204558372498,
0.18480271100997925,
0.20869815349578857,
-0.3474426865577698,
-0.05577763170003891,
-0.022632520645856857,
-0.6274707913398743,
0.4583321809768677,
0.810380756855011,
-0.7633895874023438,
-0.9683904647827148,
-0.5347056984901428,
0.1252623945474624... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
AZSXDCFV123/dataset_repository_name | AZSXDCFV123 | 2023-11-09T11:10:12Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:10:12Z | 2023-11-09T11:10:11.000Z | 2023-11-09T11:10:11 | ---
configs:
- config_name: default
data_files:
- split: train
path: data.csv
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.503662645816803,
-0.5130204558372498,
0.18480271100997925,
0.20869815349578857,
-0.3474426865577698,
-0.05577763170003891,
-0.022632520645856857,
-0.6274707913398743,
0.4583321809768677,
0.810380756855011,
-0.7633895874023438,
-0.9683904647827148,
-0.5347056984901428,
0.1252623945474624... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
thiefcat/dataset_repository_name | thiefcat | 2023-11-09T11:12:09Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:12:09Z | 2023-11-09T11:10:12.000Z | 2023-11-09T11:10:12 | ---
configs:
- config_name: default
data_files:
- split: train
path: data.csv
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.503662645816803,
-0.5130204558372498,
0.18480271100997925,
0.20869815349578857,
-0.3474426865577698,
-0.05577763170003891,
-0.022632520645856857,
-0.6274707913398743,
0.4583321809768677,
0.810380756855011,
-0.7633895874023438,
-0.9683904647827148,
-0.5347056984901428,
0.1252623945474624... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Seokeunsoo/dataset_repository_name | Seokeunsoo | 2023-11-09T11:10:12Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:10:12Z | 2023-11-09T11:10:12.000Z | 2023-11-09T11:10:12 | ---
configs:
- config_name: default
data_files:
- split: train
path: data.csv
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.503662645816803,
-0.5130204558372498,
0.18480271100997925,
0.20869815349578857,
-0.3474426865577698,
-0.05577763170003891,
-0.022632520645856857,
-0.6274707913398743,
0.4583321809768677,
0.810380756855011,
-0.7633895874023438,
-0.9683904647827148,
-0.5347056984901428,
0.1252623945474624... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
KimKimzed/dataset_repository_name | KimKimzed | 2023-11-09T11:10:39Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:10:39Z | 2023-11-09T11:10:39.000Z | 2023-11-09T11:10:39 | ---
configs:
- config_name: default
data_files:
- split: train
path: data.csv
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.503662645816803,
-0.5130204558372498,
0.18480271100997925,
0.20869815349578857,
-0.3474426865577698,
-0.05577763170003891,
-0.022632520645856857,
-0.6274707913398743,
0.4583321809768677,
0.810380756855011,
-0.7633895874023438,
-0.9683904647827148,
-0.5347056984901428,
0.1252623945474624... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ej94/dataset_repository_name | ej94 | 2023-11-09T11:15:09Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:15:09Z | 2023-11-09T11:11:21.000Z | 2023-11-09T11:11:21 | ---
configs:
- config_name: default
data_files:
- split: train
path: data.csv
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.503662645816803,
-0.5130204558372498,
0.18480271100997925,
0.20869815349578857,
-0.3474426865577698,
-0.05577763170003891,
-0.022632520645856857,
-0.6274707913398743,
0.4583321809768677,
0.810380756855011,
-0.7633895874023438,
-0.9683904647827148,
-0.5347056984901428,
0.1252623945474624... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
grang13lue/dataset_repository_name | grang13lue | 2023-11-09T11:13:16Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:13:16Z | 2023-11-09T11:13:15.000Z | 2023-11-09T11:13:15 | ---
configs:
- config_name: default
data_files:
- split: train
path: data.csv
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.503662645816803,
-0.5130204558372498,
0.18480271100997925,
0.20869815349578857,
-0.3474426865577698,
-0.05577763170003891,
-0.022632520645856857,
-0.6274707913398743,
0.4583321809768677,
0.810380756855011,
-0.7633895874023438,
-0.9683904647827148,
-0.5347056984901428,
0.1252623945474624... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
grang13lue/my_ | grang13lue | 2023-11-09T11:16:31Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:16:31Z | 2023-11-09T11:16:31.000Z | 2023-11-09T11:16:31 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Seokeunsoo/md_bbiyong | Seokeunsoo | 2023-11-09T11:20:23Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:20:23Z | 2023-11-09T11:16:48.000Z | 2023-11-09T11:16:48 | ---
configs:
- config_name: default
data_files:
- split: train
path: data.csv
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.503662645816803,
-0.5130204558372498,
0.18480271100997925,
0.20869815349578857,
-0.3474426865577698,
-0.05577763170003891,
-0.022632520645856857,
-0.6274707913398743,
0.4583321809768677,
0.810380756855011,
-0.7633895874023438,
-0.9683904647827148,
-0.5347056984901428,
0.1252623945474624... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
useoseou/md-daegu | useoseou | 2023-11-09T11:16:51Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:16:51Z | 2023-11-09T11:16:51.000Z | 2023-11-09T11:16:51 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
thiefcat/data01 | thiefcat | 2023-11-09T11:21:40Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:21:40Z | 2023-11-09T11:16:53.000Z | 2023-11-09T11:16:53 | ---
configs:
- config_name: default
data_files:
- split: train
path: data.csv
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.503662645816803,
-0.5130204558372498,
0.18480271100997925,
0.20869815349578857,
-0.3474426865577698,
-0.05577763170003891,
-0.022632520645856857,
-0.6274707913398743,
0.4583321809768677,
0.810380756855011,
-0.7633895874023438,
-0.9683904647827148,
-0.5347056984901428,
0.1252623945474624... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ej94/md-daegu231109 | ej94 | 2023-11-09T11:20:13Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:20:13Z | 2023-11-09T11:16:53.000Z | 2023-11-09T11:16:53 | ---
configs:
- config_name: default
data_files:
- split: train
path: data.csv
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.503662645816803,
-0.5130204558372498,
0.18480271100997925,
0.20869815349578857,
-0.3474426865577698,
-0.05577763170003891,
-0.022632520645856857,
-0.6274707913398743,
0.4583321809768677,
0.810380756855011,
-0.7633895874023438,
-0.9683904647827148,
-0.5347056984901428,
0.1252623945474624... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Meltyom/igonan | Meltyom | 2023-11-09T11:16:55Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:16:55Z | 2023-11-09T11:16:55.000Z | 2023-11-09T11:16:55 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
seoeunju/test1 | seoeunju | 2023-11-09T11:17:11Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:17:11Z | 2023-11-09T11:17:11.000Z | 2023-11-09T11:17:11 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
danaroth/chikusei | danaroth | 2023-11-09T15:53:41Z | 0 | 0 | null | [
"license:cc-by-3.0",
"region:us"
] | 2023-11-09T15:53:41Z | 2023-11-09T11:28:23.000Z | 2023-11-09T11:28:23 | ---
license: cc-by-3.0
---
# Description
The airborne hyperspectral dataset was taken by Headwall Hyperspec-VNIR-C imaging sensor over agricultural and urban areas in Chikusei, Ibaraki, Japan, on July 29, 2014 between the times 9:56 to 10:53 UTC+9. The central point of the scene is located at coordinates: 36.294946N, 140.008380E. The hyperspectral dataset has 128 bands in the spectral range from 363 nm to 1018 nm. The scene consists of 2517x2335 pixels and the ground sampling distance was 2.5 m. Ground truth of 19 classes was collected via a field survey and visual inspection using high-resolution color images obtained by Canon EOS 5D Mark II together with the hyperspectral data. The hyperspectral data and ground truth were made available to the scientific community in the ENVI and MATLAB formats at http://park.itc.u-tokyo.ac.jp/sal/hyperdata. More details of the experiment are presented in the technical report within the dataset.
# Quick look
<figure>
<img src= "assets/Chikusei.jpg" alt="Chikusei" width="300" />
<figcaption>Bands visualization of the Chikusei dataset.</figcaption>
</figure>
# Credits
Originally downloaded from: https://naotoyokoya.com/Download.html
In order to use the datasets, please fulfill the following three requirements:
- Giving an acknowledgement as follows:
The authors gratefully acknowledge Space Application Laboratory, Department of Advanced Interdisciplinary Studies, the University of Tokyo for providing the hyperspectral data.
- Using the following license for hyperspectral data:
http://creativecommons.org/licenses/by/3.0/
- This dataset was made public by Dr. Naoto Yokoya and Prof. Akira Iwasaki from the University of Tokyo. Please cite:
In WORD:
```
N. Yokoya and A. Iwasaki, "Airborne hyperspectral data over Chikusei," Space Appl. Lab., Univ. Tokyo, Japan, Tech. Rep., May 2016.
```
In LaTex:
```
@techreport{NYokoya2016,
author = {N. Yokoya and A. Iwasaki},
title = {Airborne hyperspectral data over Chikusei},
institution = {Space Application Laboratory, University of Tokyo},
year = 2016,
address = {Japan},
month = {May},
year = 2016,
}
``` | [
-0.7317290902137756,
-0.23049110174179077,
0.6590901017189026,
0.008513504639267921,
-0.1857042759656906,
-0.15358613431453705,
-0.17729102075099945,
-0.6727345585823059,
0.49782615900039673,
0.5374616980552673,
-0.7302129864692688,
-0.5351805090904236,
-0.511555552482605,
-0.1469760686159... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
classla/xlm-r-bertic-data | classla | 2023-11-27T10:11:20Z | 0 | 0 | null | [
"license:cc-by-sa-4.0",
"region:us"
] | 2023-11-27T10:11:20Z | 2023-11-09T11:45:41.000Z | 2023-11-09T11:45:41 | ---
license: cc-by-sa-4.0
---
# XLM-R-BERTić dataset
## Composition and usage
This dataset consists of the following splits:
* macocu_hbs
* hr_news
* bswac
* cc100_hr
* cc100_sr
* classla_sr
* classla_hr
* classla_bs
* cnrwac
* hrwac
* mC4
* riznica
* srwac
The entire dataset can be downloaded and used as follows:
```python
import datasets
dict_of_datasets = datasets.load_dataset("classla/xlm-r-bertic-data")
full_dataset = datasets.concatenate_datasets([d for d in dict_of_datasets.values()])
```
A single split can be taken as well, but note that this means all the splits will be downloaded and generated, which can take a long time:
```python
import datasets
riznica = datasets.load_dataset("classla/xlm-r-bertic-data", split="riznica")
```
To circumvent this one option is using streaming:
```python
import datasets
riznica = datasets.load_dataset("classla/xlm-r-bertic-data", split="riznica", streaming=True)
for i in riznica.take(2):
print(i)
# Output:
# {'text': 'PRAGMATIČARI DOGMATI SANJARI'}
# {'text': 'Ivica Župan'}
```
Read more on streaming [here](https://huggingface.co/docs/datasets/stream). | [
-0.42068466544151306,
-0.5604937672615051,
0.2110593169927597,
0.43424686789512634,
-0.471816748380661,
0.2091447412967682,
-0.34009838104248047,
-0.2636558413505554,
0.35911402106285095,
0.3270295560359955,
-0.7585192322731018,
-0.5311430096626282,
-0.4991875886917114,
0.47470030188560486... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
presencesw/dataset_2000_decompese_question_test | presencesw | 2023-11-09T12:40:59Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T12:40:59Z | 2023-11-09T11:48:32.000Z | 2023-11-09T11:48:32 | ---
dataset_info:
features:
- name: entities
sequence: 'null'
- name: triplets
list:
- name: question
dtype: string
- name: answer
dtype: string
- name: complex_question
dtype: string
splits:
- name: train
num_bytes: 3405
num_examples: 10
download_size: 4156
dataset_size: 3405
---
# Dataset Card for "dataset_2000_decompese_question_test"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.6911260485649109,
-0.493369460105896,
0.04338085651397705,
0.24165582656860352,
-0.1246291995048523,
-0.27313530445098877,
0.2767871618270874,
0.05832720547914505,
0.7207894921302795,
0.44343236088752747,
-0.8306340575218201,
-0.5321374535560608,
-0.42886635661125183,
0.0728440731763839... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
namok21/dataset_repository_name | namok21 | 2023-11-09T11:53:40Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T11:53:40Z | 2023-11-09T11:52:01.000Z | 2023-11-09T11:52:01 | ---
configs:
- config_name: default
data_files:
- split: train
path: data.csv
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.503662645816803,
-0.5130204558372498,
0.18480271100997925,
0.20869815349578857,
-0.3474426865577698,
-0.05577763170003891,
-0.022632520645856857,
-0.6274707913398743,
0.4583321809768677,
0.810380756855011,
-0.7633895874023438,
-0.9683904647827148,
-0.5347056984901428,
0.1252623945474624... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Phospy/rahasia | Phospy | 2023-11-09T12:07:23Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T12:07:23Z | 2023-11-09T12:07:23.000Z | 2023-11-09T12:07:23 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ajibawa-2023/Mathjson | ajibawa-2023 | 2023-11-11T07:26:48Z | 0 | 2 | null | [
"region:us"
] | 2023-11-11T07:26:48Z | 2023-11-09T12:38:59.000Z | 2023-11-09T12:38:59 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
kinit-tomassako/ver_claimdetection_demo | kinit-tomassako | 2023-11-13T08:17:53Z | 0 | 0 | null | [
"region:us"
] | 2023-11-13T08:17:53Z | 2023-11-09T12:46:48.000Z | 2023-11-09T12:46:48 | ---
configs:
- config_name: default
data_files:
- split: train
path: data.csv
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.503662645816803,
-0.5130205154418945,
0.18480272591114044,
0.20869813859462738,
-0.3474426865577698,
-0.05577763542532921,
-0.022632522508502007,
-0.6274707913398743,
0.4583321809768677,
0.8103806972503662,
-0.7633895874023438,
-0.9683905839920044,
-0.5347057580947876,
0.125262394547462... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
arieg/bw_spec_cls_80_22 | arieg | 2023-11-09T12:47:52Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T12:47:52Z | 2023-11-09T12:47:34.000Z | 2023-11-09T12:47:34 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': '55076'
'1': '55097'
'2': '55100'
'3': '55101'
'4': '55102'
'5': '55113'
'6': '55119'
'7': '55120'
'8': '55121'
'9': '55122'
'10': '55123'
'11': '55124'
'12': '55149'
'13': '55231'
'14': '55232'
'15': '55233'
'16': '55234'
'17': '55235'
'18': '55236'
'19': '55237'
'20': '55238'
'21': '55240'
'22': '55241'
'23': '55242'
'24': '55285'
'25': '55286'
'26': '55287'
'27': '55288'
'28': '55289'
'29': '55290'
'30': '55291'
'31': '55292'
'32': '55293'
'33': '55294'
'34': '55295'
'35': '55402'
'36': '55430'
'37': '55436'
'38': '55437'
'39': '55480'
'40': '55481'
'41': '55549'
'42': '55572'
'43': '55709'
'44': '55710'
'45': '55711'
'46': '55712'
'47': '55713'
'48': '55714'
'49': '55715'
'50': '55716'
'51': '55717'
'52': '55718'
'53': '55719'
'54': '55783'
'55': '55786'
'56': '55807'
'57': '55808'
'58': '55809'
'59': '55810'
'60': '55811'
'61': '55826'
'62': '55827'
'63': '55828'
'64': '55830'
'65': '55831'
'66': '55832'
'67': '55833'
'68': '55900'
'69': '56010'
'70': '56015'
'71': '56020'
'72': '56028'
'73': '56029'
'74': '56030'
'75': '56031'
'76': '56033'
'77': '56034'
'78': '56036'
'79': '56247'
splits:
- name: train
num_bytes: 88281430.4
num_examples: 1600
- name: test
num_bytes: 22107725.0
num_examples: 400
download_size: 110670044
dataset_size: 110389155.4
---
# Dataset Card for "bw_spec_cls_80_22"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.6802901029586792,
-0.24686187505722046,
0.146070197224617,
0.36949482560157776,
-0.26653873920440674,
-0.09550505876541138,
0.030580416321754456,
-0.28797945380210876,
0.5621195435523987,
0.5576102137565613,
-0.7689053416252136,
-0.7948891520500183,
-0.6108701229095459,
-0.1977168619632... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
open-llm-leaderboard/details_luffycodes__llama-shishya-7b-ep3-v1_public | open-llm-leaderboard | 2023-11-09T12:52:16Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T12:52:16Z | 2023-11-09T12:51:13.000Z | 2023-11-09T12:51:13 | ---
pretty_name: Evaluation run of luffycodes/llama-shishya-7b-ep3-v1
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [luffycodes/llama-shishya-7b-ep3-v1](https://huggingface.co/luffycodes/llama-shishya-7b-ep3-v1)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 64 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the aggregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_luffycodes__llama-shishya-7b-ep3-v1_public\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-11-09T12:48:08.068028](https://huggingface.co/datasets/open-llm-leaderboard/details_luffycodes__llama-shishya-7b-ep3-v1_public/blob/main/results_2023-11-09T12-48-08.068028.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"acc\": 0.4594923428252717,\n\
\ \"acc_stderr\": 0.03404628674654547,\n \"acc_norm\": 0.46668909375227274,\n\
\ \"acc_norm_stderr\": 0.03497039082366745,\n \"mc1\": 0.204406364749082,\n\
\ \"mc1_stderr\": 0.014117174337432616,\n \"mc2\": 0.3089869590457097,\n\
\ \"mc2_stderr\": 0.013843169413571187,\n \"em\": 0.3115562080536913,\n\
\ \"em_stderr\": 0.004742879599828378,\n \"f1\": 0.3699653942953032,\n\
\ \"f1_stderr\": 0.004671420668393907\n },\n \"harness|arc:challenge|25\"\
: {\n \"acc\": 0.45307167235494883,\n \"acc_stderr\": 0.01454689205200563,\n\
\ \"acc_norm\": 0.4803754266211604,\n \"acc_norm_stderr\": 0.014600132075947092\n\
\ },\n \"harness|hellaswag|10\": {\n \"acc\": 0.5934076877116112,\n\
\ \"acc_stderr\": 0.00490193651154613,\n \"acc_norm\": 0.7662816172077276,\n\
\ \"acc_norm_stderr\": 0.004223302177263009\n },\n \"harness|hendrycksTest-abstract_algebra|5\"\
: {\n \"acc\": 0.29,\n \"acc_stderr\": 0.04560480215720684,\n \
\ \"acc_norm\": 0.29,\n \"acc_norm_stderr\": 0.04560480215720684\n \
\ },\n \"harness|hendrycksTest-anatomy|5\": {\n \"acc\": 0.4666666666666667,\n\
\ \"acc_stderr\": 0.043097329010363554,\n \"acc_norm\": 0.4666666666666667,\n\
\ \"acc_norm_stderr\": 0.043097329010363554\n },\n \"harness|hendrycksTest-astronomy|5\"\
: {\n \"acc\": 0.4605263157894737,\n \"acc_stderr\": 0.04056242252249034,\n\
\ \"acc_norm\": 0.4605263157894737,\n \"acc_norm_stderr\": 0.04056242252249034\n\
\ },\n \"harness|hendrycksTest-business_ethics|5\": {\n \"acc\": 0.45,\n\
\ \"acc_stderr\": 0.049999999999999996,\n \"acc_norm\": 0.45,\n \
\ \"acc_norm_stderr\": 0.049999999999999996\n },\n \"harness|hendrycksTest-clinical_knowledge|5\"\
: {\n \"acc\": 0.49433962264150944,\n \"acc_stderr\": 0.030770900763851302,\n\
\ \"acc_norm\": 0.49433962264150944,\n \"acc_norm_stderr\": 0.030770900763851302\n\
\ },\n \"harness|hendrycksTest-college_biology|5\": {\n \"acc\": 0.4583333333333333,\n\
\ \"acc_stderr\": 0.04166666666666665,\n \"acc_norm\": 0.4583333333333333,\n\
\ \"acc_norm_stderr\": 0.04166666666666665\n },\n \"harness|hendrycksTest-college_chemistry|5\"\
: {\n \"acc\": 0.28,\n \"acc_stderr\": 0.04512608598542127,\n \
\ \"acc_norm\": 0.28,\n \"acc_norm_stderr\": 0.04512608598542127\n \
\ },\n \"harness|hendrycksTest-college_computer_science|5\": {\n \"acc\"\
: 0.4,\n \"acc_stderr\": 0.049236596391733084,\n \"acc_norm\": 0.4,\n\
\ \"acc_norm_stderr\": 0.049236596391733084\n },\n \"harness|hendrycksTest-college_mathematics|5\"\
: {\n \"acc\": 0.27,\n \"acc_stderr\": 0.044619604333847394,\n \
\ \"acc_norm\": 0.27,\n \"acc_norm_stderr\": 0.044619604333847394\n \
\ },\n \"harness|hendrycksTest-college_medicine|5\": {\n \"acc\": 0.3988439306358382,\n\
\ \"acc_stderr\": 0.03733626655383509,\n \"acc_norm\": 0.3988439306358382,\n\
\ \"acc_norm_stderr\": 0.03733626655383509\n },\n \"harness|hendrycksTest-college_physics|5\"\
: {\n \"acc\": 0.24509803921568626,\n \"acc_stderr\": 0.04280105837364395,\n\
\ \"acc_norm\": 0.24509803921568626,\n \"acc_norm_stderr\": 0.04280105837364395\n\
\ },\n \"harness|hendrycksTest-computer_security|5\": {\n \"acc\":\
\ 0.53,\n \"acc_stderr\": 0.05016135580465919,\n \"acc_norm\": 0.53,\n\
\ \"acc_norm_stderr\": 0.05016135580465919\n },\n \"harness|hendrycksTest-conceptual_physics|5\"\
: {\n \"acc\": 0.39574468085106385,\n \"acc_stderr\": 0.03196758697835362,\n\
\ \"acc_norm\": 0.39574468085106385,\n \"acc_norm_stderr\": 0.03196758697835362\n\
\ },\n \"harness|hendrycksTest-econometrics|5\": {\n \"acc\": 0.3333333333333333,\n\
\ \"acc_stderr\": 0.044346007015849245,\n \"acc_norm\": 0.3333333333333333,\n\
\ \"acc_norm_stderr\": 0.044346007015849245\n },\n \"harness|hendrycksTest-electrical_engineering|5\"\
: {\n \"acc\": 0.4206896551724138,\n \"acc_stderr\": 0.0411391498118926,\n\
\ \"acc_norm\": 0.4206896551724138,\n \"acc_norm_stderr\": 0.0411391498118926\n\
\ },\n \"harness|hendrycksTest-elementary_mathematics|5\": {\n \"acc\"\
: 0.30158730158730157,\n \"acc_stderr\": 0.023636975996101806,\n \"\
acc_norm\": 0.30158730158730157,\n \"acc_norm_stderr\": 0.023636975996101806\n\
\ },\n \"harness|hendrycksTest-formal_logic|5\": {\n \"acc\": 0.1984126984126984,\n\
\ \"acc_stderr\": 0.03567016675276864,\n \"acc_norm\": 0.1984126984126984,\n\
\ \"acc_norm_stderr\": 0.03567016675276864\n },\n \"harness|hendrycksTest-global_facts|5\"\
: {\n \"acc\": 0.33,\n \"acc_stderr\": 0.04725815626252604,\n \
\ \"acc_norm\": 0.33,\n \"acc_norm_stderr\": 0.04725815626252604\n \
\ },\n \"harness|hendrycksTest-high_school_biology|5\": {\n \"acc\": 0.535483870967742,\n\
\ \"acc_stderr\": 0.02837228779796293,\n \"acc_norm\": 0.535483870967742,\n\
\ \"acc_norm_stderr\": 0.02837228779796293\n },\n \"harness|hendrycksTest-high_school_chemistry|5\"\
: {\n \"acc\": 0.35960591133004927,\n \"acc_stderr\": 0.03376458246509567,\n\
\ \"acc_norm\": 0.35960591133004927,\n \"acc_norm_stderr\": 0.03376458246509567\n\
\ },\n \"harness|hendrycksTest-high_school_computer_science|5\": {\n \
\ \"acc\": 0.48,\n \"acc_stderr\": 0.050211673156867795,\n \"acc_norm\"\
: 0.48,\n \"acc_norm_stderr\": 0.050211673156867795\n },\n \"harness|hendrycksTest-high_school_european_history|5\"\
: {\n \"acc\": 0.6424242424242425,\n \"acc_stderr\": 0.037425970438065864,\n\
\ \"acc_norm\": 0.6424242424242425,\n \"acc_norm_stderr\": 0.037425970438065864\n\
\ },\n \"harness|hendrycksTest-high_school_geography|5\": {\n \"acc\"\
: 0.5454545454545454,\n \"acc_stderr\": 0.03547601494006937,\n \"\
acc_norm\": 0.5454545454545454,\n \"acc_norm_stderr\": 0.03547601494006937\n\
\ },\n \"harness|hendrycksTest-high_school_government_and_politics|5\": {\n\
\ \"acc\": 0.6735751295336787,\n \"acc_stderr\": 0.033840286211432945,\n\
\ \"acc_norm\": 0.6735751295336787,\n \"acc_norm_stderr\": 0.033840286211432945\n\
\ },\n \"harness|hendrycksTest-high_school_macroeconomics|5\": {\n \
\ \"acc\": 0.38974358974358975,\n \"acc_stderr\": 0.024726967886647078,\n\
\ \"acc_norm\": 0.38974358974358975,\n \"acc_norm_stderr\": 0.024726967886647078\n\
\ },\n \"harness|hendrycksTest-high_school_mathematics|5\": {\n \"\
acc\": 0.24444444444444444,\n \"acc_stderr\": 0.02620276653465215,\n \
\ \"acc_norm\": 0.24444444444444444,\n \"acc_norm_stderr\": 0.02620276653465215\n\
\ },\n \"harness|hendrycksTest-high_school_microeconomics|5\": {\n \
\ \"acc\": 0.4117647058823529,\n \"acc_stderr\": 0.03196876989195778,\n \
\ \"acc_norm\": 0.4117647058823529,\n \"acc_norm_stderr\": 0.03196876989195778\n\
\ },\n \"harness|hendrycksTest-high_school_physics|5\": {\n \"acc\"\
: 0.2913907284768212,\n \"acc_stderr\": 0.03710185726119995,\n \"\
acc_norm\": 0.2913907284768212,\n \"acc_norm_stderr\": 0.03710185726119995\n\
\ },\n \"harness|hendrycksTest-high_school_psychology|5\": {\n \"acc\"\
: 0.6440366972477064,\n \"acc_stderr\": 0.020528559278244214,\n \"\
acc_norm\": 0.6440366972477064,\n \"acc_norm_stderr\": 0.020528559278244214\n\
\ },\n \"harness|hendrycksTest-high_school_statistics|5\": {\n \"acc\"\
: 0.2916666666666667,\n \"acc_stderr\": 0.03099866630456053,\n \"\
acc_norm\": 0.2916666666666667,\n \"acc_norm_stderr\": 0.03099866630456053\n\
\ },\n \"harness|hendrycksTest-high_school_us_history|5\": {\n \"acc\"\
: 0.6029411764705882,\n \"acc_stderr\": 0.0343413116471913,\n \"acc_norm\"\
: 0.6029411764705882,\n \"acc_norm_stderr\": 0.0343413116471913\n },\n\
\ \"harness|hendrycksTest-high_school_world_history|5\": {\n \"acc\":\
\ 0.6455696202531646,\n \"acc_stderr\": 0.031137304297185815,\n \"\
acc_norm\": 0.6455696202531646,\n \"acc_norm_stderr\": 0.031137304297185815\n\
\ },\n \"harness|hendrycksTest-human_aging|5\": {\n \"acc\": 0.5426008968609866,\n\
\ \"acc_stderr\": 0.033435777055830646,\n \"acc_norm\": 0.5426008968609866,\n\
\ \"acc_norm_stderr\": 0.033435777055830646\n },\n \"harness|hendrycksTest-human_sexuality|5\"\
: {\n \"acc\": 0.5725190839694656,\n \"acc_stderr\": 0.04338920305792401,\n\
\ \"acc_norm\": 0.5725190839694656,\n \"acc_norm_stderr\": 0.04338920305792401\n\
\ },\n \"harness|hendrycksTest-international_law|5\": {\n \"acc\":\
\ 0.6363636363636364,\n \"acc_stderr\": 0.043913262867240704,\n \"\
acc_norm\": 0.6363636363636364,\n \"acc_norm_stderr\": 0.043913262867240704\n\
\ },\n \"harness|hendrycksTest-jurisprudence|5\": {\n \"acc\": 0.5092592592592593,\n\
\ \"acc_stderr\": 0.04832853553437055,\n \"acc_norm\": 0.5092592592592593,\n\
\ \"acc_norm_stderr\": 0.04832853553437055\n },\n \"harness|hendrycksTest-logical_fallacies|5\"\
: {\n \"acc\": 0.49693251533742333,\n \"acc_stderr\": 0.03928297078179663,\n\
\ \"acc_norm\": 0.49693251533742333,\n \"acc_norm_stderr\": 0.03928297078179663\n\
\ },\n \"harness|hendrycksTest-machine_learning|5\": {\n \"acc\": 0.23214285714285715,\n\
\ \"acc_stderr\": 0.04007341809755806,\n \"acc_norm\": 0.23214285714285715,\n\
\ \"acc_norm_stderr\": 0.04007341809755806\n },\n \"harness|hendrycksTest-management|5\"\
: {\n \"acc\": 0.6601941747572816,\n \"acc_stderr\": 0.04689765937278135,\n\
\ \"acc_norm\": 0.6601941747572816,\n \"acc_norm_stderr\": 0.04689765937278135\n\
\ },\n \"harness|hendrycksTest-marketing|5\": {\n \"acc\": 0.7307692307692307,\n\
\ \"acc_stderr\": 0.029058588303748842,\n \"acc_norm\": 0.7307692307692307,\n\
\ \"acc_norm_stderr\": 0.029058588303748842\n },\n \"harness|hendrycksTest-medical_genetics|5\"\
: {\n \"acc\": 0.49,\n \"acc_stderr\": 0.05024183937956911,\n \
\ \"acc_norm\": 0.49,\n \"acc_norm_stderr\": 0.05024183937956911\n \
\ },\n \"harness|hendrycksTest-miscellaneous|5\": {\n \"acc\": 0.6883780332056194,\n\
\ \"acc_stderr\": 0.016562433867284176,\n \"acc_norm\": 0.6883780332056194,\n\
\ \"acc_norm_stderr\": 0.016562433867284176\n },\n \"harness|hendrycksTest-moral_disputes|5\"\
: {\n \"acc\": 0.49421965317919075,\n \"acc_stderr\": 0.02691729617914911,\n\
\ \"acc_norm\": 0.49421965317919075,\n \"acc_norm_stderr\": 0.02691729617914911\n\
\ },\n \"harness|hendrycksTest-moral_scenarios|5\": {\n \"acc\": 0.2536312849162011,\n\
\ \"acc_stderr\": 0.014551553659369922,\n \"acc_norm\": 0.2536312849162011,\n\
\ \"acc_norm_stderr\": 0.014551553659369922\n },\n \"harness|hendrycksTest-nutrition|5\"\
: {\n \"acc\": 0.5130718954248366,\n \"acc_stderr\": 0.028620130800700246,\n\
\ \"acc_norm\": 0.5130718954248366,\n \"acc_norm_stderr\": 0.028620130800700246\n\
\ },\n \"harness|hendrycksTest-philosophy|5\": {\n \"acc\": 0.5594855305466238,\n\
\ \"acc_stderr\": 0.02819640057419743,\n \"acc_norm\": 0.5594855305466238,\n\
\ \"acc_norm_stderr\": 0.02819640057419743\n },\n \"harness|hendrycksTest-prehistory|5\"\
: {\n \"acc\": 0.5308641975308642,\n \"acc_stderr\": 0.027767689606833932,\n\
\ \"acc_norm\": 0.5308641975308642,\n \"acc_norm_stderr\": 0.027767689606833932\n\
\ },\n \"harness|hendrycksTest-professional_accounting|5\": {\n \"\
acc\": 0.34397163120567376,\n \"acc_stderr\": 0.028338017428611327,\n \
\ \"acc_norm\": 0.34397163120567376,\n \"acc_norm_stderr\": 0.028338017428611327\n\
\ },\n \"harness|hendrycksTest-professional_law|5\": {\n \"acc\": 0.31877444589308995,\n\
\ \"acc_stderr\": 0.011901895635786097,\n \"acc_norm\": 0.31877444589308995,\n\
\ \"acc_norm_stderr\": 0.011901895635786097\n },\n \"harness|hendrycksTest-professional_medicine|5\"\
: {\n \"acc\": 0.4264705882352941,\n \"acc_stderr\": 0.030042615832714878,\n\
\ \"acc_norm\": 0.4264705882352941,\n \"acc_norm_stderr\": 0.030042615832714878\n\
\ },\n \"harness|hendrycksTest-professional_psychology|5\": {\n \"\
acc\": 0.4493464052287582,\n \"acc_stderr\": 0.020123766528027266,\n \
\ \"acc_norm\": 0.4493464052287582,\n \"acc_norm_stderr\": 0.020123766528027266\n\
\ },\n \"harness|hendrycksTest-public_relations|5\": {\n \"acc\": 0.4727272727272727,\n\
\ \"acc_stderr\": 0.04782001791380063,\n \"acc_norm\": 0.4727272727272727,\n\
\ \"acc_norm_stderr\": 0.04782001791380063\n },\n \"harness|hendrycksTest-security_studies|5\"\
: {\n \"acc\": 0.5265306122448979,\n \"acc_stderr\": 0.03196412734523272,\n\
\ \"acc_norm\": 0.5265306122448979,\n \"acc_norm_stderr\": 0.03196412734523272\n\
\ },\n \"harness|hendrycksTest-sociology|5\": {\n \"acc\": 0.6119402985074627,\n\
\ \"acc_stderr\": 0.034457899643627506,\n \"acc_norm\": 0.6119402985074627,\n\
\ \"acc_norm_stderr\": 0.034457899643627506\n },\n \"harness|hendrycksTest-us_foreign_policy|5\"\
: {\n \"acc\": 0.67,\n \"acc_stderr\": 0.04725815626252607,\n \
\ \"acc_norm\": 0.67,\n \"acc_norm_stderr\": 0.04725815626252607\n \
\ },\n \"harness|hendrycksTest-virology|5\": {\n \"acc\": 0.4036144578313253,\n\
\ \"acc_stderr\": 0.038194861407583984,\n \"acc_norm\": 0.4036144578313253,\n\
\ \"acc_norm_stderr\": 0.038194861407583984\n },\n \"harness|hendrycksTest-world_religions|5\"\
: {\n \"acc\": 0.672514619883041,\n \"acc_stderr\": 0.035993357714560276,\n\
\ \"acc_norm\": 0.672514619883041,\n \"acc_norm_stderr\": 0.035993357714560276\n\
\ },\n \"harness|truthfulqa:mc|0\": {\n \"mc1\": 0.204406364749082,\n\
\ \"mc1_stderr\": 0.014117174337432616,\n \"mc2\": 0.3089869590457097,\n\
\ \"mc2_stderr\": 0.013843169413571187\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.6945540647198106,\n \"acc_stderr\": 0.012945038632552022\n\
\ },\n \"harness|drop|3\": {\n \"em\": 0.3115562080536913,\n \
\ \"em_stderr\": 0.004742879599828378,\n \"f1\": 0.3699653942953032,\n \
\ \"f1_stderr\": 0.004671420668393907\n },\n \"harness|gsm8k|5\": {\n\
\ \"acc\": 0.0,\n \"acc_stderr\": 0.0\n }\n}\n```"
repo_url: https://huggingface.co/luffycodes/llama-shishya-7b-ep3-v1
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|arc:challenge|25_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_drop_3
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|drop|3_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|drop|3_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|gsm8k|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hellaswag|10_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T12-48-08.068028.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T12-48-08.068028.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- '**/details_harness|winogrande|5_2023-11-09T12-48-08.068028.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-11-09T12-48-08.068028.parquet'
- config_name: results
data_files:
- split: 2023_11_09T12_48_08.068028
path:
- results_2023-11-09T12-48-08.068028.parquet
- split: latest
path:
- results_2023-11-09T12-48-08.068028.parquet
---
# Dataset Card for Evaluation run of luffycodes/llama-shishya-7b-ep3-v1
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/luffycodes/llama-shishya-7b-ep3-v1
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [luffycodes/llama-shishya-7b-ep3-v1](https://huggingface.co/luffycodes/llama-shishya-7b-ep3-v1) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 64 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_luffycodes__llama-shishya-7b-ep3-v1_public",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-11-09T12:48:08.068028](https://huggingface.co/datasets/open-llm-leaderboard/details_luffycodes__llama-shishya-7b-ep3-v1_public/blob/main/results_2023-11-09T12-48-08.068028.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"acc": 0.4594923428252717,
"acc_stderr": 0.03404628674654547,
"acc_norm": 0.46668909375227274,
"acc_norm_stderr": 0.03497039082366745,
"mc1": 0.204406364749082,
"mc1_stderr": 0.014117174337432616,
"mc2": 0.3089869590457097,
"mc2_stderr": 0.013843169413571187,
"em": 0.3115562080536913,
"em_stderr": 0.004742879599828378,
"f1": 0.3699653942953032,
"f1_stderr": 0.004671420668393907
},
"harness|arc:challenge|25": {
"acc": 0.45307167235494883,
"acc_stderr": 0.01454689205200563,
"acc_norm": 0.4803754266211604,
"acc_norm_stderr": 0.014600132075947092
},
"harness|hellaswag|10": {
"acc": 0.5934076877116112,
"acc_stderr": 0.00490193651154613,
"acc_norm": 0.7662816172077276,
"acc_norm_stderr": 0.004223302177263009
},
"harness|hendrycksTest-abstract_algebra|5": {
"acc": 0.29,
"acc_stderr": 0.04560480215720684,
"acc_norm": 0.29,
"acc_norm_stderr": 0.04560480215720684
},
"harness|hendrycksTest-anatomy|5": {
"acc": 0.4666666666666667,
"acc_stderr": 0.043097329010363554,
"acc_norm": 0.4666666666666667,
"acc_norm_stderr": 0.043097329010363554
},
"harness|hendrycksTest-astronomy|5": {
"acc": 0.4605263157894737,
"acc_stderr": 0.04056242252249034,
"acc_norm": 0.4605263157894737,
"acc_norm_stderr": 0.04056242252249034
},
"harness|hendrycksTest-business_ethics|5": {
"acc": 0.45,
"acc_stderr": 0.049999999999999996,
"acc_norm": 0.45,
"acc_norm_stderr": 0.049999999999999996
},
"harness|hendrycksTest-clinical_knowledge|5": {
"acc": 0.49433962264150944,
"acc_stderr": 0.030770900763851302,
"acc_norm": 0.49433962264150944,
"acc_norm_stderr": 0.030770900763851302
},
"harness|hendrycksTest-college_biology|5": {
"acc": 0.4583333333333333,
"acc_stderr": 0.04166666666666665,
"acc_norm": 0.4583333333333333,
"acc_norm_stderr": 0.04166666666666665
},
"harness|hendrycksTest-college_chemistry|5": {
"acc": 0.28,
"acc_stderr": 0.04512608598542127,
"acc_norm": 0.28,
"acc_norm_stderr": 0.04512608598542127
},
"harness|hendrycksTest-college_computer_science|5": {
"acc": 0.4,
"acc_stderr": 0.049236596391733084,
"acc_norm": 0.4,
"acc_norm_stderr": 0.049236596391733084
},
"harness|hendrycksTest-college_mathematics|5": {
"acc": 0.27,
"acc_stderr": 0.044619604333847394,
"acc_norm": 0.27,
"acc_norm_stderr": 0.044619604333847394
},
"harness|hendrycksTest-college_medicine|5": {
"acc": 0.3988439306358382,
"acc_stderr": 0.03733626655383509,
"acc_norm": 0.3988439306358382,
"acc_norm_stderr": 0.03733626655383509
},
"harness|hendrycksTest-college_physics|5": {
"acc": 0.24509803921568626,
"acc_stderr": 0.04280105837364395,
"acc_norm": 0.24509803921568626,
"acc_norm_stderr": 0.04280105837364395
},
"harness|hendrycksTest-computer_security|5": {
"acc": 0.53,
"acc_stderr": 0.05016135580465919,
"acc_norm": 0.53,
"acc_norm_stderr": 0.05016135580465919
},
"harness|hendrycksTest-conceptual_physics|5": {
"acc": 0.39574468085106385,
"acc_stderr": 0.03196758697835362,
"acc_norm": 0.39574468085106385,
"acc_norm_stderr": 0.03196758697835362
},
"harness|hendrycksTest-econometrics|5": {
"acc": 0.3333333333333333,
"acc_stderr": 0.044346007015849245,
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.044346007015849245
},
"harness|hendrycksTest-electrical_engineering|5": {
"acc": 0.4206896551724138,
"acc_stderr": 0.0411391498118926,
"acc_norm": 0.4206896551724138,
"acc_norm_stderr": 0.0411391498118926
},
"harness|hendrycksTest-elementary_mathematics|5": {
"acc": 0.30158730158730157,
"acc_stderr": 0.023636975996101806,
"acc_norm": 0.30158730158730157,
"acc_norm_stderr": 0.023636975996101806
},
"harness|hendrycksTest-formal_logic|5": {
"acc": 0.1984126984126984,
"acc_stderr": 0.03567016675276864,
"acc_norm": 0.1984126984126984,
"acc_norm_stderr": 0.03567016675276864
},
"harness|hendrycksTest-global_facts|5": {
"acc": 0.33,
"acc_stderr": 0.04725815626252604,
"acc_norm": 0.33,
"acc_norm_stderr": 0.04725815626252604
},
"harness|hendrycksTest-high_school_biology|5": {
"acc": 0.535483870967742,
"acc_stderr": 0.02837228779796293,
"acc_norm": 0.535483870967742,
"acc_norm_stderr": 0.02837228779796293
},
"harness|hendrycksTest-high_school_chemistry|5": {
"acc": 0.35960591133004927,
"acc_stderr": 0.03376458246509567,
"acc_norm": 0.35960591133004927,
"acc_norm_stderr": 0.03376458246509567
},
"harness|hendrycksTest-high_school_computer_science|5": {
"acc": 0.48,
"acc_stderr": 0.050211673156867795,
"acc_norm": 0.48,
"acc_norm_stderr": 0.050211673156867795
},
"harness|hendrycksTest-high_school_european_history|5": {
"acc": 0.6424242424242425,
"acc_stderr": 0.037425970438065864,
"acc_norm": 0.6424242424242425,
"acc_norm_stderr": 0.037425970438065864
},
"harness|hendrycksTest-high_school_geography|5": {
"acc": 0.5454545454545454,
"acc_stderr": 0.03547601494006937,
"acc_norm": 0.5454545454545454,
"acc_norm_stderr": 0.03547601494006937
},
"harness|hendrycksTest-high_school_government_and_politics|5": {
"acc": 0.6735751295336787,
"acc_stderr": 0.033840286211432945,
"acc_norm": 0.6735751295336787,
"acc_norm_stderr": 0.033840286211432945
},
"harness|hendrycksTest-high_school_macroeconomics|5": {
"acc": 0.38974358974358975,
"acc_stderr": 0.024726967886647078,
"acc_norm": 0.38974358974358975,
"acc_norm_stderr": 0.024726967886647078
},
"harness|hendrycksTest-high_school_mathematics|5": {
"acc": 0.24444444444444444,
"acc_stderr": 0.02620276653465215,
"acc_norm": 0.24444444444444444,
"acc_norm_stderr": 0.02620276653465215
},
"harness|hendrycksTest-high_school_microeconomics|5": {
"acc": 0.4117647058823529,
"acc_stderr": 0.03196876989195778,
"acc_norm": 0.4117647058823529,
"acc_norm_stderr": 0.03196876989195778
},
"harness|hendrycksTest-high_school_physics|5": {
"acc": 0.2913907284768212,
"acc_stderr": 0.03710185726119995,
"acc_norm": 0.2913907284768212,
"acc_norm_stderr": 0.03710185726119995
},
"harness|hendrycksTest-high_school_psychology|5": {
"acc": 0.6440366972477064,
"acc_stderr": 0.020528559278244214,
"acc_norm": 0.6440366972477064,
"acc_norm_stderr": 0.020528559278244214
},
"harness|hendrycksTest-high_school_statistics|5": {
"acc": 0.2916666666666667,
"acc_stderr": 0.03099866630456053,
"acc_norm": 0.2916666666666667,
"acc_norm_stderr": 0.03099866630456053
},
"harness|hendrycksTest-high_school_us_history|5": {
"acc": 0.6029411764705882,
"acc_stderr": 0.0343413116471913,
"acc_norm": 0.6029411764705882,
"acc_norm_stderr": 0.0343413116471913
},
"harness|hendrycksTest-high_school_world_history|5": {
"acc": 0.6455696202531646,
"acc_stderr": 0.031137304297185815,
"acc_norm": 0.6455696202531646,
"acc_norm_stderr": 0.031137304297185815
},
"harness|hendrycksTest-human_aging|5": {
"acc": 0.5426008968609866,
"acc_stderr": 0.033435777055830646,
"acc_norm": 0.5426008968609866,
"acc_norm_stderr": 0.033435777055830646
},
"harness|hendrycksTest-human_sexuality|5": {
"acc": 0.5725190839694656,
"acc_stderr": 0.04338920305792401,
"acc_norm": 0.5725190839694656,
"acc_norm_stderr": 0.04338920305792401
},
"harness|hendrycksTest-international_law|5": {
"acc": 0.6363636363636364,
"acc_stderr": 0.043913262867240704,
"acc_norm": 0.6363636363636364,
"acc_norm_stderr": 0.043913262867240704
},
"harness|hendrycksTest-jurisprudence|5": {
"acc": 0.5092592592592593,
"acc_stderr": 0.04832853553437055,
"acc_norm": 0.5092592592592593,
"acc_norm_stderr": 0.04832853553437055
},
"harness|hendrycksTest-logical_fallacies|5": {
"acc": 0.49693251533742333,
"acc_stderr": 0.03928297078179663,
"acc_norm": 0.49693251533742333,
"acc_norm_stderr": 0.03928297078179663
},
"harness|hendrycksTest-machine_learning|5": {
"acc": 0.23214285714285715,
"acc_stderr": 0.04007341809755806,
"acc_norm": 0.23214285714285715,
"acc_norm_stderr": 0.04007341809755806
},
"harness|hendrycksTest-management|5": {
"acc": 0.6601941747572816,
"acc_stderr": 0.04689765937278135,
"acc_norm": 0.6601941747572816,
"acc_norm_stderr": 0.04689765937278135
},
"harness|hendrycksTest-marketing|5": {
"acc": 0.7307692307692307,
"acc_stderr": 0.029058588303748842,
"acc_norm": 0.7307692307692307,
"acc_norm_stderr": 0.029058588303748842
},
"harness|hendrycksTest-medical_genetics|5": {
"acc": 0.49,
"acc_stderr": 0.05024183937956911,
"acc_norm": 0.49,
"acc_norm_stderr": 0.05024183937956911
},
"harness|hendrycksTest-miscellaneous|5": {
"acc": 0.6883780332056194,
"acc_stderr": 0.016562433867284176,
"acc_norm": 0.6883780332056194,
"acc_norm_stderr": 0.016562433867284176
},
"harness|hendrycksTest-moral_disputes|5": {
"acc": 0.49421965317919075,
"acc_stderr": 0.02691729617914911,
"acc_norm": 0.49421965317919075,
"acc_norm_stderr": 0.02691729617914911
},
"harness|hendrycksTest-moral_scenarios|5": {
"acc": 0.2536312849162011,
"acc_stderr": 0.014551553659369922,
"acc_norm": 0.2536312849162011,
"acc_norm_stderr": 0.014551553659369922
},
"harness|hendrycksTest-nutrition|5": {
"acc": 0.5130718954248366,
"acc_stderr": 0.028620130800700246,
"acc_norm": 0.5130718954248366,
"acc_norm_stderr": 0.028620130800700246
},
"harness|hendrycksTest-philosophy|5": {
"acc": 0.5594855305466238,
"acc_stderr": 0.02819640057419743,
"acc_norm": 0.5594855305466238,
"acc_norm_stderr": 0.02819640057419743
},
"harness|hendrycksTest-prehistory|5": {
"acc": 0.5308641975308642,
"acc_stderr": 0.027767689606833932,
"acc_norm": 0.5308641975308642,
"acc_norm_stderr": 0.027767689606833932
},
"harness|hendrycksTest-professional_accounting|5": {
"acc": 0.34397163120567376,
"acc_stderr": 0.028338017428611327,
"acc_norm": 0.34397163120567376,
"acc_norm_stderr": 0.028338017428611327
},
"harness|hendrycksTest-professional_law|5": {
"acc": 0.31877444589308995,
"acc_stderr": 0.011901895635786097,
"acc_norm": 0.31877444589308995,
"acc_norm_stderr": 0.011901895635786097
},
"harness|hendrycksTest-professional_medicine|5": {
"acc": 0.4264705882352941,
"acc_stderr": 0.030042615832714878,
"acc_norm": 0.4264705882352941,
"acc_norm_stderr": 0.030042615832714878
},
"harness|hendrycksTest-professional_psychology|5": {
"acc": 0.4493464052287582,
"acc_stderr": 0.020123766528027266,
"acc_norm": 0.4493464052287582,
"acc_norm_stderr": 0.020123766528027266
},
"harness|hendrycksTest-public_relations|5": {
"acc": 0.4727272727272727,
"acc_stderr": 0.04782001791380063,
"acc_norm": 0.4727272727272727,
"acc_norm_stderr": 0.04782001791380063
},
"harness|hendrycksTest-security_studies|5": {
"acc": 0.5265306122448979,
"acc_stderr": 0.03196412734523272,
"acc_norm": 0.5265306122448979,
"acc_norm_stderr": 0.03196412734523272
},
"harness|hendrycksTest-sociology|5": {
"acc": 0.6119402985074627,
"acc_stderr": 0.034457899643627506,
"acc_norm": 0.6119402985074627,
"acc_norm_stderr": 0.034457899643627506
},
"harness|hendrycksTest-us_foreign_policy|5": {
"acc": 0.67,
"acc_stderr": 0.04725815626252607,
"acc_norm": 0.67,
"acc_norm_stderr": 0.04725815626252607
},
"harness|hendrycksTest-virology|5": {
"acc": 0.4036144578313253,
"acc_stderr": 0.038194861407583984,
"acc_norm": 0.4036144578313253,
"acc_norm_stderr": 0.038194861407583984
},
"harness|hendrycksTest-world_religions|5": {
"acc": 0.672514619883041,
"acc_stderr": 0.035993357714560276,
"acc_norm": 0.672514619883041,
"acc_norm_stderr": 0.035993357714560276
},
"harness|truthfulqa:mc|0": {
"mc1": 0.204406364749082,
"mc1_stderr": 0.014117174337432616,
"mc2": 0.3089869590457097,
"mc2_stderr": 0.013843169413571187
},
"harness|winogrande|5": {
"acc": 0.6945540647198106,
"acc_stderr": 0.012945038632552022
},
"harness|drop|3": {
"em": 0.3115562080536913,
"em_stderr": 0.004742879599828378,
"f1": 0.3699653942953032,
"f1_stderr": 0.004671420668393907
},
"harness|gsm8k|5": {
"acc": 0.0,
"acc_stderr": 0.0
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] | [
-0.6787573099136353,
-0.8465359807014465,
0.30309465527534485,
0.22041061520576477,
-0.20234054327011108,
-0.04196353256702423,
0.005400375928729773,
-0.25119346380233765,
0.5912833213806152,
-0.012968511320650578,
-0.4790972173213959,
-0.7067990303039551,
-0.4407240152359009,
0.2757026255... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
open-llm-leaderboard/details_luffycodes__llama-shishya-7b-ep3-v2_public | open-llm-leaderboard | 2023-11-09T13:01:16Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T13:01:16Z | 2023-11-09T13:00:12.000Z | 2023-11-09T13:00:12 | ---
pretty_name: Evaluation run of luffycodes/llama-shishya-7b-ep3-v2
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [luffycodes/llama-shishya-7b-ep3-v2](https://huggingface.co/luffycodes/llama-shishya-7b-ep3-v2)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 64 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the aggregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_luffycodes__llama-shishya-7b-ep3-v2_public\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-11-09T12:57:06.707192](https://huggingface.co/datasets/open-llm-leaderboard/details_luffycodes__llama-shishya-7b-ep3-v2_public/blob/main/results_2023-11-09T12-57-06.707192.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"acc\": 0.43776292457137356,\n\
\ \"acc_stderr\": 0.03405236312139111,\n \"acc_norm\": 0.44440566326787106,\n\
\ \"acc_norm_stderr\": 0.03497626520397757,\n \"mc1\": 0.19583843329253367,\n\
\ \"mc1_stderr\": 0.01389234436774209,\n \"mc2\": 0.3016304809342682,\n\
\ \"mc2_stderr\": 0.013699598037265183,\n \"em\": 0.30557885906040266,\n\
\ \"em_stderr\": 0.004717509363446725,\n \"f1\": 0.36205327181208175,\n\
\ \"f1_stderr\": 0.004656030495449622\n },\n \"harness|arc:challenge|25\"\
: {\n \"acc\": 0.44197952218430037,\n \"acc_stderr\": 0.014512682523128345,\n\
\ \"acc_norm\": 0.4735494880546075,\n \"acc_norm_stderr\": 0.014590931358120172\n\
\ },\n \"harness|hellaswag|10\": {\n \"acc\": 0.5865365465046803,\n\
\ \"acc_stderr\": 0.004914480534533716,\n \"acc_norm\": 0.7588129854610636,\n\
\ \"acc_norm_stderr\": 0.004269291950109927\n },\n \"harness|hendrycksTest-abstract_algebra|5\"\
: {\n \"acc\": 0.26,\n \"acc_stderr\": 0.04408440022768081,\n \
\ \"acc_norm\": 0.26,\n \"acc_norm_stderr\": 0.04408440022768081\n \
\ },\n \"harness|hendrycksTest-anatomy|5\": {\n \"acc\": 0.4222222222222222,\n\
\ \"acc_stderr\": 0.04266763404099582,\n \"acc_norm\": 0.4222222222222222,\n\
\ \"acc_norm_stderr\": 0.04266763404099582\n },\n \"harness|hendrycksTest-astronomy|5\"\
: {\n \"acc\": 0.4144736842105263,\n \"acc_stderr\": 0.040089737857792046,\n\
\ \"acc_norm\": 0.4144736842105263,\n \"acc_norm_stderr\": 0.040089737857792046\n\
\ },\n \"harness|hendrycksTest-business_ethics|5\": {\n \"acc\": 0.38,\n\
\ \"acc_stderr\": 0.04878317312145632,\n \"acc_norm\": 0.38,\n \
\ \"acc_norm_stderr\": 0.04878317312145632\n },\n \"harness|hendrycksTest-clinical_knowledge|5\"\
: {\n \"acc\": 0.4716981132075472,\n \"acc_stderr\": 0.0307235352490061,\n\
\ \"acc_norm\": 0.4716981132075472,\n \"acc_norm_stderr\": 0.0307235352490061\n\
\ },\n \"harness|hendrycksTest-college_biology|5\": {\n \"acc\": 0.4236111111111111,\n\
\ \"acc_stderr\": 0.04132125019723369,\n \"acc_norm\": 0.4236111111111111,\n\
\ \"acc_norm_stderr\": 0.04132125019723369\n },\n \"harness|hendrycksTest-college_chemistry|5\"\
: {\n \"acc\": 0.28,\n \"acc_stderr\": 0.04512608598542128,\n \
\ \"acc_norm\": 0.28,\n \"acc_norm_stderr\": 0.04512608598542128\n \
\ },\n \"harness|hendrycksTest-college_computer_science|5\": {\n \"acc\"\
: 0.39,\n \"acc_stderr\": 0.04902071300001975,\n \"acc_norm\": 0.39,\n\
\ \"acc_norm_stderr\": 0.04902071300001975\n },\n \"harness|hendrycksTest-college_mathematics|5\"\
: {\n \"acc\": 0.26,\n \"acc_stderr\": 0.04408440022768077,\n \
\ \"acc_norm\": 0.26,\n \"acc_norm_stderr\": 0.04408440022768077\n \
\ },\n \"harness|hendrycksTest-college_medicine|5\": {\n \"acc\": 0.41040462427745666,\n\
\ \"acc_stderr\": 0.03750757044895537,\n \"acc_norm\": 0.41040462427745666,\n\
\ \"acc_norm_stderr\": 0.03750757044895537\n },\n \"harness|hendrycksTest-college_physics|5\"\
: {\n \"acc\": 0.23529411764705882,\n \"acc_stderr\": 0.04220773659171452,\n\
\ \"acc_norm\": 0.23529411764705882,\n \"acc_norm_stderr\": 0.04220773659171452\n\
\ },\n \"harness|hendrycksTest-computer_security|5\": {\n \"acc\":\
\ 0.54,\n \"acc_stderr\": 0.05009082659620333,\n \"acc_norm\": 0.54,\n\
\ \"acc_norm_stderr\": 0.05009082659620333\n },\n \"harness|hendrycksTest-conceptual_physics|5\"\
: {\n \"acc\": 0.39574468085106385,\n \"acc_stderr\": 0.03196758697835362,\n\
\ \"acc_norm\": 0.39574468085106385,\n \"acc_norm_stderr\": 0.03196758697835362\n\
\ },\n \"harness|hendrycksTest-econometrics|5\": {\n \"acc\": 0.2894736842105263,\n\
\ \"acc_stderr\": 0.042663394431593935,\n \"acc_norm\": 0.2894736842105263,\n\
\ \"acc_norm_stderr\": 0.042663394431593935\n },\n \"harness|hendrycksTest-electrical_engineering|5\"\
: {\n \"acc\": 0.3931034482758621,\n \"acc_stderr\": 0.0407032901370707,\n\
\ \"acc_norm\": 0.3931034482758621,\n \"acc_norm_stderr\": 0.0407032901370707\n\
\ },\n \"harness|hendrycksTest-elementary_mathematics|5\": {\n \"acc\"\
: 0.30423280423280424,\n \"acc_stderr\": 0.023695415009463087,\n \"\
acc_norm\": 0.30423280423280424,\n \"acc_norm_stderr\": 0.023695415009463087\n\
\ },\n \"harness|hendrycksTest-formal_logic|5\": {\n \"acc\": 0.19047619047619047,\n\
\ \"acc_stderr\": 0.035122074123020514,\n \"acc_norm\": 0.19047619047619047,\n\
\ \"acc_norm_stderr\": 0.035122074123020514\n },\n \"harness|hendrycksTest-global_facts|5\"\
: {\n \"acc\": 0.33,\n \"acc_stderr\": 0.04725815626252604,\n \
\ \"acc_norm\": 0.33,\n \"acc_norm_stderr\": 0.04725815626252604\n \
\ },\n \"harness|hendrycksTest-high_school_biology|5\": {\n \"acc\": 0.5096774193548387,\n\
\ \"acc_stderr\": 0.02843867799890955,\n \"acc_norm\": 0.5096774193548387,\n\
\ \"acc_norm_stderr\": 0.02843867799890955\n },\n \"harness|hendrycksTest-high_school_chemistry|5\"\
: {\n \"acc\": 0.32019704433497537,\n \"acc_stderr\": 0.032826493853041504,\n\
\ \"acc_norm\": 0.32019704433497537,\n \"acc_norm_stderr\": 0.032826493853041504\n\
\ },\n \"harness|hendrycksTest-high_school_computer_science|5\": {\n \
\ \"acc\": 0.44,\n \"acc_stderr\": 0.04988876515698589,\n \"acc_norm\"\
: 0.44,\n \"acc_norm_stderr\": 0.04988876515698589\n },\n \"harness|hendrycksTest-high_school_european_history|5\"\
: {\n \"acc\": 0.5696969696969697,\n \"acc_stderr\": 0.03866225962879077,\n\
\ \"acc_norm\": 0.5696969696969697,\n \"acc_norm_stderr\": 0.03866225962879077\n\
\ },\n \"harness|hendrycksTest-high_school_geography|5\": {\n \"acc\"\
: 0.51010101010101,\n \"acc_stderr\": 0.035616254886737454,\n \"acc_norm\"\
: 0.51010101010101,\n \"acc_norm_stderr\": 0.035616254886737454\n },\n\
\ \"harness|hendrycksTest-high_school_government_and_politics|5\": {\n \
\ \"acc\": 0.5854922279792746,\n \"acc_stderr\": 0.035553003195576686,\n\
\ \"acc_norm\": 0.5854922279792746,\n \"acc_norm_stderr\": 0.035553003195576686\n\
\ },\n \"harness|hendrycksTest-high_school_macroeconomics|5\": {\n \
\ \"acc\": 0.37435897435897436,\n \"acc_stderr\": 0.024537591572830517,\n\
\ \"acc_norm\": 0.37435897435897436,\n \"acc_norm_stderr\": 0.024537591572830517\n\
\ },\n \"harness|hendrycksTest-high_school_mathematics|5\": {\n \"\
acc\": 0.27037037037037037,\n \"acc_stderr\": 0.027080372815145665,\n \
\ \"acc_norm\": 0.27037037037037037,\n \"acc_norm_stderr\": 0.027080372815145665\n\
\ },\n \"harness|hendrycksTest-high_school_microeconomics|5\": {\n \
\ \"acc\": 0.42016806722689076,\n \"acc_stderr\": 0.03206183783236153,\n\
\ \"acc_norm\": 0.42016806722689076,\n \"acc_norm_stderr\": 0.03206183783236153\n\
\ },\n \"harness|hendrycksTest-high_school_physics|5\": {\n \"acc\"\
: 0.31788079470198677,\n \"acc_stderr\": 0.03802039760107903,\n \"\
acc_norm\": 0.31788079470198677,\n \"acc_norm_stderr\": 0.03802039760107903\n\
\ },\n \"harness|hendrycksTest-high_school_psychology|5\": {\n \"acc\"\
: 0.5944954128440367,\n \"acc_stderr\": 0.021050997991896834,\n \"\
acc_norm\": 0.5944954128440367,\n \"acc_norm_stderr\": 0.021050997991896834\n\
\ },\n \"harness|hendrycksTest-high_school_statistics|5\": {\n \"acc\"\
: 0.2175925925925926,\n \"acc_stderr\": 0.028139689444859672,\n \"\
acc_norm\": 0.2175925925925926,\n \"acc_norm_stderr\": 0.028139689444859672\n\
\ },\n \"harness|hendrycksTest-high_school_us_history|5\": {\n \"acc\"\
: 0.5784313725490197,\n \"acc_stderr\": 0.03465868196380762,\n \"\
acc_norm\": 0.5784313725490197,\n \"acc_norm_stderr\": 0.03465868196380762\n\
\ },\n \"harness|hendrycksTest-high_school_world_history|5\": {\n \"\
acc\": 0.6835443037974683,\n \"acc_stderr\": 0.03027497488021898,\n \
\ \"acc_norm\": 0.6835443037974683,\n \"acc_norm_stderr\": 0.03027497488021898\n\
\ },\n \"harness|hendrycksTest-human_aging|5\": {\n \"acc\": 0.5605381165919282,\n\
\ \"acc_stderr\": 0.03331092511038179,\n \"acc_norm\": 0.5605381165919282,\n\
\ \"acc_norm_stderr\": 0.03331092511038179\n },\n \"harness|hendrycksTest-human_sexuality|5\"\
: {\n \"acc\": 0.48091603053435117,\n \"acc_stderr\": 0.04382094705550988,\n\
\ \"acc_norm\": 0.48091603053435117,\n \"acc_norm_stderr\": 0.04382094705550988\n\
\ },\n \"harness|hendrycksTest-international_law|5\": {\n \"acc\":\
\ 0.628099173553719,\n \"acc_stderr\": 0.04412015806624504,\n \"acc_norm\"\
: 0.628099173553719,\n \"acc_norm_stderr\": 0.04412015806624504\n },\n\
\ \"harness|hendrycksTest-jurisprudence|5\": {\n \"acc\": 0.49074074074074076,\n\
\ \"acc_stderr\": 0.04832853553437055,\n \"acc_norm\": 0.49074074074074076,\n\
\ \"acc_norm_stderr\": 0.04832853553437055\n },\n \"harness|hendrycksTest-logical_fallacies|5\"\
: {\n \"acc\": 0.44171779141104295,\n \"acc_stderr\": 0.039015918258361836,\n\
\ \"acc_norm\": 0.44171779141104295,\n \"acc_norm_stderr\": 0.039015918258361836\n\
\ },\n \"harness|hendrycksTest-machine_learning|5\": {\n \"acc\": 0.32142857142857145,\n\
\ \"acc_stderr\": 0.044328040552915185,\n \"acc_norm\": 0.32142857142857145,\n\
\ \"acc_norm_stderr\": 0.044328040552915185\n },\n \"harness|hendrycksTest-management|5\"\
: {\n \"acc\": 0.6116504854368932,\n \"acc_stderr\": 0.0482572933735639,\n\
\ \"acc_norm\": 0.6116504854368932,\n \"acc_norm_stderr\": 0.0482572933735639\n\
\ },\n \"harness|hendrycksTest-marketing|5\": {\n \"acc\": 0.6923076923076923,\n\
\ \"acc_stderr\": 0.03023638994217308,\n \"acc_norm\": 0.6923076923076923,\n\
\ \"acc_norm_stderr\": 0.03023638994217308\n },\n \"harness|hendrycksTest-medical_genetics|5\"\
: {\n \"acc\": 0.47,\n \"acc_stderr\": 0.050161355804659205,\n \
\ \"acc_norm\": 0.47,\n \"acc_norm_stderr\": 0.050161355804659205\n \
\ },\n \"harness|hendrycksTest-miscellaneous|5\": {\n \"acc\": 0.6590038314176245,\n\
\ \"acc_stderr\": 0.016951781383223313,\n \"acc_norm\": 0.6590038314176245,\n\
\ \"acc_norm_stderr\": 0.016951781383223313\n },\n \"harness|hendrycksTest-moral_disputes|5\"\
: {\n \"acc\": 0.49421965317919075,\n \"acc_stderr\": 0.026917296179149116,\n\
\ \"acc_norm\": 0.49421965317919075,\n \"acc_norm_stderr\": 0.026917296179149116\n\
\ },\n \"harness|hendrycksTest-moral_scenarios|5\": {\n \"acc\": 0.26145251396648045,\n\
\ \"acc_stderr\": 0.014696599650364546,\n \"acc_norm\": 0.26145251396648045,\n\
\ \"acc_norm_stderr\": 0.014696599650364546\n },\n \"harness|hendrycksTest-nutrition|5\"\
: {\n \"acc\": 0.46405228758169936,\n \"acc_stderr\": 0.02855582751652878,\n\
\ \"acc_norm\": 0.46405228758169936,\n \"acc_norm_stderr\": 0.02855582751652878\n\
\ },\n \"harness|hendrycksTest-philosophy|5\": {\n \"acc\": 0.5369774919614148,\n\
\ \"acc_stderr\": 0.028320325830105908,\n \"acc_norm\": 0.5369774919614148,\n\
\ \"acc_norm_stderr\": 0.028320325830105908\n },\n \"harness|hendrycksTest-prehistory|5\"\
: {\n \"acc\": 0.4845679012345679,\n \"acc_stderr\": 0.02780749004427621,\n\
\ \"acc_norm\": 0.4845679012345679,\n \"acc_norm_stderr\": 0.02780749004427621\n\
\ },\n \"harness|hendrycksTest-professional_accounting|5\": {\n \"\
acc\": 0.36524822695035464,\n \"acc_stderr\": 0.02872386385328128,\n \
\ \"acc_norm\": 0.36524822695035464,\n \"acc_norm_stderr\": 0.02872386385328128\n\
\ },\n \"harness|hendrycksTest-professional_law|5\": {\n \"acc\": 0.3044328552803129,\n\
\ \"acc_stderr\": 0.01175287759259757,\n \"acc_norm\": 0.3044328552803129,\n\
\ \"acc_norm_stderr\": 0.01175287759259757\n },\n \"harness|hendrycksTest-professional_medicine|5\"\
: {\n \"acc\": 0.33088235294117646,\n \"acc_stderr\": 0.02858270975389844,\n\
\ \"acc_norm\": 0.33088235294117646,\n \"acc_norm_stderr\": 0.02858270975389844\n\
\ },\n \"harness|hendrycksTest-professional_psychology|5\": {\n \"\
acc\": 0.4199346405228758,\n \"acc_stderr\": 0.019966811178256483,\n \
\ \"acc_norm\": 0.4199346405228758,\n \"acc_norm_stderr\": 0.019966811178256483\n\
\ },\n \"harness|hendrycksTest-public_relations|5\": {\n \"acc\": 0.4909090909090909,\n\
\ \"acc_stderr\": 0.04788339768702861,\n \"acc_norm\": 0.4909090909090909,\n\
\ \"acc_norm_stderr\": 0.04788339768702861\n },\n \"harness|hendrycksTest-security_studies|5\"\
: {\n \"acc\": 0.46530612244897956,\n \"acc_stderr\": 0.03193207024425314,\n\
\ \"acc_norm\": 0.46530612244897956,\n \"acc_norm_stderr\": 0.03193207024425314\n\
\ },\n \"harness|hendrycksTest-sociology|5\": {\n \"acc\": 0.5223880597014925,\n\
\ \"acc_stderr\": 0.03531987930208731,\n \"acc_norm\": 0.5223880597014925,\n\
\ \"acc_norm_stderr\": 0.03531987930208731\n },\n \"harness|hendrycksTest-us_foreign_policy|5\"\
: {\n \"acc\": 0.65,\n \"acc_stderr\": 0.047937248544110196,\n \
\ \"acc_norm\": 0.65,\n \"acc_norm_stderr\": 0.047937248544110196\n \
\ },\n \"harness|hendrycksTest-virology|5\": {\n \"acc\": 0.3855421686746988,\n\
\ \"acc_stderr\": 0.037891344246115496,\n \"acc_norm\": 0.3855421686746988,\n\
\ \"acc_norm_stderr\": 0.037891344246115496\n },\n \"harness|hendrycksTest-world_religions|5\"\
: {\n \"acc\": 0.672514619883041,\n \"acc_stderr\": 0.03599335771456027,\n\
\ \"acc_norm\": 0.672514619883041,\n \"acc_norm_stderr\": 0.03599335771456027\n\
\ },\n \"harness|truthfulqa:mc|0\": {\n \"mc1\": 0.19583843329253367,\n\
\ \"mc1_stderr\": 0.01389234436774209,\n \"mc2\": 0.3016304809342682,\n\
\ \"mc2_stderr\": 0.013699598037265183\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.6874506708760852,\n \"acc_stderr\": 0.013027563620748835\n\
\ },\n \"harness|drop|3\": {\n \"em\": 0.30557885906040266,\n \
\ \"em_stderr\": 0.004717509363446725,\n \"f1\": 0.36205327181208175,\n\
\ \"f1_stderr\": 0.004656030495449622\n },\n \"harness|gsm8k|5\": {\n\
\ \"acc\": 0.0,\n \"acc_stderr\": 0.0\n }\n}\n```"
repo_url: https://huggingface.co/luffycodes/llama-shishya-7b-ep3-v2
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|arc:challenge|25_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_drop_3
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|drop|3_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|drop|3_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|gsm8k|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hellaswag|10_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T12-57-06.707192.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T12-57-06.707192.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- '**/details_harness|winogrande|5_2023-11-09T12-57-06.707192.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-11-09T12-57-06.707192.parquet'
- config_name: results
data_files:
- split: 2023_11_09T12_57_06.707192
path:
- results_2023-11-09T12-57-06.707192.parquet
- split: latest
path:
- results_2023-11-09T12-57-06.707192.parquet
---
# Dataset Card for Evaluation run of luffycodes/llama-shishya-7b-ep3-v2
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/luffycodes/llama-shishya-7b-ep3-v2
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [luffycodes/llama-shishya-7b-ep3-v2](https://huggingface.co/luffycodes/llama-shishya-7b-ep3-v2) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 64 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_luffycodes__llama-shishya-7b-ep3-v2_public",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-11-09T12:57:06.707192](https://huggingface.co/datasets/open-llm-leaderboard/details_luffycodes__llama-shishya-7b-ep3-v2_public/blob/main/results_2023-11-09T12-57-06.707192.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"acc": 0.43776292457137356,
"acc_stderr": 0.03405236312139111,
"acc_norm": 0.44440566326787106,
"acc_norm_stderr": 0.03497626520397757,
"mc1": 0.19583843329253367,
"mc1_stderr": 0.01389234436774209,
"mc2": 0.3016304809342682,
"mc2_stderr": 0.013699598037265183,
"em": 0.30557885906040266,
"em_stderr": 0.004717509363446725,
"f1": 0.36205327181208175,
"f1_stderr": 0.004656030495449622
},
"harness|arc:challenge|25": {
"acc": 0.44197952218430037,
"acc_stderr": 0.014512682523128345,
"acc_norm": 0.4735494880546075,
"acc_norm_stderr": 0.014590931358120172
},
"harness|hellaswag|10": {
"acc": 0.5865365465046803,
"acc_stderr": 0.004914480534533716,
"acc_norm": 0.7588129854610636,
"acc_norm_stderr": 0.004269291950109927
},
"harness|hendrycksTest-abstract_algebra|5": {
"acc": 0.26,
"acc_stderr": 0.04408440022768081,
"acc_norm": 0.26,
"acc_norm_stderr": 0.04408440022768081
},
"harness|hendrycksTest-anatomy|5": {
"acc": 0.4222222222222222,
"acc_stderr": 0.04266763404099582,
"acc_norm": 0.4222222222222222,
"acc_norm_stderr": 0.04266763404099582
},
"harness|hendrycksTest-astronomy|5": {
"acc": 0.4144736842105263,
"acc_stderr": 0.040089737857792046,
"acc_norm": 0.4144736842105263,
"acc_norm_stderr": 0.040089737857792046
},
"harness|hendrycksTest-business_ethics|5": {
"acc": 0.38,
"acc_stderr": 0.04878317312145632,
"acc_norm": 0.38,
"acc_norm_stderr": 0.04878317312145632
},
"harness|hendrycksTest-clinical_knowledge|5": {
"acc": 0.4716981132075472,
"acc_stderr": 0.0307235352490061,
"acc_norm": 0.4716981132075472,
"acc_norm_stderr": 0.0307235352490061
},
"harness|hendrycksTest-college_biology|5": {
"acc": 0.4236111111111111,
"acc_stderr": 0.04132125019723369,
"acc_norm": 0.4236111111111111,
"acc_norm_stderr": 0.04132125019723369
},
"harness|hendrycksTest-college_chemistry|5": {
"acc": 0.28,
"acc_stderr": 0.04512608598542128,
"acc_norm": 0.28,
"acc_norm_stderr": 0.04512608598542128
},
"harness|hendrycksTest-college_computer_science|5": {
"acc": 0.39,
"acc_stderr": 0.04902071300001975,
"acc_norm": 0.39,
"acc_norm_stderr": 0.04902071300001975
},
"harness|hendrycksTest-college_mathematics|5": {
"acc": 0.26,
"acc_stderr": 0.04408440022768077,
"acc_norm": 0.26,
"acc_norm_stderr": 0.04408440022768077
},
"harness|hendrycksTest-college_medicine|5": {
"acc": 0.41040462427745666,
"acc_stderr": 0.03750757044895537,
"acc_norm": 0.41040462427745666,
"acc_norm_stderr": 0.03750757044895537
},
"harness|hendrycksTest-college_physics|5": {
"acc": 0.23529411764705882,
"acc_stderr": 0.04220773659171452,
"acc_norm": 0.23529411764705882,
"acc_norm_stderr": 0.04220773659171452
},
"harness|hendrycksTest-computer_security|5": {
"acc": 0.54,
"acc_stderr": 0.05009082659620333,
"acc_norm": 0.54,
"acc_norm_stderr": 0.05009082659620333
},
"harness|hendrycksTest-conceptual_physics|5": {
"acc": 0.39574468085106385,
"acc_stderr": 0.03196758697835362,
"acc_norm": 0.39574468085106385,
"acc_norm_stderr": 0.03196758697835362
},
"harness|hendrycksTest-econometrics|5": {
"acc": 0.2894736842105263,
"acc_stderr": 0.042663394431593935,
"acc_norm": 0.2894736842105263,
"acc_norm_stderr": 0.042663394431593935
},
"harness|hendrycksTest-electrical_engineering|5": {
"acc": 0.3931034482758621,
"acc_stderr": 0.0407032901370707,
"acc_norm": 0.3931034482758621,
"acc_norm_stderr": 0.0407032901370707
},
"harness|hendrycksTest-elementary_mathematics|5": {
"acc": 0.30423280423280424,
"acc_stderr": 0.023695415009463087,
"acc_norm": 0.30423280423280424,
"acc_norm_stderr": 0.023695415009463087
},
"harness|hendrycksTest-formal_logic|5": {
"acc": 0.19047619047619047,
"acc_stderr": 0.035122074123020514,
"acc_norm": 0.19047619047619047,
"acc_norm_stderr": 0.035122074123020514
},
"harness|hendrycksTest-global_facts|5": {
"acc": 0.33,
"acc_stderr": 0.04725815626252604,
"acc_norm": 0.33,
"acc_norm_stderr": 0.04725815626252604
},
"harness|hendrycksTest-high_school_biology|5": {
"acc": 0.5096774193548387,
"acc_stderr": 0.02843867799890955,
"acc_norm": 0.5096774193548387,
"acc_norm_stderr": 0.02843867799890955
},
"harness|hendrycksTest-high_school_chemistry|5": {
"acc": 0.32019704433497537,
"acc_stderr": 0.032826493853041504,
"acc_norm": 0.32019704433497537,
"acc_norm_stderr": 0.032826493853041504
},
"harness|hendrycksTest-high_school_computer_science|5": {
"acc": 0.44,
"acc_stderr": 0.04988876515698589,
"acc_norm": 0.44,
"acc_norm_stderr": 0.04988876515698589
},
"harness|hendrycksTest-high_school_european_history|5": {
"acc": 0.5696969696969697,
"acc_stderr": 0.03866225962879077,
"acc_norm": 0.5696969696969697,
"acc_norm_stderr": 0.03866225962879077
},
"harness|hendrycksTest-high_school_geography|5": {
"acc": 0.51010101010101,
"acc_stderr": 0.035616254886737454,
"acc_norm": 0.51010101010101,
"acc_norm_stderr": 0.035616254886737454
},
"harness|hendrycksTest-high_school_government_and_politics|5": {
"acc": 0.5854922279792746,
"acc_stderr": 0.035553003195576686,
"acc_norm": 0.5854922279792746,
"acc_norm_stderr": 0.035553003195576686
},
"harness|hendrycksTest-high_school_macroeconomics|5": {
"acc": 0.37435897435897436,
"acc_stderr": 0.024537591572830517,
"acc_norm": 0.37435897435897436,
"acc_norm_stderr": 0.024537591572830517
},
"harness|hendrycksTest-high_school_mathematics|5": {
"acc": 0.27037037037037037,
"acc_stderr": 0.027080372815145665,
"acc_norm": 0.27037037037037037,
"acc_norm_stderr": 0.027080372815145665
},
"harness|hendrycksTest-high_school_microeconomics|5": {
"acc": 0.42016806722689076,
"acc_stderr": 0.03206183783236153,
"acc_norm": 0.42016806722689076,
"acc_norm_stderr": 0.03206183783236153
},
"harness|hendrycksTest-high_school_physics|5": {
"acc": 0.31788079470198677,
"acc_stderr": 0.03802039760107903,
"acc_norm": 0.31788079470198677,
"acc_norm_stderr": 0.03802039760107903
},
"harness|hendrycksTest-high_school_psychology|5": {
"acc": 0.5944954128440367,
"acc_stderr": 0.021050997991896834,
"acc_norm": 0.5944954128440367,
"acc_norm_stderr": 0.021050997991896834
},
"harness|hendrycksTest-high_school_statistics|5": {
"acc": 0.2175925925925926,
"acc_stderr": 0.028139689444859672,
"acc_norm": 0.2175925925925926,
"acc_norm_stderr": 0.028139689444859672
},
"harness|hendrycksTest-high_school_us_history|5": {
"acc": 0.5784313725490197,
"acc_stderr": 0.03465868196380762,
"acc_norm": 0.5784313725490197,
"acc_norm_stderr": 0.03465868196380762
},
"harness|hendrycksTest-high_school_world_history|5": {
"acc": 0.6835443037974683,
"acc_stderr": 0.03027497488021898,
"acc_norm": 0.6835443037974683,
"acc_norm_stderr": 0.03027497488021898
},
"harness|hendrycksTest-human_aging|5": {
"acc": 0.5605381165919282,
"acc_stderr": 0.03331092511038179,
"acc_norm": 0.5605381165919282,
"acc_norm_stderr": 0.03331092511038179
},
"harness|hendrycksTest-human_sexuality|5": {
"acc": 0.48091603053435117,
"acc_stderr": 0.04382094705550988,
"acc_norm": 0.48091603053435117,
"acc_norm_stderr": 0.04382094705550988
},
"harness|hendrycksTest-international_law|5": {
"acc": 0.628099173553719,
"acc_stderr": 0.04412015806624504,
"acc_norm": 0.628099173553719,
"acc_norm_stderr": 0.04412015806624504
},
"harness|hendrycksTest-jurisprudence|5": {
"acc": 0.49074074074074076,
"acc_stderr": 0.04832853553437055,
"acc_norm": 0.49074074074074076,
"acc_norm_stderr": 0.04832853553437055
},
"harness|hendrycksTest-logical_fallacies|5": {
"acc": 0.44171779141104295,
"acc_stderr": 0.039015918258361836,
"acc_norm": 0.44171779141104295,
"acc_norm_stderr": 0.039015918258361836
},
"harness|hendrycksTest-machine_learning|5": {
"acc": 0.32142857142857145,
"acc_stderr": 0.044328040552915185,
"acc_norm": 0.32142857142857145,
"acc_norm_stderr": 0.044328040552915185
},
"harness|hendrycksTest-management|5": {
"acc": 0.6116504854368932,
"acc_stderr": 0.0482572933735639,
"acc_norm": 0.6116504854368932,
"acc_norm_stderr": 0.0482572933735639
},
"harness|hendrycksTest-marketing|5": {
"acc": 0.6923076923076923,
"acc_stderr": 0.03023638994217308,
"acc_norm": 0.6923076923076923,
"acc_norm_stderr": 0.03023638994217308
},
"harness|hendrycksTest-medical_genetics|5": {
"acc": 0.47,
"acc_stderr": 0.050161355804659205,
"acc_norm": 0.47,
"acc_norm_stderr": 0.050161355804659205
},
"harness|hendrycksTest-miscellaneous|5": {
"acc": 0.6590038314176245,
"acc_stderr": 0.016951781383223313,
"acc_norm": 0.6590038314176245,
"acc_norm_stderr": 0.016951781383223313
},
"harness|hendrycksTest-moral_disputes|5": {
"acc": 0.49421965317919075,
"acc_stderr": 0.026917296179149116,
"acc_norm": 0.49421965317919075,
"acc_norm_stderr": 0.026917296179149116
},
"harness|hendrycksTest-moral_scenarios|5": {
"acc": 0.26145251396648045,
"acc_stderr": 0.014696599650364546,
"acc_norm": 0.26145251396648045,
"acc_norm_stderr": 0.014696599650364546
},
"harness|hendrycksTest-nutrition|5": {
"acc": 0.46405228758169936,
"acc_stderr": 0.02855582751652878,
"acc_norm": 0.46405228758169936,
"acc_norm_stderr": 0.02855582751652878
},
"harness|hendrycksTest-philosophy|5": {
"acc": 0.5369774919614148,
"acc_stderr": 0.028320325830105908,
"acc_norm": 0.5369774919614148,
"acc_norm_stderr": 0.028320325830105908
},
"harness|hendrycksTest-prehistory|5": {
"acc": 0.4845679012345679,
"acc_stderr": 0.02780749004427621,
"acc_norm": 0.4845679012345679,
"acc_norm_stderr": 0.02780749004427621
},
"harness|hendrycksTest-professional_accounting|5": {
"acc": 0.36524822695035464,
"acc_stderr": 0.02872386385328128,
"acc_norm": 0.36524822695035464,
"acc_norm_stderr": 0.02872386385328128
},
"harness|hendrycksTest-professional_law|5": {
"acc": 0.3044328552803129,
"acc_stderr": 0.01175287759259757,
"acc_norm": 0.3044328552803129,
"acc_norm_stderr": 0.01175287759259757
},
"harness|hendrycksTest-professional_medicine|5": {
"acc": 0.33088235294117646,
"acc_stderr": 0.02858270975389844,
"acc_norm": 0.33088235294117646,
"acc_norm_stderr": 0.02858270975389844
},
"harness|hendrycksTest-professional_psychology|5": {
"acc": 0.4199346405228758,
"acc_stderr": 0.019966811178256483,
"acc_norm": 0.4199346405228758,
"acc_norm_stderr": 0.019966811178256483
},
"harness|hendrycksTest-public_relations|5": {
"acc": 0.4909090909090909,
"acc_stderr": 0.04788339768702861,
"acc_norm": 0.4909090909090909,
"acc_norm_stderr": 0.04788339768702861
},
"harness|hendrycksTest-security_studies|5": {
"acc": 0.46530612244897956,
"acc_stderr": 0.03193207024425314,
"acc_norm": 0.46530612244897956,
"acc_norm_stderr": 0.03193207024425314
},
"harness|hendrycksTest-sociology|5": {
"acc": 0.5223880597014925,
"acc_stderr": 0.03531987930208731,
"acc_norm": 0.5223880597014925,
"acc_norm_stderr": 0.03531987930208731
},
"harness|hendrycksTest-us_foreign_policy|5": {
"acc": 0.65,
"acc_stderr": 0.047937248544110196,
"acc_norm": 0.65,
"acc_norm_stderr": 0.047937248544110196
},
"harness|hendrycksTest-virology|5": {
"acc": 0.3855421686746988,
"acc_stderr": 0.037891344246115496,
"acc_norm": 0.3855421686746988,
"acc_norm_stderr": 0.037891344246115496
},
"harness|hendrycksTest-world_religions|5": {
"acc": 0.672514619883041,
"acc_stderr": 0.03599335771456027,
"acc_norm": 0.672514619883041,
"acc_norm_stderr": 0.03599335771456027
},
"harness|truthfulqa:mc|0": {
"mc1": 0.19583843329253367,
"mc1_stderr": 0.01389234436774209,
"mc2": 0.3016304809342682,
"mc2_stderr": 0.013699598037265183
},
"harness|winogrande|5": {
"acc": 0.6874506708760852,
"acc_stderr": 0.013027563620748835
},
"harness|drop|3": {
"em": 0.30557885906040266,
"em_stderr": 0.004717509363446725,
"f1": 0.36205327181208175,
"f1_stderr": 0.004656030495449622
},
"harness|gsm8k|5": {
"acc": 0.0,
"acc_stderr": 0.0
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] | [
-0.6817039251327515,
-0.8550171256065369,
0.29265034198760986,
0.21951784193515778,
-0.21226823329925537,
-0.04404621571302414,
0.011274543590843678,
-0.26080435514450073,
0.604394793510437,
-0.010770346969366074,
-0.4796794354915619,
-0.707273006439209,
-0.41850998997688293,
0.26189863681... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
jhhon80/jhonathan | jhhon80 | 2023-11-09T13:19:00Z | 0 | 0 | null | [
"license:openrail",
"region:us"
] | 2023-11-09T13:19:00Z | 2023-11-09T13:17:53.000Z | 2023-11-09T13:17:53 | ---
license: openrail
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
open-llm-leaderboard/details_psmathur__model_007_public | open-llm-leaderboard | 2023-11-09T13:26:43Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T13:26:43Z | 2023-11-09T13:26:34.000Z | 2023-11-09T13:26:34 | ---
pretty_name: Evaluation run of psmathur/model_007
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [psmathur/model_007](https://huggingface.co/psmathur/model_007) on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 3 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the aggregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_psmathur__model_007_public\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-11-09T13:26:16.051201](https://huggingface.co/datasets/open-llm-leaderboard/details_psmathur__model_007_public/blob/main/results_2023-11-09T13-26-16.051201.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"em\": 0.13276006711409397,\n\
\ \"em_stderr\": 0.0034749056446198375,\n \"f1\": 0.31045721476510313,\n\
\ \"f1_stderr\": 0.003655086215890851,\n \"acc\": 0.602479216693903,\n\
\ \"acc_stderr\": 0.011890317786243781\n },\n \"harness|drop|3\": {\n\
\ \"em\": 0.13276006711409397,\n \"em_stderr\": 0.0034749056446198375,\n\
\ \"f1\": 0.31045721476510313,\n \"f1_stderr\": 0.003655086215890851\n\
\ },\n \"harness|gsm8k|5\": {\n \"acc\": 0.37149355572403336,\n \
\ \"acc_stderr\": 0.01330983907570648\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.8334648776637726,\n \"acc_stderr\": 0.010470796496781083\n\
\ }\n}\n```"
repo_url: https://huggingface.co/psmathur/model_007
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_drop_3
data_files:
- split: 2023_11_09T13_26_16.051201
path:
- '**/details_harness|drop|3_2023-11-09T13-26-16.051201.parquet'
- split: latest
path:
- '**/details_harness|drop|3_2023-11-09T13-26-16.051201.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_11_09T13_26_16.051201
path:
- '**/details_harness|gsm8k|5_2023-11-09T13-26-16.051201.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-11-09T13-26-16.051201.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_11_09T13_26_16.051201
path:
- '**/details_harness|winogrande|5_2023-11-09T13-26-16.051201.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-11-09T13-26-16.051201.parquet'
- config_name: results
data_files:
- split: 2023_11_09T13_26_16.051201
path:
- results_2023-11-09T13-26-16.051201.parquet
- split: latest
path:
- results_2023-11-09T13-26-16.051201.parquet
---
# Dataset Card for Evaluation run of psmathur/model_007
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/psmathur/model_007
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [psmathur/model_007](https://huggingface.co/psmathur/model_007) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 3 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_psmathur__model_007_public",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-11-09T13:26:16.051201](https://huggingface.co/datasets/open-llm-leaderboard/details_psmathur__model_007_public/blob/main/results_2023-11-09T13-26-16.051201.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"em": 0.13276006711409397,
"em_stderr": 0.0034749056446198375,
"f1": 0.31045721476510313,
"f1_stderr": 0.003655086215890851,
"acc": 0.602479216693903,
"acc_stderr": 0.011890317786243781
},
"harness|drop|3": {
"em": 0.13276006711409397,
"em_stderr": 0.0034749056446198375,
"f1": 0.31045721476510313,
"f1_stderr": 0.003655086215890851
},
"harness|gsm8k|5": {
"acc": 0.37149355572403336,
"acc_stderr": 0.01330983907570648
},
"harness|winogrande|5": {
"acc": 0.8334648776637726,
"acc_stderr": 0.010470796496781083
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] | [
-0.36684277653694153,
-0.5705538988113403,
0.28696927428245544,
0.22378383576869965,
-0.2822236716747284,
0.16335292160511017,
-0.3365996181964874,
-0.0399344339966774,
0.3802487552165985,
0.5301393866539001,
-0.7355412244796753,
-0.8914458155632019,
-0.7145884037017822,
0.1985425651073455... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
open-llm-leaderboard/details_luffycodes__vicuna-shishya-7b-ep3-v1_public | open-llm-leaderboard | 2023-11-09T13:28:58Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T13:28:58Z | 2023-11-09T13:27:53.000Z | 2023-11-09T13:27:53 | ---
pretty_name: Evaluation run of luffycodes/vicuna-shishya-7b-ep3-v1
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [luffycodes/vicuna-shishya-7b-ep3-v1](https://huggingface.co/luffycodes/vicuna-shishya-7b-ep3-v1)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 64 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the aggregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_luffycodes__vicuna-shishya-7b-ep3-v1_public\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-11-09T13:24:49.230828](https://huggingface.co/datasets/open-llm-leaderboard/details_luffycodes__vicuna-shishya-7b-ep3-v1_public/blob/main/results_2023-11-09T13-24-49.230828.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"acc\": 0.49601218028660454,\n\
\ \"acc_stderr\": 0.03399727784474729,\n \"acc_norm\": 0.5041920928165192,\n\
\ \"acc_norm_stderr\": 0.03492449912034474,\n \"mc1\": 0.2631578947368421,\n\
\ \"mc1_stderr\": 0.015415241740237012,\n \"mc2\": 0.4032424062517679,\n\
\ \"mc2_stderr\": 0.014739501986326583,\n \"em\": 0.2950922818791946,\n\
\ \"em_stderr\": 0.004670729426706436,\n \"f1\": 0.3578932466442965,\n\
\ \"f1_stderr\": 0.004607902070294773\n },\n \"harness|arc:challenge|25\"\
: {\n \"acc\": 0.439419795221843,\n \"acc_stderr\": 0.014503747823580129,\n\
\ \"acc_norm\": 0.4590443686006826,\n \"acc_norm_stderr\": 0.014562291073601234\n\
\ },\n \"harness|hellaswag|10\": {\n \"acc\": 0.5751842262497511,\n\
\ \"acc_stderr\": 0.004933047726996794,\n \"acc_norm\": 0.7635929097789285,\n\
\ \"acc_norm_stderr\": 0.004240066898702511\n },\n \"harness|hendrycksTest-abstract_algebra|5\"\
: {\n \"acc\": 0.29,\n \"acc_stderr\": 0.04560480215720684,\n \
\ \"acc_norm\": 0.29,\n \"acc_norm_stderr\": 0.04560480215720684\n \
\ },\n \"harness|hendrycksTest-anatomy|5\": {\n \"acc\": 0.4740740740740741,\n\
\ \"acc_stderr\": 0.04313531696750574,\n \"acc_norm\": 0.4740740740740741,\n\
\ \"acc_norm_stderr\": 0.04313531696750574\n },\n \"harness|hendrycksTest-astronomy|5\"\
: {\n \"acc\": 0.4868421052631579,\n \"acc_stderr\": 0.04067533136309172,\n\
\ \"acc_norm\": 0.4868421052631579,\n \"acc_norm_stderr\": 0.04067533136309172\n\
\ },\n \"harness|hendrycksTest-business_ethics|5\": {\n \"acc\": 0.47,\n\
\ \"acc_stderr\": 0.050161355804659205,\n \"acc_norm\": 0.47,\n \
\ \"acc_norm_stderr\": 0.050161355804659205\n },\n \"harness|hendrycksTest-clinical_knowledge|5\"\
: {\n \"acc\": 0.5169811320754717,\n \"acc_stderr\": 0.030755120364119905,\n\
\ \"acc_norm\": 0.5169811320754717,\n \"acc_norm_stderr\": 0.030755120364119905\n\
\ },\n \"harness|hendrycksTest-college_biology|5\": {\n \"acc\": 0.4861111111111111,\n\
\ \"acc_stderr\": 0.041795966175810016,\n \"acc_norm\": 0.4861111111111111,\n\
\ \"acc_norm_stderr\": 0.041795966175810016\n },\n \"harness|hendrycksTest-college_chemistry|5\"\
: {\n \"acc\": 0.31,\n \"acc_stderr\": 0.04648231987117316,\n \
\ \"acc_norm\": 0.31,\n \"acc_norm_stderr\": 0.04648231987117316\n \
\ },\n \"harness|hendrycksTest-college_computer_science|5\": {\n \"acc\"\
: 0.41,\n \"acc_stderr\": 0.049431107042371025,\n \"acc_norm\": 0.41,\n\
\ \"acc_norm_stderr\": 0.049431107042371025\n },\n \"harness|hendrycksTest-college_mathematics|5\"\
: {\n \"acc\": 0.35,\n \"acc_stderr\": 0.047937248544110196,\n \
\ \"acc_norm\": 0.35,\n \"acc_norm_stderr\": 0.047937248544110196\n \
\ },\n \"harness|hendrycksTest-college_medicine|5\": {\n \"acc\": 0.47398843930635837,\n\
\ \"acc_stderr\": 0.03807301726504511,\n \"acc_norm\": 0.47398843930635837,\n\
\ \"acc_norm_stderr\": 0.03807301726504511\n },\n \"harness|hendrycksTest-college_physics|5\"\
: {\n \"acc\": 0.16666666666666666,\n \"acc_stderr\": 0.03708284662416544,\n\
\ \"acc_norm\": 0.16666666666666666,\n \"acc_norm_stderr\": 0.03708284662416544\n\
\ },\n \"harness|hendrycksTest-computer_security|5\": {\n \"acc\":\
\ 0.63,\n \"acc_stderr\": 0.04852365870939099,\n \"acc_norm\": 0.63,\n\
\ \"acc_norm_stderr\": 0.04852365870939099\n },\n \"harness|hendrycksTest-conceptual_physics|5\"\
: {\n \"acc\": 0.4553191489361702,\n \"acc_stderr\": 0.03255525359340355,\n\
\ \"acc_norm\": 0.4553191489361702,\n \"acc_norm_stderr\": 0.03255525359340355\n\
\ },\n \"harness|hendrycksTest-econometrics|5\": {\n \"acc\": 0.2894736842105263,\n\
\ \"acc_stderr\": 0.04266339443159393,\n \"acc_norm\": 0.2894736842105263,\n\
\ \"acc_norm_stderr\": 0.04266339443159393\n },\n \"harness|hendrycksTest-electrical_engineering|5\"\
: {\n \"acc\": 0.45517241379310347,\n \"acc_stderr\": 0.04149886942192117,\n\
\ \"acc_norm\": 0.45517241379310347,\n \"acc_norm_stderr\": 0.04149886942192117\n\
\ },\n \"harness|hendrycksTest-elementary_mathematics|5\": {\n \"acc\"\
: 0.30158730158730157,\n \"acc_stderr\": 0.023636975996101806,\n \"\
acc_norm\": 0.30158730158730157,\n \"acc_norm_stderr\": 0.023636975996101806\n\
\ },\n \"harness|hendrycksTest-formal_logic|5\": {\n \"acc\": 0.31746031746031744,\n\
\ \"acc_stderr\": 0.04163453031302859,\n \"acc_norm\": 0.31746031746031744,\n\
\ \"acc_norm_stderr\": 0.04163453031302859\n },\n \"harness|hendrycksTest-global_facts|5\"\
: {\n \"acc\": 0.35,\n \"acc_stderr\": 0.047937248544110196,\n \
\ \"acc_norm\": 0.35,\n \"acc_norm_stderr\": 0.047937248544110196\n \
\ },\n \"harness|hendrycksTest-high_school_biology|5\": {\n \"acc\"\
: 0.535483870967742,\n \"acc_stderr\": 0.02837228779796293,\n \"acc_norm\"\
: 0.535483870967742,\n \"acc_norm_stderr\": 0.02837228779796293\n },\n\
\ \"harness|hendrycksTest-high_school_chemistry|5\": {\n \"acc\": 0.3891625615763547,\n\
\ \"acc_stderr\": 0.034304624161038716,\n \"acc_norm\": 0.3891625615763547,\n\
\ \"acc_norm_stderr\": 0.034304624161038716\n },\n \"harness|hendrycksTest-high_school_computer_science|5\"\
: {\n \"acc\": 0.47,\n \"acc_stderr\": 0.050161355804659205,\n \
\ \"acc_norm\": 0.47,\n \"acc_norm_stderr\": 0.050161355804659205\n \
\ },\n \"harness|hendrycksTest-high_school_european_history|5\": {\n \
\ \"acc\": 0.6424242424242425,\n \"acc_stderr\": 0.037425970438065864,\n\
\ \"acc_norm\": 0.6424242424242425,\n \"acc_norm_stderr\": 0.037425970438065864\n\
\ },\n \"harness|hendrycksTest-high_school_geography|5\": {\n \"acc\"\
: 0.5909090909090909,\n \"acc_stderr\": 0.03502975799413007,\n \"\
acc_norm\": 0.5909090909090909,\n \"acc_norm_stderr\": 0.03502975799413007\n\
\ },\n \"harness|hendrycksTest-high_school_government_and_politics|5\": {\n\
\ \"acc\": 0.7305699481865285,\n \"acc_stderr\": 0.03201867122877794,\n\
\ \"acc_norm\": 0.7305699481865285,\n \"acc_norm_stderr\": 0.03201867122877794\n\
\ },\n \"harness|hendrycksTest-high_school_macroeconomics|5\": {\n \
\ \"acc\": 0.4307692307692308,\n \"acc_stderr\": 0.02510682066053975,\n \
\ \"acc_norm\": 0.4307692307692308,\n \"acc_norm_stderr\": 0.02510682066053975\n\
\ },\n \"harness|hendrycksTest-high_school_mathematics|5\": {\n \"\
acc\": 0.25555555555555554,\n \"acc_stderr\": 0.02659393910184408,\n \
\ \"acc_norm\": 0.25555555555555554,\n \"acc_norm_stderr\": 0.02659393910184408\n\
\ },\n \"harness|hendrycksTest-high_school_microeconomics|5\": {\n \
\ \"acc\": 0.4495798319327731,\n \"acc_stderr\": 0.03231293497137707,\n \
\ \"acc_norm\": 0.4495798319327731,\n \"acc_norm_stderr\": 0.03231293497137707\n\
\ },\n \"harness|hendrycksTest-high_school_physics|5\": {\n \"acc\"\
: 0.2781456953642384,\n \"acc_stderr\": 0.03658603262763743,\n \"\
acc_norm\": 0.2781456953642384,\n \"acc_norm_stderr\": 0.03658603262763743\n\
\ },\n \"harness|hendrycksTest-high_school_psychology|5\": {\n \"acc\"\
: 0.6862385321100918,\n \"acc_stderr\": 0.019894723341469116,\n \"\
acc_norm\": 0.6862385321100918,\n \"acc_norm_stderr\": 0.019894723341469116\n\
\ },\n \"harness|hendrycksTest-high_school_statistics|5\": {\n \"acc\"\
: 0.3611111111111111,\n \"acc_stderr\": 0.032757734861009996,\n \"\
acc_norm\": 0.3611111111111111,\n \"acc_norm_stderr\": 0.032757734861009996\n\
\ },\n \"harness|hendrycksTest-high_school_us_history|5\": {\n \"acc\"\
: 0.6813725490196079,\n \"acc_stderr\": 0.03270287181482081,\n \"\
acc_norm\": 0.6813725490196079,\n \"acc_norm_stderr\": 0.03270287181482081\n\
\ },\n \"harness|hendrycksTest-high_school_world_history|5\": {\n \"\
acc\": 0.7341772151898734,\n \"acc_stderr\": 0.02875679962965834,\n \
\ \"acc_norm\": 0.7341772151898734,\n \"acc_norm_stderr\": 0.02875679962965834\n\
\ },\n \"harness|hendrycksTest-human_aging|5\": {\n \"acc\": 0.6188340807174888,\n\
\ \"acc_stderr\": 0.03259625118416827,\n \"acc_norm\": 0.6188340807174888,\n\
\ \"acc_norm_stderr\": 0.03259625118416827\n },\n \"harness|hendrycksTest-human_sexuality|5\"\
: {\n \"acc\": 0.6412213740458015,\n \"acc_stderr\": 0.04206739313864908,\n\
\ \"acc_norm\": 0.6412213740458015,\n \"acc_norm_stderr\": 0.04206739313864908\n\
\ },\n \"harness|hendrycksTest-international_law|5\": {\n \"acc\":\
\ 0.6198347107438017,\n \"acc_stderr\": 0.04431324501968431,\n \"\
acc_norm\": 0.6198347107438017,\n \"acc_norm_stderr\": 0.04431324501968431\n\
\ },\n \"harness|hendrycksTest-jurisprudence|5\": {\n \"acc\": 0.5648148148148148,\n\
\ \"acc_stderr\": 0.04792898170907061,\n \"acc_norm\": 0.5648148148148148,\n\
\ \"acc_norm_stderr\": 0.04792898170907061\n },\n \"harness|hendrycksTest-logical_fallacies|5\"\
: {\n \"acc\": 0.50920245398773,\n \"acc_stderr\": 0.03927705600787443,\n\
\ \"acc_norm\": 0.50920245398773,\n \"acc_norm_stderr\": 0.03927705600787443\n\
\ },\n \"harness|hendrycksTest-machine_learning|5\": {\n \"acc\": 0.41964285714285715,\n\
\ \"acc_stderr\": 0.04684099321077106,\n \"acc_norm\": 0.41964285714285715,\n\
\ \"acc_norm_stderr\": 0.04684099321077106\n },\n \"harness|hendrycksTest-management|5\"\
: {\n \"acc\": 0.6019417475728155,\n \"acc_stderr\": 0.0484674825397724,\n\
\ \"acc_norm\": 0.6019417475728155,\n \"acc_norm_stderr\": 0.0484674825397724\n\
\ },\n \"harness|hendrycksTest-marketing|5\": {\n \"acc\": 0.7393162393162394,\n\
\ \"acc_stderr\": 0.028760348956523414,\n \"acc_norm\": 0.7393162393162394,\n\
\ \"acc_norm_stderr\": 0.028760348956523414\n },\n \"harness|hendrycksTest-medical_genetics|5\"\
: {\n \"acc\": 0.55,\n \"acc_stderr\": 0.049999999999999996,\n \
\ \"acc_norm\": 0.55,\n \"acc_norm_stderr\": 0.049999999999999996\n \
\ },\n \"harness|hendrycksTest-miscellaneous|5\": {\n \"acc\": 0.6934865900383141,\n\
\ \"acc_stderr\": 0.016486952893041504,\n \"acc_norm\": 0.6934865900383141,\n\
\ \"acc_norm_stderr\": 0.016486952893041504\n },\n \"harness|hendrycksTest-moral_disputes|5\"\
: {\n \"acc\": 0.5375722543352601,\n \"acc_stderr\": 0.026842985519615375,\n\
\ \"acc_norm\": 0.5375722543352601,\n \"acc_norm_stderr\": 0.026842985519615375\n\
\ },\n \"harness|hendrycksTest-moral_scenarios|5\": {\n \"acc\": 0.2636871508379888,\n\
\ \"acc_stderr\": 0.014736926383761987,\n \"acc_norm\": 0.2636871508379888,\n\
\ \"acc_norm_stderr\": 0.014736926383761987\n },\n \"harness|hendrycksTest-nutrition|5\"\
: {\n \"acc\": 0.5686274509803921,\n \"acc_stderr\": 0.028358956313423545,\n\
\ \"acc_norm\": 0.5686274509803921,\n \"acc_norm_stderr\": 0.028358956313423545\n\
\ },\n \"harness|hendrycksTest-philosophy|5\": {\n \"acc\": 0.5980707395498392,\n\
\ \"acc_stderr\": 0.02784647600593047,\n \"acc_norm\": 0.5980707395498392,\n\
\ \"acc_norm_stderr\": 0.02784647600593047\n },\n \"harness|hendrycksTest-prehistory|5\"\
: {\n \"acc\": 0.5679012345679012,\n \"acc_stderr\": 0.02756301097160668,\n\
\ \"acc_norm\": 0.5679012345679012,\n \"acc_norm_stderr\": 0.02756301097160668\n\
\ },\n \"harness|hendrycksTest-professional_accounting|5\": {\n \"\
acc\": 0.35815602836879434,\n \"acc_stderr\": 0.028602085862759415,\n \
\ \"acc_norm\": 0.35815602836879434,\n \"acc_norm_stderr\": 0.028602085862759415\n\
\ },\n \"harness|hendrycksTest-professional_law|5\": {\n \"acc\": 0.36897001303780963,\n\
\ \"acc_stderr\": 0.01232393665017486,\n \"acc_norm\": 0.36897001303780963,\n\
\ \"acc_norm_stderr\": 0.01232393665017486\n },\n \"harness|hendrycksTest-professional_medicine|5\"\
: {\n \"acc\": 0.49264705882352944,\n \"acc_stderr\": 0.030369552523902173,\n\
\ \"acc_norm\": 0.49264705882352944,\n \"acc_norm_stderr\": 0.030369552523902173\n\
\ },\n \"harness|hendrycksTest-professional_psychology|5\": {\n \"\
acc\": 0.49836601307189543,\n \"acc_stderr\": 0.020227726838150124,\n \
\ \"acc_norm\": 0.49836601307189543,\n \"acc_norm_stderr\": 0.020227726838150124\n\
\ },\n \"harness|hendrycksTest-public_relations|5\": {\n \"acc\": 0.5727272727272728,\n\
\ \"acc_stderr\": 0.047381987035454834,\n \"acc_norm\": 0.5727272727272728,\n\
\ \"acc_norm_stderr\": 0.047381987035454834\n },\n \"harness|hendrycksTest-security_studies|5\"\
: {\n \"acc\": 0.6204081632653061,\n \"acc_stderr\": 0.031067211262872485,\n\
\ \"acc_norm\": 0.6204081632653061,\n \"acc_norm_stderr\": 0.031067211262872485\n\
\ },\n \"harness|hendrycksTest-sociology|5\": {\n \"acc\": 0.7114427860696517,\n\
\ \"acc_stderr\": 0.03203841040213322,\n \"acc_norm\": 0.7114427860696517,\n\
\ \"acc_norm_stderr\": 0.03203841040213322\n },\n \"harness|hendrycksTest-us_foreign_policy|5\"\
: {\n \"acc\": 0.75,\n \"acc_stderr\": 0.04351941398892446,\n \
\ \"acc_norm\": 0.75,\n \"acc_norm_stderr\": 0.04351941398892446\n \
\ },\n \"harness|hendrycksTest-virology|5\": {\n \"acc\": 0.45180722891566266,\n\
\ \"acc_stderr\": 0.03874371556587953,\n \"acc_norm\": 0.45180722891566266,\n\
\ \"acc_norm_stderr\": 0.03874371556587953\n },\n \"harness|hendrycksTest-world_religions|5\"\
: {\n \"acc\": 0.7368421052631579,\n \"acc_stderr\": 0.03377310252209205,\n\
\ \"acc_norm\": 0.7368421052631579,\n \"acc_norm_stderr\": 0.03377310252209205\n\
\ },\n \"harness|truthfulqa:mc|0\": {\n \"mc1\": 0.2631578947368421,\n\
\ \"mc1_stderr\": 0.015415241740237012,\n \"mc2\": 0.4032424062517679,\n\
\ \"mc2_stderr\": 0.014739501986326583\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.7174427782162589,\n \"acc_stderr\": 0.012654062850971405\n\
\ },\n \"harness|drop|3\": {\n \"em\": 0.2950922818791946,\n \
\ \"em_stderr\": 0.004670729426706436,\n \"f1\": 0.3578932466442965,\n \
\ \"f1_stderr\": 0.004607902070294773\n },\n \"harness|gsm8k|5\": {\n\
\ \"acc\": 0.0,\n \"acc_stderr\": 0.0\n }\n}\n```"
repo_url: https://huggingface.co/luffycodes/vicuna-shishya-7b-ep3-v1
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|arc:challenge|25_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_drop_3
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|drop|3_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|drop|3_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|gsm8k|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hellaswag|10_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T13-24-49.230828.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T13-24-49.230828.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- '**/details_harness|winogrande|5_2023-11-09T13-24-49.230828.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-11-09T13-24-49.230828.parquet'
- config_name: results
data_files:
- split: 2023_11_09T13_24_49.230828
path:
- results_2023-11-09T13-24-49.230828.parquet
- split: latest
path:
- results_2023-11-09T13-24-49.230828.parquet
---
# Dataset Card for Evaluation run of luffycodes/vicuna-shishya-7b-ep3-v1
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/luffycodes/vicuna-shishya-7b-ep3-v1
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [luffycodes/vicuna-shishya-7b-ep3-v1](https://huggingface.co/luffycodes/vicuna-shishya-7b-ep3-v1) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 64 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_luffycodes__vicuna-shishya-7b-ep3-v1_public",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-11-09T13:24:49.230828](https://huggingface.co/datasets/open-llm-leaderboard/details_luffycodes__vicuna-shishya-7b-ep3-v1_public/blob/main/results_2023-11-09T13-24-49.230828.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"acc": 0.49601218028660454,
"acc_stderr": 0.03399727784474729,
"acc_norm": 0.5041920928165192,
"acc_norm_stderr": 0.03492449912034474,
"mc1": 0.2631578947368421,
"mc1_stderr": 0.015415241740237012,
"mc2": 0.4032424062517679,
"mc2_stderr": 0.014739501986326583,
"em": 0.2950922818791946,
"em_stderr": 0.004670729426706436,
"f1": 0.3578932466442965,
"f1_stderr": 0.004607902070294773
},
"harness|arc:challenge|25": {
"acc": 0.439419795221843,
"acc_stderr": 0.014503747823580129,
"acc_norm": 0.4590443686006826,
"acc_norm_stderr": 0.014562291073601234
},
"harness|hellaswag|10": {
"acc": 0.5751842262497511,
"acc_stderr": 0.004933047726996794,
"acc_norm": 0.7635929097789285,
"acc_norm_stderr": 0.004240066898702511
},
"harness|hendrycksTest-abstract_algebra|5": {
"acc": 0.29,
"acc_stderr": 0.04560480215720684,
"acc_norm": 0.29,
"acc_norm_stderr": 0.04560480215720684
},
"harness|hendrycksTest-anatomy|5": {
"acc": 0.4740740740740741,
"acc_stderr": 0.04313531696750574,
"acc_norm": 0.4740740740740741,
"acc_norm_stderr": 0.04313531696750574
},
"harness|hendrycksTest-astronomy|5": {
"acc": 0.4868421052631579,
"acc_stderr": 0.04067533136309172,
"acc_norm": 0.4868421052631579,
"acc_norm_stderr": 0.04067533136309172
},
"harness|hendrycksTest-business_ethics|5": {
"acc": 0.47,
"acc_stderr": 0.050161355804659205,
"acc_norm": 0.47,
"acc_norm_stderr": 0.050161355804659205
},
"harness|hendrycksTest-clinical_knowledge|5": {
"acc": 0.5169811320754717,
"acc_stderr": 0.030755120364119905,
"acc_norm": 0.5169811320754717,
"acc_norm_stderr": 0.030755120364119905
},
"harness|hendrycksTest-college_biology|5": {
"acc": 0.4861111111111111,
"acc_stderr": 0.041795966175810016,
"acc_norm": 0.4861111111111111,
"acc_norm_stderr": 0.041795966175810016
},
"harness|hendrycksTest-college_chemistry|5": {
"acc": 0.31,
"acc_stderr": 0.04648231987117316,
"acc_norm": 0.31,
"acc_norm_stderr": 0.04648231987117316
},
"harness|hendrycksTest-college_computer_science|5": {
"acc": 0.41,
"acc_stderr": 0.049431107042371025,
"acc_norm": 0.41,
"acc_norm_stderr": 0.049431107042371025
},
"harness|hendrycksTest-college_mathematics|5": {
"acc": 0.35,
"acc_stderr": 0.047937248544110196,
"acc_norm": 0.35,
"acc_norm_stderr": 0.047937248544110196
},
"harness|hendrycksTest-college_medicine|5": {
"acc": 0.47398843930635837,
"acc_stderr": 0.03807301726504511,
"acc_norm": 0.47398843930635837,
"acc_norm_stderr": 0.03807301726504511
},
"harness|hendrycksTest-college_physics|5": {
"acc": 0.16666666666666666,
"acc_stderr": 0.03708284662416544,
"acc_norm": 0.16666666666666666,
"acc_norm_stderr": 0.03708284662416544
},
"harness|hendrycksTest-computer_security|5": {
"acc": 0.63,
"acc_stderr": 0.04852365870939099,
"acc_norm": 0.63,
"acc_norm_stderr": 0.04852365870939099
},
"harness|hendrycksTest-conceptual_physics|5": {
"acc": 0.4553191489361702,
"acc_stderr": 0.03255525359340355,
"acc_norm": 0.4553191489361702,
"acc_norm_stderr": 0.03255525359340355
},
"harness|hendrycksTest-econometrics|5": {
"acc": 0.2894736842105263,
"acc_stderr": 0.04266339443159393,
"acc_norm": 0.2894736842105263,
"acc_norm_stderr": 0.04266339443159393
},
"harness|hendrycksTest-electrical_engineering|5": {
"acc": 0.45517241379310347,
"acc_stderr": 0.04149886942192117,
"acc_norm": 0.45517241379310347,
"acc_norm_stderr": 0.04149886942192117
},
"harness|hendrycksTest-elementary_mathematics|5": {
"acc": 0.30158730158730157,
"acc_stderr": 0.023636975996101806,
"acc_norm": 0.30158730158730157,
"acc_norm_stderr": 0.023636975996101806
},
"harness|hendrycksTest-formal_logic|5": {
"acc": 0.31746031746031744,
"acc_stderr": 0.04163453031302859,
"acc_norm": 0.31746031746031744,
"acc_norm_stderr": 0.04163453031302859
},
"harness|hendrycksTest-global_facts|5": {
"acc": 0.35,
"acc_stderr": 0.047937248544110196,
"acc_norm": 0.35,
"acc_norm_stderr": 0.047937248544110196
},
"harness|hendrycksTest-high_school_biology|5": {
"acc": 0.535483870967742,
"acc_stderr": 0.02837228779796293,
"acc_norm": 0.535483870967742,
"acc_norm_stderr": 0.02837228779796293
},
"harness|hendrycksTest-high_school_chemistry|5": {
"acc": 0.3891625615763547,
"acc_stderr": 0.034304624161038716,
"acc_norm": 0.3891625615763547,
"acc_norm_stderr": 0.034304624161038716
},
"harness|hendrycksTest-high_school_computer_science|5": {
"acc": 0.47,
"acc_stderr": 0.050161355804659205,
"acc_norm": 0.47,
"acc_norm_stderr": 0.050161355804659205
},
"harness|hendrycksTest-high_school_european_history|5": {
"acc": 0.6424242424242425,
"acc_stderr": 0.037425970438065864,
"acc_norm": 0.6424242424242425,
"acc_norm_stderr": 0.037425970438065864
},
"harness|hendrycksTest-high_school_geography|5": {
"acc": 0.5909090909090909,
"acc_stderr": 0.03502975799413007,
"acc_norm": 0.5909090909090909,
"acc_norm_stderr": 0.03502975799413007
},
"harness|hendrycksTest-high_school_government_and_politics|5": {
"acc": 0.7305699481865285,
"acc_stderr": 0.03201867122877794,
"acc_norm": 0.7305699481865285,
"acc_norm_stderr": 0.03201867122877794
},
"harness|hendrycksTest-high_school_macroeconomics|5": {
"acc": 0.4307692307692308,
"acc_stderr": 0.02510682066053975,
"acc_norm": 0.4307692307692308,
"acc_norm_stderr": 0.02510682066053975
},
"harness|hendrycksTest-high_school_mathematics|5": {
"acc": 0.25555555555555554,
"acc_stderr": 0.02659393910184408,
"acc_norm": 0.25555555555555554,
"acc_norm_stderr": 0.02659393910184408
},
"harness|hendrycksTest-high_school_microeconomics|5": {
"acc": 0.4495798319327731,
"acc_stderr": 0.03231293497137707,
"acc_norm": 0.4495798319327731,
"acc_norm_stderr": 0.03231293497137707
},
"harness|hendrycksTest-high_school_physics|5": {
"acc": 0.2781456953642384,
"acc_stderr": 0.03658603262763743,
"acc_norm": 0.2781456953642384,
"acc_norm_stderr": 0.03658603262763743
},
"harness|hendrycksTest-high_school_psychology|5": {
"acc": 0.6862385321100918,
"acc_stderr": 0.019894723341469116,
"acc_norm": 0.6862385321100918,
"acc_norm_stderr": 0.019894723341469116
},
"harness|hendrycksTest-high_school_statistics|5": {
"acc": 0.3611111111111111,
"acc_stderr": 0.032757734861009996,
"acc_norm": 0.3611111111111111,
"acc_norm_stderr": 0.032757734861009996
},
"harness|hendrycksTest-high_school_us_history|5": {
"acc": 0.6813725490196079,
"acc_stderr": 0.03270287181482081,
"acc_norm": 0.6813725490196079,
"acc_norm_stderr": 0.03270287181482081
},
"harness|hendrycksTest-high_school_world_history|5": {
"acc": 0.7341772151898734,
"acc_stderr": 0.02875679962965834,
"acc_norm": 0.7341772151898734,
"acc_norm_stderr": 0.02875679962965834
},
"harness|hendrycksTest-human_aging|5": {
"acc": 0.6188340807174888,
"acc_stderr": 0.03259625118416827,
"acc_norm": 0.6188340807174888,
"acc_norm_stderr": 0.03259625118416827
},
"harness|hendrycksTest-human_sexuality|5": {
"acc": 0.6412213740458015,
"acc_stderr": 0.04206739313864908,
"acc_norm": 0.6412213740458015,
"acc_norm_stderr": 0.04206739313864908
},
"harness|hendrycksTest-international_law|5": {
"acc": 0.6198347107438017,
"acc_stderr": 0.04431324501968431,
"acc_norm": 0.6198347107438017,
"acc_norm_stderr": 0.04431324501968431
},
"harness|hendrycksTest-jurisprudence|5": {
"acc": 0.5648148148148148,
"acc_stderr": 0.04792898170907061,
"acc_norm": 0.5648148148148148,
"acc_norm_stderr": 0.04792898170907061
},
"harness|hendrycksTest-logical_fallacies|5": {
"acc": 0.50920245398773,
"acc_stderr": 0.03927705600787443,
"acc_norm": 0.50920245398773,
"acc_norm_stderr": 0.03927705600787443
},
"harness|hendrycksTest-machine_learning|5": {
"acc": 0.41964285714285715,
"acc_stderr": 0.04684099321077106,
"acc_norm": 0.41964285714285715,
"acc_norm_stderr": 0.04684099321077106
},
"harness|hendrycksTest-management|5": {
"acc": 0.6019417475728155,
"acc_stderr": 0.0484674825397724,
"acc_norm": 0.6019417475728155,
"acc_norm_stderr": 0.0484674825397724
},
"harness|hendrycksTest-marketing|5": {
"acc": 0.7393162393162394,
"acc_stderr": 0.028760348956523414,
"acc_norm": 0.7393162393162394,
"acc_norm_stderr": 0.028760348956523414
},
"harness|hendrycksTest-medical_genetics|5": {
"acc": 0.55,
"acc_stderr": 0.049999999999999996,
"acc_norm": 0.55,
"acc_norm_stderr": 0.049999999999999996
},
"harness|hendrycksTest-miscellaneous|5": {
"acc": 0.6934865900383141,
"acc_stderr": 0.016486952893041504,
"acc_norm": 0.6934865900383141,
"acc_norm_stderr": 0.016486952893041504
},
"harness|hendrycksTest-moral_disputes|5": {
"acc": 0.5375722543352601,
"acc_stderr": 0.026842985519615375,
"acc_norm": 0.5375722543352601,
"acc_norm_stderr": 0.026842985519615375
},
"harness|hendrycksTest-moral_scenarios|5": {
"acc": 0.2636871508379888,
"acc_stderr": 0.014736926383761987,
"acc_norm": 0.2636871508379888,
"acc_norm_stderr": 0.014736926383761987
},
"harness|hendrycksTest-nutrition|5": {
"acc": 0.5686274509803921,
"acc_stderr": 0.028358956313423545,
"acc_norm": 0.5686274509803921,
"acc_norm_stderr": 0.028358956313423545
},
"harness|hendrycksTest-philosophy|5": {
"acc": 0.5980707395498392,
"acc_stderr": 0.02784647600593047,
"acc_norm": 0.5980707395498392,
"acc_norm_stderr": 0.02784647600593047
},
"harness|hendrycksTest-prehistory|5": {
"acc": 0.5679012345679012,
"acc_stderr": 0.02756301097160668,
"acc_norm": 0.5679012345679012,
"acc_norm_stderr": 0.02756301097160668
},
"harness|hendrycksTest-professional_accounting|5": {
"acc": 0.35815602836879434,
"acc_stderr": 0.028602085862759415,
"acc_norm": 0.35815602836879434,
"acc_norm_stderr": 0.028602085862759415
},
"harness|hendrycksTest-professional_law|5": {
"acc": 0.36897001303780963,
"acc_stderr": 0.01232393665017486,
"acc_norm": 0.36897001303780963,
"acc_norm_stderr": 0.01232393665017486
},
"harness|hendrycksTest-professional_medicine|5": {
"acc": 0.49264705882352944,
"acc_stderr": 0.030369552523902173,
"acc_norm": 0.49264705882352944,
"acc_norm_stderr": 0.030369552523902173
},
"harness|hendrycksTest-professional_psychology|5": {
"acc": 0.49836601307189543,
"acc_stderr": 0.020227726838150124,
"acc_norm": 0.49836601307189543,
"acc_norm_stderr": 0.020227726838150124
},
"harness|hendrycksTest-public_relations|5": {
"acc": 0.5727272727272728,
"acc_stderr": 0.047381987035454834,
"acc_norm": 0.5727272727272728,
"acc_norm_stderr": 0.047381987035454834
},
"harness|hendrycksTest-security_studies|5": {
"acc": 0.6204081632653061,
"acc_stderr": 0.031067211262872485,
"acc_norm": 0.6204081632653061,
"acc_norm_stderr": 0.031067211262872485
},
"harness|hendrycksTest-sociology|5": {
"acc": 0.7114427860696517,
"acc_stderr": 0.03203841040213322,
"acc_norm": 0.7114427860696517,
"acc_norm_stderr": 0.03203841040213322
},
"harness|hendrycksTest-us_foreign_policy|5": {
"acc": 0.75,
"acc_stderr": 0.04351941398892446,
"acc_norm": 0.75,
"acc_norm_stderr": 0.04351941398892446
},
"harness|hendrycksTest-virology|5": {
"acc": 0.45180722891566266,
"acc_stderr": 0.03874371556587953,
"acc_norm": 0.45180722891566266,
"acc_norm_stderr": 0.03874371556587953
},
"harness|hendrycksTest-world_religions|5": {
"acc": 0.7368421052631579,
"acc_stderr": 0.03377310252209205,
"acc_norm": 0.7368421052631579,
"acc_norm_stderr": 0.03377310252209205
},
"harness|truthfulqa:mc|0": {
"mc1": 0.2631578947368421,
"mc1_stderr": 0.015415241740237012,
"mc2": 0.4032424062517679,
"mc2_stderr": 0.014739501986326583
},
"harness|winogrande|5": {
"acc": 0.7174427782162589,
"acc_stderr": 0.012654062850971405
},
"harness|drop|3": {
"em": 0.2950922818791946,
"em_stderr": 0.004670729426706436,
"f1": 0.3578932466442965,
"f1_stderr": 0.004607902070294773
},
"harness|gsm8k|5": {
"acc": 0.0,
"acc_stderr": 0.0
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] | [
-0.687960147857666,
-0.8568715453147888,
0.29805776476860046,
0.20959773659706116,
-0.2136719822883606,
-0.10236594080924988,
0.0013543427921831608,
-0.23921041190624237,
0.5950338840484619,
-0.0002795749460346997,
-0.4878493547439575,
-0.7204574346542358,
-0.41581282019615173,
0.251854360... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
mpachauri/TrainingDatasetNew | mpachauri | 2023-11-09T13:37:44Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T13:37:44Z | 2023-11-09T13:34:01.000Z | 2023-11-09T13:34:01 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
cleanrl/benchmark | cleanrl | 2023-11-13T21:04:48Z | 0 | 0 | null | [
"region:us"
] | 2023-11-13T21:04:48Z | 2023-11-09T13:35:19.000Z | 2023-11-09T13:35:19 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
open-llm-leaderboard/details_psmathur__orca_mini_v3_70b_public | open-llm-leaderboard | 2023-11-09T13:41:05Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T13:41:05Z | 2023-11-09T13:40:56.000Z | 2023-11-09T13:40:56 | ---
pretty_name: Evaluation run of psmathur/orca_mini_v3_70b
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [psmathur/orca_mini_v3_70b](https://huggingface.co/psmathur/orca_mini_v3_70b)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 3 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the aggregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_psmathur__orca_mini_v3_70b_public\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-11-09T13:40:37.998536](https://huggingface.co/datasets/open-llm-leaderboard/details_psmathur__orca_mini_v3_70b_public/blob/main/results_2023-11-09T13-40-37.998536.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"em\": 0.31061241610738255,\n\
\ \"em_stderr\": 0.004738935370907925,\n \"f1\": 0.4017103607382563,\n\
\ \"f1_stderr\": 0.004555690324539627,\n \"acc\": 0.6178968305495601,\n\
\ \"acc_stderr\": 0.012083802131657148\n },\n \"harness|drop|3\": {\n\
\ \"em\": 0.31061241610738255,\n \"em_stderr\": 0.004738935370907925,\n\
\ \"f1\": 0.4017103607382563,\n \"f1_stderr\": 0.004555690324539627\n\
\ },\n \"harness|gsm8k|5\": {\n \"acc\": 0.4086429112964367,\n \
\ \"acc_stderr\": 0.01354063973334243\n },\n \"harness|winogrande|5\":\
\ {\n \"acc\": 0.8271507498026835,\n \"acc_stderr\": 0.010626964529971864\n\
\ }\n}\n```"
repo_url: https://huggingface.co/psmathur/orca_mini_v3_70b
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_drop_3
data_files:
- split: 2023_11_09T13_40_37.998536
path:
- '**/details_harness|drop|3_2023-11-09T13-40-37.998536.parquet'
- split: latest
path:
- '**/details_harness|drop|3_2023-11-09T13-40-37.998536.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_11_09T13_40_37.998536
path:
- '**/details_harness|gsm8k|5_2023-11-09T13-40-37.998536.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-11-09T13-40-37.998536.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_11_09T13_40_37.998536
path:
- '**/details_harness|winogrande|5_2023-11-09T13-40-37.998536.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-11-09T13-40-37.998536.parquet'
- config_name: results
data_files:
- split: 2023_11_09T13_40_37.998536
path:
- results_2023-11-09T13-40-37.998536.parquet
- split: latest
path:
- results_2023-11-09T13-40-37.998536.parquet
---
# Dataset Card for Evaluation run of psmathur/orca_mini_v3_70b
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/psmathur/orca_mini_v3_70b
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [psmathur/orca_mini_v3_70b](https://huggingface.co/psmathur/orca_mini_v3_70b) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 3 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_psmathur__orca_mini_v3_70b_public",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-11-09T13:40:37.998536](https://huggingface.co/datasets/open-llm-leaderboard/details_psmathur__orca_mini_v3_70b_public/blob/main/results_2023-11-09T13-40-37.998536.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"em": 0.31061241610738255,
"em_stderr": 0.004738935370907925,
"f1": 0.4017103607382563,
"f1_stderr": 0.004555690324539627,
"acc": 0.6178968305495601,
"acc_stderr": 0.012083802131657148
},
"harness|drop|3": {
"em": 0.31061241610738255,
"em_stderr": 0.004738935370907925,
"f1": 0.4017103607382563,
"f1_stderr": 0.004555690324539627
},
"harness|gsm8k|5": {
"acc": 0.4086429112964367,
"acc_stderr": 0.01354063973334243
},
"harness|winogrande|5": {
"acc": 0.8271507498026835,
"acc_stderr": 0.010626964529971864
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] | [
-0.3835715651512146,
-0.6409476399421692,
0.19595688581466675,
0.10468635708093643,
-0.280229777097702,
0.10589656978845596,
-0.32181236147880554,
-0.1743643879890442,
0.47015300393104553,
0.5166938304901123,
-0.7326904535293579,
-0.9124718308448792,
-0.6742269992828369,
0.1549297422170639... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Jinsns/flk | Jinsns | 2023-11-09T15:42:43Z | 0 | 0 | null | [
"license:mit",
"region:us"
] | 2023-11-09T15:42:43Z | 2023-11-09T13:55:30.000Z | 2023-11-09T13:55:30 | ---
license: mit
---
# 法律法规 #
从国家法律法规数据库(https://flk.npc.gov.cn/) 下载的法律法规
解压后得到:
law_list.tsv 法律法规的信息列表
law_docs/ 目录下有五个文件夹,分别装有不同状态的法律法规。
status 1 有效 ,3 尚未生效 ,5 已修改(有对应的1),7 两种:【有关法律问题和重大问题的决定】或【修改、废止的决定】,9 已废止
txt_files/ 用脚本处理 status1 中的非扫描件,生成的txt文件,每一行是形式是
```
某法 第n章 第n条 法条内容
```
laws_vector_store/ 是FAISS向量数据库,embedding模型采用text2vec (https://huggingface.co/GanymedeNil/text2vec-large-chinese)
向量数据库的每一条数据是txt的一行(向量数据库的范围是所有txt_files)
| [
-0.06411316245794296,
-0.7345316410064697,
0.15400639176368713,
0.533113956451416,
-0.8872972130775452,
-0.3988828957080841,
-0.07356996834278107,
-0.07975536584854126,
0.19727204740047455,
0.5430484414100647,
0.15323582291603088,
-0.6214050650596619,
-0.8482459187507629,
0.088907741010189... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
open-llm-leaderboard/details_willnguyen__lacda-2-7B-chat-v0.1_public | open-llm-leaderboard | 2023-11-09T13:57:18Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T13:57:18Z | 2023-11-09T13:56:17.000Z | 2023-11-09T13:56:17 | ---
pretty_name: Evaluation run of willnguyen/lacda-2-7B-chat-v0.1
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [willnguyen/lacda-2-7B-chat-v0.1](https://huggingface.co/willnguyen/lacda-2-7B-chat-v0.1)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 64 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the aggregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_willnguyen__lacda-2-7B-chat-v0.1_public\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-11-09T13:53:53.211938](https://huggingface.co/datasets/open-llm-leaderboard/details_willnguyen__lacda-2-7B-chat-v0.1_public/blob/main/results_2023-11-09T13-53-53.211938.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"acc\": 0.46065725811605934,\n\
\ \"acc_stderr\": 0.034477280778802896,\n \"acc_norm\": 0.4668080345369505,\n\
\ \"acc_norm_stderr\": 0.035310968004727446,\n \"mc1\": 0.3011015911872705,\n\
\ \"mc1_stderr\": 0.016058999026100612,\n \"mc2\": 0.4456721895962505,\n\
\ \"mc2_stderr\": 0.014265726453599933,\n \"em\": 0.001363255033557047,\n\
\ \"em_stderr\": 0.0003778609196460794,\n \"f1\": 0.05649014261744978,\n\
\ \"f1_stderr\": 0.0013342363586640303\n },\n \"harness|arc:challenge|25\"\
: {\n \"acc\": 0.4803754266211604,\n \"acc_stderr\": 0.014600132075947087,\n\
\ \"acc_norm\": 0.5307167235494881,\n \"acc_norm_stderr\": 0.014583792546304038\n\
\ },\n \"harness|hellaswag|10\": {\n \"acc\": 0.5796654052977495,\n\
\ \"acc_stderr\": 0.0049260381977145225,\n \"acc_norm\": 0.7757418840868353,\n\
\ \"acc_norm_stderr\": 0.0041624039148053385\n },\n \"harness|hendrycksTest-abstract_algebra|5\"\
: {\n \"acc\": 0.34,\n \"acc_stderr\": 0.047609522856952365,\n \
\ \"acc_norm\": 0.34,\n \"acc_norm_stderr\": 0.047609522856952365\n \
\ },\n \"harness|hendrycksTest-anatomy|5\": {\n \"acc\": 0.43703703703703706,\n\
\ \"acc_stderr\": 0.04284958639753399,\n \"acc_norm\": 0.43703703703703706,\n\
\ \"acc_norm_stderr\": 0.04284958639753399\n },\n \"harness|hendrycksTest-astronomy|5\"\
: {\n \"acc\": 0.48026315789473684,\n \"acc_stderr\": 0.040657710025626036,\n\
\ \"acc_norm\": 0.48026315789473684,\n \"acc_norm_stderr\": 0.040657710025626036\n\
\ },\n \"harness|hendrycksTest-business_ethics|5\": {\n \"acc\": 0.46,\n\
\ \"acc_stderr\": 0.05009082659620333,\n \"acc_norm\": 0.46,\n \
\ \"acc_norm_stderr\": 0.05009082659620333\n },\n \"harness|hendrycksTest-clinical_knowledge|5\"\
: {\n \"acc\": 0.45660377358490567,\n \"acc_stderr\": 0.030656748696739438,\n\
\ \"acc_norm\": 0.45660377358490567,\n \"acc_norm_stderr\": 0.030656748696739438\n\
\ },\n \"harness|hendrycksTest-college_biology|5\": {\n \"acc\": 0.4375,\n\
\ \"acc_stderr\": 0.04148415739394154,\n \"acc_norm\": 0.4375,\n \
\ \"acc_norm_stderr\": 0.04148415739394154\n },\n \"harness|hendrycksTest-college_chemistry|5\"\
: {\n \"acc\": 0.34,\n \"acc_stderr\": 0.04760952285695235,\n \
\ \"acc_norm\": 0.34,\n \"acc_norm_stderr\": 0.04760952285695235\n \
\ },\n \"harness|hendrycksTest-college_computer_science|5\": {\n \"acc\"\
: 0.35,\n \"acc_stderr\": 0.047937248544110196,\n \"acc_norm\": 0.35,\n\
\ \"acc_norm_stderr\": 0.047937248544110196\n },\n \"harness|hendrycksTest-college_mathematics|5\"\
: {\n \"acc\": 0.28,\n \"acc_stderr\": 0.04512608598542128,\n \
\ \"acc_norm\": 0.28,\n \"acc_norm_stderr\": 0.04512608598542128\n \
\ },\n \"harness|hendrycksTest-college_medicine|5\": {\n \"acc\": 0.4277456647398844,\n\
\ \"acc_stderr\": 0.037724468575180255,\n \"acc_norm\": 0.4277456647398844,\n\
\ \"acc_norm_stderr\": 0.037724468575180255\n },\n \"harness|hendrycksTest-college_physics|5\"\
: {\n \"acc\": 0.24509803921568626,\n \"acc_stderr\": 0.04280105837364395,\n\
\ \"acc_norm\": 0.24509803921568626,\n \"acc_norm_stderr\": 0.04280105837364395\n\
\ },\n \"harness|hendrycksTest-computer_security|5\": {\n \"acc\":\
\ 0.63,\n \"acc_stderr\": 0.048523658709391,\n \"acc_norm\": 0.63,\n\
\ \"acc_norm_stderr\": 0.048523658709391\n },\n \"harness|hendrycksTest-conceptual_physics|5\"\
: {\n \"acc\": 0.39574468085106385,\n \"acc_stderr\": 0.03196758697835362,\n\
\ \"acc_norm\": 0.39574468085106385,\n \"acc_norm_stderr\": 0.03196758697835362\n\
\ },\n \"harness|hendrycksTest-econometrics|5\": {\n \"acc\": 0.2631578947368421,\n\
\ \"acc_stderr\": 0.041424397194893624,\n \"acc_norm\": 0.2631578947368421,\n\
\ \"acc_norm_stderr\": 0.041424397194893624\n },\n \"harness|hendrycksTest-electrical_engineering|5\"\
: {\n \"acc\": 0.5241379310344828,\n \"acc_stderr\": 0.041618085035015295,\n\
\ \"acc_norm\": 0.5241379310344828,\n \"acc_norm_stderr\": 0.041618085035015295\n\
\ },\n \"harness|hendrycksTest-elementary_mathematics|5\": {\n \"acc\"\
: 0.2619047619047619,\n \"acc_stderr\": 0.02264421261552521,\n \"\
acc_norm\": 0.2619047619047619,\n \"acc_norm_stderr\": 0.02264421261552521\n\
\ },\n \"harness|hendrycksTest-formal_logic|5\": {\n \"acc\": 0.30158730158730157,\n\
\ \"acc_stderr\": 0.04104947269903394,\n \"acc_norm\": 0.30158730158730157,\n\
\ \"acc_norm_stderr\": 0.04104947269903394\n },\n \"harness|hendrycksTest-global_facts|5\"\
: {\n \"acc\": 0.38,\n \"acc_stderr\": 0.048783173121456316,\n \
\ \"acc_norm\": 0.38,\n \"acc_norm_stderr\": 0.048783173121456316\n \
\ },\n \"harness|hendrycksTest-high_school_biology|5\": {\n \"acc\"\
: 0.4645161290322581,\n \"acc_stderr\": 0.028372287797962956,\n \"\
acc_norm\": 0.4645161290322581,\n \"acc_norm_stderr\": 0.028372287797962956\n\
\ },\n \"harness|hendrycksTest-high_school_chemistry|5\": {\n \"acc\"\
: 0.35467980295566504,\n \"acc_stderr\": 0.0336612448905145,\n \"\
acc_norm\": 0.35467980295566504,\n \"acc_norm_stderr\": 0.0336612448905145\n\
\ },\n \"harness|hendrycksTest-high_school_computer_science|5\": {\n \
\ \"acc\": 0.43,\n \"acc_stderr\": 0.049756985195624284,\n \"acc_norm\"\
: 0.43,\n \"acc_norm_stderr\": 0.049756985195624284\n },\n \"harness|hendrycksTest-high_school_european_history|5\"\
: {\n \"acc\": 0.6303030303030303,\n \"acc_stderr\": 0.03769430314512566,\n\
\ \"acc_norm\": 0.6303030303030303,\n \"acc_norm_stderr\": 0.03769430314512566\n\
\ },\n \"harness|hendrycksTest-high_school_geography|5\": {\n \"acc\"\
: 0.5,\n \"acc_stderr\": 0.035623524993954825,\n \"acc_norm\": 0.5,\n\
\ \"acc_norm_stderr\": 0.035623524993954825\n },\n \"harness|hendrycksTest-high_school_government_and_politics|5\"\
: {\n \"acc\": 0.6062176165803109,\n \"acc_stderr\": 0.035260770955482405,\n\
\ \"acc_norm\": 0.6062176165803109,\n \"acc_norm_stderr\": 0.035260770955482405\n\
\ },\n \"harness|hendrycksTest-high_school_macroeconomics|5\": {\n \
\ \"acc\": 0.4461538461538462,\n \"acc_stderr\": 0.02520357177302833,\n \
\ \"acc_norm\": 0.4461538461538462,\n \"acc_norm_stderr\": 0.02520357177302833\n\
\ },\n \"harness|hendrycksTest-high_school_mathematics|5\": {\n \"\
acc\": 0.2851851851851852,\n \"acc_stderr\": 0.027528599210340496,\n \
\ \"acc_norm\": 0.2851851851851852,\n \"acc_norm_stderr\": 0.027528599210340496\n\
\ },\n \"harness|hendrycksTest-high_school_microeconomics|5\": {\n \
\ \"acc\": 0.46218487394957986,\n \"acc_stderr\": 0.032385469487589795,\n\
\ \"acc_norm\": 0.46218487394957986,\n \"acc_norm_stderr\": 0.032385469487589795\n\
\ },\n \"harness|hendrycksTest-high_school_physics|5\": {\n \"acc\"\
: 0.31125827814569534,\n \"acc_stderr\": 0.03780445850526733,\n \"\
acc_norm\": 0.31125827814569534,\n \"acc_norm_stderr\": 0.03780445850526733\n\
\ },\n \"harness|hendrycksTest-high_school_psychology|5\": {\n \"acc\"\
: 0.6165137614678899,\n \"acc_stderr\": 0.020847156641915977,\n \"\
acc_norm\": 0.6165137614678899,\n \"acc_norm_stderr\": 0.020847156641915977\n\
\ },\n \"harness|hendrycksTest-high_school_statistics|5\": {\n \"acc\"\
: 0.23148148148148148,\n \"acc_stderr\": 0.028765111718046955,\n \"\
acc_norm\": 0.23148148148148148,\n \"acc_norm_stderr\": 0.028765111718046955\n\
\ },\n \"harness|hendrycksTest-high_school_us_history|5\": {\n \"acc\"\
: 0.5,\n \"acc_stderr\": 0.03509312031717982,\n \"acc_norm\": 0.5,\n\
\ \"acc_norm_stderr\": 0.03509312031717982\n },\n \"harness|hendrycksTest-high_school_world_history|5\"\
: {\n \"acc\": 0.5780590717299579,\n \"acc_stderr\": 0.032148146302403695,\n\
\ \"acc_norm\": 0.5780590717299579,\n \"acc_norm_stderr\": 0.032148146302403695\n\
\ },\n \"harness|hendrycksTest-human_aging|5\": {\n \"acc\": 0.5560538116591929,\n\
\ \"acc_stderr\": 0.03334625674242728,\n \"acc_norm\": 0.5560538116591929,\n\
\ \"acc_norm_stderr\": 0.03334625674242728\n },\n \"harness|hendrycksTest-human_sexuality|5\"\
: {\n \"acc\": 0.5725190839694656,\n \"acc_stderr\": 0.04338920305792401,\n\
\ \"acc_norm\": 0.5725190839694656,\n \"acc_norm_stderr\": 0.04338920305792401\n\
\ },\n \"harness|hendrycksTest-international_law|5\": {\n \"acc\":\
\ 0.6198347107438017,\n \"acc_stderr\": 0.04431324501968431,\n \"\
acc_norm\": 0.6198347107438017,\n \"acc_norm_stderr\": 0.04431324501968431\n\
\ },\n \"harness|hendrycksTest-jurisprudence|5\": {\n \"acc\": 0.5370370370370371,\n\
\ \"acc_stderr\": 0.04820403072760628,\n \"acc_norm\": 0.5370370370370371,\n\
\ \"acc_norm_stderr\": 0.04820403072760628\n },\n \"harness|hendrycksTest-logical_fallacies|5\"\
: {\n \"acc\": 0.49079754601226994,\n \"acc_stderr\": 0.039277056007874414,\n\
\ \"acc_norm\": 0.49079754601226994,\n \"acc_norm_stderr\": 0.039277056007874414\n\
\ },\n \"harness|hendrycksTest-machine_learning|5\": {\n \"acc\": 0.375,\n\
\ \"acc_stderr\": 0.04595091388086298,\n \"acc_norm\": 0.375,\n \
\ \"acc_norm_stderr\": 0.04595091388086298\n },\n \"harness|hendrycksTest-management|5\"\
: {\n \"acc\": 0.6019417475728155,\n \"acc_stderr\": 0.048467482539772386,\n\
\ \"acc_norm\": 0.6019417475728155,\n \"acc_norm_stderr\": 0.048467482539772386\n\
\ },\n \"harness|hendrycksTest-marketing|5\": {\n \"acc\": 0.6709401709401709,\n\
\ \"acc_stderr\": 0.030782321577688173,\n \"acc_norm\": 0.6709401709401709,\n\
\ \"acc_norm_stderr\": 0.030782321577688173\n },\n \"harness|hendrycksTest-medical_genetics|5\"\
: {\n \"acc\": 0.52,\n \"acc_stderr\": 0.050211673156867795,\n \
\ \"acc_norm\": 0.52,\n \"acc_norm_stderr\": 0.050211673156867795\n \
\ },\n \"harness|hendrycksTest-miscellaneous|5\": {\n \"acc\": 0.6143039591315453,\n\
\ \"acc_stderr\": 0.017406476619212907,\n \"acc_norm\": 0.6143039591315453,\n\
\ \"acc_norm_stderr\": 0.017406476619212907\n },\n \"harness|hendrycksTest-moral_disputes|5\"\
: {\n \"acc\": 0.4884393063583815,\n \"acc_stderr\": 0.026911898686377913,\n\
\ \"acc_norm\": 0.4884393063583815,\n \"acc_norm_stderr\": 0.026911898686377913\n\
\ },\n \"harness|hendrycksTest-moral_scenarios|5\": {\n \"acc\": 0.23798882681564246,\n\
\ \"acc_stderr\": 0.014242630070574915,\n \"acc_norm\": 0.23798882681564246,\n\
\ \"acc_norm_stderr\": 0.014242630070574915\n },\n \"harness|hendrycksTest-nutrition|5\"\
: {\n \"acc\": 0.4673202614379085,\n \"acc_stderr\": 0.028568699752225875,\n\
\ \"acc_norm\": 0.4673202614379085,\n \"acc_norm_stderr\": 0.028568699752225875\n\
\ },\n \"harness|hendrycksTest-philosophy|5\": {\n \"acc\": 0.5884244372990354,\n\
\ \"acc_stderr\": 0.02795048149440127,\n \"acc_norm\": 0.5884244372990354,\n\
\ \"acc_norm_stderr\": 0.02795048149440127\n },\n \"harness|hendrycksTest-prehistory|5\"\
: {\n \"acc\": 0.4783950617283951,\n \"acc_stderr\": 0.027794760105008746,\n\
\ \"acc_norm\": 0.4783950617283951,\n \"acc_norm_stderr\": 0.027794760105008746\n\
\ },\n \"harness|hendrycksTest-professional_accounting|5\": {\n \"\
acc\": 0.3475177304964539,\n \"acc_stderr\": 0.028406627809590954,\n \
\ \"acc_norm\": 0.3475177304964539,\n \"acc_norm_stderr\": 0.028406627809590954\n\
\ },\n \"harness|hendrycksTest-professional_law|5\": {\n \"acc\": 0.35723598435462844,\n\
\ \"acc_stderr\": 0.012238615750316503,\n \"acc_norm\": 0.35723598435462844,\n\
\ \"acc_norm_stderr\": 0.012238615750316503\n },\n \"harness|hendrycksTest-professional_medicine|5\"\
: {\n \"acc\": 0.49264705882352944,\n \"acc_stderr\": 0.030369552523902173,\n\
\ \"acc_norm\": 0.49264705882352944,\n \"acc_norm_stderr\": 0.030369552523902173\n\
\ },\n \"harness|hendrycksTest-professional_psychology|5\": {\n \"\
acc\": 0.43137254901960786,\n \"acc_stderr\": 0.020036393768352638,\n \
\ \"acc_norm\": 0.43137254901960786,\n \"acc_norm_stderr\": 0.020036393768352638\n\
\ },\n \"harness|hendrycksTest-public_relations|5\": {\n \"acc\": 0.5818181818181818,\n\
\ \"acc_stderr\": 0.04724577405731572,\n \"acc_norm\": 0.5818181818181818,\n\
\ \"acc_norm_stderr\": 0.04724577405731572\n },\n \"harness|hendrycksTest-security_studies|5\"\
: {\n \"acc\": 0.46122448979591835,\n \"acc_stderr\": 0.03191282052669277,\n\
\ \"acc_norm\": 0.46122448979591835,\n \"acc_norm_stderr\": 0.03191282052669277\n\
\ },\n \"harness|hendrycksTest-sociology|5\": {\n \"acc\": 0.6069651741293532,\n\
\ \"acc_stderr\": 0.0345368246603156,\n \"acc_norm\": 0.6069651741293532,\n\
\ \"acc_norm_stderr\": 0.0345368246603156\n },\n \"harness|hendrycksTest-us_foreign_policy|5\"\
: {\n \"acc\": 0.64,\n \"acc_stderr\": 0.04824181513244218,\n \
\ \"acc_norm\": 0.64,\n \"acc_norm_stderr\": 0.04824181513244218\n \
\ },\n \"harness|hendrycksTest-virology|5\": {\n \"acc\": 0.39759036144578314,\n\
\ \"acc_stderr\": 0.038099730845402184,\n \"acc_norm\": 0.39759036144578314,\n\
\ \"acc_norm_stderr\": 0.038099730845402184\n },\n \"harness|hendrycksTest-world_religions|5\"\
: {\n \"acc\": 0.672514619883041,\n \"acc_stderr\": 0.03599335771456027,\n\
\ \"acc_norm\": 0.672514619883041,\n \"acc_norm_stderr\": 0.03599335771456027\n\
\ },\n \"harness|truthfulqa:mc|0\": {\n \"mc1\": 0.3011015911872705,\n\
\ \"mc1_stderr\": 0.016058999026100612,\n \"mc2\": 0.4456721895962505,\n\
\ \"mc2_stderr\": 0.014265726453599933\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.7419100236779794,\n \"acc_stderr\": 0.012298278833972397\n\
\ },\n \"harness|drop|3\": {\n \"em\": 0.001363255033557047,\n \
\ \"em_stderr\": 0.0003778609196460794,\n \"f1\": 0.05649014261744978,\n\
\ \"f1_stderr\": 0.0013342363586640303\n },\n \"harness|gsm8k|5\":\
\ {\n \"acc\": 0.06292645943896892,\n \"acc_stderr\": 0.0066887625815327395\n\
\ }\n}\n```"
repo_url: https://huggingface.co/willnguyen/lacda-2-7B-chat-v0.1
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|arc:challenge|25_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_drop_3
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|drop|3_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|drop|3_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|gsm8k|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hellaswag|10_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T13-53-53.211938.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T13-53-53.211938.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- '**/details_harness|winogrande|5_2023-11-09T13-53-53.211938.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-11-09T13-53-53.211938.parquet'
- config_name: results
data_files:
- split: 2023_11_09T13_53_53.211938
path:
- results_2023-11-09T13-53-53.211938.parquet
- split: latest
path:
- results_2023-11-09T13-53-53.211938.parquet
---
# Dataset Card for Evaluation run of willnguyen/lacda-2-7B-chat-v0.1
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/willnguyen/lacda-2-7B-chat-v0.1
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [willnguyen/lacda-2-7B-chat-v0.1](https://huggingface.co/willnguyen/lacda-2-7B-chat-v0.1) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 64 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_willnguyen__lacda-2-7B-chat-v0.1_public",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-11-09T13:53:53.211938](https://huggingface.co/datasets/open-llm-leaderboard/details_willnguyen__lacda-2-7B-chat-v0.1_public/blob/main/results_2023-11-09T13-53-53.211938.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"acc": 0.46065725811605934,
"acc_stderr": 0.034477280778802896,
"acc_norm": 0.4668080345369505,
"acc_norm_stderr": 0.035310968004727446,
"mc1": 0.3011015911872705,
"mc1_stderr": 0.016058999026100612,
"mc2": 0.4456721895962505,
"mc2_stderr": 0.014265726453599933,
"em": 0.001363255033557047,
"em_stderr": 0.0003778609196460794,
"f1": 0.05649014261744978,
"f1_stderr": 0.0013342363586640303
},
"harness|arc:challenge|25": {
"acc": 0.4803754266211604,
"acc_stderr": 0.014600132075947087,
"acc_norm": 0.5307167235494881,
"acc_norm_stderr": 0.014583792546304038
},
"harness|hellaswag|10": {
"acc": 0.5796654052977495,
"acc_stderr": 0.0049260381977145225,
"acc_norm": 0.7757418840868353,
"acc_norm_stderr": 0.0041624039148053385
},
"harness|hendrycksTest-abstract_algebra|5": {
"acc": 0.34,
"acc_stderr": 0.047609522856952365,
"acc_norm": 0.34,
"acc_norm_stderr": 0.047609522856952365
},
"harness|hendrycksTest-anatomy|5": {
"acc": 0.43703703703703706,
"acc_stderr": 0.04284958639753399,
"acc_norm": 0.43703703703703706,
"acc_norm_stderr": 0.04284958639753399
},
"harness|hendrycksTest-astronomy|5": {
"acc": 0.48026315789473684,
"acc_stderr": 0.040657710025626036,
"acc_norm": 0.48026315789473684,
"acc_norm_stderr": 0.040657710025626036
},
"harness|hendrycksTest-business_ethics|5": {
"acc": 0.46,
"acc_stderr": 0.05009082659620333,
"acc_norm": 0.46,
"acc_norm_stderr": 0.05009082659620333
},
"harness|hendrycksTest-clinical_knowledge|5": {
"acc": 0.45660377358490567,
"acc_stderr": 0.030656748696739438,
"acc_norm": 0.45660377358490567,
"acc_norm_stderr": 0.030656748696739438
},
"harness|hendrycksTest-college_biology|5": {
"acc": 0.4375,
"acc_stderr": 0.04148415739394154,
"acc_norm": 0.4375,
"acc_norm_stderr": 0.04148415739394154
},
"harness|hendrycksTest-college_chemistry|5": {
"acc": 0.34,
"acc_stderr": 0.04760952285695235,
"acc_norm": 0.34,
"acc_norm_stderr": 0.04760952285695235
},
"harness|hendrycksTest-college_computer_science|5": {
"acc": 0.35,
"acc_stderr": 0.047937248544110196,
"acc_norm": 0.35,
"acc_norm_stderr": 0.047937248544110196
},
"harness|hendrycksTest-college_mathematics|5": {
"acc": 0.28,
"acc_stderr": 0.04512608598542128,
"acc_norm": 0.28,
"acc_norm_stderr": 0.04512608598542128
},
"harness|hendrycksTest-college_medicine|5": {
"acc": 0.4277456647398844,
"acc_stderr": 0.037724468575180255,
"acc_norm": 0.4277456647398844,
"acc_norm_stderr": 0.037724468575180255
},
"harness|hendrycksTest-college_physics|5": {
"acc": 0.24509803921568626,
"acc_stderr": 0.04280105837364395,
"acc_norm": 0.24509803921568626,
"acc_norm_stderr": 0.04280105837364395
},
"harness|hendrycksTest-computer_security|5": {
"acc": 0.63,
"acc_stderr": 0.048523658709391,
"acc_norm": 0.63,
"acc_norm_stderr": 0.048523658709391
},
"harness|hendrycksTest-conceptual_physics|5": {
"acc": 0.39574468085106385,
"acc_stderr": 0.03196758697835362,
"acc_norm": 0.39574468085106385,
"acc_norm_stderr": 0.03196758697835362
},
"harness|hendrycksTest-econometrics|5": {
"acc": 0.2631578947368421,
"acc_stderr": 0.041424397194893624,
"acc_norm": 0.2631578947368421,
"acc_norm_stderr": 0.041424397194893624
},
"harness|hendrycksTest-electrical_engineering|5": {
"acc": 0.5241379310344828,
"acc_stderr": 0.041618085035015295,
"acc_norm": 0.5241379310344828,
"acc_norm_stderr": 0.041618085035015295
},
"harness|hendrycksTest-elementary_mathematics|5": {
"acc": 0.2619047619047619,
"acc_stderr": 0.02264421261552521,
"acc_norm": 0.2619047619047619,
"acc_norm_stderr": 0.02264421261552521
},
"harness|hendrycksTest-formal_logic|5": {
"acc": 0.30158730158730157,
"acc_stderr": 0.04104947269903394,
"acc_norm": 0.30158730158730157,
"acc_norm_stderr": 0.04104947269903394
},
"harness|hendrycksTest-global_facts|5": {
"acc": 0.38,
"acc_stderr": 0.048783173121456316,
"acc_norm": 0.38,
"acc_norm_stderr": 0.048783173121456316
},
"harness|hendrycksTest-high_school_biology|5": {
"acc": 0.4645161290322581,
"acc_stderr": 0.028372287797962956,
"acc_norm": 0.4645161290322581,
"acc_norm_stderr": 0.028372287797962956
},
"harness|hendrycksTest-high_school_chemistry|5": {
"acc": 0.35467980295566504,
"acc_stderr": 0.0336612448905145,
"acc_norm": 0.35467980295566504,
"acc_norm_stderr": 0.0336612448905145
},
"harness|hendrycksTest-high_school_computer_science|5": {
"acc": 0.43,
"acc_stderr": 0.049756985195624284,
"acc_norm": 0.43,
"acc_norm_stderr": 0.049756985195624284
},
"harness|hendrycksTest-high_school_european_history|5": {
"acc": 0.6303030303030303,
"acc_stderr": 0.03769430314512566,
"acc_norm": 0.6303030303030303,
"acc_norm_stderr": 0.03769430314512566
},
"harness|hendrycksTest-high_school_geography|5": {
"acc": 0.5,
"acc_stderr": 0.035623524993954825,
"acc_norm": 0.5,
"acc_norm_stderr": 0.035623524993954825
},
"harness|hendrycksTest-high_school_government_and_politics|5": {
"acc": 0.6062176165803109,
"acc_stderr": 0.035260770955482405,
"acc_norm": 0.6062176165803109,
"acc_norm_stderr": 0.035260770955482405
},
"harness|hendrycksTest-high_school_macroeconomics|5": {
"acc": 0.4461538461538462,
"acc_stderr": 0.02520357177302833,
"acc_norm": 0.4461538461538462,
"acc_norm_stderr": 0.02520357177302833
},
"harness|hendrycksTest-high_school_mathematics|5": {
"acc": 0.2851851851851852,
"acc_stderr": 0.027528599210340496,
"acc_norm": 0.2851851851851852,
"acc_norm_stderr": 0.027528599210340496
},
"harness|hendrycksTest-high_school_microeconomics|5": {
"acc": 0.46218487394957986,
"acc_stderr": 0.032385469487589795,
"acc_norm": 0.46218487394957986,
"acc_norm_stderr": 0.032385469487589795
},
"harness|hendrycksTest-high_school_physics|5": {
"acc": 0.31125827814569534,
"acc_stderr": 0.03780445850526733,
"acc_norm": 0.31125827814569534,
"acc_norm_stderr": 0.03780445850526733
},
"harness|hendrycksTest-high_school_psychology|5": {
"acc": 0.6165137614678899,
"acc_stderr": 0.020847156641915977,
"acc_norm": 0.6165137614678899,
"acc_norm_stderr": 0.020847156641915977
},
"harness|hendrycksTest-high_school_statistics|5": {
"acc": 0.23148148148148148,
"acc_stderr": 0.028765111718046955,
"acc_norm": 0.23148148148148148,
"acc_norm_stderr": 0.028765111718046955
},
"harness|hendrycksTest-high_school_us_history|5": {
"acc": 0.5,
"acc_stderr": 0.03509312031717982,
"acc_norm": 0.5,
"acc_norm_stderr": 0.03509312031717982
},
"harness|hendrycksTest-high_school_world_history|5": {
"acc": 0.5780590717299579,
"acc_stderr": 0.032148146302403695,
"acc_norm": 0.5780590717299579,
"acc_norm_stderr": 0.032148146302403695
},
"harness|hendrycksTest-human_aging|5": {
"acc": 0.5560538116591929,
"acc_stderr": 0.03334625674242728,
"acc_norm": 0.5560538116591929,
"acc_norm_stderr": 0.03334625674242728
},
"harness|hendrycksTest-human_sexuality|5": {
"acc": 0.5725190839694656,
"acc_stderr": 0.04338920305792401,
"acc_norm": 0.5725190839694656,
"acc_norm_stderr": 0.04338920305792401
},
"harness|hendrycksTest-international_law|5": {
"acc": 0.6198347107438017,
"acc_stderr": 0.04431324501968431,
"acc_norm": 0.6198347107438017,
"acc_norm_stderr": 0.04431324501968431
},
"harness|hendrycksTest-jurisprudence|5": {
"acc": 0.5370370370370371,
"acc_stderr": 0.04820403072760628,
"acc_norm": 0.5370370370370371,
"acc_norm_stderr": 0.04820403072760628
},
"harness|hendrycksTest-logical_fallacies|5": {
"acc": 0.49079754601226994,
"acc_stderr": 0.039277056007874414,
"acc_norm": 0.49079754601226994,
"acc_norm_stderr": 0.039277056007874414
},
"harness|hendrycksTest-machine_learning|5": {
"acc": 0.375,
"acc_stderr": 0.04595091388086298,
"acc_norm": 0.375,
"acc_norm_stderr": 0.04595091388086298
},
"harness|hendrycksTest-management|5": {
"acc": 0.6019417475728155,
"acc_stderr": 0.048467482539772386,
"acc_norm": 0.6019417475728155,
"acc_norm_stderr": 0.048467482539772386
},
"harness|hendrycksTest-marketing|5": {
"acc": 0.6709401709401709,
"acc_stderr": 0.030782321577688173,
"acc_norm": 0.6709401709401709,
"acc_norm_stderr": 0.030782321577688173
},
"harness|hendrycksTest-medical_genetics|5": {
"acc": 0.52,
"acc_stderr": 0.050211673156867795,
"acc_norm": 0.52,
"acc_norm_stderr": 0.050211673156867795
},
"harness|hendrycksTest-miscellaneous|5": {
"acc": 0.6143039591315453,
"acc_stderr": 0.017406476619212907,
"acc_norm": 0.6143039591315453,
"acc_norm_stderr": 0.017406476619212907
},
"harness|hendrycksTest-moral_disputes|5": {
"acc": 0.4884393063583815,
"acc_stderr": 0.026911898686377913,
"acc_norm": 0.4884393063583815,
"acc_norm_stderr": 0.026911898686377913
},
"harness|hendrycksTest-moral_scenarios|5": {
"acc": 0.23798882681564246,
"acc_stderr": 0.014242630070574915,
"acc_norm": 0.23798882681564246,
"acc_norm_stderr": 0.014242630070574915
},
"harness|hendrycksTest-nutrition|5": {
"acc": 0.4673202614379085,
"acc_stderr": 0.028568699752225875,
"acc_norm": 0.4673202614379085,
"acc_norm_stderr": 0.028568699752225875
},
"harness|hendrycksTest-philosophy|5": {
"acc": 0.5884244372990354,
"acc_stderr": 0.02795048149440127,
"acc_norm": 0.5884244372990354,
"acc_norm_stderr": 0.02795048149440127
},
"harness|hendrycksTest-prehistory|5": {
"acc": 0.4783950617283951,
"acc_stderr": 0.027794760105008746,
"acc_norm": 0.4783950617283951,
"acc_norm_stderr": 0.027794760105008746
},
"harness|hendrycksTest-professional_accounting|5": {
"acc": 0.3475177304964539,
"acc_stderr": 0.028406627809590954,
"acc_norm": 0.3475177304964539,
"acc_norm_stderr": 0.028406627809590954
},
"harness|hendrycksTest-professional_law|5": {
"acc": 0.35723598435462844,
"acc_stderr": 0.012238615750316503,
"acc_norm": 0.35723598435462844,
"acc_norm_stderr": 0.012238615750316503
},
"harness|hendrycksTest-professional_medicine|5": {
"acc": 0.49264705882352944,
"acc_stderr": 0.030369552523902173,
"acc_norm": 0.49264705882352944,
"acc_norm_stderr": 0.030369552523902173
},
"harness|hendrycksTest-professional_psychology|5": {
"acc": 0.43137254901960786,
"acc_stderr": 0.020036393768352638,
"acc_norm": 0.43137254901960786,
"acc_norm_stderr": 0.020036393768352638
},
"harness|hendrycksTest-public_relations|5": {
"acc": 0.5818181818181818,
"acc_stderr": 0.04724577405731572,
"acc_norm": 0.5818181818181818,
"acc_norm_stderr": 0.04724577405731572
},
"harness|hendrycksTest-security_studies|5": {
"acc": 0.46122448979591835,
"acc_stderr": 0.03191282052669277,
"acc_norm": 0.46122448979591835,
"acc_norm_stderr": 0.03191282052669277
},
"harness|hendrycksTest-sociology|5": {
"acc": 0.6069651741293532,
"acc_stderr": 0.0345368246603156,
"acc_norm": 0.6069651741293532,
"acc_norm_stderr": 0.0345368246603156
},
"harness|hendrycksTest-us_foreign_policy|5": {
"acc": 0.64,
"acc_stderr": 0.04824181513244218,
"acc_norm": 0.64,
"acc_norm_stderr": 0.04824181513244218
},
"harness|hendrycksTest-virology|5": {
"acc": 0.39759036144578314,
"acc_stderr": 0.038099730845402184,
"acc_norm": 0.39759036144578314,
"acc_norm_stderr": 0.038099730845402184
},
"harness|hendrycksTest-world_religions|5": {
"acc": 0.672514619883041,
"acc_stderr": 0.03599335771456027,
"acc_norm": 0.672514619883041,
"acc_norm_stderr": 0.03599335771456027
},
"harness|truthfulqa:mc|0": {
"mc1": 0.3011015911872705,
"mc1_stderr": 0.016058999026100612,
"mc2": 0.4456721895962505,
"mc2_stderr": 0.014265726453599933
},
"harness|winogrande|5": {
"acc": 0.7419100236779794,
"acc_stderr": 0.012298278833972397
},
"harness|drop|3": {
"em": 0.001363255033557047,
"em_stderr": 0.0003778609196460794,
"f1": 0.05649014261744978,
"f1_stderr": 0.0013342363586640303
},
"harness|gsm8k|5": {
"acc": 0.06292645943896892,
"acc_stderr": 0.0066887625815327395
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] | [
-0.6914949417114258,
-0.8606988191604614,
0.23809479176998138,
0.2551316022872925,
-0.19105802476406097,
-0.052804943174123764,
0.017448384314775467,
-0.23784257471561432,
0.6127488613128662,
-0.020069487392902374,
-0.5123556852340698,
-0.6848430633544922,
-0.44687047600746155,
0.210112005... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
arieg/bw_spec_cls_80_25 | arieg | 2023-11-09T14:05:33Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T14:05:33Z | 2023-11-09T14:05:15.000Z | 2023-11-09T14:05:15 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': '59702'
'1': '59706'
'2': '59707'
'3': '59708'
'4': '59709'
'5': '59710'
'6': '59719'
'7': '59720'
'8': '59721'
'9': '59723'
'10': '59724'
'11': '59725'
'12': '59726'
'13': '59727'
'14': '59823'
'15': '59876'
'16': '59930'
'17': '60037'
'18': '60038'
'19': '60041'
'20': '60042'
'21': '60045'
'22': '60048'
'23': '60074'
'24': '60143'
'25': '60144'
'26': '60145'
'27': '60146'
'28': '60170'
'29': '60317'
'30': '60472'
'31': '60474'
'32': '60477'
'33': '60478'
'34': '60510'
'35': '60544'
'36': '60547'
'37': '60548'
'38': '60549'
'39': '60736'
'40': '60753'
'41': '60754'
'42': '60755'
'43': '60756'
'44': '60757'
'45': '60758'
'46': '60775'
'47': '60776'
'48': '60777'
'49': '60857'
'50': '60864'
'51': '60865'
'52': '60994'
'53': '61006'
'54': '61007'
'55': '61008'
'56': '61010'
'57': '61011'
'58': '61012'
'59': '61013'
'60': '61159'
'61': '61160'
'62': '61161'
'63': '61172'
'64': '61174'
'65': '61175'
'66': '61452'
'67': '61453'
'68': '61491'
'69': '61492'
'70': '61493'
'71': '61587'
'72': '61589'
'73': '61591'
'74': '61592'
'75': '61668'
'76': '61670'
'77': '61679'
'78': '61814'
'79': '61884'
splits:
- name: train
num_bytes: 93110896.0
num_examples: 1600
- name: test
num_bytes: 22653803.0
num_examples: 400
download_size: 113211430
dataset_size: 115764699.0
---
# Dataset Card for "bw_spec_cls_80_25"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.7564918994903564,
-0.15674267709255219,
0.16004414856433868,
0.4123137295246124,
-0.2835460901260376,
-0.11639276146888733,
0.01473055686801672,
-0.3027874529361725,
0.5583688020706177,
0.5254092216491699,
-0.8148316740989685,
-0.8411582708358765,
-0.5278752446174622,
-0.231248259544372... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
mesolitica/translated-glaive-function-call | mesolitica | 2023-11-09T14:17:12Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T14:17:12Z | 2023-11-09T14:15:54.000Z | 2023-11-09T14:15:54 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
open-llm-leaderboard/details_Weyaxi__Dolphin2.1-OpenOrca-7B_public | open-llm-leaderboard | 2023-11-09T14:25:33Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T14:25:33Z | 2023-11-09T14:16:24.000Z | 2023-11-09T14:16:24 | ---
pretty_name: Evaluation run of Weyaxi/Dolphin2.1-OpenOrca-7B
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [Weyaxi/Dolphin2.1-OpenOrca-7B](https://huggingface.co/Weyaxi/Dolphin2.1-OpenOrca-7B)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 64 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 2 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the aggregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_Weyaxi__Dolphin2.1-OpenOrca-7B_public\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-11-09T14:21:27.933712](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__Dolphin2.1-OpenOrca-7B_public/blob/main/results_2023-11-09T14-21-27.933712.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"acc\": 0.6224567348896717,\n\
\ \"acc_stderr\": 0.032466479047476085,\n \"acc_norm\": 0.6308724361156662,\n\
\ \"acc_norm_stderr\": 0.033159611933737225,\n \"mc1\": 0.36107711138310894,\n\
\ \"mc1_stderr\": 0.016814312844836886,\n \"mc2\": 0.538254375639854,\n\
\ \"mc2_stderr\": 0.015244755693358225,\n \"em\": 0.0030411073825503355,\n\
\ \"em_stderr\": 0.0005638896908753155,\n \"f1\": 0.08151740771812048,\n\
\ \"f1_stderr\": 0.0016591952257614033\n },\n \"harness|arc:challenge|25\"\
: {\n \"acc\": 0.6083617747440273,\n \"acc_stderr\": 0.01426412212493821,\n\
\ \"acc_norm\": 0.6416382252559727,\n \"acc_norm_stderr\": 0.014012883334859857\n\
\ },\n \"harness|hellaswag|10\": {\n \"acc\": 0.650368452499502,\n\
\ \"acc_stderr\": 0.004758790172436687,\n \"acc_norm\": 0.8424616610237005,\n\
\ \"acc_norm_stderr\": 0.0036356303524759065\n },\n \"harness|hendrycksTest-abstract_algebra|5\"\
: {\n \"acc\": 0.32,\n \"acc_stderr\": 0.046882617226215034,\n \
\ \"acc_norm\": 0.32,\n \"acc_norm_stderr\": 0.046882617226215034\n \
\ },\n \"harness|hendrycksTest-anatomy|5\": {\n \"acc\": 0.6,\n \
\ \"acc_stderr\": 0.04232073695151589,\n \"acc_norm\": 0.6,\n \"\
acc_norm_stderr\": 0.04232073695151589\n },\n \"harness|hendrycksTest-astronomy|5\"\
: {\n \"acc\": 0.6710526315789473,\n \"acc_stderr\": 0.03823428969926605,\n\
\ \"acc_norm\": 0.6710526315789473,\n \"acc_norm_stderr\": 0.03823428969926605\n\
\ },\n \"harness|hendrycksTest-business_ethics|5\": {\n \"acc\": 0.59,\n\
\ \"acc_stderr\": 0.04943110704237102,\n \"acc_norm\": 0.59,\n \
\ \"acc_norm_stderr\": 0.04943110704237102\n },\n \"harness|hendrycksTest-clinical_knowledge|5\"\
: {\n \"acc\": 0.6981132075471698,\n \"acc_stderr\": 0.02825420034443866,\n\
\ \"acc_norm\": 0.6981132075471698,\n \"acc_norm_stderr\": 0.02825420034443866\n\
\ },\n \"harness|hendrycksTest-college_biology|5\": {\n \"acc\": 0.7222222222222222,\n\
\ \"acc_stderr\": 0.037455547914624555,\n \"acc_norm\": 0.7222222222222222,\n\
\ \"acc_norm_stderr\": 0.037455547914624555\n },\n \"harness|hendrycksTest-college_chemistry|5\"\
: {\n \"acc\": 0.48,\n \"acc_stderr\": 0.050211673156867795,\n \
\ \"acc_norm\": 0.48,\n \"acc_norm_stderr\": 0.050211673156867795\n \
\ },\n \"harness|hendrycksTest-college_computer_science|5\": {\n \"\
acc\": 0.54,\n \"acc_stderr\": 0.05009082659620333,\n \"acc_norm\"\
: 0.54,\n \"acc_norm_stderr\": 0.05009082659620333\n },\n \"harness|hendrycksTest-college_mathematics|5\"\
: {\n \"acc\": 0.37,\n \"acc_stderr\": 0.04852365870939099,\n \
\ \"acc_norm\": 0.37,\n \"acc_norm_stderr\": 0.04852365870939099\n \
\ },\n \"harness|hendrycksTest-college_medicine|5\": {\n \"acc\": 0.5953757225433526,\n\
\ \"acc_stderr\": 0.03742461193887248,\n \"acc_norm\": 0.5953757225433526,\n\
\ \"acc_norm_stderr\": 0.03742461193887248\n },\n \"harness|hendrycksTest-college_physics|5\"\
: {\n \"acc\": 0.3333333333333333,\n \"acc_stderr\": 0.04690650298201943,\n\
\ \"acc_norm\": 0.3333333333333333,\n \"acc_norm_stderr\": 0.04690650298201943\n\
\ },\n \"harness|hendrycksTest-computer_security|5\": {\n \"acc\":\
\ 0.79,\n \"acc_stderr\": 0.040936018074033256,\n \"acc_norm\": 0.79,\n\
\ \"acc_norm_stderr\": 0.040936018074033256\n },\n \"harness|hendrycksTest-conceptual_physics|5\"\
: {\n \"acc\": 0.5531914893617021,\n \"acc_stderr\": 0.0325005368436584,\n\
\ \"acc_norm\": 0.5531914893617021,\n \"acc_norm_stderr\": 0.0325005368436584\n\
\ },\n \"harness|hendrycksTest-econometrics|5\": {\n \"acc\": 0.4298245614035088,\n\
\ \"acc_stderr\": 0.04657047260594963,\n \"acc_norm\": 0.4298245614035088,\n\
\ \"acc_norm_stderr\": 0.04657047260594963\n },\n \"harness|hendrycksTest-electrical_engineering|5\"\
: {\n \"acc\": 0.5793103448275863,\n \"acc_stderr\": 0.0411391498118926,\n\
\ \"acc_norm\": 0.5793103448275863,\n \"acc_norm_stderr\": 0.0411391498118926\n\
\ },\n \"harness|hendrycksTest-elementary_mathematics|5\": {\n \"acc\"\
: 0.3968253968253968,\n \"acc_stderr\": 0.025197101074246487,\n \"\
acc_norm\": 0.3968253968253968,\n \"acc_norm_stderr\": 0.025197101074246487\n\
\ },\n \"harness|hendrycksTest-formal_logic|5\": {\n \"acc\": 0.42063492063492064,\n\
\ \"acc_stderr\": 0.04415438226743744,\n \"acc_norm\": 0.42063492063492064,\n\
\ \"acc_norm_stderr\": 0.04415438226743744\n },\n \"harness|hendrycksTest-global_facts|5\"\
: {\n \"acc\": 0.3,\n \"acc_stderr\": 0.046056618647183814,\n \
\ \"acc_norm\": 0.3,\n \"acc_norm_stderr\": 0.046056618647183814\n \
\ },\n \"harness|hendrycksTest-high_school_biology|5\": {\n \"acc\": 0.7645161290322581,\n\
\ \"acc_stderr\": 0.02413763242933771,\n \"acc_norm\": 0.7645161290322581,\n\
\ \"acc_norm_stderr\": 0.02413763242933771\n },\n \"harness|hendrycksTest-high_school_chemistry|5\"\
: {\n \"acc\": 0.5024630541871922,\n \"acc_stderr\": 0.03517945038691063,\n\
\ \"acc_norm\": 0.5024630541871922,\n \"acc_norm_stderr\": 0.03517945038691063\n\
\ },\n \"harness|hendrycksTest-high_school_computer_science|5\": {\n \
\ \"acc\": 0.69,\n \"acc_stderr\": 0.04648231987117316,\n \"acc_norm\"\
: 0.69,\n \"acc_norm_stderr\": 0.04648231987117316\n },\n \"harness|hendrycksTest-high_school_european_history|5\"\
: {\n \"acc\": 0.7515151515151515,\n \"acc_stderr\": 0.03374402644139403,\n\
\ \"acc_norm\": 0.7515151515151515,\n \"acc_norm_stderr\": 0.03374402644139403\n\
\ },\n \"harness|hendrycksTest-high_school_geography|5\": {\n \"acc\"\
: 0.7727272727272727,\n \"acc_stderr\": 0.029857515673386414,\n \"\
acc_norm\": 0.7727272727272727,\n \"acc_norm_stderr\": 0.029857515673386414\n\
\ },\n \"harness|hendrycksTest-high_school_government_and_politics|5\": {\n\
\ \"acc\": 0.8652849740932642,\n \"acc_stderr\": 0.02463978909770944,\n\
\ \"acc_norm\": 0.8652849740932642,\n \"acc_norm_stderr\": 0.02463978909770944\n\
\ },\n \"harness|hendrycksTest-high_school_macroeconomics|5\": {\n \
\ \"acc\": 0.6333333333333333,\n \"acc_stderr\": 0.02443301646605246,\n \
\ \"acc_norm\": 0.6333333333333333,\n \"acc_norm_stderr\": 0.02443301646605246\n\
\ },\n \"harness|hendrycksTest-high_school_mathematics|5\": {\n \"\
acc\": 0.32592592592592595,\n \"acc_stderr\": 0.02857834836547308,\n \
\ \"acc_norm\": 0.32592592592592595,\n \"acc_norm_stderr\": 0.02857834836547308\n\
\ },\n \"harness|hendrycksTest-high_school_microeconomics|5\": {\n \
\ \"acc\": 0.6596638655462185,\n \"acc_stderr\": 0.03077805742293167,\n \
\ \"acc_norm\": 0.6596638655462185,\n \"acc_norm_stderr\": 0.03077805742293167\n\
\ },\n \"harness|hendrycksTest-high_school_physics|5\": {\n \"acc\"\
: 0.3509933774834437,\n \"acc_stderr\": 0.03896981964257375,\n \"\
acc_norm\": 0.3509933774834437,\n \"acc_norm_stderr\": 0.03896981964257375\n\
\ },\n \"harness|hendrycksTest-high_school_psychology|5\": {\n \"acc\"\
: 0.8128440366972477,\n \"acc_stderr\": 0.016722684526200148,\n \"\
acc_norm\": 0.8128440366972477,\n \"acc_norm_stderr\": 0.016722684526200148\n\
\ },\n \"harness|hendrycksTest-high_school_statistics|5\": {\n \"acc\"\
: 0.49074074074074076,\n \"acc_stderr\": 0.034093869469927006,\n \"\
acc_norm\": 0.49074074074074076,\n \"acc_norm_stderr\": 0.034093869469927006\n\
\ },\n \"harness|hendrycksTest-high_school_us_history|5\": {\n \"acc\"\
: 0.7990196078431373,\n \"acc_stderr\": 0.02812597226565437,\n \"\
acc_norm\": 0.7990196078431373,\n \"acc_norm_stderr\": 0.02812597226565437\n\
\ },\n \"harness|hendrycksTest-high_school_world_history|5\": {\n \"\
acc\": 0.7974683544303798,\n \"acc_stderr\": 0.026160568246601436,\n \
\ \"acc_norm\": 0.7974683544303798,\n \"acc_norm_stderr\": 0.026160568246601436\n\
\ },\n \"harness|hendrycksTest-human_aging|5\": {\n \"acc\": 0.6681614349775785,\n\
\ \"acc_stderr\": 0.03160295143776679,\n \"acc_norm\": 0.6681614349775785,\n\
\ \"acc_norm_stderr\": 0.03160295143776679\n },\n \"harness|hendrycksTest-human_sexuality|5\"\
: {\n \"acc\": 0.7709923664122137,\n \"acc_stderr\": 0.036853466317118506,\n\
\ \"acc_norm\": 0.7709923664122137,\n \"acc_norm_stderr\": 0.036853466317118506\n\
\ },\n \"harness|hendrycksTest-international_law|5\": {\n \"acc\":\
\ 0.7768595041322314,\n \"acc_stderr\": 0.03800754475228733,\n \"\
acc_norm\": 0.7768595041322314,\n \"acc_norm_stderr\": 0.03800754475228733\n\
\ },\n \"harness|hendrycksTest-jurisprudence|5\": {\n \"acc\": 0.7592592592592593,\n\
\ \"acc_stderr\": 0.04133119440243838,\n \"acc_norm\": 0.7592592592592593,\n\
\ \"acc_norm_stderr\": 0.04133119440243838\n },\n \"harness|hendrycksTest-logical_fallacies|5\"\
: {\n \"acc\": 0.7423312883435583,\n \"acc_stderr\": 0.03436150827846917,\n\
\ \"acc_norm\": 0.7423312883435583,\n \"acc_norm_stderr\": 0.03436150827846917\n\
\ },\n \"harness|hendrycksTest-machine_learning|5\": {\n \"acc\": 0.4642857142857143,\n\
\ \"acc_stderr\": 0.04733667890053756,\n \"acc_norm\": 0.4642857142857143,\n\
\ \"acc_norm_stderr\": 0.04733667890053756\n },\n \"harness|hendrycksTest-management|5\"\
: {\n \"acc\": 0.8058252427184466,\n \"acc_stderr\": 0.039166677628225836,\n\
\ \"acc_norm\": 0.8058252427184466,\n \"acc_norm_stderr\": 0.039166677628225836\n\
\ },\n \"harness|hendrycksTest-marketing|5\": {\n \"acc\": 0.8675213675213675,\n\
\ \"acc_stderr\": 0.02220930907316561,\n \"acc_norm\": 0.8675213675213675,\n\
\ \"acc_norm_stderr\": 0.02220930907316561\n },\n \"harness|hendrycksTest-medical_genetics|5\"\
: {\n \"acc\": 0.72,\n \"acc_stderr\": 0.04512608598542128,\n \
\ \"acc_norm\": 0.72,\n \"acc_norm_stderr\": 0.04512608598542128\n \
\ },\n \"harness|hendrycksTest-miscellaneous|5\": {\n \"acc\": 0.8058748403575989,\n\
\ \"acc_stderr\": 0.014143970276657567,\n \"acc_norm\": 0.8058748403575989,\n\
\ \"acc_norm_stderr\": 0.014143970276657567\n },\n \"harness|hendrycksTest-moral_disputes|5\"\
: {\n \"acc\": 0.7109826589595376,\n \"acc_stderr\": 0.024405173935783234,\n\
\ \"acc_norm\": 0.7109826589595376,\n \"acc_norm_stderr\": 0.024405173935783234\n\
\ },\n \"harness|hendrycksTest-moral_scenarios|5\": {\n \"acc\": 0.3340782122905028,\n\
\ \"acc_stderr\": 0.015774911422381625,\n \"acc_norm\": 0.3340782122905028,\n\
\ \"acc_norm_stderr\": 0.015774911422381625\n },\n \"harness|hendrycksTest-nutrition|5\"\
: {\n \"acc\": 0.6993464052287581,\n \"acc_stderr\": 0.026256053835718964,\n\
\ \"acc_norm\": 0.6993464052287581,\n \"acc_norm_stderr\": 0.026256053835718964\n\
\ },\n \"harness|hendrycksTest-philosophy|5\": {\n \"acc\": 0.684887459807074,\n\
\ \"acc_stderr\": 0.02638527370346449,\n \"acc_norm\": 0.684887459807074,\n\
\ \"acc_norm_stderr\": 0.02638527370346449\n },\n \"harness|hendrycksTest-prehistory|5\"\
: {\n \"acc\": 0.7191358024691358,\n \"acc_stderr\": 0.02500646975579921,\n\
\ \"acc_norm\": 0.7191358024691358,\n \"acc_norm_stderr\": 0.02500646975579921\n\
\ },\n \"harness|hendrycksTest-professional_accounting|5\": {\n \"\
acc\": 0.44680851063829785,\n \"acc_stderr\": 0.029658235097666904,\n \
\ \"acc_norm\": 0.44680851063829785,\n \"acc_norm_stderr\": 0.029658235097666904\n\
\ },\n \"harness|hendrycksTest-professional_law|5\": {\n \"acc\": 0.4621903520208605,\n\
\ \"acc_stderr\": 0.012733671880342507,\n \"acc_norm\": 0.4621903520208605,\n\
\ \"acc_norm_stderr\": 0.012733671880342507\n },\n \"harness|hendrycksTest-professional_medicine|5\"\
: {\n \"acc\": 0.6176470588235294,\n \"acc_stderr\": 0.029520095697687765,\n\
\ \"acc_norm\": 0.6176470588235294,\n \"acc_norm_stderr\": 0.029520095697687765\n\
\ },\n \"harness|hendrycksTest-professional_psychology|5\": {\n \"\
acc\": 0.6470588235294118,\n \"acc_stderr\": 0.01933314202079716,\n \
\ \"acc_norm\": 0.6470588235294118,\n \"acc_norm_stderr\": 0.01933314202079716\n\
\ },\n \"harness|hendrycksTest-public_relations|5\": {\n \"acc\": 0.6727272727272727,\n\
\ \"acc_stderr\": 0.0449429086625209,\n \"acc_norm\": 0.6727272727272727,\n\
\ \"acc_norm_stderr\": 0.0449429086625209\n },\n \"harness|hendrycksTest-security_studies|5\"\
: {\n \"acc\": 0.7183673469387755,\n \"acc_stderr\": 0.028795185574291296,\n\
\ \"acc_norm\": 0.7183673469387755,\n \"acc_norm_stderr\": 0.028795185574291296\n\
\ },\n \"harness|hendrycksTest-sociology|5\": {\n \"acc\": 0.845771144278607,\n\
\ \"acc_stderr\": 0.025538433368578337,\n \"acc_norm\": 0.845771144278607,\n\
\ \"acc_norm_stderr\": 0.025538433368578337\n },\n \"harness|hendrycksTest-us_foreign_policy|5\"\
: {\n \"acc\": 0.83,\n \"acc_stderr\": 0.0377525168068637,\n \
\ \"acc_norm\": 0.83,\n \"acc_norm_stderr\": 0.0377525168068637\n },\n\
\ \"harness|hendrycksTest-virology|5\": {\n \"acc\": 0.5421686746987951,\n\
\ \"acc_stderr\": 0.0387862677100236,\n \"acc_norm\": 0.5421686746987951,\n\
\ \"acc_norm_stderr\": 0.0387862677100236\n },\n \"harness|hendrycksTest-world_religions|5\"\
: {\n \"acc\": 0.8187134502923976,\n \"acc_stderr\": 0.029547741687640038,\n\
\ \"acc_norm\": 0.8187134502923976,\n \"acc_norm_stderr\": 0.029547741687640038\n\
\ },\n \"harness|truthfulqa:mc|0\": {\n \"mc1\": 0.36107711138310894,\n\
\ \"mc1_stderr\": 0.016814312844836886,\n \"mc2\": 0.538254375639854,\n\
\ \"mc2_stderr\": 0.015244755693358225\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.77663772691397,\n \"acc_stderr\": 0.011705697565205193\n\
\ },\n \"harness|drop|3\": {\n \"em\": 0.0030411073825503355,\n \
\ \"em_stderr\": 0.0005638896908753155,\n \"f1\": 0.08151740771812048,\n\
\ \"f1_stderr\": 0.0016591952257614033\n },\n \"harness|gsm8k|5\":\
\ {\n \"acc\": 0.19711902956785443,\n \"acc_stderr\": 0.01095802163030062\n\
\ }\n}\n```"
repo_url: https://huggingface.co/Weyaxi/Dolphin2.1-OpenOrca-7B
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|arc:challenge|25_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|arc:challenge|25_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_drop_3
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|drop|3_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|drop|3_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|drop|3_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|gsm8k|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|gsm8k|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hellaswag|10_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hellaswag|10_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-13-23.628272.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-21-27.933712.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T14-21-27.933712.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- '**/details_harness|winogrande|5_2023-11-09T14-13-23.628272.parquet'
- split: 2023_11_09T14_21_27.933712
path:
- '**/details_harness|winogrande|5_2023-11-09T14-21-27.933712.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-11-09T14-21-27.933712.parquet'
- config_name: results
data_files:
- split: 2023_11_09T14_13_23.628272
path:
- results_2023-11-09T14-13-23.628272.parquet
- split: 2023_11_09T14_21_27.933712
path:
- results_2023-11-09T14-21-27.933712.parquet
- split: latest
path:
- results_2023-11-09T14-21-27.933712.parquet
---
# Dataset Card for Evaluation run of Weyaxi/Dolphin2.1-OpenOrca-7B
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/Weyaxi/Dolphin2.1-OpenOrca-7B
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [Weyaxi/Dolphin2.1-OpenOrca-7B](https://huggingface.co/Weyaxi/Dolphin2.1-OpenOrca-7B) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 64 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 2 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_Weyaxi__Dolphin2.1-OpenOrca-7B_public",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-11-09T14:21:27.933712](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__Dolphin2.1-OpenOrca-7B_public/blob/main/results_2023-11-09T14-21-27.933712.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"acc": 0.6224567348896717,
"acc_stderr": 0.032466479047476085,
"acc_norm": 0.6308724361156662,
"acc_norm_stderr": 0.033159611933737225,
"mc1": 0.36107711138310894,
"mc1_stderr": 0.016814312844836886,
"mc2": 0.538254375639854,
"mc2_stderr": 0.015244755693358225,
"em": 0.0030411073825503355,
"em_stderr": 0.0005638896908753155,
"f1": 0.08151740771812048,
"f1_stderr": 0.0016591952257614033
},
"harness|arc:challenge|25": {
"acc": 0.6083617747440273,
"acc_stderr": 0.01426412212493821,
"acc_norm": 0.6416382252559727,
"acc_norm_stderr": 0.014012883334859857
},
"harness|hellaswag|10": {
"acc": 0.650368452499502,
"acc_stderr": 0.004758790172436687,
"acc_norm": 0.8424616610237005,
"acc_norm_stderr": 0.0036356303524759065
},
"harness|hendrycksTest-abstract_algebra|5": {
"acc": 0.32,
"acc_stderr": 0.046882617226215034,
"acc_norm": 0.32,
"acc_norm_stderr": 0.046882617226215034
},
"harness|hendrycksTest-anatomy|5": {
"acc": 0.6,
"acc_stderr": 0.04232073695151589,
"acc_norm": 0.6,
"acc_norm_stderr": 0.04232073695151589
},
"harness|hendrycksTest-astronomy|5": {
"acc": 0.6710526315789473,
"acc_stderr": 0.03823428969926605,
"acc_norm": 0.6710526315789473,
"acc_norm_stderr": 0.03823428969926605
},
"harness|hendrycksTest-business_ethics|5": {
"acc": 0.59,
"acc_stderr": 0.04943110704237102,
"acc_norm": 0.59,
"acc_norm_stderr": 0.04943110704237102
},
"harness|hendrycksTest-clinical_knowledge|5": {
"acc": 0.6981132075471698,
"acc_stderr": 0.02825420034443866,
"acc_norm": 0.6981132075471698,
"acc_norm_stderr": 0.02825420034443866
},
"harness|hendrycksTest-college_biology|5": {
"acc": 0.7222222222222222,
"acc_stderr": 0.037455547914624555,
"acc_norm": 0.7222222222222222,
"acc_norm_stderr": 0.037455547914624555
},
"harness|hendrycksTest-college_chemistry|5": {
"acc": 0.48,
"acc_stderr": 0.050211673156867795,
"acc_norm": 0.48,
"acc_norm_stderr": 0.050211673156867795
},
"harness|hendrycksTest-college_computer_science|5": {
"acc": 0.54,
"acc_stderr": 0.05009082659620333,
"acc_norm": 0.54,
"acc_norm_stderr": 0.05009082659620333
},
"harness|hendrycksTest-college_mathematics|5": {
"acc": 0.37,
"acc_stderr": 0.04852365870939099,
"acc_norm": 0.37,
"acc_norm_stderr": 0.04852365870939099
},
"harness|hendrycksTest-college_medicine|5": {
"acc": 0.5953757225433526,
"acc_stderr": 0.03742461193887248,
"acc_norm": 0.5953757225433526,
"acc_norm_stderr": 0.03742461193887248
},
"harness|hendrycksTest-college_physics|5": {
"acc": 0.3333333333333333,
"acc_stderr": 0.04690650298201943,
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.04690650298201943
},
"harness|hendrycksTest-computer_security|5": {
"acc": 0.79,
"acc_stderr": 0.040936018074033256,
"acc_norm": 0.79,
"acc_norm_stderr": 0.040936018074033256
},
"harness|hendrycksTest-conceptual_physics|5": {
"acc": 0.5531914893617021,
"acc_stderr": 0.0325005368436584,
"acc_norm": 0.5531914893617021,
"acc_norm_stderr": 0.0325005368436584
},
"harness|hendrycksTest-econometrics|5": {
"acc": 0.4298245614035088,
"acc_stderr": 0.04657047260594963,
"acc_norm": 0.4298245614035088,
"acc_norm_stderr": 0.04657047260594963
},
"harness|hendrycksTest-electrical_engineering|5": {
"acc": 0.5793103448275863,
"acc_stderr": 0.0411391498118926,
"acc_norm": 0.5793103448275863,
"acc_norm_stderr": 0.0411391498118926
},
"harness|hendrycksTest-elementary_mathematics|5": {
"acc": 0.3968253968253968,
"acc_stderr": 0.025197101074246487,
"acc_norm": 0.3968253968253968,
"acc_norm_stderr": 0.025197101074246487
},
"harness|hendrycksTest-formal_logic|5": {
"acc": 0.42063492063492064,
"acc_stderr": 0.04415438226743744,
"acc_norm": 0.42063492063492064,
"acc_norm_stderr": 0.04415438226743744
},
"harness|hendrycksTest-global_facts|5": {
"acc": 0.3,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.3,
"acc_norm_stderr": 0.046056618647183814
},
"harness|hendrycksTest-high_school_biology|5": {
"acc": 0.7645161290322581,
"acc_stderr": 0.02413763242933771,
"acc_norm": 0.7645161290322581,
"acc_norm_stderr": 0.02413763242933771
},
"harness|hendrycksTest-high_school_chemistry|5": {
"acc": 0.5024630541871922,
"acc_stderr": 0.03517945038691063,
"acc_norm": 0.5024630541871922,
"acc_norm_stderr": 0.03517945038691063
},
"harness|hendrycksTest-high_school_computer_science|5": {
"acc": 0.69,
"acc_stderr": 0.04648231987117316,
"acc_norm": 0.69,
"acc_norm_stderr": 0.04648231987117316
},
"harness|hendrycksTest-high_school_european_history|5": {
"acc": 0.7515151515151515,
"acc_stderr": 0.03374402644139403,
"acc_norm": 0.7515151515151515,
"acc_norm_stderr": 0.03374402644139403
},
"harness|hendrycksTest-high_school_geography|5": {
"acc": 0.7727272727272727,
"acc_stderr": 0.029857515673386414,
"acc_norm": 0.7727272727272727,
"acc_norm_stderr": 0.029857515673386414
},
"harness|hendrycksTest-high_school_government_and_politics|5": {
"acc": 0.8652849740932642,
"acc_stderr": 0.02463978909770944,
"acc_norm": 0.8652849740932642,
"acc_norm_stderr": 0.02463978909770944
},
"harness|hendrycksTest-high_school_macroeconomics|5": {
"acc": 0.6333333333333333,
"acc_stderr": 0.02443301646605246,
"acc_norm": 0.6333333333333333,
"acc_norm_stderr": 0.02443301646605246
},
"harness|hendrycksTest-high_school_mathematics|5": {
"acc": 0.32592592592592595,
"acc_stderr": 0.02857834836547308,
"acc_norm": 0.32592592592592595,
"acc_norm_stderr": 0.02857834836547308
},
"harness|hendrycksTest-high_school_microeconomics|5": {
"acc": 0.6596638655462185,
"acc_stderr": 0.03077805742293167,
"acc_norm": 0.6596638655462185,
"acc_norm_stderr": 0.03077805742293167
},
"harness|hendrycksTest-high_school_physics|5": {
"acc": 0.3509933774834437,
"acc_stderr": 0.03896981964257375,
"acc_norm": 0.3509933774834437,
"acc_norm_stderr": 0.03896981964257375
},
"harness|hendrycksTest-high_school_psychology|5": {
"acc": 0.8128440366972477,
"acc_stderr": 0.016722684526200148,
"acc_norm": 0.8128440366972477,
"acc_norm_stderr": 0.016722684526200148
},
"harness|hendrycksTest-high_school_statistics|5": {
"acc": 0.49074074074074076,
"acc_stderr": 0.034093869469927006,
"acc_norm": 0.49074074074074076,
"acc_norm_stderr": 0.034093869469927006
},
"harness|hendrycksTest-high_school_us_history|5": {
"acc": 0.7990196078431373,
"acc_stderr": 0.02812597226565437,
"acc_norm": 0.7990196078431373,
"acc_norm_stderr": 0.02812597226565437
},
"harness|hendrycksTest-high_school_world_history|5": {
"acc": 0.7974683544303798,
"acc_stderr": 0.026160568246601436,
"acc_norm": 0.7974683544303798,
"acc_norm_stderr": 0.026160568246601436
},
"harness|hendrycksTest-human_aging|5": {
"acc": 0.6681614349775785,
"acc_stderr": 0.03160295143776679,
"acc_norm": 0.6681614349775785,
"acc_norm_stderr": 0.03160295143776679
},
"harness|hendrycksTest-human_sexuality|5": {
"acc": 0.7709923664122137,
"acc_stderr": 0.036853466317118506,
"acc_norm": 0.7709923664122137,
"acc_norm_stderr": 0.036853466317118506
},
"harness|hendrycksTest-international_law|5": {
"acc": 0.7768595041322314,
"acc_stderr": 0.03800754475228733,
"acc_norm": 0.7768595041322314,
"acc_norm_stderr": 0.03800754475228733
},
"harness|hendrycksTest-jurisprudence|5": {
"acc": 0.7592592592592593,
"acc_stderr": 0.04133119440243838,
"acc_norm": 0.7592592592592593,
"acc_norm_stderr": 0.04133119440243838
},
"harness|hendrycksTest-logical_fallacies|5": {
"acc": 0.7423312883435583,
"acc_stderr": 0.03436150827846917,
"acc_norm": 0.7423312883435583,
"acc_norm_stderr": 0.03436150827846917
},
"harness|hendrycksTest-machine_learning|5": {
"acc": 0.4642857142857143,
"acc_stderr": 0.04733667890053756,
"acc_norm": 0.4642857142857143,
"acc_norm_stderr": 0.04733667890053756
},
"harness|hendrycksTest-management|5": {
"acc": 0.8058252427184466,
"acc_stderr": 0.039166677628225836,
"acc_norm": 0.8058252427184466,
"acc_norm_stderr": 0.039166677628225836
},
"harness|hendrycksTest-marketing|5": {
"acc": 0.8675213675213675,
"acc_stderr": 0.02220930907316561,
"acc_norm": 0.8675213675213675,
"acc_norm_stderr": 0.02220930907316561
},
"harness|hendrycksTest-medical_genetics|5": {
"acc": 0.72,
"acc_stderr": 0.04512608598542128,
"acc_norm": 0.72,
"acc_norm_stderr": 0.04512608598542128
},
"harness|hendrycksTest-miscellaneous|5": {
"acc": 0.8058748403575989,
"acc_stderr": 0.014143970276657567,
"acc_norm": 0.8058748403575989,
"acc_norm_stderr": 0.014143970276657567
},
"harness|hendrycksTest-moral_disputes|5": {
"acc": 0.7109826589595376,
"acc_stderr": 0.024405173935783234,
"acc_norm": 0.7109826589595376,
"acc_norm_stderr": 0.024405173935783234
},
"harness|hendrycksTest-moral_scenarios|5": {
"acc": 0.3340782122905028,
"acc_stderr": 0.015774911422381625,
"acc_norm": 0.3340782122905028,
"acc_norm_stderr": 0.015774911422381625
},
"harness|hendrycksTest-nutrition|5": {
"acc": 0.6993464052287581,
"acc_stderr": 0.026256053835718964,
"acc_norm": 0.6993464052287581,
"acc_norm_stderr": 0.026256053835718964
},
"harness|hendrycksTest-philosophy|5": {
"acc": 0.684887459807074,
"acc_stderr": 0.02638527370346449,
"acc_norm": 0.684887459807074,
"acc_norm_stderr": 0.02638527370346449
},
"harness|hendrycksTest-prehistory|5": {
"acc": 0.7191358024691358,
"acc_stderr": 0.02500646975579921,
"acc_norm": 0.7191358024691358,
"acc_norm_stderr": 0.02500646975579921
},
"harness|hendrycksTest-professional_accounting|5": {
"acc": 0.44680851063829785,
"acc_stderr": 0.029658235097666904,
"acc_norm": 0.44680851063829785,
"acc_norm_stderr": 0.029658235097666904
},
"harness|hendrycksTest-professional_law|5": {
"acc": 0.4621903520208605,
"acc_stderr": 0.012733671880342507,
"acc_norm": 0.4621903520208605,
"acc_norm_stderr": 0.012733671880342507
},
"harness|hendrycksTest-professional_medicine|5": {
"acc": 0.6176470588235294,
"acc_stderr": 0.029520095697687765,
"acc_norm": 0.6176470588235294,
"acc_norm_stderr": 0.029520095697687765
},
"harness|hendrycksTest-professional_psychology|5": {
"acc": 0.6470588235294118,
"acc_stderr": 0.01933314202079716,
"acc_norm": 0.6470588235294118,
"acc_norm_stderr": 0.01933314202079716
},
"harness|hendrycksTest-public_relations|5": {
"acc": 0.6727272727272727,
"acc_stderr": 0.0449429086625209,
"acc_norm": 0.6727272727272727,
"acc_norm_stderr": 0.0449429086625209
},
"harness|hendrycksTest-security_studies|5": {
"acc": 0.7183673469387755,
"acc_stderr": 0.028795185574291296,
"acc_norm": 0.7183673469387755,
"acc_norm_stderr": 0.028795185574291296
},
"harness|hendrycksTest-sociology|5": {
"acc": 0.845771144278607,
"acc_stderr": 0.025538433368578337,
"acc_norm": 0.845771144278607,
"acc_norm_stderr": 0.025538433368578337
},
"harness|hendrycksTest-us_foreign_policy|5": {
"acc": 0.83,
"acc_stderr": 0.0377525168068637,
"acc_norm": 0.83,
"acc_norm_stderr": 0.0377525168068637
},
"harness|hendrycksTest-virology|5": {
"acc": 0.5421686746987951,
"acc_stderr": 0.0387862677100236,
"acc_norm": 0.5421686746987951,
"acc_norm_stderr": 0.0387862677100236
},
"harness|hendrycksTest-world_religions|5": {
"acc": 0.8187134502923976,
"acc_stderr": 0.029547741687640038,
"acc_norm": 0.8187134502923976,
"acc_norm_stderr": 0.029547741687640038
},
"harness|truthfulqa:mc|0": {
"mc1": 0.36107711138310894,
"mc1_stderr": 0.016814312844836886,
"mc2": 0.538254375639854,
"mc2_stderr": 0.015244755693358225
},
"harness|winogrande|5": {
"acc": 0.77663772691397,
"acc_stderr": 0.011705697565205193
},
"harness|drop|3": {
"em": 0.0030411073825503355,
"em_stderr": 0.0005638896908753155,
"f1": 0.08151740771812048,
"f1_stderr": 0.0016591952257614033
},
"harness|gsm8k|5": {
"acc": 0.19711902956785443,
"acc_stderr": 0.01095802163030062
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] | [
-0.7714576721191406,
-0.8367005586624146,
0.2529853582382202,
0.22413581609725952,
-0.22336626052856445,
-0.0807865783572197,
0.0032354784198105335,
-0.2755500376224518,
0.5612590909004211,
-0.0015003372682258487,
-0.483015239238739,
-0.7057119011878967,
-0.4548150897026062,
0.229652643203... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
arieg/bw_spec_cls_160 | arieg | 2023-11-10T12:25:59Z | 0 | 0 | null | [
"region:us"
] | 2023-11-10T12:25:59Z | 2023-11-09T14:19:12.000Z | 2023-11-09T14:19:12 | ---
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': '56248'
'1': '56249'
'2': '56273'
'3': '56274'
'4': '56275'
'5': '56465'
'6': '56466'
'7': '56467'
'8': '56468'
'9': '56469'
'10': '56470'
'11': '56471'
'12': '56472'
'13': '56474'
'14': '56493'
'15': '56495'
'16': '56496'
'17': '56497'
'18': '56498'
'19': '56499'
'20': '56516'
'21': '56517'
'22': '56518'
'23': '56519'
'24': '56520'
'25': '56521'
'26': '56639'
'27': '56640'
'28': '56641'
'29': '56645'
'30': '56646'
'31': '56648'
'32': '56649'
'33': '56650'
'34': '56651'
'35': '56686'
'36': '56687'
'37': '56688'
'38': '56689'
'39': '56690'
'40': '56691'
'41': '56692'
'42': '56693'
'43': '56694'
'44': '56695'
'45': '56696'
'46': '56795'
'47': '56796'
'48': '56797'
'49': '56798'
'50': '56799'
'51': '56800'
'52': '56801'
'53': '56802'
'54': '56803'
'55': '56804'
'56': '56805'
'57': '56888'
'58': '57164'
'59': '57175'
'60': '57176'
'61': '57177'
'62': '57178'
'63': '57179'
'64': '57180'
'65': '57344'
'66': '57360'
'67': '57371'
'68': '57417'
'69': '57418'
'70': '57440'
'71': '57442'
'72': '57500'
'73': '57569'
'74': '57626'
'75': '57627'
'76': '57628'
'77': '57629'
'78': '57630'
'79': '57639'
'80': '57640'
'81': '57648'
'82': '57658'
'83': '57661'
'84': '57662'
'85': '57663'
'86': '57665'
'87': '57691'
'88': '57697'
'89': '57819'
'90': '57820'
'91': '57821'
'92': '57822'
'93': '57823'
'94': '57936'
'95': '57937'
'96': '57938'
'97': '57939'
'98': '57943'
'99': '57968'
'100': '58052'
'101': '58053'
'102': '58054'
'103': '58060'
'104': '58061'
'105': '58063'
'106': '58068'
'107': '58070'
'108': '58115'
'109': '58116'
'110': '58117'
'111': '58135'
'112': '58140'
'113': '58161'
'114': '58162'
'115': '58164'
'116': '58166'
'117': '58169'
'118': '58170'
'119': '58173'
'120': '58174'
'121': '58212'
'122': '58213'
'123': '58215'
'124': '58221'
'125': '58225'
'126': '58341'
'127': '58474'
'128': '59078'
'129': '59373'
'130': '59374'
'131': '59561'
'132': '59653'
'133': '59654'
'134': '59656'
'135': '59657'
'136': '59658'
'137': '59659'
'138': '59660'
'139': '59663'
'140': '59664'
'141': '59666'
'142': '59667'
'143': '59669'
'144': '59671'
'145': '59673'
'146': '59675'
'147': '59676'
'148': '59677'
'149': '59678'
'150': '59679'
'151': '59680'
'152': '59681'
'153': '59682'
'154': '59683'
'155': '59684'
'156': '59685'
'157': '59686'
'158': '59687'
'159': '59688'
splits:
- name: train
num_bytes: 179214128.0
num_examples: 3200
download_size: 179008943
dataset_size: 179214128.0
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "bw_spec_cls_160"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.7950762510299683,
-0.09063873440027237,
0.182620570063591,
0.34082117676734924,
-0.26321759819984436,
-0.17038708925247192,
-0.08298292011022568,
-0.2777201533317566,
0.6963418126106262,
0.42784714698791504,
-0.8451219797134399,
-0.7450318932533264,
-0.5596488118171692,
-0.2055472582578... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
lawallanre/geo_nlp_tweets | lawallanre | 2023-11-10T04:09:12Z | 0 | 1 | null | [
"task_categories:text-classification",
"size_categories:1K<n<10K",
"language:en",
"license:ofl-1.1",
"region:us"
] | 2023-11-10T04:09:12Z | 2023-11-09T14:28:48.000Z | 2023-11-09T14:28:48 | ---
configs:
- config_name: default
data_files:
- split: train
path: "data/train,tsv"
- split: test
path: "data.tsv"
- split: dev
path: "data/dev.tsv"
license: ofl-1.1
task_categories:
- text-classification
language:
- en
pretty_name: lanre
size_categories:
- 1K<n<10K
--- | [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
thealy168/solaratest | thealy168 | 2023-11-09T14:30:04Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T14:30:04Z | 2023-11-09T14:29:34.000Z | 2023-11-09T14:29:34 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
arieg/bw_spec_cls_80_26 | arieg | 2023-11-09T14:32:02Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T14:32:02Z | 2023-11-09T14:31:46.000Z | 2023-11-09T14:31:46 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': '62001'
'1': '62003'
'2': '62005'
'3': '62007'
'4': '62163'
'5': '62164'
'6': '62165'
'7': '62180'
'8': '62183'
'9': '62185'
'10': '62186'
'11': '62187'
'12': '62188'
'13': '62189'
'14': '62190'
'15': '62191'
'16': '62192'
'17': '62193'
'18': '62194'
'19': '62195'
'20': '62196'
'21': '62337'
'22': '62426'
'23': '62436'
'24': '62445'
'25': '62446'
'26': '62448'
'27': '62449'
'28': '62450'
'29': '62452'
'30': '62458'
'31': '62525'
'32': '62526'
'33': '62527'
'34': '62528'
'35': '62529'
'36': '62531'
'37': '62532'
'38': '62533'
'39': '62534'
'40': '62586'
'41': '62589'
'42': '62591'
'43': '62592'
'44': '62594'
'45': '62595'
'46': '62596'
'47': '62655'
'48': '62671'
'49': '62742'
'50': '62748'
'51': '62749'
'52': '62750'
'53': '62751'
'54': '62753'
'55': '63043'
'56': '63044'
'57': '63045'
'58': '63117'
'59': '63191'
'60': '63208'
'61': '63224'
'62': '63226'
'63': '63287'
'64': '63289'
'65': '63290'
'66': '63291'
'67': '63292'
'68': '63470'
'69': '63471'
'70': '63472'
'71': '63626'
'72': '63655'
'73': '63733'
'74': '63747'
'75': '63755'
'76': '63757'
'77': '63770'
'78': '63789'
'79': '63803'
splits:
- name: train
num_bytes: 89137873.6
num_examples: 1600
- name: test
num_bytes: 22127983.0
num_examples: 400
download_size: 110364015
dataset_size: 111265856.6
---
# Dataset Card for "bw_spec_cls_80_26"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.7584282159805298,
-0.18570773303508759,
0.16150814294815063,
0.40829697251319885,
-0.29163771867752075,
-0.15308676660060883,
0.016579505056142807,
-0.27851009368896484,
0.6211332678794861,
0.5020564794540405,
-0.7990765571594238,
-0.785754382610321,
-0.6016364097595215,
-0.206377789378... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
open-llm-leaderboard/details_CobraMamba__mamba-gpt-7b-v1_public | open-llm-leaderboard | 2023-11-09T14:38:25Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T14:38:25Z | 2023-11-09T14:37:23.000Z | 2023-11-09T14:37:23 | ---
pretty_name: Evaluation run of CobraMamba/mamba-gpt-7b-v1
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [CobraMamba/mamba-gpt-7b-v1](https://huggingface.co/CobraMamba/mamba-gpt-7b-v1)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 64 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the aggregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_CobraMamba__mamba-gpt-7b-v1_public\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-11-09T14:34:23.926109](https://huggingface.co/datasets/open-llm-leaderboard/details_CobraMamba__mamba-gpt-7b-v1_public/blob/main/results_2023-11-09T14-34-23.926109.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"acc\": 0.6286909633628079,\n\
\ \"acc_stderr\": 0.03215522070353069,\n \"acc_norm\": 0.6377478775248846,\n\
\ \"acc_norm_stderr\": 0.032851877291432414,\n \"mc1\": 0.3084455324357405,\n\
\ \"mc1_stderr\": 0.01616803938315687,\n \"mc2\": 0.4634199786351567,\n\
\ \"mc2_stderr\": 0.014481061527331505,\n \"em\": 0.2679320469798658,\n\
\ \"em_stderr\": 0.004535526201164825,\n \"f1\": 0.31668204697986585,\n\
\ \"f1_stderr\": 0.004459593071277455\n },\n \"harness|arc:challenge|25\"\
: {\n \"acc\": 0.575938566552901,\n \"acc_stderr\": 0.014441889627464396,\n\
\ \"acc_norm\": 0.6126279863481229,\n \"acc_norm_stderr\": 0.01423587248790987\n\
\ },\n \"harness|hellaswag|10\": {\n \"acc\": 0.6354311890061741,\n\
\ \"acc_stderr\": 0.004803253812881043,\n \"acc_norm\": 0.8409679346743677,\n\
\ \"acc_norm_stderr\": 0.003649585852821192\n },\n \"harness|hendrycksTest-abstract_algebra|5\"\
: {\n \"acc\": 0.26,\n \"acc_stderr\": 0.0440844002276808,\n \
\ \"acc_norm\": 0.26,\n \"acc_norm_stderr\": 0.0440844002276808\n },\n\
\ \"harness|hendrycksTest-anatomy|5\": {\n \"acc\": 0.6370370370370371,\n\
\ \"acc_stderr\": 0.04153948404742398,\n \"acc_norm\": 0.6370370370370371,\n\
\ \"acc_norm_stderr\": 0.04153948404742398\n },\n \"harness|hendrycksTest-astronomy|5\"\
: {\n \"acc\": 0.631578947368421,\n \"acc_stderr\": 0.03925523381052932,\n\
\ \"acc_norm\": 0.631578947368421,\n \"acc_norm_stderr\": 0.03925523381052932\n\
\ },\n \"harness|hendrycksTest-business_ethics|5\": {\n \"acc\": 0.61,\n\
\ \"acc_stderr\": 0.04902071300001975,\n \"acc_norm\": 0.61,\n \
\ \"acc_norm_stderr\": 0.04902071300001975\n },\n \"harness|hendrycksTest-clinical_knowledge|5\"\
: {\n \"acc\": 0.690566037735849,\n \"acc_stderr\": 0.028450154794118637,\n\
\ \"acc_norm\": 0.690566037735849,\n \"acc_norm_stderr\": 0.028450154794118637\n\
\ },\n \"harness|hendrycksTest-college_biology|5\": {\n \"acc\": 0.75,\n\
\ \"acc_stderr\": 0.03621034121889507,\n \"acc_norm\": 0.75,\n \
\ \"acc_norm_stderr\": 0.03621034121889507\n },\n \"harness|hendrycksTest-college_chemistry|5\"\
: {\n \"acc\": 0.48,\n \"acc_stderr\": 0.050211673156867795,\n \
\ \"acc_norm\": 0.48,\n \"acc_norm_stderr\": 0.050211673156867795\n \
\ },\n \"harness|hendrycksTest-college_computer_science|5\": {\n \"\
acc\": 0.57,\n \"acc_stderr\": 0.04975698519562428,\n \"acc_norm\"\
: 0.57,\n \"acc_norm_stderr\": 0.04975698519562428\n },\n \"harness|hendrycksTest-college_mathematics|5\"\
: {\n \"acc\": 0.35,\n \"acc_stderr\": 0.047937248544110196,\n \
\ \"acc_norm\": 0.35,\n \"acc_norm_stderr\": 0.047937248544110196\n \
\ },\n \"harness|hendrycksTest-college_medicine|5\": {\n \"acc\": 0.630057803468208,\n\
\ \"acc_stderr\": 0.0368122963339432,\n \"acc_norm\": 0.630057803468208,\n\
\ \"acc_norm_stderr\": 0.0368122963339432\n },\n \"harness|hendrycksTest-college_physics|5\"\
: {\n \"acc\": 0.45098039215686275,\n \"acc_stderr\": 0.04951218252396264,\n\
\ \"acc_norm\": 0.45098039215686275,\n \"acc_norm_stderr\": 0.04951218252396264\n\
\ },\n \"harness|hendrycksTest-computer_security|5\": {\n \"acc\":\
\ 0.81,\n \"acc_stderr\": 0.03942772444036624,\n \"acc_norm\": 0.81,\n\
\ \"acc_norm_stderr\": 0.03942772444036624\n },\n \"harness|hendrycksTest-conceptual_physics|5\"\
: {\n \"acc\": 0.5404255319148936,\n \"acc_stderr\": 0.03257901482099835,\n\
\ \"acc_norm\": 0.5404255319148936,\n \"acc_norm_stderr\": 0.03257901482099835\n\
\ },\n \"harness|hendrycksTest-econometrics|5\": {\n \"acc\": 0.5,\n\
\ \"acc_stderr\": 0.047036043419179864,\n \"acc_norm\": 0.5,\n \
\ \"acc_norm_stderr\": 0.047036043419179864\n },\n \"harness|hendrycksTest-electrical_engineering|5\"\
: {\n \"acc\": 0.5793103448275863,\n \"acc_stderr\": 0.0411391498118926,\n\
\ \"acc_norm\": 0.5793103448275863,\n \"acc_norm_stderr\": 0.0411391498118926\n\
\ },\n \"harness|hendrycksTest-elementary_mathematics|5\": {\n \"acc\"\
: 0.3994708994708995,\n \"acc_stderr\": 0.02522545028406788,\n \"\
acc_norm\": 0.3994708994708995,\n \"acc_norm_stderr\": 0.02522545028406788\n\
\ },\n \"harness|hendrycksTest-formal_logic|5\": {\n \"acc\": 0.38095238095238093,\n\
\ \"acc_stderr\": 0.043435254289490965,\n \"acc_norm\": 0.38095238095238093,\n\
\ \"acc_norm_stderr\": 0.043435254289490965\n },\n \"harness|hendrycksTest-global_facts|5\"\
: {\n \"acc\": 0.35,\n \"acc_stderr\": 0.047937248544110196,\n \
\ \"acc_norm\": 0.35,\n \"acc_norm_stderr\": 0.047937248544110196\n \
\ },\n \"harness|hendrycksTest-high_school_biology|5\": {\n \"acc\"\
: 0.7612903225806451,\n \"acc_stderr\": 0.02425107126220884,\n \"\
acc_norm\": 0.7612903225806451,\n \"acc_norm_stderr\": 0.02425107126220884\n\
\ },\n \"harness|hendrycksTest-high_school_chemistry|5\": {\n \"acc\"\
: 0.47783251231527096,\n \"acc_stderr\": 0.03514528562175007,\n \"\
acc_norm\": 0.47783251231527096,\n \"acc_norm_stderr\": 0.03514528562175007\n\
\ },\n \"harness|hendrycksTest-high_school_computer_science|5\": {\n \
\ \"acc\": 0.71,\n \"acc_stderr\": 0.045604802157206845,\n \"acc_norm\"\
: 0.71,\n \"acc_norm_stderr\": 0.045604802157206845\n },\n \"harness|hendrycksTest-high_school_european_history|5\"\
: {\n \"acc\": 0.7454545454545455,\n \"acc_stderr\": 0.03401506715249039,\n\
\ \"acc_norm\": 0.7454545454545455,\n \"acc_norm_stderr\": 0.03401506715249039\n\
\ },\n \"harness|hendrycksTest-high_school_geography|5\": {\n \"acc\"\
: 0.7878787878787878,\n \"acc_stderr\": 0.029126522834586808,\n \"\
acc_norm\": 0.7878787878787878,\n \"acc_norm_stderr\": 0.029126522834586808\n\
\ },\n \"harness|hendrycksTest-high_school_government_and_politics|5\": {\n\
\ \"acc\": 0.8756476683937824,\n \"acc_stderr\": 0.02381447708659355,\n\
\ \"acc_norm\": 0.8756476683937824,\n \"acc_norm_stderr\": 0.02381447708659355\n\
\ },\n \"harness|hendrycksTest-high_school_macroeconomics|5\": {\n \
\ \"acc\": 0.6461538461538462,\n \"acc_stderr\": 0.02424378399406216,\n \
\ \"acc_norm\": 0.6461538461538462,\n \"acc_norm_stderr\": 0.02424378399406216\n\
\ },\n \"harness|hendrycksTest-high_school_mathematics|5\": {\n \"\
acc\": 0.3592592592592593,\n \"acc_stderr\": 0.02925290592725198,\n \
\ \"acc_norm\": 0.3592592592592593,\n \"acc_norm_stderr\": 0.02925290592725198\n\
\ },\n \"harness|hendrycksTest-high_school_microeconomics|5\": {\n \
\ \"acc\": 0.6470588235294118,\n \"acc_stderr\": 0.03104194130405929,\n \
\ \"acc_norm\": 0.6470588235294118,\n \"acc_norm_stderr\": 0.03104194130405929\n\
\ },\n \"harness|hendrycksTest-high_school_physics|5\": {\n \"acc\"\
: 0.2847682119205298,\n \"acc_stderr\": 0.03684881521389023,\n \"\
acc_norm\": 0.2847682119205298,\n \"acc_norm_stderr\": 0.03684881521389023\n\
\ },\n \"harness|hendrycksTest-high_school_psychology|5\": {\n \"acc\"\
: 0.8238532110091743,\n \"acc_stderr\": 0.016332882393431385,\n \"\
acc_norm\": 0.8238532110091743,\n \"acc_norm_stderr\": 0.016332882393431385\n\
\ },\n \"harness|hendrycksTest-high_school_statistics|5\": {\n \"acc\"\
: 0.49537037037037035,\n \"acc_stderr\": 0.03409825519163572,\n \"\
acc_norm\": 0.49537037037037035,\n \"acc_norm_stderr\": 0.03409825519163572\n\
\ },\n \"harness|hendrycksTest-high_school_us_history|5\": {\n \"acc\"\
: 0.7843137254901961,\n \"acc_stderr\": 0.028867431449849316,\n \"\
acc_norm\": 0.7843137254901961,\n \"acc_norm_stderr\": 0.028867431449849316\n\
\ },\n \"harness|hendrycksTest-high_school_world_history|5\": {\n \"\
acc\": 0.7932489451476793,\n \"acc_stderr\": 0.0263616516683891,\n \
\ \"acc_norm\": 0.7932489451476793,\n \"acc_norm_stderr\": 0.0263616516683891\n\
\ },\n \"harness|hendrycksTest-human_aging|5\": {\n \"acc\": 0.6771300448430493,\n\
\ \"acc_stderr\": 0.031381476375754995,\n \"acc_norm\": 0.6771300448430493,\n\
\ \"acc_norm_stderr\": 0.031381476375754995\n },\n \"harness|hendrycksTest-human_sexuality|5\"\
: {\n \"acc\": 0.7938931297709924,\n \"acc_stderr\": 0.03547771004159464,\n\
\ \"acc_norm\": 0.7938931297709924,\n \"acc_norm_stderr\": 0.03547771004159464\n\
\ },\n \"harness|hendrycksTest-international_law|5\": {\n \"acc\":\
\ 0.8264462809917356,\n \"acc_stderr\": 0.0345727283691767,\n \"acc_norm\"\
: 0.8264462809917356,\n \"acc_norm_stderr\": 0.0345727283691767\n },\n\
\ \"harness|hendrycksTest-jurisprudence|5\": {\n \"acc\": 0.75,\n \
\ \"acc_stderr\": 0.04186091791394607,\n \"acc_norm\": 0.75,\n \
\ \"acc_norm_stderr\": 0.04186091791394607\n },\n \"harness|hendrycksTest-logical_fallacies|5\"\
: {\n \"acc\": 0.7975460122699386,\n \"acc_stderr\": 0.03157065078911901,\n\
\ \"acc_norm\": 0.7975460122699386,\n \"acc_norm_stderr\": 0.03157065078911901\n\
\ },\n \"harness|hendrycksTest-machine_learning|5\": {\n \"acc\": 0.49107142857142855,\n\
\ \"acc_stderr\": 0.04745033255489123,\n \"acc_norm\": 0.49107142857142855,\n\
\ \"acc_norm_stderr\": 0.04745033255489123\n },\n \"harness|hendrycksTest-management|5\"\
: {\n \"acc\": 0.7864077669902912,\n \"acc_stderr\": 0.04058042015646034,\n\
\ \"acc_norm\": 0.7864077669902912,\n \"acc_norm_stderr\": 0.04058042015646034\n\
\ },\n \"harness|hendrycksTest-marketing|5\": {\n \"acc\": 0.8760683760683761,\n\
\ \"acc_stderr\": 0.02158649400128137,\n \"acc_norm\": 0.8760683760683761,\n\
\ \"acc_norm_stderr\": 0.02158649400128137\n },\n \"harness|hendrycksTest-medical_genetics|5\"\
: {\n \"acc\": 0.74,\n \"acc_stderr\": 0.04408440022768078,\n \
\ \"acc_norm\": 0.74,\n \"acc_norm_stderr\": 0.04408440022768078\n \
\ },\n \"harness|hendrycksTest-miscellaneous|5\": {\n \"acc\": 0.8148148148148148,\n\
\ \"acc_stderr\": 0.013890862162876163,\n \"acc_norm\": 0.8148148148148148,\n\
\ \"acc_norm_stderr\": 0.013890862162876163\n },\n \"harness|hendrycksTest-moral_disputes|5\"\
: {\n \"acc\": 0.708092485549133,\n \"acc_stderr\": 0.02447699407624734,\n\
\ \"acc_norm\": 0.708092485549133,\n \"acc_norm_stderr\": 0.02447699407624734\n\
\ },\n \"harness|hendrycksTest-moral_scenarios|5\": {\n \"acc\": 0.24692737430167597,\n\
\ \"acc_stderr\": 0.01442229220480884,\n \"acc_norm\": 0.24692737430167597,\n\
\ \"acc_norm_stderr\": 0.01442229220480884\n },\n \"harness|hendrycksTest-nutrition|5\"\
: {\n \"acc\": 0.7516339869281046,\n \"acc_stderr\": 0.02473998135511359,\n\
\ \"acc_norm\": 0.7516339869281046,\n \"acc_norm_stderr\": 0.02473998135511359\n\
\ },\n \"harness|hendrycksTest-philosophy|5\": {\n \"acc\": 0.6913183279742765,\n\
\ \"acc_stderr\": 0.026236965881153262,\n \"acc_norm\": 0.6913183279742765,\n\
\ \"acc_norm_stderr\": 0.026236965881153262\n },\n \"harness|hendrycksTest-prehistory|5\"\
: {\n \"acc\": 0.7129629629629629,\n \"acc_stderr\": 0.025171041915309684,\n\
\ \"acc_norm\": 0.7129629629629629,\n \"acc_norm_stderr\": 0.025171041915309684\n\
\ },\n \"harness|hendrycksTest-professional_accounting|5\": {\n \"\
acc\": 0.4858156028368794,\n \"acc_stderr\": 0.02981549448368206,\n \
\ \"acc_norm\": 0.4858156028368794,\n \"acc_norm_stderr\": 0.02981549448368206\n\
\ },\n \"harness|hendrycksTest-professional_law|5\": {\n \"acc\": 0.44784876140808344,\n\
\ \"acc_stderr\": 0.012700582404768221,\n \"acc_norm\": 0.44784876140808344,\n\
\ \"acc_norm_stderr\": 0.012700582404768221\n },\n \"harness|hendrycksTest-professional_medicine|5\"\
: {\n \"acc\": 0.6691176470588235,\n \"acc_stderr\": 0.028582709753898445,\n\
\ \"acc_norm\": 0.6691176470588235,\n \"acc_norm_stderr\": 0.028582709753898445\n\
\ },\n \"harness|hendrycksTest-professional_psychology|5\": {\n \"\
acc\": 0.6584967320261438,\n \"acc_stderr\": 0.019184639328092487,\n \
\ \"acc_norm\": 0.6584967320261438,\n \"acc_norm_stderr\": 0.019184639328092487\n\
\ },\n \"harness|hendrycksTest-public_relations|5\": {\n \"acc\": 0.6636363636363637,\n\
\ \"acc_stderr\": 0.04525393596302506,\n \"acc_norm\": 0.6636363636363637,\n\
\ \"acc_norm_stderr\": 0.04525393596302506\n },\n \"harness|hendrycksTest-security_studies|5\"\
: {\n \"acc\": 0.7183673469387755,\n \"acc_stderr\": 0.02879518557429129,\n\
\ \"acc_norm\": 0.7183673469387755,\n \"acc_norm_stderr\": 0.02879518557429129\n\
\ },\n \"harness|hendrycksTest-sociology|5\": {\n \"acc\": 0.8308457711442786,\n\
\ \"acc_stderr\": 0.026508590656233268,\n \"acc_norm\": 0.8308457711442786,\n\
\ \"acc_norm_stderr\": 0.026508590656233268\n },\n \"harness|hendrycksTest-us_foreign_policy|5\"\
: {\n \"acc\": 0.85,\n \"acc_stderr\": 0.0358870281282637,\n \
\ \"acc_norm\": 0.85,\n \"acc_norm_stderr\": 0.0358870281282637\n },\n\
\ \"harness|hendrycksTest-virology|5\": {\n \"acc\": 0.5421686746987951,\n\
\ \"acc_stderr\": 0.0387862677100236,\n \"acc_norm\": 0.5421686746987951,\n\
\ \"acc_norm_stderr\": 0.0387862677100236\n },\n \"harness|hendrycksTest-world_religions|5\"\
: {\n \"acc\": 0.8304093567251462,\n \"acc_stderr\": 0.02878210810540171,\n\
\ \"acc_norm\": 0.8304093567251462,\n \"acc_norm_stderr\": 0.02878210810540171\n\
\ },\n \"harness|truthfulqa:mc|0\": {\n \"mc1\": 0.3084455324357405,\n\
\ \"mc1_stderr\": 0.01616803938315687,\n \"mc2\": 0.4634199786351567,\n\
\ \"mc2_stderr\": 0.014481061527331505\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.7916337805840569,\n \"acc_stderr\": 0.01141455439998773\n\
\ },\n \"harness|drop|3\": {\n \"em\": 0.2679320469798658,\n \
\ \"em_stderr\": 0.004535526201164825,\n \"f1\": 0.31668204697986585,\n\
\ \"f1_stderr\": 0.004459593071277455\n },\n \"harness|gsm8k|5\": {\n\
\ \"acc\": 0.17361637604245642,\n \"acc_stderr\": 0.01043346322125763\n\
\ }\n}\n```"
repo_url: https://huggingface.co/CobraMamba/mamba-gpt-7b-v1
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|arc:challenge|25_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_drop_3
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|drop|3_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|drop|3_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|gsm8k|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hellaswag|10_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-34-23.926109.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T14-34-23.926109.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- '**/details_harness|winogrande|5_2023-11-09T14-34-23.926109.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-11-09T14-34-23.926109.parquet'
- config_name: results
data_files:
- split: 2023_11_09T14_34_23.926109
path:
- results_2023-11-09T14-34-23.926109.parquet
- split: latest
path:
- results_2023-11-09T14-34-23.926109.parquet
---
# Dataset Card for Evaluation run of CobraMamba/mamba-gpt-7b-v1
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/CobraMamba/mamba-gpt-7b-v1
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [CobraMamba/mamba-gpt-7b-v1](https://huggingface.co/CobraMamba/mamba-gpt-7b-v1) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 64 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_CobraMamba__mamba-gpt-7b-v1_public",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-11-09T14:34:23.926109](https://huggingface.co/datasets/open-llm-leaderboard/details_CobraMamba__mamba-gpt-7b-v1_public/blob/main/results_2023-11-09T14-34-23.926109.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"acc": 0.6286909633628079,
"acc_stderr": 0.03215522070353069,
"acc_norm": 0.6377478775248846,
"acc_norm_stderr": 0.032851877291432414,
"mc1": 0.3084455324357405,
"mc1_stderr": 0.01616803938315687,
"mc2": 0.4634199786351567,
"mc2_stderr": 0.014481061527331505,
"em": 0.2679320469798658,
"em_stderr": 0.004535526201164825,
"f1": 0.31668204697986585,
"f1_stderr": 0.004459593071277455
},
"harness|arc:challenge|25": {
"acc": 0.575938566552901,
"acc_stderr": 0.014441889627464396,
"acc_norm": 0.6126279863481229,
"acc_norm_stderr": 0.01423587248790987
},
"harness|hellaswag|10": {
"acc": 0.6354311890061741,
"acc_stderr": 0.004803253812881043,
"acc_norm": 0.8409679346743677,
"acc_norm_stderr": 0.003649585852821192
},
"harness|hendrycksTest-abstract_algebra|5": {
"acc": 0.26,
"acc_stderr": 0.0440844002276808,
"acc_norm": 0.26,
"acc_norm_stderr": 0.0440844002276808
},
"harness|hendrycksTest-anatomy|5": {
"acc": 0.6370370370370371,
"acc_stderr": 0.04153948404742398,
"acc_norm": 0.6370370370370371,
"acc_norm_stderr": 0.04153948404742398
},
"harness|hendrycksTest-astronomy|5": {
"acc": 0.631578947368421,
"acc_stderr": 0.03925523381052932,
"acc_norm": 0.631578947368421,
"acc_norm_stderr": 0.03925523381052932
},
"harness|hendrycksTest-business_ethics|5": {
"acc": 0.61,
"acc_stderr": 0.04902071300001975,
"acc_norm": 0.61,
"acc_norm_stderr": 0.04902071300001975
},
"harness|hendrycksTest-clinical_knowledge|5": {
"acc": 0.690566037735849,
"acc_stderr": 0.028450154794118637,
"acc_norm": 0.690566037735849,
"acc_norm_stderr": 0.028450154794118637
},
"harness|hendrycksTest-college_biology|5": {
"acc": 0.75,
"acc_stderr": 0.03621034121889507,
"acc_norm": 0.75,
"acc_norm_stderr": 0.03621034121889507
},
"harness|hendrycksTest-college_chemistry|5": {
"acc": 0.48,
"acc_stderr": 0.050211673156867795,
"acc_norm": 0.48,
"acc_norm_stderr": 0.050211673156867795
},
"harness|hendrycksTest-college_computer_science|5": {
"acc": 0.57,
"acc_stderr": 0.04975698519562428,
"acc_norm": 0.57,
"acc_norm_stderr": 0.04975698519562428
},
"harness|hendrycksTest-college_mathematics|5": {
"acc": 0.35,
"acc_stderr": 0.047937248544110196,
"acc_norm": 0.35,
"acc_norm_stderr": 0.047937248544110196
},
"harness|hendrycksTest-college_medicine|5": {
"acc": 0.630057803468208,
"acc_stderr": 0.0368122963339432,
"acc_norm": 0.630057803468208,
"acc_norm_stderr": 0.0368122963339432
},
"harness|hendrycksTest-college_physics|5": {
"acc": 0.45098039215686275,
"acc_stderr": 0.04951218252396264,
"acc_norm": 0.45098039215686275,
"acc_norm_stderr": 0.04951218252396264
},
"harness|hendrycksTest-computer_security|5": {
"acc": 0.81,
"acc_stderr": 0.03942772444036624,
"acc_norm": 0.81,
"acc_norm_stderr": 0.03942772444036624
},
"harness|hendrycksTest-conceptual_physics|5": {
"acc": 0.5404255319148936,
"acc_stderr": 0.03257901482099835,
"acc_norm": 0.5404255319148936,
"acc_norm_stderr": 0.03257901482099835
},
"harness|hendrycksTest-econometrics|5": {
"acc": 0.5,
"acc_stderr": 0.047036043419179864,
"acc_norm": 0.5,
"acc_norm_stderr": 0.047036043419179864
},
"harness|hendrycksTest-electrical_engineering|5": {
"acc": 0.5793103448275863,
"acc_stderr": 0.0411391498118926,
"acc_norm": 0.5793103448275863,
"acc_norm_stderr": 0.0411391498118926
},
"harness|hendrycksTest-elementary_mathematics|5": {
"acc": 0.3994708994708995,
"acc_stderr": 0.02522545028406788,
"acc_norm": 0.3994708994708995,
"acc_norm_stderr": 0.02522545028406788
},
"harness|hendrycksTest-formal_logic|5": {
"acc": 0.38095238095238093,
"acc_stderr": 0.043435254289490965,
"acc_norm": 0.38095238095238093,
"acc_norm_stderr": 0.043435254289490965
},
"harness|hendrycksTest-global_facts|5": {
"acc": 0.35,
"acc_stderr": 0.047937248544110196,
"acc_norm": 0.35,
"acc_norm_stderr": 0.047937248544110196
},
"harness|hendrycksTest-high_school_biology|5": {
"acc": 0.7612903225806451,
"acc_stderr": 0.02425107126220884,
"acc_norm": 0.7612903225806451,
"acc_norm_stderr": 0.02425107126220884
},
"harness|hendrycksTest-high_school_chemistry|5": {
"acc": 0.47783251231527096,
"acc_stderr": 0.03514528562175007,
"acc_norm": 0.47783251231527096,
"acc_norm_stderr": 0.03514528562175007
},
"harness|hendrycksTest-high_school_computer_science|5": {
"acc": 0.71,
"acc_stderr": 0.045604802157206845,
"acc_norm": 0.71,
"acc_norm_stderr": 0.045604802157206845
},
"harness|hendrycksTest-high_school_european_history|5": {
"acc": 0.7454545454545455,
"acc_stderr": 0.03401506715249039,
"acc_norm": 0.7454545454545455,
"acc_norm_stderr": 0.03401506715249039
},
"harness|hendrycksTest-high_school_geography|5": {
"acc": 0.7878787878787878,
"acc_stderr": 0.029126522834586808,
"acc_norm": 0.7878787878787878,
"acc_norm_stderr": 0.029126522834586808
},
"harness|hendrycksTest-high_school_government_and_politics|5": {
"acc": 0.8756476683937824,
"acc_stderr": 0.02381447708659355,
"acc_norm": 0.8756476683937824,
"acc_norm_stderr": 0.02381447708659355
},
"harness|hendrycksTest-high_school_macroeconomics|5": {
"acc": 0.6461538461538462,
"acc_stderr": 0.02424378399406216,
"acc_norm": 0.6461538461538462,
"acc_norm_stderr": 0.02424378399406216
},
"harness|hendrycksTest-high_school_mathematics|5": {
"acc": 0.3592592592592593,
"acc_stderr": 0.02925290592725198,
"acc_norm": 0.3592592592592593,
"acc_norm_stderr": 0.02925290592725198
},
"harness|hendrycksTest-high_school_microeconomics|5": {
"acc": 0.6470588235294118,
"acc_stderr": 0.03104194130405929,
"acc_norm": 0.6470588235294118,
"acc_norm_stderr": 0.03104194130405929
},
"harness|hendrycksTest-high_school_physics|5": {
"acc": 0.2847682119205298,
"acc_stderr": 0.03684881521389023,
"acc_norm": 0.2847682119205298,
"acc_norm_stderr": 0.03684881521389023
},
"harness|hendrycksTest-high_school_psychology|5": {
"acc": 0.8238532110091743,
"acc_stderr": 0.016332882393431385,
"acc_norm": 0.8238532110091743,
"acc_norm_stderr": 0.016332882393431385
},
"harness|hendrycksTest-high_school_statistics|5": {
"acc": 0.49537037037037035,
"acc_stderr": 0.03409825519163572,
"acc_norm": 0.49537037037037035,
"acc_norm_stderr": 0.03409825519163572
},
"harness|hendrycksTest-high_school_us_history|5": {
"acc": 0.7843137254901961,
"acc_stderr": 0.028867431449849316,
"acc_norm": 0.7843137254901961,
"acc_norm_stderr": 0.028867431449849316
},
"harness|hendrycksTest-high_school_world_history|5": {
"acc": 0.7932489451476793,
"acc_stderr": 0.0263616516683891,
"acc_norm": 0.7932489451476793,
"acc_norm_stderr": 0.0263616516683891
},
"harness|hendrycksTest-human_aging|5": {
"acc": 0.6771300448430493,
"acc_stderr": 0.031381476375754995,
"acc_norm": 0.6771300448430493,
"acc_norm_stderr": 0.031381476375754995
},
"harness|hendrycksTest-human_sexuality|5": {
"acc": 0.7938931297709924,
"acc_stderr": 0.03547771004159464,
"acc_norm": 0.7938931297709924,
"acc_norm_stderr": 0.03547771004159464
},
"harness|hendrycksTest-international_law|5": {
"acc": 0.8264462809917356,
"acc_stderr": 0.0345727283691767,
"acc_norm": 0.8264462809917356,
"acc_norm_stderr": 0.0345727283691767
},
"harness|hendrycksTest-jurisprudence|5": {
"acc": 0.75,
"acc_stderr": 0.04186091791394607,
"acc_norm": 0.75,
"acc_norm_stderr": 0.04186091791394607
},
"harness|hendrycksTest-logical_fallacies|5": {
"acc": 0.7975460122699386,
"acc_stderr": 0.03157065078911901,
"acc_norm": 0.7975460122699386,
"acc_norm_stderr": 0.03157065078911901
},
"harness|hendrycksTest-machine_learning|5": {
"acc": 0.49107142857142855,
"acc_stderr": 0.04745033255489123,
"acc_norm": 0.49107142857142855,
"acc_norm_stderr": 0.04745033255489123
},
"harness|hendrycksTest-management|5": {
"acc": 0.7864077669902912,
"acc_stderr": 0.04058042015646034,
"acc_norm": 0.7864077669902912,
"acc_norm_stderr": 0.04058042015646034
},
"harness|hendrycksTest-marketing|5": {
"acc": 0.8760683760683761,
"acc_stderr": 0.02158649400128137,
"acc_norm": 0.8760683760683761,
"acc_norm_stderr": 0.02158649400128137
},
"harness|hendrycksTest-medical_genetics|5": {
"acc": 0.74,
"acc_stderr": 0.04408440022768078,
"acc_norm": 0.74,
"acc_norm_stderr": 0.04408440022768078
},
"harness|hendrycksTest-miscellaneous|5": {
"acc": 0.8148148148148148,
"acc_stderr": 0.013890862162876163,
"acc_norm": 0.8148148148148148,
"acc_norm_stderr": 0.013890862162876163
},
"harness|hendrycksTest-moral_disputes|5": {
"acc": 0.708092485549133,
"acc_stderr": 0.02447699407624734,
"acc_norm": 0.708092485549133,
"acc_norm_stderr": 0.02447699407624734
},
"harness|hendrycksTest-moral_scenarios|5": {
"acc": 0.24692737430167597,
"acc_stderr": 0.01442229220480884,
"acc_norm": 0.24692737430167597,
"acc_norm_stderr": 0.01442229220480884
},
"harness|hendrycksTest-nutrition|5": {
"acc": 0.7516339869281046,
"acc_stderr": 0.02473998135511359,
"acc_norm": 0.7516339869281046,
"acc_norm_stderr": 0.02473998135511359
},
"harness|hendrycksTest-philosophy|5": {
"acc": 0.6913183279742765,
"acc_stderr": 0.026236965881153262,
"acc_norm": 0.6913183279742765,
"acc_norm_stderr": 0.026236965881153262
},
"harness|hendrycksTest-prehistory|5": {
"acc": 0.7129629629629629,
"acc_stderr": 0.025171041915309684,
"acc_norm": 0.7129629629629629,
"acc_norm_stderr": 0.025171041915309684
},
"harness|hendrycksTest-professional_accounting|5": {
"acc": 0.4858156028368794,
"acc_stderr": 0.02981549448368206,
"acc_norm": 0.4858156028368794,
"acc_norm_stderr": 0.02981549448368206
},
"harness|hendrycksTest-professional_law|5": {
"acc": 0.44784876140808344,
"acc_stderr": 0.012700582404768221,
"acc_norm": 0.44784876140808344,
"acc_norm_stderr": 0.012700582404768221
},
"harness|hendrycksTest-professional_medicine|5": {
"acc": 0.6691176470588235,
"acc_stderr": 0.028582709753898445,
"acc_norm": 0.6691176470588235,
"acc_norm_stderr": 0.028582709753898445
},
"harness|hendrycksTest-professional_psychology|5": {
"acc": 0.6584967320261438,
"acc_stderr": 0.019184639328092487,
"acc_norm": 0.6584967320261438,
"acc_norm_stderr": 0.019184639328092487
},
"harness|hendrycksTest-public_relations|5": {
"acc": 0.6636363636363637,
"acc_stderr": 0.04525393596302506,
"acc_norm": 0.6636363636363637,
"acc_norm_stderr": 0.04525393596302506
},
"harness|hendrycksTest-security_studies|5": {
"acc": 0.7183673469387755,
"acc_stderr": 0.02879518557429129,
"acc_norm": 0.7183673469387755,
"acc_norm_stderr": 0.02879518557429129
},
"harness|hendrycksTest-sociology|5": {
"acc": 0.8308457711442786,
"acc_stderr": 0.026508590656233268,
"acc_norm": 0.8308457711442786,
"acc_norm_stderr": 0.026508590656233268
},
"harness|hendrycksTest-us_foreign_policy|5": {
"acc": 0.85,
"acc_stderr": 0.0358870281282637,
"acc_norm": 0.85,
"acc_norm_stderr": 0.0358870281282637
},
"harness|hendrycksTest-virology|5": {
"acc": 0.5421686746987951,
"acc_stderr": 0.0387862677100236,
"acc_norm": 0.5421686746987951,
"acc_norm_stderr": 0.0387862677100236
},
"harness|hendrycksTest-world_religions|5": {
"acc": 0.8304093567251462,
"acc_stderr": 0.02878210810540171,
"acc_norm": 0.8304093567251462,
"acc_norm_stderr": 0.02878210810540171
},
"harness|truthfulqa:mc|0": {
"mc1": 0.3084455324357405,
"mc1_stderr": 0.01616803938315687,
"mc2": 0.4634199786351567,
"mc2_stderr": 0.014481061527331505
},
"harness|winogrande|5": {
"acc": 0.7916337805840569,
"acc_stderr": 0.01141455439998773
},
"harness|drop|3": {
"em": 0.2679320469798658,
"em_stderr": 0.004535526201164825,
"f1": 0.31668204697986585,
"f1_stderr": 0.004459593071277455
},
"harness|gsm8k|5": {
"acc": 0.17361637604245642,
"acc_stderr": 0.01043346322125763
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] | [
-0.7672438025474548,
-0.8967578411102295,
0.2513273060321808,
0.19400064647197723,
-0.2396737039089203,
-0.03243205323815346,
0.054524075239896774,
-0.19431810081005096,
0.6111423373222351,
-0.05694881081581116,
-0.49031350016593933,
-0.6822000741958618,
-0.4746376872062683,
0.216163039207... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
danaroth/washington_dc_mall | danaroth | 2023-11-10T16:15:44Z | 0 | 0 | null | [
"license:unknown",
"region:us"
] | 2023-11-10T16:15:44Z | 2023-11-09T14:41:28.000Z | 2023-11-09T14:41:28 | ---
license: unknown
---
# Description
This dataset contains airborne hyperspectral data flightline over the Washington DC Mall provided with the permission of Spectral Information Technology Application Center of Virginia who was responsible for its collection. The sensor system HYDICE used in this case measured pixel response in 210 bands in the 0.4 to 2.4 μm region of the visible and infrared spectrum. Bands in the 0.9 and 1.4 μm region where the atmosphere is opaque have been omitted from the data set, leaving 191 bands. The data set contains 1208 scan lines with 307 pixels in each scan line. It totals approximately 150 Megabytes.
# Characteristics
Washington DC Mall data set classes, labels and the number of samples.
| # | Class | Samples |
|---|----------------|---------|
| 1 | Roofs | 21419 |
| 2 | Street | 9834 |
| 3 | Grass | 22873 |
| 4 | Trees | 6882 |
| 5 | Path | 1105 |
| 6 | Water | 11063 |
| 7 | Shadow | 3061 |
# Quick look
<figure>
<img src= "assets/1771082.gif" alt="Washington DC Mall" width="300" />
<figcaption>Fake color visualization of the Washington DC Mall dataset, with bands 60, 27, 17 for red, green, blue respectively.</figcaption>
</figure>
<figure>
<img src= "assets/4264435.gif" alt="Indian Pines gt" width="300" />
<figcaption>Groundtruth of Washington DC Mall dataset.</figcaption>
</figure>
# Credits
Dataset originally available as part of the Multispec project at: https://engineering.purdue.edu/~biehl/MultiSpec/hyperspectral.html
Copyright (C) 1994-2020 Purdue Research Foundation.
Work leading to MultiSpec was funded in part by NASA Grants NAGW-925, NAGW-3924 and NAGW5-3975.
Supported by AmericaView (www.americaview.org)
The hyperspectral data set (dc.tif) of the Washington, DC mall area is provided with the permission of Spectral Information Technology Application Center of Virginia who was responsible for its collection.
| [
-0.6316325664520264,
-0.20571161806583405,
0.2872912585735321,
-0.1971430480480194,
-0.16766588389873505,
0.21120496094226837,
0.15830686688423157,
-0.1924549788236618,
0.4321642220020294,
0.29867827892303467,
-0.6998411417007446,
-0.4884980320930481,
-0.16598786413669586,
-0.2146312445402... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Bilgi-arayan-Aslan/Prayer-times-Vienna_Gebetszeiten-Wien-GOE | Bilgi-arayan-Aslan | 2023-11-09T16:22:50Z | 0 | 1 | null | [
"region:us"
] | 2023-11-09T16:22:50Z | 2023-11-09T14:43:53.000Z | 2023-11-09T14:43:53 | # Gebetszeiten-Datensatz für Wien
Dieser Datensatz enthält die Gebetszeiten für Wien für das gesamte Jahr. Die Zeiten sind in der lokalen Wiener Zeit angegeben und enthalten die folgenden Gebete: Fajr, Shuruq, Dhuhr, Assr, Maghrib und Ishaa.
## Struktur
Jeder Tag ist im `MM-DD`-Format angegeben und enthält ein JSON-Objekt mit den Gebetszeiten für diesen Tag.
## Nutzung
Dieser Datensatz ist für Forschung, Bildung oder persönlichen Gebrauch bestimmt. Bitte stellen Sie sicher, dass Sie die Zeiten überprüfen, bevor Sie sie für religiöse Zwecke nutzen.
## Beitrag
Wenn Sie Fehler finden oder Verbesserungen vorschlagen möchten, erstellen Sie bitte ein Issue oder einen Pull-Request im Repository.
# Prayer Times Dataset for Vienna
This dataset contains the prayer times for Vienna for the entire year. Times are provided in local Vienna time and include the following prayers: Fajr, Shuruq, Dhuhr, Assr, Maghrib, and Ishaa.
## Structure
Each day is listed in `MM-DD` format and contains a JSON object with the prayer times for that day.
## Usage
This dataset is intended for research, educational, or personal use. Please ensure to verify the times before using them for religious purposes.
## Contributions
If you find any errors or would like to suggest improvements, please create an issue or a pull request in the repository.
# مجموعة بيانات أوقات الصلاة لمدينة فيينا
تحتوي هذه المجموعة على أوقات الصلاة لمدينة فيينا للعام كامل. الأوقات مُقدمة بتوقيت فيينا المحلي وتشمل الصلوات التالية: الفجر، الشروق، الظهر، العصر، المغرب والعشاء.
## الهيكل
يُسرد كل يوم بتنسيق `MM-DD` ويحتوي على كائن JSON بأوقات الصلاة لذلك اليوم.
## الاستخدام
هذه المجموعة مُعدة للبحث، الأغراض التعليمية، أو الاستخدام الشخصي. يُرجى التأكد من التحقق من الأوقات قبل استخدامها لأغراض دينية.
## المساهمات
إذا وجدت أي أخطاء أو أردت اقتراح تحسينات، الرجاء إنشاء مُشكلة أو طلب سحب في المستودع.
### Credits:
The data comes from this source: https://islamiccentre.at/goe/
---
license: other
license_name: islamic-mit-license
license_link: LICENSE
---
| [
-0.5605645775794983,
-0.827768862247467,
0.35233232378959656,
0.34216928482055664,
-0.5860269665718079,
-0.2938080132007599,
0.07205575704574585,
-0.4799289107322693,
0.43757951259613037,
0.611514151096344,
-0.5666208863258362,
-0.9839429259300232,
-0.5971713066101074,
0.4725368320941925,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
open-llm-leaderboard/details_migtissera__SynthIA-7B-v1.5_public | open-llm-leaderboard | 2023-11-09T14:46:00Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T14:46:00Z | 2023-11-09T14:44:56.000Z | 2023-11-09T14:44:56 | ---
pretty_name: Evaluation run of migtissera/SynthIA-7B-v1.5
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [migtissera/SynthIA-7B-v1.5](https://huggingface.co/migtissera/SynthIA-7B-v1.5)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 64 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the aggregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_migtissera__SynthIA-7B-v1.5_public\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-11-09T14:41:56.883085](https://huggingface.co/datasets/open-llm-leaderboard/details_migtissera__SynthIA-7B-v1.5_public/blob/main/results_2023-11-09T14-41-56.883085.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"acc\": 0.6291968571108129,\n\
\ \"acc_stderr\": 0.03252538162461919,\n \"acc_norm\": 0.63804599014876,\n\
\ \"acc_norm_stderr\": 0.03323519542303871,\n \"mc1\": 0.35128518971848227,\n\
\ \"mc1_stderr\": 0.016711358163544403,\n \"mc2\": 0.5131996962275648,\n\
\ \"mc2_stderr\": 0.015337988977122931,\n \"em\": 0.1875,\n \
\ \"em_stderr\": 0.003997164044486006,\n \"f1\": 0.26010591442953035,\n\
\ \"f1_stderr\": 0.004042449995216609\n },\n \"harness|arc:challenge|25\"\
: {\n \"acc\": 0.5870307167235495,\n \"acc_stderr\": 0.014388344935398324,\n\
\ \"acc_norm\": 0.6271331058020477,\n \"acc_norm_stderr\": 0.014131176760131172\n\
\ },\n \"harness|hellaswag|10\": {\n \"acc\": 0.6432981477793268,\n\
\ \"acc_stderr\": 0.0047804672709117705,\n \"acc_norm\": 0.833698466440948,\n\
\ \"acc_norm_stderr\": 0.0037159010850549967\n },\n \"harness|hendrycksTest-abstract_algebra|5\"\
: {\n \"acc\": 0.34,\n \"acc_stderr\": 0.04760952285695236,\n \
\ \"acc_norm\": 0.34,\n \"acc_norm_stderr\": 0.04760952285695236\n \
\ },\n \"harness|hendrycksTest-anatomy|5\": {\n \"acc\": 0.6148148148148148,\n\
\ \"acc_stderr\": 0.04203921040156279,\n \"acc_norm\": 0.6148148148148148,\n\
\ \"acc_norm_stderr\": 0.04203921040156279\n },\n \"harness|hendrycksTest-astronomy|5\"\
: {\n \"acc\": 0.6644736842105263,\n \"acc_stderr\": 0.038424985593952694,\n\
\ \"acc_norm\": 0.6644736842105263,\n \"acc_norm_stderr\": 0.038424985593952694\n\
\ },\n \"harness|hendrycksTest-business_ethics|5\": {\n \"acc\": 0.61,\n\
\ \"acc_stderr\": 0.04902071300001975,\n \"acc_norm\": 0.61,\n \
\ \"acc_norm_stderr\": 0.04902071300001975\n },\n \"harness|hendrycksTest-clinical_knowledge|5\"\
: {\n \"acc\": 0.6981132075471698,\n \"acc_stderr\": 0.02825420034443866,\n\
\ \"acc_norm\": 0.6981132075471698,\n \"acc_norm_stderr\": 0.02825420034443866\n\
\ },\n \"harness|hendrycksTest-college_biology|5\": {\n \"acc\": 0.6944444444444444,\n\
\ \"acc_stderr\": 0.03852084696008534,\n \"acc_norm\": 0.6944444444444444,\n\
\ \"acc_norm_stderr\": 0.03852084696008534\n },\n \"harness|hendrycksTest-college_chemistry|5\"\
: {\n \"acc\": 0.49,\n \"acc_stderr\": 0.05024183937956912,\n \
\ \"acc_norm\": 0.49,\n \"acc_norm_stderr\": 0.05024183937956912\n \
\ },\n \"harness|hendrycksTest-college_computer_science|5\": {\n \"acc\"\
: 0.52,\n \"acc_stderr\": 0.050211673156867795,\n \"acc_norm\": 0.52,\n\
\ \"acc_norm_stderr\": 0.050211673156867795\n },\n \"harness|hendrycksTest-college_mathematics|5\"\
: {\n \"acc\": 0.4,\n \"acc_stderr\": 0.04923659639173309,\n \
\ \"acc_norm\": 0.4,\n \"acc_norm_stderr\": 0.04923659639173309\n },\n\
\ \"harness|hendrycksTest-college_medicine|5\": {\n \"acc\": 0.6473988439306358,\n\
\ \"acc_stderr\": 0.036430371689585475,\n \"acc_norm\": 0.6473988439306358,\n\
\ \"acc_norm_stderr\": 0.036430371689585475\n },\n \"harness|hendrycksTest-college_physics|5\"\
: {\n \"acc\": 0.38235294117647056,\n \"acc_stderr\": 0.04835503696107223,\n\
\ \"acc_norm\": 0.38235294117647056,\n \"acc_norm_stderr\": 0.04835503696107223\n\
\ },\n \"harness|hendrycksTest-computer_security|5\": {\n \"acc\":\
\ 0.76,\n \"acc_stderr\": 0.042923469599092816,\n \"acc_norm\": 0.76,\n\
\ \"acc_norm_stderr\": 0.042923469599092816\n },\n \"harness|hendrycksTest-conceptual_physics|5\"\
: {\n \"acc\": 0.5957446808510638,\n \"acc_stderr\": 0.03208115750788684,\n\
\ \"acc_norm\": 0.5957446808510638,\n \"acc_norm_stderr\": 0.03208115750788684\n\
\ },\n \"harness|hendrycksTest-econometrics|5\": {\n \"acc\": 0.43859649122807015,\n\
\ \"acc_stderr\": 0.04668000738510455,\n \"acc_norm\": 0.43859649122807015,\n\
\ \"acc_norm_stderr\": 0.04668000738510455\n },\n \"harness|hendrycksTest-electrical_engineering|5\"\
: {\n \"acc\": 0.5448275862068965,\n \"acc_stderr\": 0.04149886942192117,\n\
\ \"acc_norm\": 0.5448275862068965,\n \"acc_norm_stderr\": 0.04149886942192117\n\
\ },\n \"harness|hendrycksTest-elementary_mathematics|5\": {\n \"acc\"\
: 0.3941798941798942,\n \"acc_stderr\": 0.02516798233389414,\n \"\
acc_norm\": 0.3941798941798942,\n \"acc_norm_stderr\": 0.02516798233389414\n\
\ },\n \"harness|hendrycksTest-formal_logic|5\": {\n \"acc\": 0.4126984126984127,\n\
\ \"acc_stderr\": 0.04403438954768176,\n \"acc_norm\": 0.4126984126984127,\n\
\ \"acc_norm_stderr\": 0.04403438954768176\n },\n \"harness|hendrycksTest-global_facts|5\"\
: {\n \"acc\": 0.39,\n \"acc_stderr\": 0.04902071300001975,\n \
\ \"acc_norm\": 0.39,\n \"acc_norm_stderr\": 0.04902071300001975\n \
\ },\n \"harness|hendrycksTest-high_school_biology|5\": {\n \"acc\": 0.7645161290322581,\n\
\ \"acc_stderr\": 0.02413763242933771,\n \"acc_norm\": 0.7645161290322581,\n\
\ \"acc_norm_stderr\": 0.02413763242933771\n },\n \"harness|hendrycksTest-high_school_chemistry|5\"\
: {\n \"acc\": 0.4975369458128079,\n \"acc_stderr\": 0.03517945038691063,\n\
\ \"acc_norm\": 0.4975369458128079,\n \"acc_norm_stderr\": 0.03517945038691063\n\
\ },\n \"harness|hendrycksTest-high_school_computer_science|5\": {\n \
\ \"acc\": 0.68,\n \"acc_stderr\": 0.04688261722621505,\n \"acc_norm\"\
: 0.68,\n \"acc_norm_stderr\": 0.04688261722621505\n },\n \"harness|hendrycksTest-high_school_european_history|5\"\
: {\n \"acc\": 0.7575757575757576,\n \"acc_stderr\": 0.03346409881055953,\n\
\ \"acc_norm\": 0.7575757575757576,\n \"acc_norm_stderr\": 0.03346409881055953\n\
\ },\n \"harness|hendrycksTest-high_school_geography|5\": {\n \"acc\"\
: 0.7878787878787878,\n \"acc_stderr\": 0.029126522834586818,\n \"\
acc_norm\": 0.7878787878787878,\n \"acc_norm_stderr\": 0.029126522834586818\n\
\ },\n \"harness|hendrycksTest-high_school_government_and_politics|5\": {\n\
\ \"acc\": 0.8652849740932642,\n \"acc_stderr\": 0.02463978909770944,\n\
\ \"acc_norm\": 0.8652849740932642,\n \"acc_norm_stderr\": 0.02463978909770944\n\
\ },\n \"harness|hendrycksTest-high_school_macroeconomics|5\": {\n \
\ \"acc\": 0.6641025641025641,\n \"acc_stderr\": 0.023946724741563973,\n\
\ \"acc_norm\": 0.6641025641025641,\n \"acc_norm_stderr\": 0.023946724741563973\n\
\ },\n \"harness|hendrycksTest-high_school_mathematics|5\": {\n \"\
acc\": 0.3592592592592593,\n \"acc_stderr\": 0.029252905927251976,\n \
\ \"acc_norm\": 0.3592592592592593,\n \"acc_norm_stderr\": 0.029252905927251976\n\
\ },\n \"harness|hendrycksTest-high_school_microeconomics|5\": {\n \
\ \"acc\": 0.7058823529411765,\n \"acc_stderr\": 0.02959732973097809,\n \
\ \"acc_norm\": 0.7058823529411765,\n \"acc_norm_stderr\": 0.02959732973097809\n\
\ },\n \"harness|hendrycksTest-high_school_physics|5\": {\n \"acc\"\
: 0.32450331125827814,\n \"acc_stderr\": 0.03822746937658753,\n \"\
acc_norm\": 0.32450331125827814,\n \"acc_norm_stderr\": 0.03822746937658753\n\
\ },\n \"harness|hendrycksTest-high_school_psychology|5\": {\n \"acc\"\
: 0.818348623853211,\n \"acc_stderr\": 0.016530617409266875,\n \"\
acc_norm\": 0.818348623853211,\n \"acc_norm_stderr\": 0.016530617409266875\n\
\ },\n \"harness|hendrycksTest-high_school_statistics|5\": {\n \"acc\"\
: 0.5231481481481481,\n \"acc_stderr\": 0.03406315360711507,\n \"\
acc_norm\": 0.5231481481481481,\n \"acc_norm_stderr\": 0.03406315360711507\n\
\ },\n \"harness|hendrycksTest-high_school_us_history|5\": {\n \"acc\"\
: 0.7990196078431373,\n \"acc_stderr\": 0.02812597226565438,\n \"\
acc_norm\": 0.7990196078431373,\n \"acc_norm_stderr\": 0.02812597226565438\n\
\ },\n \"harness|hendrycksTest-high_school_world_history|5\": {\n \"\
acc\": 0.7932489451476793,\n \"acc_stderr\": 0.026361651668389094,\n \
\ \"acc_norm\": 0.7932489451476793,\n \"acc_norm_stderr\": 0.026361651668389094\n\
\ },\n \"harness|hendrycksTest-human_aging|5\": {\n \"acc\": 0.6816143497757847,\n\
\ \"acc_stderr\": 0.03126580522513713,\n \"acc_norm\": 0.6816143497757847,\n\
\ \"acc_norm_stderr\": 0.03126580522513713\n },\n \"harness|hendrycksTest-human_sexuality|5\"\
: {\n \"acc\": 0.7633587786259542,\n \"acc_stderr\": 0.03727673575596914,\n\
\ \"acc_norm\": 0.7633587786259542,\n \"acc_norm_stderr\": 0.03727673575596914\n\
\ },\n \"harness|hendrycksTest-international_law|5\": {\n \"acc\":\
\ 0.7851239669421488,\n \"acc_stderr\": 0.037494924487096966,\n \"\
acc_norm\": 0.7851239669421488,\n \"acc_norm_stderr\": 0.037494924487096966\n\
\ },\n \"harness|hendrycksTest-jurisprudence|5\": {\n \"acc\": 0.7870370370370371,\n\
\ \"acc_stderr\": 0.0395783547198098,\n \"acc_norm\": 0.7870370370370371,\n\
\ \"acc_norm_stderr\": 0.0395783547198098\n },\n \"harness|hendrycksTest-logical_fallacies|5\"\
: {\n \"acc\": 0.7607361963190185,\n \"acc_stderr\": 0.03351953879521272,\n\
\ \"acc_norm\": 0.7607361963190185,\n \"acc_norm_stderr\": 0.03351953879521272\n\
\ },\n \"harness|hendrycksTest-machine_learning|5\": {\n \"acc\": 0.44642857142857145,\n\
\ \"acc_stderr\": 0.047184714852195886,\n \"acc_norm\": 0.44642857142857145,\n\
\ \"acc_norm_stderr\": 0.047184714852195886\n },\n \"harness|hendrycksTest-management|5\"\
: {\n \"acc\": 0.7669902912621359,\n \"acc_stderr\": 0.04185832598928315,\n\
\ \"acc_norm\": 0.7669902912621359,\n \"acc_norm_stderr\": 0.04185832598928315\n\
\ },\n \"harness|hendrycksTest-marketing|5\": {\n \"acc\": 0.8760683760683761,\n\
\ \"acc_stderr\": 0.02158649400128138,\n \"acc_norm\": 0.8760683760683761,\n\
\ \"acc_norm_stderr\": 0.02158649400128138\n },\n \"harness|hendrycksTest-medical_genetics|5\"\
: {\n \"acc\": 0.69,\n \"acc_stderr\": 0.04648231987117316,\n \
\ \"acc_norm\": 0.69,\n \"acc_norm_stderr\": 0.04648231987117316\n \
\ },\n \"harness|hendrycksTest-miscellaneous|5\": {\n \"acc\": 0.8199233716475096,\n\
\ \"acc_stderr\": 0.013740797258579828,\n \"acc_norm\": 0.8199233716475096,\n\
\ \"acc_norm_stderr\": 0.013740797258579828\n },\n \"harness|hendrycksTest-moral_disputes|5\"\
: {\n \"acc\": 0.6994219653179191,\n \"acc_stderr\": 0.0246853168672578,\n\
\ \"acc_norm\": 0.6994219653179191,\n \"acc_norm_stderr\": 0.0246853168672578\n\
\ },\n \"harness|hendrycksTest-moral_scenarios|5\": {\n \"acc\": 0.376536312849162,\n\
\ \"acc_stderr\": 0.016204672385106596,\n \"acc_norm\": 0.376536312849162,\n\
\ \"acc_norm_stderr\": 0.016204672385106596\n },\n \"harness|hendrycksTest-nutrition|5\"\
: {\n \"acc\": 0.7156862745098039,\n \"acc_stderr\": 0.02582916327275748,\n\
\ \"acc_norm\": 0.7156862745098039,\n \"acc_norm_stderr\": 0.02582916327275748\n\
\ },\n \"harness|hendrycksTest-philosophy|5\": {\n \"acc\": 0.7041800643086816,\n\
\ \"acc_stderr\": 0.02592237178881876,\n \"acc_norm\": 0.7041800643086816,\n\
\ \"acc_norm_stderr\": 0.02592237178881876\n },\n \"harness|hendrycksTest-prehistory|5\"\
: {\n \"acc\": 0.7283950617283951,\n \"acc_stderr\": 0.02474862449053737,\n\
\ \"acc_norm\": 0.7283950617283951,\n \"acc_norm_stderr\": 0.02474862449053737\n\
\ },\n \"harness|hendrycksTest-professional_accounting|5\": {\n \"\
acc\": 0.4716312056737589,\n \"acc_stderr\": 0.029779450957303062,\n \
\ \"acc_norm\": 0.4716312056737589,\n \"acc_norm_stderr\": 0.029779450957303062\n\
\ },\n \"harness|hendrycksTest-professional_law|5\": {\n \"acc\": 0.4589308996088657,\n\
\ \"acc_stderr\": 0.012727084826799797,\n \"acc_norm\": 0.4589308996088657,\n\
\ \"acc_norm_stderr\": 0.012727084826799797\n },\n \"harness|hendrycksTest-professional_medicine|5\"\
: {\n \"acc\": 0.6580882352941176,\n \"acc_stderr\": 0.028814722422254184,\n\
\ \"acc_norm\": 0.6580882352941176,\n \"acc_norm_stderr\": 0.028814722422254184\n\
\ },\n \"harness|hendrycksTest-professional_psychology|5\": {\n \"\
acc\": 0.6584967320261438,\n \"acc_stderr\": 0.019184639328092487,\n \
\ \"acc_norm\": 0.6584967320261438,\n \"acc_norm_stderr\": 0.019184639328092487\n\
\ },\n \"harness|hendrycksTest-public_relations|5\": {\n \"acc\": 0.6727272727272727,\n\
\ \"acc_stderr\": 0.0449429086625209,\n \"acc_norm\": 0.6727272727272727,\n\
\ \"acc_norm_stderr\": 0.0449429086625209\n },\n \"harness|hendrycksTest-security_studies|5\"\
: {\n \"acc\": 0.6938775510204082,\n \"acc_stderr\": 0.02950489645459596,\n\
\ \"acc_norm\": 0.6938775510204082,\n \"acc_norm_stderr\": 0.02950489645459596\n\
\ },\n \"harness|hendrycksTest-sociology|5\": {\n \"acc\": 0.835820895522388,\n\
\ \"acc_stderr\": 0.026193923544454132,\n \"acc_norm\": 0.835820895522388,\n\
\ \"acc_norm_stderr\": 0.026193923544454132\n },\n \"harness|hendrycksTest-us_foreign_policy|5\"\
: {\n \"acc\": 0.84,\n \"acc_stderr\": 0.03684529491774711,\n \
\ \"acc_norm\": 0.84,\n \"acc_norm_stderr\": 0.03684529491774711\n \
\ },\n \"harness|hendrycksTest-virology|5\": {\n \"acc\": 0.536144578313253,\n\
\ \"acc_stderr\": 0.03882310850890594,\n \"acc_norm\": 0.536144578313253,\n\
\ \"acc_norm_stderr\": 0.03882310850890594\n },\n \"harness|hendrycksTest-world_religions|5\"\
: {\n \"acc\": 0.8187134502923976,\n \"acc_stderr\": 0.029547741687640038,\n\
\ \"acc_norm\": 0.8187134502923976,\n \"acc_norm_stderr\": 0.029547741687640038\n\
\ },\n \"harness|truthfulqa:mc|0\": {\n \"mc1\": 0.35128518971848227,\n\
\ \"mc1_stderr\": 0.016711358163544403,\n \"mc2\": 0.5131996962275648,\n\
\ \"mc2_stderr\": 0.015337988977122931\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.7924230465666929,\n \"acc_stderr\": 0.01139859341938678\n\
\ },\n \"harness|drop|3\": {\n \"em\": 0.1875,\n \"em_stderr\"\
: 0.003997164044486006,\n \"f1\": 0.26010591442953035,\n \"f1_stderr\"\
: 0.004042449995216609\n },\n \"harness|gsm8k|5\": {\n \"acc\": 0.17437452615617893,\n\
\ \"acc_stderr\": 0.010451421361976231\n }\n}\n```"
repo_url: https://huggingface.co/migtissera/SynthIA-7B-v1.5
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|arc:challenge|25_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_drop_3
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|drop|3_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|drop|3_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|gsm8k|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hellaswag|10_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-41-56.883085.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T14-41-56.883085.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- '**/details_harness|winogrande|5_2023-11-09T14-41-56.883085.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-11-09T14-41-56.883085.parquet'
- config_name: results
data_files:
- split: 2023_11_09T14_41_56.883085
path:
- results_2023-11-09T14-41-56.883085.parquet
- split: latest
path:
- results_2023-11-09T14-41-56.883085.parquet
---
# Dataset Card for Evaluation run of migtissera/SynthIA-7B-v1.5
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/migtissera/SynthIA-7B-v1.5
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [migtissera/SynthIA-7B-v1.5](https://huggingface.co/migtissera/SynthIA-7B-v1.5) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 64 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_migtissera__SynthIA-7B-v1.5_public",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-11-09T14:41:56.883085](https://huggingface.co/datasets/open-llm-leaderboard/details_migtissera__SynthIA-7B-v1.5_public/blob/main/results_2023-11-09T14-41-56.883085.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"acc": 0.6291968571108129,
"acc_stderr": 0.03252538162461919,
"acc_norm": 0.63804599014876,
"acc_norm_stderr": 0.03323519542303871,
"mc1": 0.35128518971848227,
"mc1_stderr": 0.016711358163544403,
"mc2": 0.5131996962275648,
"mc2_stderr": 0.015337988977122931,
"em": 0.1875,
"em_stderr": 0.003997164044486006,
"f1": 0.26010591442953035,
"f1_stderr": 0.004042449995216609
},
"harness|arc:challenge|25": {
"acc": 0.5870307167235495,
"acc_stderr": 0.014388344935398324,
"acc_norm": 0.6271331058020477,
"acc_norm_stderr": 0.014131176760131172
},
"harness|hellaswag|10": {
"acc": 0.6432981477793268,
"acc_stderr": 0.0047804672709117705,
"acc_norm": 0.833698466440948,
"acc_norm_stderr": 0.0037159010850549967
},
"harness|hendrycksTest-abstract_algebra|5": {
"acc": 0.34,
"acc_stderr": 0.04760952285695236,
"acc_norm": 0.34,
"acc_norm_stderr": 0.04760952285695236
},
"harness|hendrycksTest-anatomy|5": {
"acc": 0.6148148148148148,
"acc_stderr": 0.04203921040156279,
"acc_norm": 0.6148148148148148,
"acc_norm_stderr": 0.04203921040156279
},
"harness|hendrycksTest-astronomy|5": {
"acc": 0.6644736842105263,
"acc_stderr": 0.038424985593952694,
"acc_norm": 0.6644736842105263,
"acc_norm_stderr": 0.038424985593952694
},
"harness|hendrycksTest-business_ethics|5": {
"acc": 0.61,
"acc_stderr": 0.04902071300001975,
"acc_norm": 0.61,
"acc_norm_stderr": 0.04902071300001975
},
"harness|hendrycksTest-clinical_knowledge|5": {
"acc": 0.6981132075471698,
"acc_stderr": 0.02825420034443866,
"acc_norm": 0.6981132075471698,
"acc_norm_stderr": 0.02825420034443866
},
"harness|hendrycksTest-college_biology|5": {
"acc": 0.6944444444444444,
"acc_stderr": 0.03852084696008534,
"acc_norm": 0.6944444444444444,
"acc_norm_stderr": 0.03852084696008534
},
"harness|hendrycksTest-college_chemistry|5": {
"acc": 0.49,
"acc_stderr": 0.05024183937956912,
"acc_norm": 0.49,
"acc_norm_stderr": 0.05024183937956912
},
"harness|hendrycksTest-college_computer_science|5": {
"acc": 0.52,
"acc_stderr": 0.050211673156867795,
"acc_norm": 0.52,
"acc_norm_stderr": 0.050211673156867795
},
"harness|hendrycksTest-college_mathematics|5": {
"acc": 0.4,
"acc_stderr": 0.04923659639173309,
"acc_norm": 0.4,
"acc_norm_stderr": 0.04923659639173309
},
"harness|hendrycksTest-college_medicine|5": {
"acc": 0.6473988439306358,
"acc_stderr": 0.036430371689585475,
"acc_norm": 0.6473988439306358,
"acc_norm_stderr": 0.036430371689585475
},
"harness|hendrycksTest-college_physics|5": {
"acc": 0.38235294117647056,
"acc_stderr": 0.04835503696107223,
"acc_norm": 0.38235294117647056,
"acc_norm_stderr": 0.04835503696107223
},
"harness|hendrycksTest-computer_security|5": {
"acc": 0.76,
"acc_stderr": 0.042923469599092816,
"acc_norm": 0.76,
"acc_norm_stderr": 0.042923469599092816
},
"harness|hendrycksTest-conceptual_physics|5": {
"acc": 0.5957446808510638,
"acc_stderr": 0.03208115750788684,
"acc_norm": 0.5957446808510638,
"acc_norm_stderr": 0.03208115750788684
},
"harness|hendrycksTest-econometrics|5": {
"acc": 0.43859649122807015,
"acc_stderr": 0.04668000738510455,
"acc_norm": 0.43859649122807015,
"acc_norm_stderr": 0.04668000738510455
},
"harness|hendrycksTest-electrical_engineering|5": {
"acc": 0.5448275862068965,
"acc_stderr": 0.04149886942192117,
"acc_norm": 0.5448275862068965,
"acc_norm_stderr": 0.04149886942192117
},
"harness|hendrycksTest-elementary_mathematics|5": {
"acc": 0.3941798941798942,
"acc_stderr": 0.02516798233389414,
"acc_norm": 0.3941798941798942,
"acc_norm_stderr": 0.02516798233389414
},
"harness|hendrycksTest-formal_logic|5": {
"acc": 0.4126984126984127,
"acc_stderr": 0.04403438954768176,
"acc_norm": 0.4126984126984127,
"acc_norm_stderr": 0.04403438954768176
},
"harness|hendrycksTest-global_facts|5": {
"acc": 0.39,
"acc_stderr": 0.04902071300001975,
"acc_norm": 0.39,
"acc_norm_stderr": 0.04902071300001975
},
"harness|hendrycksTest-high_school_biology|5": {
"acc": 0.7645161290322581,
"acc_stderr": 0.02413763242933771,
"acc_norm": 0.7645161290322581,
"acc_norm_stderr": 0.02413763242933771
},
"harness|hendrycksTest-high_school_chemistry|5": {
"acc": 0.4975369458128079,
"acc_stderr": 0.03517945038691063,
"acc_norm": 0.4975369458128079,
"acc_norm_stderr": 0.03517945038691063
},
"harness|hendrycksTest-high_school_computer_science|5": {
"acc": 0.68,
"acc_stderr": 0.04688261722621505,
"acc_norm": 0.68,
"acc_norm_stderr": 0.04688261722621505
},
"harness|hendrycksTest-high_school_european_history|5": {
"acc": 0.7575757575757576,
"acc_stderr": 0.03346409881055953,
"acc_norm": 0.7575757575757576,
"acc_norm_stderr": 0.03346409881055953
},
"harness|hendrycksTest-high_school_geography|5": {
"acc": 0.7878787878787878,
"acc_stderr": 0.029126522834586818,
"acc_norm": 0.7878787878787878,
"acc_norm_stderr": 0.029126522834586818
},
"harness|hendrycksTest-high_school_government_and_politics|5": {
"acc": 0.8652849740932642,
"acc_stderr": 0.02463978909770944,
"acc_norm": 0.8652849740932642,
"acc_norm_stderr": 0.02463978909770944
},
"harness|hendrycksTest-high_school_macroeconomics|5": {
"acc": 0.6641025641025641,
"acc_stderr": 0.023946724741563973,
"acc_norm": 0.6641025641025641,
"acc_norm_stderr": 0.023946724741563973
},
"harness|hendrycksTest-high_school_mathematics|5": {
"acc": 0.3592592592592593,
"acc_stderr": 0.029252905927251976,
"acc_norm": 0.3592592592592593,
"acc_norm_stderr": 0.029252905927251976
},
"harness|hendrycksTest-high_school_microeconomics|5": {
"acc": 0.7058823529411765,
"acc_stderr": 0.02959732973097809,
"acc_norm": 0.7058823529411765,
"acc_norm_stderr": 0.02959732973097809
},
"harness|hendrycksTest-high_school_physics|5": {
"acc": 0.32450331125827814,
"acc_stderr": 0.03822746937658753,
"acc_norm": 0.32450331125827814,
"acc_norm_stderr": 0.03822746937658753
},
"harness|hendrycksTest-high_school_psychology|5": {
"acc": 0.818348623853211,
"acc_stderr": 0.016530617409266875,
"acc_norm": 0.818348623853211,
"acc_norm_stderr": 0.016530617409266875
},
"harness|hendrycksTest-high_school_statistics|5": {
"acc": 0.5231481481481481,
"acc_stderr": 0.03406315360711507,
"acc_norm": 0.5231481481481481,
"acc_norm_stderr": 0.03406315360711507
},
"harness|hendrycksTest-high_school_us_history|5": {
"acc": 0.7990196078431373,
"acc_stderr": 0.02812597226565438,
"acc_norm": 0.7990196078431373,
"acc_norm_stderr": 0.02812597226565438
},
"harness|hendrycksTest-high_school_world_history|5": {
"acc": 0.7932489451476793,
"acc_stderr": 0.026361651668389094,
"acc_norm": 0.7932489451476793,
"acc_norm_stderr": 0.026361651668389094
},
"harness|hendrycksTest-human_aging|5": {
"acc": 0.6816143497757847,
"acc_stderr": 0.03126580522513713,
"acc_norm": 0.6816143497757847,
"acc_norm_stderr": 0.03126580522513713
},
"harness|hendrycksTest-human_sexuality|5": {
"acc": 0.7633587786259542,
"acc_stderr": 0.03727673575596914,
"acc_norm": 0.7633587786259542,
"acc_norm_stderr": 0.03727673575596914
},
"harness|hendrycksTest-international_law|5": {
"acc": 0.7851239669421488,
"acc_stderr": 0.037494924487096966,
"acc_norm": 0.7851239669421488,
"acc_norm_stderr": 0.037494924487096966
},
"harness|hendrycksTest-jurisprudence|5": {
"acc": 0.7870370370370371,
"acc_stderr": 0.0395783547198098,
"acc_norm": 0.7870370370370371,
"acc_norm_stderr": 0.0395783547198098
},
"harness|hendrycksTest-logical_fallacies|5": {
"acc": 0.7607361963190185,
"acc_stderr": 0.03351953879521272,
"acc_norm": 0.7607361963190185,
"acc_norm_stderr": 0.03351953879521272
},
"harness|hendrycksTest-machine_learning|5": {
"acc": 0.44642857142857145,
"acc_stderr": 0.047184714852195886,
"acc_norm": 0.44642857142857145,
"acc_norm_stderr": 0.047184714852195886
},
"harness|hendrycksTest-management|5": {
"acc": 0.7669902912621359,
"acc_stderr": 0.04185832598928315,
"acc_norm": 0.7669902912621359,
"acc_norm_stderr": 0.04185832598928315
},
"harness|hendrycksTest-marketing|5": {
"acc": 0.8760683760683761,
"acc_stderr": 0.02158649400128138,
"acc_norm": 0.8760683760683761,
"acc_norm_stderr": 0.02158649400128138
},
"harness|hendrycksTest-medical_genetics|5": {
"acc": 0.69,
"acc_stderr": 0.04648231987117316,
"acc_norm": 0.69,
"acc_norm_stderr": 0.04648231987117316
},
"harness|hendrycksTest-miscellaneous|5": {
"acc": 0.8199233716475096,
"acc_stderr": 0.013740797258579828,
"acc_norm": 0.8199233716475096,
"acc_norm_stderr": 0.013740797258579828
},
"harness|hendrycksTest-moral_disputes|5": {
"acc": 0.6994219653179191,
"acc_stderr": 0.0246853168672578,
"acc_norm": 0.6994219653179191,
"acc_norm_stderr": 0.0246853168672578
},
"harness|hendrycksTest-moral_scenarios|5": {
"acc": 0.376536312849162,
"acc_stderr": 0.016204672385106596,
"acc_norm": 0.376536312849162,
"acc_norm_stderr": 0.016204672385106596
},
"harness|hendrycksTest-nutrition|5": {
"acc": 0.7156862745098039,
"acc_stderr": 0.02582916327275748,
"acc_norm": 0.7156862745098039,
"acc_norm_stderr": 0.02582916327275748
},
"harness|hendrycksTest-philosophy|5": {
"acc": 0.7041800643086816,
"acc_stderr": 0.02592237178881876,
"acc_norm": 0.7041800643086816,
"acc_norm_stderr": 0.02592237178881876
},
"harness|hendrycksTest-prehistory|5": {
"acc": 0.7283950617283951,
"acc_stderr": 0.02474862449053737,
"acc_norm": 0.7283950617283951,
"acc_norm_stderr": 0.02474862449053737
},
"harness|hendrycksTest-professional_accounting|5": {
"acc": 0.4716312056737589,
"acc_stderr": 0.029779450957303062,
"acc_norm": 0.4716312056737589,
"acc_norm_stderr": 0.029779450957303062
},
"harness|hendrycksTest-professional_law|5": {
"acc": 0.4589308996088657,
"acc_stderr": 0.012727084826799797,
"acc_norm": 0.4589308996088657,
"acc_norm_stderr": 0.012727084826799797
},
"harness|hendrycksTest-professional_medicine|5": {
"acc": 0.6580882352941176,
"acc_stderr": 0.028814722422254184,
"acc_norm": 0.6580882352941176,
"acc_norm_stderr": 0.028814722422254184
},
"harness|hendrycksTest-professional_psychology|5": {
"acc": 0.6584967320261438,
"acc_stderr": 0.019184639328092487,
"acc_norm": 0.6584967320261438,
"acc_norm_stderr": 0.019184639328092487
},
"harness|hendrycksTest-public_relations|5": {
"acc": 0.6727272727272727,
"acc_stderr": 0.0449429086625209,
"acc_norm": 0.6727272727272727,
"acc_norm_stderr": 0.0449429086625209
},
"harness|hendrycksTest-security_studies|5": {
"acc": 0.6938775510204082,
"acc_stderr": 0.02950489645459596,
"acc_norm": 0.6938775510204082,
"acc_norm_stderr": 0.02950489645459596
},
"harness|hendrycksTest-sociology|5": {
"acc": 0.835820895522388,
"acc_stderr": 0.026193923544454132,
"acc_norm": 0.835820895522388,
"acc_norm_stderr": 0.026193923544454132
},
"harness|hendrycksTest-us_foreign_policy|5": {
"acc": 0.84,
"acc_stderr": 0.03684529491774711,
"acc_norm": 0.84,
"acc_norm_stderr": 0.03684529491774711
},
"harness|hendrycksTest-virology|5": {
"acc": 0.536144578313253,
"acc_stderr": 0.03882310850890594,
"acc_norm": 0.536144578313253,
"acc_norm_stderr": 0.03882310850890594
},
"harness|hendrycksTest-world_religions|5": {
"acc": 0.8187134502923976,
"acc_stderr": 0.029547741687640038,
"acc_norm": 0.8187134502923976,
"acc_norm_stderr": 0.029547741687640038
},
"harness|truthfulqa:mc|0": {
"mc1": 0.35128518971848227,
"mc1_stderr": 0.016711358163544403,
"mc2": 0.5131996962275648,
"mc2_stderr": 0.015337988977122931
},
"harness|winogrande|5": {
"acc": 0.7924230465666929,
"acc_stderr": 0.01139859341938678
},
"harness|drop|3": {
"em": 0.1875,
"em_stderr": 0.003997164044486006,
"f1": 0.26010591442953035,
"f1_stderr": 0.004042449995216609
},
"harness|gsm8k|5": {
"acc": 0.17437452615617893,
"acc_stderr": 0.010451421361976231
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] | [
-0.7448519468307495,
-0.8373547792434692,
0.3236550986766815,
0.2275657057762146,
-0.1883840411901474,
-0.04911923408508301,
0.019352992996573448,
-0.24485768377780914,
0.5842747092247009,
-0.06284689903259277,
-0.4906705915927887,
-0.744883120059967,
-0.4455663859844208,
0.267535358667373... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
open-llm-leaderboard/details_CobraMamba__mamba-gpt-7b-v2_public | open-llm-leaderboard | 2023-11-09T14:46:50Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T14:46:50Z | 2023-11-09T14:45:45.000Z | 2023-11-09T14:45:45 | ---
pretty_name: Evaluation run of CobraMamba/mamba-gpt-7b-v2
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [CobraMamba/mamba-gpt-7b-v2](https://huggingface.co/CobraMamba/mamba-gpt-7b-v2)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 64 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the aggregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_CobraMamba__mamba-gpt-7b-v2_public\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-11-09T14:42:44.506385](https://huggingface.co/datasets/open-llm-leaderboard/details_CobraMamba__mamba-gpt-7b-v2_public/blob/main/results_2023-11-09T14-42-44.506385.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"acc\": 0.6125048552997057,\n\
\ \"acc_stderr\": 0.03288150582791299,\n \"acc_norm\": 0.621215728198735,\n\
\ \"acc_norm_stderr\": 0.03360029488770885,\n \"mc1\": 0.30599755201958384,\n\
\ \"mc1_stderr\": 0.016132229728155045,\n \"mc2\": 0.466285204838536,\n\
\ \"mc2_stderr\": 0.014482857157517471,\n \"em\": 0.2946728187919463,\n\
\ \"em_stderr\": 0.004668797098936446,\n \"f1\": 0.3407151845637583,\n\
\ \"f1_stderr\": 0.004587411171504163\n },\n \"harness|arc:challenge|25\"\
: {\n \"acc\": 0.5750853242320819,\n \"acc_stderr\": 0.014445698968520769,\n\
\ \"acc_norm\": 0.6194539249146758,\n \"acc_norm_stderr\": 0.01418827771234981\n\
\ },\n \"harness|hellaswag|10\": {\n \"acc\": 0.6363274248157738,\n\
\ \"acc_stderr\": 0.004800728138792391,\n \"acc_norm\": 0.8382792272455686,\n\
\ \"acc_norm_stderr\": 0.00367441979935367\n },\n \"harness|hendrycksTest-abstract_algebra|5\"\
: {\n \"acc\": 0.33,\n \"acc_stderr\": 0.04725815626252606,\n \
\ \"acc_norm\": 0.33,\n \"acc_norm_stderr\": 0.04725815626252606\n \
\ },\n \"harness|hendrycksTest-anatomy|5\": {\n \"acc\": 0.5777777777777777,\n\
\ \"acc_stderr\": 0.04266763404099582,\n \"acc_norm\": 0.5777777777777777,\n\
\ \"acc_norm_stderr\": 0.04266763404099582\n },\n \"harness|hendrycksTest-astronomy|5\"\
: {\n \"acc\": 0.631578947368421,\n \"acc_stderr\": 0.03925523381052932,\n\
\ \"acc_norm\": 0.631578947368421,\n \"acc_norm_stderr\": 0.03925523381052932\n\
\ },\n \"harness|hendrycksTest-business_ethics|5\": {\n \"acc\": 0.57,\n\
\ \"acc_stderr\": 0.049756985195624284,\n \"acc_norm\": 0.57,\n \
\ \"acc_norm_stderr\": 0.049756985195624284\n },\n \"harness|hendrycksTest-clinical_knowledge|5\"\
: {\n \"acc\": 0.660377358490566,\n \"acc_stderr\": 0.029146904747798328,\n\
\ \"acc_norm\": 0.660377358490566,\n \"acc_norm_stderr\": 0.029146904747798328\n\
\ },\n \"harness|hendrycksTest-college_biology|5\": {\n \"acc\": 0.7152777777777778,\n\
\ \"acc_stderr\": 0.03773809990686934,\n \"acc_norm\": 0.7152777777777778,\n\
\ \"acc_norm_stderr\": 0.03773809990686934\n },\n \"harness|hendrycksTest-college_chemistry|5\"\
: {\n \"acc\": 0.45,\n \"acc_stderr\": 0.05,\n \"acc_norm\"\
: 0.45,\n \"acc_norm_stderr\": 0.05\n },\n \"harness|hendrycksTest-college_computer_science|5\"\
: {\n \"acc\": 0.51,\n \"acc_stderr\": 0.05024183937956911,\n \
\ \"acc_norm\": 0.51,\n \"acc_norm_stderr\": 0.05024183937956911\n \
\ },\n \"harness|hendrycksTest-college_mathematics|5\": {\n \"acc\": 0.41,\n\
\ \"acc_stderr\": 0.04943110704237102,\n \"acc_norm\": 0.41,\n \
\ \"acc_norm_stderr\": 0.04943110704237102\n },\n \"harness|hendrycksTest-college_medicine|5\"\
: {\n \"acc\": 0.5895953757225434,\n \"acc_stderr\": 0.03750757044895537,\n\
\ \"acc_norm\": 0.5895953757225434,\n \"acc_norm_stderr\": 0.03750757044895537\n\
\ },\n \"harness|hendrycksTest-college_physics|5\": {\n \"acc\": 0.38235294117647056,\n\
\ \"acc_stderr\": 0.04835503696107223,\n \"acc_norm\": 0.38235294117647056,\n\
\ \"acc_norm_stderr\": 0.04835503696107223\n },\n \"harness|hendrycksTest-computer_security|5\"\
: {\n \"acc\": 0.74,\n \"acc_stderr\": 0.0440844002276808,\n \
\ \"acc_norm\": 0.74,\n \"acc_norm_stderr\": 0.0440844002276808\n },\n\
\ \"harness|hendrycksTest-conceptual_physics|5\": {\n \"acc\": 0.5191489361702127,\n\
\ \"acc_stderr\": 0.032662042990646796,\n \"acc_norm\": 0.5191489361702127,\n\
\ \"acc_norm_stderr\": 0.032662042990646796\n },\n \"harness|hendrycksTest-econometrics|5\"\
: {\n \"acc\": 0.49122807017543857,\n \"acc_stderr\": 0.04702880432049615,\n\
\ \"acc_norm\": 0.49122807017543857,\n \"acc_norm_stderr\": 0.04702880432049615\n\
\ },\n \"harness|hendrycksTest-electrical_engineering|5\": {\n \"acc\"\
: 0.5586206896551724,\n \"acc_stderr\": 0.04137931034482758,\n \"\
acc_norm\": 0.5586206896551724,\n \"acc_norm_stderr\": 0.04137931034482758\n\
\ },\n \"harness|hendrycksTest-elementary_mathematics|5\": {\n \"acc\"\
: 0.4074074074074074,\n \"acc_stderr\": 0.025305906241590632,\n \"\
acc_norm\": 0.4074074074074074,\n \"acc_norm_stderr\": 0.025305906241590632\n\
\ },\n \"harness|hendrycksTest-formal_logic|5\": {\n \"acc\": 0.3888888888888889,\n\
\ \"acc_stderr\": 0.04360314860077459,\n \"acc_norm\": 0.3888888888888889,\n\
\ \"acc_norm_stderr\": 0.04360314860077459\n },\n \"harness|hendrycksTest-global_facts|5\"\
: {\n \"acc\": 0.37,\n \"acc_stderr\": 0.048523658709391,\n \
\ \"acc_norm\": 0.37,\n \"acc_norm_stderr\": 0.048523658709391\n },\n\
\ \"harness|hendrycksTest-high_school_biology|5\": {\n \"acc\": 0.7225806451612903,\n\
\ \"acc_stderr\": 0.025470196835900055,\n \"acc_norm\": 0.7225806451612903,\n\
\ \"acc_norm_stderr\": 0.025470196835900055\n },\n \"harness|hendrycksTest-high_school_chemistry|5\"\
: {\n \"acc\": 0.4630541871921182,\n \"acc_stderr\": 0.035083705204426656,\n\
\ \"acc_norm\": 0.4630541871921182,\n \"acc_norm_stderr\": 0.035083705204426656\n\
\ },\n \"harness|hendrycksTest-high_school_computer_science|5\": {\n \
\ \"acc\": 0.66,\n \"acc_stderr\": 0.04760952285695237,\n \"acc_norm\"\
: 0.66,\n \"acc_norm_stderr\": 0.04760952285695237\n },\n \"harness|hendrycksTest-high_school_european_history|5\"\
: {\n \"acc\": 0.7636363636363637,\n \"acc_stderr\": 0.03317505930009181,\n\
\ \"acc_norm\": 0.7636363636363637,\n \"acc_norm_stderr\": 0.03317505930009181\n\
\ },\n \"harness|hendrycksTest-high_school_geography|5\": {\n \"acc\"\
: 0.803030303030303,\n \"acc_stderr\": 0.028335609732463362,\n \"\
acc_norm\": 0.803030303030303,\n \"acc_norm_stderr\": 0.028335609732463362\n\
\ },\n \"harness|hendrycksTest-high_school_government_and_politics|5\": {\n\
\ \"acc\": 0.8186528497409327,\n \"acc_stderr\": 0.02780703236068609,\n\
\ \"acc_norm\": 0.8186528497409327,\n \"acc_norm_stderr\": 0.02780703236068609\n\
\ },\n \"harness|hendrycksTest-high_school_macroeconomics|5\": {\n \
\ \"acc\": 0.6128205128205129,\n \"acc_stderr\": 0.024697216930878937,\n\
\ \"acc_norm\": 0.6128205128205129,\n \"acc_norm_stderr\": 0.024697216930878937\n\
\ },\n \"harness|hendrycksTest-high_school_mathematics|5\": {\n \"\
acc\": 0.34444444444444444,\n \"acc_stderr\": 0.02897264888484427,\n \
\ \"acc_norm\": 0.34444444444444444,\n \"acc_norm_stderr\": 0.02897264888484427\n\
\ },\n \"harness|hendrycksTest-high_school_microeconomics|5\": {\n \
\ \"acc\": 0.6092436974789915,\n \"acc_stderr\": 0.03169380235712997,\n \
\ \"acc_norm\": 0.6092436974789915,\n \"acc_norm_stderr\": 0.03169380235712997\n\
\ },\n \"harness|hendrycksTest-high_school_physics|5\": {\n \"acc\"\
: 0.2980132450331126,\n \"acc_stderr\": 0.037345356767871984,\n \"\
acc_norm\": 0.2980132450331126,\n \"acc_norm_stderr\": 0.037345356767871984\n\
\ },\n \"harness|hendrycksTest-high_school_psychology|5\": {\n \"acc\"\
: 0.7779816513761468,\n \"acc_stderr\": 0.01781884956479664,\n \"\
acc_norm\": 0.7779816513761468,\n \"acc_norm_stderr\": 0.01781884956479664\n\
\ },\n \"harness|hendrycksTest-high_school_statistics|5\": {\n \"acc\"\
: 0.5185185185185185,\n \"acc_stderr\": 0.03407632093854051,\n \"\
acc_norm\": 0.5185185185185185,\n \"acc_norm_stderr\": 0.03407632093854051\n\
\ },\n \"harness|hendrycksTest-high_school_us_history|5\": {\n \"acc\"\
: 0.7450980392156863,\n \"acc_stderr\": 0.03058759135160425,\n \"\
acc_norm\": 0.7450980392156863,\n \"acc_norm_stderr\": 0.03058759135160425\n\
\ },\n \"harness|hendrycksTest-high_school_world_history|5\": {\n \"\
acc\": 0.7763713080168776,\n \"acc_stderr\": 0.027123298205229966,\n \
\ \"acc_norm\": 0.7763713080168776,\n \"acc_norm_stderr\": 0.027123298205229966\n\
\ },\n \"harness|hendrycksTest-human_aging|5\": {\n \"acc\": 0.6636771300448431,\n\
\ \"acc_stderr\": 0.031708824268455,\n \"acc_norm\": 0.6636771300448431,\n\
\ \"acc_norm_stderr\": 0.031708824268455\n },\n \"harness|hendrycksTest-human_sexuality|5\"\
: {\n \"acc\": 0.7480916030534351,\n \"acc_stderr\": 0.03807387116306086,\n\
\ \"acc_norm\": 0.7480916030534351,\n \"acc_norm_stderr\": 0.03807387116306086\n\
\ },\n \"harness|hendrycksTest-international_law|5\": {\n \"acc\":\
\ 0.7851239669421488,\n \"acc_stderr\": 0.03749492448709698,\n \"\
acc_norm\": 0.7851239669421488,\n \"acc_norm_stderr\": 0.03749492448709698\n\
\ },\n \"harness|hendrycksTest-jurisprudence|5\": {\n \"acc\": 0.7314814814814815,\n\
\ \"acc_stderr\": 0.042844679680521934,\n \"acc_norm\": 0.7314814814814815,\n\
\ \"acc_norm_stderr\": 0.042844679680521934\n },\n \"harness|hendrycksTest-logical_fallacies|5\"\
: {\n \"acc\": 0.7361963190184049,\n \"acc_stderr\": 0.03462419931615624,\n\
\ \"acc_norm\": 0.7361963190184049,\n \"acc_norm_stderr\": 0.03462419931615624\n\
\ },\n \"harness|hendrycksTest-machine_learning|5\": {\n \"acc\": 0.4642857142857143,\n\
\ \"acc_stderr\": 0.04733667890053756,\n \"acc_norm\": 0.4642857142857143,\n\
\ \"acc_norm_stderr\": 0.04733667890053756\n },\n \"harness|hendrycksTest-management|5\"\
: {\n \"acc\": 0.7766990291262136,\n \"acc_stderr\": 0.04123553189891431,\n\
\ \"acc_norm\": 0.7766990291262136,\n \"acc_norm_stderr\": 0.04123553189891431\n\
\ },\n \"harness|hendrycksTest-marketing|5\": {\n \"acc\": 0.8547008547008547,\n\
\ \"acc_stderr\": 0.023086635086841407,\n \"acc_norm\": 0.8547008547008547,\n\
\ \"acc_norm_stderr\": 0.023086635086841407\n },\n \"harness|hendrycksTest-medical_genetics|5\"\
: {\n \"acc\": 0.74,\n \"acc_stderr\": 0.04408440022768079,\n \
\ \"acc_norm\": 0.74,\n \"acc_norm_stderr\": 0.04408440022768079\n \
\ },\n \"harness|hendrycksTest-miscellaneous|5\": {\n \"acc\": 0.7918263090676884,\n\
\ \"acc_stderr\": 0.014518592248904033,\n \"acc_norm\": 0.7918263090676884,\n\
\ \"acc_norm_stderr\": 0.014518592248904033\n },\n \"harness|hendrycksTest-moral_disputes|5\"\
: {\n \"acc\": 0.6763005780346821,\n \"acc_stderr\": 0.025190181327608405,\n\
\ \"acc_norm\": 0.6763005780346821,\n \"acc_norm_stderr\": 0.025190181327608405\n\
\ },\n \"harness|hendrycksTest-moral_scenarios|5\": {\n \"acc\": 0.358659217877095,\n\
\ \"acc_stderr\": 0.016040454426164474,\n \"acc_norm\": 0.358659217877095,\n\
\ \"acc_norm_stderr\": 0.016040454426164474\n },\n \"harness|hendrycksTest-nutrition|5\"\
: {\n \"acc\": 0.7254901960784313,\n \"acc_stderr\": 0.025553169991826528,\n\
\ \"acc_norm\": 0.7254901960784313,\n \"acc_norm_stderr\": 0.025553169991826528\n\
\ },\n \"harness|hendrycksTest-philosophy|5\": {\n \"acc\": 0.6913183279742765,\n\
\ \"acc_stderr\": 0.026236965881153266,\n \"acc_norm\": 0.6913183279742765,\n\
\ \"acc_norm_stderr\": 0.026236965881153266\n },\n \"harness|hendrycksTest-prehistory|5\"\
: {\n \"acc\": 0.6604938271604939,\n \"acc_stderr\": 0.02634856441201162,\n\
\ \"acc_norm\": 0.6604938271604939,\n \"acc_norm_stderr\": 0.02634856441201162\n\
\ },\n \"harness|hendrycksTest-professional_accounting|5\": {\n \"\
acc\": 0.4397163120567376,\n \"acc_stderr\": 0.029609912075594106,\n \
\ \"acc_norm\": 0.4397163120567376,\n \"acc_norm_stderr\": 0.029609912075594106\n\
\ },\n \"harness|hendrycksTest-professional_law|5\": {\n \"acc\": 0.44002607561929596,\n\
\ \"acc_stderr\": 0.012678037478574513,\n \"acc_norm\": 0.44002607561929596,\n\
\ \"acc_norm_stderr\": 0.012678037478574513\n },\n \"harness|hendrycksTest-professional_medicine|5\"\
: {\n \"acc\": 0.6397058823529411,\n \"acc_stderr\": 0.02916312857067073,\n\
\ \"acc_norm\": 0.6397058823529411,\n \"acc_norm_stderr\": 0.02916312857067073\n\
\ },\n \"harness|hendrycksTest-professional_psychology|5\": {\n \"\
acc\": 0.6372549019607843,\n \"acc_stderr\": 0.019450768432505514,\n \
\ \"acc_norm\": 0.6372549019607843,\n \"acc_norm_stderr\": 0.019450768432505514\n\
\ },\n \"harness|hendrycksTest-public_relations|5\": {\n \"acc\": 0.6545454545454545,\n\
\ \"acc_stderr\": 0.04554619617541054,\n \"acc_norm\": 0.6545454545454545,\n\
\ \"acc_norm_stderr\": 0.04554619617541054\n },\n \"harness|hendrycksTest-security_studies|5\"\
: {\n \"acc\": 0.6979591836734694,\n \"acc_stderr\": 0.029393609319879804,\n\
\ \"acc_norm\": 0.6979591836734694,\n \"acc_norm_stderr\": 0.029393609319879804\n\
\ },\n \"harness|hendrycksTest-sociology|5\": {\n \"acc\": 0.7960199004975125,\n\
\ \"acc_stderr\": 0.02849317624532607,\n \"acc_norm\": 0.7960199004975125,\n\
\ \"acc_norm_stderr\": 0.02849317624532607\n },\n \"harness|hendrycksTest-us_foreign_policy|5\"\
: {\n \"acc\": 0.86,\n \"acc_stderr\": 0.03487350880197771,\n \
\ \"acc_norm\": 0.86,\n \"acc_norm_stderr\": 0.03487350880197771\n \
\ },\n \"harness|hendrycksTest-virology|5\": {\n \"acc\": 0.5783132530120482,\n\
\ \"acc_stderr\": 0.038444531817709175,\n \"acc_norm\": 0.5783132530120482,\n\
\ \"acc_norm_stderr\": 0.038444531817709175\n },\n \"harness|hendrycksTest-world_religions|5\"\
: {\n \"acc\": 0.8304093567251462,\n \"acc_stderr\": 0.02878210810540171,\n\
\ \"acc_norm\": 0.8304093567251462,\n \"acc_norm_stderr\": 0.02878210810540171\n\
\ },\n \"harness|truthfulqa:mc|0\": {\n \"mc1\": 0.30599755201958384,\n\
\ \"mc1_stderr\": 0.016132229728155045,\n \"mc2\": 0.466285204838536,\n\
\ \"mc2_stderr\": 0.014482857157517471\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.7845303867403315,\n \"acc_stderr\": 0.011555295286059282\n\
\ },\n \"harness|drop|3\": {\n \"em\": 0.2946728187919463,\n \
\ \"em_stderr\": 0.004668797098936446,\n \"f1\": 0.3407151845637583,\n \
\ \"f1_stderr\": 0.004587411171504163\n },\n \"harness|gsm8k|5\": {\n\
\ \"acc\": 0.1728582259287339,\n \"acc_stderr\": 0.010415432246200569\n\
\ }\n}\n```"
repo_url: https://huggingface.co/CobraMamba/mamba-gpt-7b-v2
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|arc:challenge|25_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_drop_3
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|drop|3_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|drop|3_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|gsm8k|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hellaswag|10_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-42-44.506385.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2023-11-09T14-42-44.506385.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- '**/details_harness|winogrande|5_2023-11-09T14-42-44.506385.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-11-09T14-42-44.506385.parquet'
- config_name: results
data_files:
- split: 2023_11_09T14_42_44.506385
path:
- results_2023-11-09T14-42-44.506385.parquet
- split: latest
path:
- results_2023-11-09T14-42-44.506385.parquet
---
# Dataset Card for Evaluation run of CobraMamba/mamba-gpt-7b-v2
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/CobraMamba/mamba-gpt-7b-v2
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [CobraMamba/mamba-gpt-7b-v2](https://huggingface.co/CobraMamba/mamba-gpt-7b-v2) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 64 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_CobraMamba__mamba-gpt-7b-v2_public",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-11-09T14:42:44.506385](https://huggingface.co/datasets/open-llm-leaderboard/details_CobraMamba__mamba-gpt-7b-v2_public/blob/main/results_2023-11-09T14-42-44.506385.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"acc": 0.6125048552997057,
"acc_stderr": 0.03288150582791299,
"acc_norm": 0.621215728198735,
"acc_norm_stderr": 0.03360029488770885,
"mc1": 0.30599755201958384,
"mc1_stderr": 0.016132229728155045,
"mc2": 0.466285204838536,
"mc2_stderr": 0.014482857157517471,
"em": 0.2946728187919463,
"em_stderr": 0.004668797098936446,
"f1": 0.3407151845637583,
"f1_stderr": 0.004587411171504163
},
"harness|arc:challenge|25": {
"acc": 0.5750853242320819,
"acc_stderr": 0.014445698968520769,
"acc_norm": 0.6194539249146758,
"acc_norm_stderr": 0.01418827771234981
},
"harness|hellaswag|10": {
"acc": 0.6363274248157738,
"acc_stderr": 0.004800728138792391,
"acc_norm": 0.8382792272455686,
"acc_norm_stderr": 0.00367441979935367
},
"harness|hendrycksTest-abstract_algebra|5": {
"acc": 0.33,
"acc_stderr": 0.04725815626252606,
"acc_norm": 0.33,
"acc_norm_stderr": 0.04725815626252606
},
"harness|hendrycksTest-anatomy|5": {
"acc": 0.5777777777777777,
"acc_stderr": 0.04266763404099582,
"acc_norm": 0.5777777777777777,
"acc_norm_stderr": 0.04266763404099582
},
"harness|hendrycksTest-astronomy|5": {
"acc": 0.631578947368421,
"acc_stderr": 0.03925523381052932,
"acc_norm": 0.631578947368421,
"acc_norm_stderr": 0.03925523381052932
},
"harness|hendrycksTest-business_ethics|5": {
"acc": 0.57,
"acc_stderr": 0.049756985195624284,
"acc_norm": 0.57,
"acc_norm_stderr": 0.049756985195624284
},
"harness|hendrycksTest-clinical_knowledge|5": {
"acc": 0.660377358490566,
"acc_stderr": 0.029146904747798328,
"acc_norm": 0.660377358490566,
"acc_norm_stderr": 0.029146904747798328
},
"harness|hendrycksTest-college_biology|5": {
"acc": 0.7152777777777778,
"acc_stderr": 0.03773809990686934,
"acc_norm": 0.7152777777777778,
"acc_norm_stderr": 0.03773809990686934
},
"harness|hendrycksTest-college_chemistry|5": {
"acc": 0.45,
"acc_stderr": 0.05,
"acc_norm": 0.45,
"acc_norm_stderr": 0.05
},
"harness|hendrycksTest-college_computer_science|5": {
"acc": 0.51,
"acc_stderr": 0.05024183937956911,
"acc_norm": 0.51,
"acc_norm_stderr": 0.05024183937956911
},
"harness|hendrycksTest-college_mathematics|5": {
"acc": 0.41,
"acc_stderr": 0.04943110704237102,
"acc_norm": 0.41,
"acc_norm_stderr": 0.04943110704237102
},
"harness|hendrycksTest-college_medicine|5": {
"acc": 0.5895953757225434,
"acc_stderr": 0.03750757044895537,
"acc_norm": 0.5895953757225434,
"acc_norm_stderr": 0.03750757044895537
},
"harness|hendrycksTest-college_physics|5": {
"acc": 0.38235294117647056,
"acc_stderr": 0.04835503696107223,
"acc_norm": 0.38235294117647056,
"acc_norm_stderr": 0.04835503696107223
},
"harness|hendrycksTest-computer_security|5": {
"acc": 0.74,
"acc_stderr": 0.0440844002276808,
"acc_norm": 0.74,
"acc_norm_stderr": 0.0440844002276808
},
"harness|hendrycksTest-conceptual_physics|5": {
"acc": 0.5191489361702127,
"acc_stderr": 0.032662042990646796,
"acc_norm": 0.5191489361702127,
"acc_norm_stderr": 0.032662042990646796
},
"harness|hendrycksTest-econometrics|5": {
"acc": 0.49122807017543857,
"acc_stderr": 0.04702880432049615,
"acc_norm": 0.49122807017543857,
"acc_norm_stderr": 0.04702880432049615
},
"harness|hendrycksTest-electrical_engineering|5": {
"acc": 0.5586206896551724,
"acc_stderr": 0.04137931034482758,
"acc_norm": 0.5586206896551724,
"acc_norm_stderr": 0.04137931034482758
},
"harness|hendrycksTest-elementary_mathematics|5": {
"acc": 0.4074074074074074,
"acc_stderr": 0.025305906241590632,
"acc_norm": 0.4074074074074074,
"acc_norm_stderr": 0.025305906241590632
},
"harness|hendrycksTest-formal_logic|5": {
"acc": 0.3888888888888889,
"acc_stderr": 0.04360314860077459,
"acc_norm": 0.3888888888888889,
"acc_norm_stderr": 0.04360314860077459
},
"harness|hendrycksTest-global_facts|5": {
"acc": 0.37,
"acc_stderr": 0.048523658709391,
"acc_norm": 0.37,
"acc_norm_stderr": 0.048523658709391
},
"harness|hendrycksTest-high_school_biology|5": {
"acc": 0.7225806451612903,
"acc_stderr": 0.025470196835900055,
"acc_norm": 0.7225806451612903,
"acc_norm_stderr": 0.025470196835900055
},
"harness|hendrycksTest-high_school_chemistry|5": {
"acc": 0.4630541871921182,
"acc_stderr": 0.035083705204426656,
"acc_norm": 0.4630541871921182,
"acc_norm_stderr": 0.035083705204426656
},
"harness|hendrycksTest-high_school_computer_science|5": {
"acc": 0.66,
"acc_stderr": 0.04760952285695237,
"acc_norm": 0.66,
"acc_norm_stderr": 0.04760952285695237
},
"harness|hendrycksTest-high_school_european_history|5": {
"acc": 0.7636363636363637,
"acc_stderr": 0.03317505930009181,
"acc_norm": 0.7636363636363637,
"acc_norm_stderr": 0.03317505930009181
},
"harness|hendrycksTest-high_school_geography|5": {
"acc": 0.803030303030303,
"acc_stderr": 0.028335609732463362,
"acc_norm": 0.803030303030303,
"acc_norm_stderr": 0.028335609732463362
},
"harness|hendrycksTest-high_school_government_and_politics|5": {
"acc": 0.8186528497409327,
"acc_stderr": 0.02780703236068609,
"acc_norm": 0.8186528497409327,
"acc_norm_stderr": 0.02780703236068609
},
"harness|hendrycksTest-high_school_macroeconomics|5": {
"acc": 0.6128205128205129,
"acc_stderr": 0.024697216930878937,
"acc_norm": 0.6128205128205129,
"acc_norm_stderr": 0.024697216930878937
},
"harness|hendrycksTest-high_school_mathematics|5": {
"acc": 0.34444444444444444,
"acc_stderr": 0.02897264888484427,
"acc_norm": 0.34444444444444444,
"acc_norm_stderr": 0.02897264888484427
},
"harness|hendrycksTest-high_school_microeconomics|5": {
"acc": 0.6092436974789915,
"acc_stderr": 0.03169380235712997,
"acc_norm": 0.6092436974789915,
"acc_norm_stderr": 0.03169380235712997
},
"harness|hendrycksTest-high_school_physics|5": {
"acc": 0.2980132450331126,
"acc_stderr": 0.037345356767871984,
"acc_norm": 0.2980132450331126,
"acc_norm_stderr": 0.037345356767871984
},
"harness|hendrycksTest-high_school_psychology|5": {
"acc": 0.7779816513761468,
"acc_stderr": 0.01781884956479664,
"acc_norm": 0.7779816513761468,
"acc_norm_stderr": 0.01781884956479664
},
"harness|hendrycksTest-high_school_statistics|5": {
"acc": 0.5185185185185185,
"acc_stderr": 0.03407632093854051,
"acc_norm": 0.5185185185185185,
"acc_norm_stderr": 0.03407632093854051
},
"harness|hendrycksTest-high_school_us_history|5": {
"acc": 0.7450980392156863,
"acc_stderr": 0.03058759135160425,
"acc_norm": 0.7450980392156863,
"acc_norm_stderr": 0.03058759135160425
},
"harness|hendrycksTest-high_school_world_history|5": {
"acc": 0.7763713080168776,
"acc_stderr": 0.027123298205229966,
"acc_norm": 0.7763713080168776,
"acc_norm_stderr": 0.027123298205229966
},
"harness|hendrycksTest-human_aging|5": {
"acc": 0.6636771300448431,
"acc_stderr": 0.031708824268455,
"acc_norm": 0.6636771300448431,
"acc_norm_stderr": 0.031708824268455
},
"harness|hendrycksTest-human_sexuality|5": {
"acc": 0.7480916030534351,
"acc_stderr": 0.03807387116306086,
"acc_norm": 0.7480916030534351,
"acc_norm_stderr": 0.03807387116306086
},
"harness|hendrycksTest-international_law|5": {
"acc": 0.7851239669421488,
"acc_stderr": 0.03749492448709698,
"acc_norm": 0.7851239669421488,
"acc_norm_stderr": 0.03749492448709698
},
"harness|hendrycksTest-jurisprudence|5": {
"acc": 0.7314814814814815,
"acc_stderr": 0.042844679680521934,
"acc_norm": 0.7314814814814815,
"acc_norm_stderr": 0.042844679680521934
},
"harness|hendrycksTest-logical_fallacies|5": {
"acc": 0.7361963190184049,
"acc_stderr": 0.03462419931615624,
"acc_norm": 0.7361963190184049,
"acc_norm_stderr": 0.03462419931615624
},
"harness|hendrycksTest-machine_learning|5": {
"acc": 0.4642857142857143,
"acc_stderr": 0.04733667890053756,
"acc_norm": 0.4642857142857143,
"acc_norm_stderr": 0.04733667890053756
},
"harness|hendrycksTest-management|5": {
"acc": 0.7766990291262136,
"acc_stderr": 0.04123553189891431,
"acc_norm": 0.7766990291262136,
"acc_norm_stderr": 0.04123553189891431
},
"harness|hendrycksTest-marketing|5": {
"acc": 0.8547008547008547,
"acc_stderr": 0.023086635086841407,
"acc_norm": 0.8547008547008547,
"acc_norm_stderr": 0.023086635086841407
},
"harness|hendrycksTest-medical_genetics|5": {
"acc": 0.74,
"acc_stderr": 0.04408440022768079,
"acc_norm": 0.74,
"acc_norm_stderr": 0.04408440022768079
},
"harness|hendrycksTest-miscellaneous|5": {
"acc": 0.7918263090676884,
"acc_stderr": 0.014518592248904033,
"acc_norm": 0.7918263090676884,
"acc_norm_stderr": 0.014518592248904033
},
"harness|hendrycksTest-moral_disputes|5": {
"acc": 0.6763005780346821,
"acc_stderr": 0.025190181327608405,
"acc_norm": 0.6763005780346821,
"acc_norm_stderr": 0.025190181327608405
},
"harness|hendrycksTest-moral_scenarios|5": {
"acc": 0.358659217877095,
"acc_stderr": 0.016040454426164474,
"acc_norm": 0.358659217877095,
"acc_norm_stderr": 0.016040454426164474
},
"harness|hendrycksTest-nutrition|5": {
"acc": 0.7254901960784313,
"acc_stderr": 0.025553169991826528,
"acc_norm": 0.7254901960784313,
"acc_norm_stderr": 0.025553169991826528
},
"harness|hendrycksTest-philosophy|5": {
"acc": 0.6913183279742765,
"acc_stderr": 0.026236965881153266,
"acc_norm": 0.6913183279742765,
"acc_norm_stderr": 0.026236965881153266
},
"harness|hendrycksTest-prehistory|5": {
"acc": 0.6604938271604939,
"acc_stderr": 0.02634856441201162,
"acc_norm": 0.6604938271604939,
"acc_norm_stderr": 0.02634856441201162
},
"harness|hendrycksTest-professional_accounting|5": {
"acc": 0.4397163120567376,
"acc_stderr": 0.029609912075594106,
"acc_norm": 0.4397163120567376,
"acc_norm_stderr": 0.029609912075594106
},
"harness|hendrycksTest-professional_law|5": {
"acc": 0.44002607561929596,
"acc_stderr": 0.012678037478574513,
"acc_norm": 0.44002607561929596,
"acc_norm_stderr": 0.012678037478574513
},
"harness|hendrycksTest-professional_medicine|5": {
"acc": 0.6397058823529411,
"acc_stderr": 0.02916312857067073,
"acc_norm": 0.6397058823529411,
"acc_norm_stderr": 0.02916312857067073
},
"harness|hendrycksTest-professional_psychology|5": {
"acc": 0.6372549019607843,
"acc_stderr": 0.019450768432505514,
"acc_norm": 0.6372549019607843,
"acc_norm_stderr": 0.019450768432505514
},
"harness|hendrycksTest-public_relations|5": {
"acc": 0.6545454545454545,
"acc_stderr": 0.04554619617541054,
"acc_norm": 0.6545454545454545,
"acc_norm_stderr": 0.04554619617541054
},
"harness|hendrycksTest-security_studies|5": {
"acc": 0.6979591836734694,
"acc_stderr": 0.029393609319879804,
"acc_norm": 0.6979591836734694,
"acc_norm_stderr": 0.029393609319879804
},
"harness|hendrycksTest-sociology|5": {
"acc": 0.7960199004975125,
"acc_stderr": 0.02849317624532607,
"acc_norm": 0.7960199004975125,
"acc_norm_stderr": 0.02849317624532607
},
"harness|hendrycksTest-us_foreign_policy|5": {
"acc": 0.86,
"acc_stderr": 0.03487350880197771,
"acc_norm": 0.86,
"acc_norm_stderr": 0.03487350880197771
},
"harness|hendrycksTest-virology|5": {
"acc": 0.5783132530120482,
"acc_stderr": 0.038444531817709175,
"acc_norm": 0.5783132530120482,
"acc_norm_stderr": 0.038444531817709175
},
"harness|hendrycksTest-world_religions|5": {
"acc": 0.8304093567251462,
"acc_stderr": 0.02878210810540171,
"acc_norm": 0.8304093567251462,
"acc_norm_stderr": 0.02878210810540171
},
"harness|truthfulqa:mc|0": {
"mc1": 0.30599755201958384,
"mc1_stderr": 0.016132229728155045,
"mc2": 0.466285204838536,
"mc2_stderr": 0.014482857157517471
},
"harness|winogrande|5": {
"acc": 0.7845303867403315,
"acc_stderr": 0.011555295286059282
},
"harness|drop|3": {
"em": 0.2946728187919463,
"em_stderr": 0.004668797098936446,
"f1": 0.3407151845637583,
"f1_stderr": 0.004587411171504163
},
"harness|gsm8k|5": {
"acc": 0.1728582259287339,
"acc_stderr": 0.010415432246200569
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] | [
-0.7607290744781494,
-0.8917080760002136,
0.2466440200805664,
0.1831502616405487,
-0.2506023645401001,
-0.0446934811770916,
0.06420076638460159,
-0.19955208897590637,
0.5959251523017883,
-0.0570576973259449,
-0.4731307923793793,
-0.689109206199646,
-0.4778085947036743,
0.2227730005979538,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
efkanozkan41/captcha-devex | efkanozkan41 | 2023-11-09T14:48:13Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T14:48:13Z | 2023-11-09T14:48:13.000Z | 2023-11-09T14:48:13 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Back-up/qa-temp | Back-up | 2023-11-09T14:51:20Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T14:51:20Z | 2023-11-09T14:51:05.000Z | 2023-11-09T14:51:05 | ---
dataset_info:
features:
- name: id
dtype: string
- name: title
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: response
struct:
- name: response
dtype: string
- name: answers
struct:
- name: answer_start
sequence: int64
- name: text
sequence: string
- name: instruction
dtype: string
- name: prompt_name
dtype: string
splits:
- name: train
num_bytes: 276860
num_examples: 101
download_size: 144679
dataset_size: 276860
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "qa-temp"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.520023763179779,
-0.15336845815181732,
0.3803879916667938,
0.035392653197050095,
-0.40651199221611023,
0.0452619232237339,
0.44345971941947937,
-0.014189035631716251,
0.8905845284461975,
0.3408973515033722,
-0.6927614808082581,
-0.7631391286849976,
-0.32362887263298035,
-0.2754004001617... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
male-2/training_v2-public | male-2 | 2023-11-09T14:55:26Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T14:55:26Z | 2023-11-09T14:55:23.000Z | 2023-11-09T14:55:23 | ---
dataset_info:
features:
- name: conversation
dtype: string
- name: type
dtype: string
splits:
- name: train
num_bytes: 1091
num_examples: 1
download_size: 8505
dataset_size: 1091
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "training_v2-public"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.4464028477668762,
0.011462578549981117,
0.09084419906139374,
0.34425869584083557,
-0.1631014347076416,
-0.12800516188144684,
0.25826704502105713,
-0.23381195962429047,
0.5693694353103638,
0.4296196699142456,
-0.7658221125602722,
-0.729041337966919,
-0.7393747568130493,
-0.25943592190742... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
arieg/bw_spec_cls_80_27 | arieg | 2023-11-09T14:57:14Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T14:57:14Z | 2023-11-09T14:56:58.000Z | 2023-11-09T14:56:58 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': '63804'
'1': '63805'
'2': '63874'
'3': '63900'
'4': '63908'
'5': '63922'
'6': '63936'
'7': '63999'
'8': '64005'
'9': '64006'
'10': '64007'
'11': '64008'
'12': '64009'
'13': '64035'
'14': '64078'
'15': '64091'
'16': '64093'
'17': '64247'
'18': '64248'
'19': '64249'
'20': '64252'
'21': '64253'
'22': '64331'
'23': '64332'
'24': '64333'
'25': '64334'
'26': '64338'
'27': '64364'
'28': '64365'
'29': '64366'
'30': '64407'
'31': '64409'
'32': '64410'
'33': '64535'
'34': '64536'
'35': '64537'
'36': '64538'
'37': '64542'
'38': '64553'
'39': '64556'
'40': '64567'
'41': '64594'
'42': '64601'
'43': '64604'
'44': '64659'
'45': '64787'
'46': '64788'
'47': '64789'
'48': '64796'
'49': '64809'
'50': '64834'
'51': '64840'
'52': '64841'
'53': '64854'
'54': '64855'
'55': '64856'
'56': '64857'
'57': '64858'
'58': '64859'
'59': '64860'
'60': '64861'
'61': '64862'
'62': '64863'
'63': '64864'
'64': '64865'
'65': '64866'
'66': '64893'
'67': '64895'
'68': '64896'
'69': '64918'
'70': '64919'
'71': '64988'
'72': '64989'
'73': '64990'
'74': '64991'
'75': '64992'
'76': '64993'
'77': '64994'
'78': '64995'
'79': '65063'
splits:
- name: train
num_bytes: 88333251.2
num_examples: 1600
- name: test
num_bytes: 22046259.0
num_examples: 400
download_size: 110321369
dataset_size: 110379510.2
---
# Dataset Card for "bw_spec_cls_80_27"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.7769212126731873,
-0.14979401230812073,
0.1627131700515747,
0.4082741141319275,
-0.26394423842430115,
-0.16657009720802307,
0.016960367560386658,
-0.26151230931282043,
0.6047933101654053,
0.5480143427848816,
-0.7970455288887024,
-0.8018385767936707,
-0.5971016883850098,
-0.1936725229024... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Back-up/qa-with-answer | Back-up | 2023-11-09T15:00:00Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T15:00:00Z | 2023-11-09T14:59:58.000Z | 2023-11-09T14:59:58 | ---
dataset_info:
features:
- name: id
dtype: string
- name: title
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
struct:
- name: answer_start
sequence: int64
- name: text
sequence: string
- name: is_impossible
dtype: bool
splits:
- name: train
num_bytes: 24536185.07080859
num_examples: 19240
download_size: 4197812
dataset_size: 24536185.07080859
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "qa-with-answer"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.6821112632751465,
-0.40649425983428955,
0.37745338678359985,
0.12727604806423187,
-0.21328866481781006,
0.11401280760765076,
0.4574277997016907,
-0.195969820022583,
0.9819688200950623,
0.33685192465782166,
-0.7867239117622375,
-0.5050672888755798,
-0.41391924023628235,
-0.24368920922279... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Back-up/qa-no-answer | Back-up | 2023-11-09T15:00:03Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T15:00:03Z | 2023-11-09T15:00:01.000Z | 2023-11-09T15:00:01 | ---
dataset_info:
features:
- name: id
dtype: string
- name: title
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
struct:
- name: answer_start
sequence: int64
- name: text
sequence: string
- name: is_impossible
dtype: bool
splits:
- name: train
num_bytes: 11754158.929191412
num_examples: 9217
download_size: 2677376
dataset_size: 11754158.929191412
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "qa-no-answer"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.6486272215843201,
-0.4847179055213928,
0.4794023633003235,
0.1136813685297966,
-0.25101780891418457,
0.02397979609668255,
0.596357524394989,
0.018824543803930283,
1.0305471420288086,
0.5181039571762085,
-0.9143060445785522,
-0.6634995937347412,
-0.30541354417800903,
-0.12370362132787704... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
03jshaye/test | 03jshaye | 2023-11-09T15:01:59Z | 0 | 0 | null | [
"license:unknown",
"region:us"
] | 2023-11-09T15:01:59Z | 2023-11-09T15:01:58.000Z | 2023-11-09T15:01:58 | ---
license: unknown
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
arieg/bw_spec_cls_80_28 | arieg | 2023-11-09T15:23:48Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T15:23:48Z | 2023-11-09T15:23:31.000Z | 2023-11-09T15:23:31 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': '65064'
'1': '65073'
'2': '65076'
'3': '65077'
'4': '65090'
'5': '65234'
'6': '65488'
'7': '65619'
'8': '65685'
'9': '65752'
'10': '65755'
'11': '65756'
'12': '65893'
'13': '66058'
'14': '66073'
'15': '66074'
'16': '66075'
'17': '66076'
'18': '66180'
'19': '66187'
'20': '66390'
'21': '66405'
'22': '66469'
'23': '66482'
'24': '66483'
'25': '66525'
'26': '66636'
'27': '66637'
'28': '66638'
'29': '66641'
'30': '66643'
'31': '66644'
'32': '66646'
'33': '66648'
'34': '66649'
'35': '66650'
'36': '66757'
'37': '67007'
'38': '67010'
'39': '67011'
'40': '67016'
'41': '67017'
'42': '67121'
'43': '67163'
'44': '67232'
'45': '67233'
'46': '67235'
'47': '67308'
'48': '67357'
'49': '67358'
'50': '67359'
'51': '67360'
'52': '67361'
'53': '67362'
'54': '67363'
'55': '67366'
'56': '67367'
'57': '67368'
'58': '67412'
'59': '67470'
'60': '67500'
'61': '67553'
'62': '67556'
'63': '67557'
'64': '67558'
'65': '67597'
'66': '67598'
'67': '67600'
'68': '67637'
'69': '67639'
'70': '67640'
'71': '67660'
'72': '67661'
'73': '67673'
'74': '67707'
'75': '67760'
'76': '67763'
'77': '67764'
'78': '67765'
'79': '67766'
splits:
- name: train
num_bytes: 87471356.8
num_examples: 1600
- name: test
num_bytes: 21888454.0
num_examples: 400
download_size: 109587336
dataset_size: 109359810.8
---
# Dataset Card for "bw_spec_cls_80_28"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.807960033416748,
-0.1538601666688919,
0.16525867581367493,
0.3710475265979767,
-0.31234225630760193,
-0.16121430695056915,
0.047746576368808746,
-0.28589075803756714,
0.5521849989891052,
0.596014142036438,
-0.7638887166976929,
-0.8282160758972168,
-0.5956875681877136,
-0.179379358887672... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
materials-toolkits/oqmd | materials-toolkits | 2023-11-09T15:38:10Z | 0 | 0 | null | [
"license:cc-by-4.0",
"region:us"
] | 2023-11-09T15:38:10Z | 2023-11-09T15:38:10.000Z | 2023-11-09T15:38:10 | ---
license: cc-by-4.0
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
arieg/bw_spec_cls_80_29 | arieg | 2023-11-09T15:49:08Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T15:49:08Z | 2023-11-09T15:48:51.000Z | 2023-11-09T15:48:51 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': '67784'
'1': '67793'
'2': '67829'
'3': '68353'
'4': '68354'
'5': '68355'
'6': '68356'
'7': '68407'
'8': '68410'
'9': '68444'
'10': '68531'
'11': '68539'
'12': '68540'
'13': '68541'
'14': '68543'
'15': '68549'
'16': '68573'
'17': '68579'
'18': '68592'
'19': '68600'
'20': '68601'
'21': '68680'
'22': '68682'
'23': '68683'
'24': '68820'
'25': '68821'
'26': '68837'
'27': '68838'
'28': '68839'
'29': '68840'
'30': '68841'
'31': '68842'
'32': '68843'
'33': '68844'
'34': '68851'
'35': '68852'
'36': '68853'
'37': '68854'
'38': '68860'
'39': '68861'
'40': '68862'
'41': '68869'
'42': '68872'
'43': '68875'
'44': '69001'
'45': '69002'
'46': '69170'
'47': '69181'
'48': '69182'
'49': '69188'
'50': '69193'
'51': '69195'
'52': '69196'
'53': '69197'
'54': '69198'
'55': '69199'
'56': '69200'
'57': '69201'
'58': '69202'
'59': '69203'
'60': '69204'
'61': '69205'
'62': '69206'
'63': '69207'
'64': '69208'
'65': '69209'
'66': '69210'
'67': '69211'
'68': '69554'
'69': '69555'
'70': '69561'
'71': '69563'
'72': '69564'
'73': '69567'
'74': '69682'
'75': '69723'
'76': '69726'
'77': '69727'
'78': '69732'
'79': '69744'
splits:
- name: train
num_bytes: 88025524.8
num_examples: 1600
- name: test
num_bytes: 21927703.0
num_examples: 400
download_size: 109110671
dataset_size: 109953227.8
---
# Dataset Card for "bw_spec_cls_80_29"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.7508549690246582,
-0.14780616760253906,
0.15002481639385223,
0.4053829610347748,
-0.2827664017677307,
-0.0919933170080185,
0.008720614947378635,
-0.2620096802711487,
0.5750346183776855,
0.5123918056488037,
-0.7236824631690979,
-0.800897479057312,
-0.5918912291526794,
-0.2118739634752273... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
nlplabtdtu/Extractive-QA-type-2 | nlplabtdtu | 2023-11-09T15:49:04Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T15:49:04Z | 2023-11-09T15:49:03.000Z | 2023-11-09T15:49:03 | ---
dataset_info:
features:
- name: id
dtype: string
- name: title
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
struct:
- name: answer_start
sequence: int64
- name: text
sequence: string
- name: is_impossible
dtype: bool
- name: instruction
dtype: string
- name: prompt_name
dtype: string
splits:
- name: train
num_bytes: 22373994
num_examples: 9217
download_size: 5375276
dataset_size: 22373994
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "Extractive-QA-type-2"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.29088324308395386,
-0.391913503408432,
0.21470573544502258,
0.2555055022239685,
-0.3461480438709259,
0.10215668380260468,
0.49164247512817383,
-0.25752484798431396,
0.7082347273826599,
0.3449534475803375,
-0.5670895576477051,
-0.5522832870483398,
-0.7032645344734192,
-0.2418133020401001... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
kmichiru/NikaidoChat | kmichiru | 2023-11-09T15:57:02Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T15:57:02Z | 2023-11-09T15:51:29.000Z | 2023-11-09T15:51:29 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
nhuquynh10/fever | nhuquynh10 | 2023-11-09T15:56:31Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T15:56:31Z | 2023-11-09T15:56:31.000Z | 2023-11-09T15:56:31 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
pradeeptac/RSFinetuning | pradeeptac | 2023-11-09T15:59:17Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T15:59:17Z | 2023-11-09T15:59:17.000Z | 2023-11-09T15:59:17 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
snake8c/test | snake8c | 2023-11-09T16:27:49Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T16:27:49Z | 2023-11-09T16:27:49.000Z | 2023-11-09T16:27:49 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
principeroxido/data | principeroxido | 2023-11-09T16:34:46Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T16:34:46Z | 2023-11-09T16:34:46.000Z | 2023-11-09T16:34:46 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Vatsalyarajta/Nehaal | Vatsalyarajta | 2023-11-09T16:59:25Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T16:59:25Z | 2023-11-09T16:40:01.000Z | 2023-11-09T16:40:01 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
arieg/bw_spec_cls_80_31 | arieg | 2023-11-09T16:40:54Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T16:40:54Z | 2023-11-09T16:40:38.000Z | 2023-11-09T16:40:38 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': '71509'
'1': '71510'
'2': '71511'
'3': '71512'
'4': '71513'
'5': '71514'
'6': '71515'
'7': '71516'
'8': '71617'
'9': '71620'
'10': '71622'
'11': '71690'
'12': '71691'
'13': '71692'
'14': '71693'
'15': '71694'
'16': '71695'
'17': '71709'
'18': '71714'
'19': '71715'
'20': '71719'
'21': '71721'
'22': '71822'
'23': '71884'
'24': '71885'
'25': '71937'
'26': '71938'
'27': '72046'
'28': '72047'
'29': '72050'
'30': '72056'
'31': '72058'
'32': '72059'
'33': '72064'
'34': '72067'
'35': '72068'
'36': '72069'
'37': '72070'
'38': '72071'
'39': '72072'
'40': '72073'
'41': '72074'
'42': '72075'
'43': '72076'
'44': '72129'
'45': '72130'
'46': '72131'
'47': '72134'
'48': '72135'
'49': '72136'
'50': '72146'
'51': '72149'
'52': '72200'
'53': '72206'
'54': '72210'
'55': '72215'
'56': '72232'
'57': '72233'
'58': '72234'
'59': '72287'
'60': '72288'
'61': '72289'
'62': '72290'
'63': '72456'
'64': '72468'
'65': '72476'
'66': '72477'
'67': '72562'
'68': '72565'
'69': '72570'
'70': '72604'
'71': '72605'
'72': '72607'
'73': '72612'
'74': '72738'
'75': '72781'
'76': '72782'
'77': '72783'
'78': '72784'
'79': '72785'
splits:
- name: train
num_bytes: 85390192.0
num_examples: 1600
- name: test
num_bytes: 21745201.0
num_examples: 400
download_size: 108634813
dataset_size: 107135393.0
---
# Dataset Card for "bw_spec_cls_80_31"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.7832074165344238,
-0.06951367110013962,
0.16644462943077087,
0.36505526304244995,
-0.2941475212574005,
-0.13488245010375977,
0.008954070508480072,
-0.2966289818286896,
0.5896204710006714,
0.48584163188934326,
-0.7720692753791809,
-0.8219717741012573,
-0.5973066687583923,
-0.186532884836... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
mtkinit/short_slovak_sentiment | mtkinit | 2023-11-09T16:41:02Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T16:41:02Z | 2023-11-09T16:40:46.000Z | 2023-11-09T16:40:46 | ---
pretty_name: short-slovak-sentiment
---
# short-slovak-sentiment
Created from AIOD platform | [
-0.3831780254840851,
-0.32680150866508484,
0.3124564290046692,
0.8976507782936096,
-0.9162772297859192,
0.22899429500102997,
0.004519171547144651,
-0.12118010967969894,
1.1899895668029785,
0.5659729242324829,
-0.5348101854324341,
-0.5957553386688232,
-0.5647960901260376,
0.0933953151106834... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
mtkinit/another_short_slovak_dataset | mtkinit | 2023-11-09T16:42:09Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T16:42:09Z | 2023-11-09T16:42:09.000Z | 2023-11-09T16:42:09 | ---
pretty_name: another-short-slovak-dataset
---
# another-short-slovak-dataset
Created from AIOD platform | [
-0.3878781199455261,
-0.21948441863059998,
0.2728722393512726,
0.33139491081237793,
-0.5881621241569519,
0.12121076881885529,
0.03041936457157135,
-0.10407960414886475,
1.0201528072357178,
0.9148393273353577,
-0.7846208810806274,
-0.8642552495002747,
-0.4899473488330841,
-0.062748536467552... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Wikit/nlu-covid | Wikit | 2023-11-09T17:04:37Z | 0 | 0 | null | [
"task_categories:text-classification",
"language:fr",
"license:apache-2.0",
"region:us"
] | 2023-11-09T17:04:37Z | 2023-11-09T16:48:46.000Z | 2023-11-09T16:48:46 | ---
license: apache-2.0
task_categories:
- text-classification
language:
- fr
---
French benchmark of NLU services for employee support use case during covid-19 pandemic.
These datasets were created by the Wikit team in order to compare the performances of NLU tools on the French language.
The dataset use case is employee support during the covid 19 pandemic. The intents were defined to answer department employees' questions on the evolution of work conditions related to the crisis.
- The training_dataset.csv file contains training utterances with associated intent used to train NLU services.
- The test_dataset.csv file contains test utterances with associated intent used to test NLU services.
To use this work, please cite :
> Marion Schaeffer, Christophe Bouvard. Comparaison des solutions de NLU sur un corpus français pour un chatbot de support COVID-19. IC 2022 - Journées francophones d’Ingénierie des Connaissances, Plate-Forme Intelligence Artificielle (PFIA'22), Jun 2022, Saint-Etienne, France. pp.199-208. ⟨hal-03727958⟩
| [
-0.46289899945259094,
-0.4392467141151428,
-0.08971690386533737,
0.6649605631828308,
0.10902749747037888,
-0.12125244736671448,
-0.18278397619724274,
-0.4093570411205292,
0.2308795303106308,
0.5135073661804199,
-0.6690905690193176,
-0.07151897251605988,
-0.24245712161064148,
0.567428648471... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
arieg/bw_spec_cls_80_32 | arieg | 2023-11-09T17:06:39Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T17:06:39Z | 2023-11-09T17:06:24.000Z | 2023-11-09T17:06:24 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': '72786'
'1': '72787'
'2': '72788'
'3': '72789'
'4': '72790'
'5': '72926'
'6': '72927'
'7': '72928'
'8': '72930'
'9': '73099'
'10': '73100'
'11': '73123'
'12': '73124'
'13': '73125'
'14': '73169'
'15': '73170'
'16': '73171'
'17': '73172'
'18': '73174'
'19': '73175'
'20': '73192'
'21': '73193'
'22': '73306'
'23': '73309'
'24': '73318'
'25': '73335'
'26': '73340'
'27': '73341'
'28': '73342'
'29': '73343'
'30': '73344'
'31': '73363'
'32': '73365'
'33': '73366'
'34': '73367'
'35': '73368'
'36': '73369'
'37': '73370'
'38': '73371'
'39': '73372'
'40': '73465'
'41': '73466'
'42': '73467'
'43': '73468'
'44': '73469'
'45': '73486'
'46': '73495'
'47': '73550'
'48': '73551'
'49': '73566'
'50': '73568'
'51': '73572'
'52': '73573'
'53': '73580'
'54': '73584'
'55': '73585'
'56': '73587'
'57': '73658'
'58': '73675'
'59': '73760'
'60': '73761'
'61': '73762'
'62': '73764'
'63': '73765'
'64': '73766'
'65': '73767'
'66': '73768'
'67': '73769'
'68': '73770'
'69': '73771'
'70': '73772'
'71': '73774'
'72': '73778'
'73': '73792'
'74': '73797'
'75': '73819'
'76': '73820'
'77': '73821'
'78': '73822'
'79': '73921'
splits:
- name: train
num_bytes: 85147582.4
num_examples: 1600
- name: test
num_bytes: 21417107.0
num_examples: 400
download_size: 107224330
dataset_size: 106564689.4
---
# Dataset Card for "bw_spec_cls_80_32"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.757181704044342,
-0.09688924252986908,
0.19754040241241455,
0.36448344588279724,
-0.3023640215396881,
-0.1116906926035881,
0.01039532944560051,
-0.29008421301841736,
0.5572066307067871,
0.48085924983024597,
-0.7296229004859924,
-0.8123128414154053,
-0.6161960959434509,
-0.22259914875030... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
danaroth/jasper_ridge | danaroth | 2023-11-10T08:22:05Z | 0 | 0 | null | [
"license:unknown",
"arxiv:1403.4682",
"arxiv:1409.0685",
"arxiv:1305.7311",
"arxiv:1708.05125",
"region:us"
] | 2023-11-10T08:22:05Z | 2023-11-09T17:29:18.000Z | 2023-11-09T17:29:18 | ---
license: unknown
---
# Description
Jasper Ridge is a popular hyperspectral data used in [[enviTutorials](http://www.cossa.csiro.au/hswww/Overview.htm), [SS-NMF](http://arxiv.org/abs/1403.4682), [DgS-NMF](http://www.sciencedirect.com/science/article/pii/S0924271613002761), [RRLbS](http://arxiv.org/abs/1409.0685), [L1-CENMF](http://arxiv.org/abs/1305.7311)]. There are 512 x 614 pixels in it. Each pixel is recorded at 224 channels ranging from 380 nm to 2500 nm. The spectral resolution is up to 9.46nm. Since this hyperspectral image is too complex to get the ground truth, we consider a subimage of 100 x 100 pixels. The first pixel starts from the (105,269)-th pixel in the original image. After removing the channels 1--3, 108--112, 154--166 and 220--224 (due to dense water vapor and atmospheric effects), 198 channels are left (this is a common preprocess for HU analyses). There are four endmembers latent in this data: "#1 Road", "#2 Soil", "#3 Water" and "#4 Tree".
# Quick look
<figure>
<img src= "assets/1908991_orig.jpg" alt="Jasper Ridge" width="500" />
<figcaption>Jasper Ridge and its ground truth (GT:abundances and GT:endmembers).</figcaption>
</figure>
# Credits
Dataset originally collected by Feiyun Zhu and originally available at: http://www.escience.cn/people/feiyunZHU/Dataset_GT.html
To use this dataset, cite the associated paper:
```
@misc{zhu2017hyperspectral,
title={Hyperspectral Unmixing: Ground Truth Labeling, Datasets, Benchmark Performances and Survey},
author={Feiyun Zhu},
year={2017},
eprint={1708.05125},
archivePrefix={arXiv},
primaryClass={cs.CV}
}
``` | [
-0.602279782295227,
-0.26619452238082886,
0.5578780770301819,
0.006177427247166634,
-0.4376757740974426,
-0.12432537972927094,
-0.15516215562820435,
-0.09758874773979187,
0.4781437814235687,
0.46318280696868896,
-0.3486926555633545,
-0.7255200147628784,
-0.6556859612464905,
-0.057522118091... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
arieg/bw_spec_cls_80_33 | arieg | 2023-11-09T17:32:41Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T17:32:41Z | 2023-11-09T17:32:21.000Z | 2023-11-09T17:32:21 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': '74302'
'1': '74347'
'2': '74348'
'3': '74362'
'4': '74365'
'5': '74370'
'6': '74371'
'7': '74372'
'8': '74373'
'9': '74375'
'10': '74377'
'11': '74378'
'12': '74380'
'13': '74381'
'14': '74382'
'15': '74383'
'16': '74384'
'17': '74385'
'18': '74386'
'19': '74387'
'20': '74388'
'21': '74390'
'22': '74392'
'23': '74421'
'24': '74445'
'25': '74546'
'26': '74669'
'27': '74671'
'28': '74706'
'29': '74908'
'30': '74942'
'31': '74954'
'32': '74955'
'33': '74959'
'34': '74960'
'35': '75194'
'36': '75221'
'37': '75230'
'38': '75304'
'39': '75310'
'40': '75314'
'41': '75317'
'42': '75429'
'43': '75430'
'44': '75431'
'45': '75432'
'46': '75433'
'47': '75434'
'48': '75435'
'49': '75436'
'50': '75437'
'51': '75438'
'52': '75439'
'53': '75440'
'54': '75441'
'55': '75442'
'56': '75443'
'57': '75607'
'58': '75612'
'59': '75692'
'60': '75762'
'61': '75763'
'62': '75764'
'63': '75782'
'64': '75783'
'65': '75784'
'66': '75785'
'67': '75786'
'68': '75787'
'69': '75788'
'70': '75844'
'71': '75862'
'72': '75866'
'73': '75869'
'74': '75883'
'75': '75903'
'76': '75908'
'77': '75925'
'78': '75926'
'79': '75927'
splits:
- name: train
num_bytes: 88794100.8
num_examples: 1600
- name: test
num_bytes: 22341388.0
num_examples: 400
download_size: 111396696
dataset_size: 111135488.8
---
# Dataset Card for "bw_spec_cls_80_33"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.7818379402160645,
-0.11079391092061996,
0.20223098993301392,
0.32841891050338745,
-0.2773447036743164,
-0.11314045637845993,
-0.004527751822024584,
-0.2897574007511139,
0.5100889205932617,
0.49065983295440674,
-0.7855525016784668,
-0.8165704011917114,
-0.5531187653541565,
-0.18009008467... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
jhhon80/jhon | jhhon80 | 2023-11-09T17:40:39Z | 0 | 0 | null | [
"license:openrail",
"region:us"
] | 2023-11-09T17:40:39Z | 2023-11-09T17:39:48.000Z | 2023-11-09T17:39:48 | ---
license: openrail
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
allison221/pai | allison221 | 2023-11-09T22:18:24Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T22:18:24Z | 2023-11-09T17:45:32.000Z | 2023-11-09T17:45:32 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
pcuenq/gists | pcuenq | 2023-11-13T12:32:25Z | 0 | 4 | null | [
"region:us"
] | 2023-11-13T12:32:25Z | 2023-11-09T17:49:11.000Z | 2023-11-09T17:49:11 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
arieg/bw_spec_cls_80_34 | arieg | 2023-11-09T17:58:39Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T17:58:39Z | 2023-11-09T17:58:22.000Z | 2023-11-09T17:58:22 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': '75928'
'1': '75929'
'2': '75930'
'3': '75931'
'4': '75932'
'5': '75933'
'6': '75935'
'7': '75936'
'8': '75937'
'9': '75975'
'10': '76036'
'11': '76069'
'12': '76071'
'13': '76072'
'14': '76073'
'15': '76074'
'16': '76075'
'17': '76076'
'18': '76077'
'19': '76078'
'20': '76079'
'21': '76121'
'22': '76375'
'23': '76381'
'24': '76437'
'25': '76440'
'26': '76654'
'27': '76659'
'28': '77517'
'29': '77519'
'30': '77521'
'31': '77522'
'32': '77523'
'33': '77564'
'34': '77571'
'35': '77572'
'36': '77952'
'37': '78038'
'38': '78156'
'39': '78213'
'40': '78516'
'41': '78833'
'42': '78834'
'43': '78839'
'44': '78841'
'45': '78843'
'46': '78845'
'47': '78847'
'48': '78848'
'49': '78849'
'50': '78850'
'51': '78851'
'52': '78852'
'53': '78984'
'54': '78998'
'55': '79087'
'56': '79575'
'57': '79593'
'58': '79605'
'59': '79606'
'60': '79610'
'61': '79616'
'62': '79741'
'63': '79973'
'64': '79975'
'65': '79977'
'66': '79978'
'67': '79985'
'68': '79986'
'69': '79988'
'70': '79990'
'71': '79995'
'72': '80035'
'73': '80293'
'74': '80341'
'75': '80351'
'76': '80389'
'77': '80402'
'78': '80515'
'79': '80516'
splits:
- name: train
num_bytes: 88501139.2
num_examples: 1600
- name: test
num_bytes: 21775350.0
num_examples: 400
download_size: 109195616
dataset_size: 110276489.2
---
# Dataset Card for "bw_spec_cls_80_34"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.7508752346038818,
-0.06011578068137169,
0.17650024592876434,
0.3692382872104645,
-0.29800447821617126,
-0.11584567278623581,
-0.0018922502640634775,
-0.25966712832450867,
0.5170286893844604,
0.4697507917881012,
-0.7984611392021179,
-0.7917879819869995,
-0.5591444969177246,
-0.1882717162... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
misterytoon/katara | misterytoon | 2023-11-09T18:00:44Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T18:00:44Z | 2023-11-09T17:59:03.000Z | 2023-11-09T17:59:03 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
jhhon80/jhhon | jhhon80 | 2023-11-09T18:10:09Z | 0 | 0 | null | [
"license:openrail",
"region:us"
] | 2023-11-09T18:10:09Z | 2023-11-09T18:09:41.000Z | 2023-11-09T18:09:41 | ---
license: openrail
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Nunofofo/rr | Nunofofo | 2023-11-09T18:14:12Z | 0 | 0 | null | [
"license:openrail",
"region:us"
] | 2023-11-09T18:14:12Z | 2023-11-09T18:13:31.000Z | 2023-11-09T18:13:31 | ---
license: openrail
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
arieg/bw_spec_cls_80_35 | arieg | 2023-11-09T18:25:04Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T18:25:04Z | 2023-11-09T18:24:44.000Z | 2023-11-09T18:24:44 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': '80517'
'1': '80518'
'2': '80519'
'3': '80520'
'4': '80693'
'5': '80694'
'6': '80695'
'7': '80696'
'8': '80697'
'9': '80751'
'10': '80753'
'11': '80754'
'12': '80755'
'13': '80756'
'14': '80758'
'15': '80765'
'16': '80766'
'17': '80772'
'18': '80773'
'19': '80774'
'20': '80775'
'21': '80776'
'22': '80793'
'23': '80833'
'24': '80834'
'25': '80835'
'26': '80836'
'27': '81033'
'28': '81037'
'29': '81082'
'30': '81083'
'31': '81084'
'32': '81085'
'33': '81189'
'34': '81193'
'35': '81194'
'36': '81195'
'37': '81362'
'38': '81365'
'39': '81436'
'40': '81457'
'41': '81485'
'42': '81491'
'43': '81512'
'44': '81523'
'45': '81543'
'46': '81554'
'47': '81555'
'48': '81565'
'49': '81576'
'50': '81586'
'51': '81600'
'52': '81612'
'53': '81613'
'54': '81623'
'55': '81638'
'56': '81650'
'57': '81660'
'58': '81781'
'59': '81782'
'60': '81792'
'61': '81802'
'62': '81803'
'63': '81814'
'64': '81868'
'65': '81938'
'66': '81945'
'67': '81946'
'68': '81988'
'69': '81999'
'70': '82157'
'71': '82231'
'72': '82237'
'73': '82242'
'74': '82250'
'75': '82410'
'76': '82505'
'77': '82507'
'78': '82628'
'79': '82629'
splits:
- name: train
num_bytes: 90214491.2
num_examples: 1600
- name: test
num_bytes: 22067286.0
num_examples: 400
download_size: 110421965
dataset_size: 112281777.2
---
# Dataset Card for "bw_spec_cls_80_35"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.7758931517601013,
-0.10796183347702026,
0.18643243610858917,
0.3463329076766968,
-0.31640365719795227,
-0.08013816922903061,
-0.02627328783273697,
-0.28796467185020447,
0.5062318444252014,
0.5099354982376099,
-0.7436660528182983,
-0.8414629697799683,
-0.5256196856498718,
-0.184878051280... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ylacombe/benchmark-comparison | ylacombe | 2023-11-09T18:37:57Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T18:37:57Z | 2023-11-09T18:37:43.000Z | 2023-11-09T18:37:43 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
JoyeJiang/PwC4KPG | JoyeJiang | 2023-11-09T19:27:08Z | 0 | 0 | null | [
"task_categories:text-generation",
"size_categories:1K<n<10K",
"language:en",
"region:us"
] | 2023-11-09T19:27:08Z | 2023-11-09T18:40:40.000Z | 2023-11-09T18:40:40 | ---
task_categories:
- text-generation
language:
- en
size_categories:
- 1K<n<10K
---
## PwC4KPG dataset
Due to the strict copyright restriction, the dataset is only available for non-commercial research use ONLY.
Currently it requires manual approval for access. Please send an email to yijiang@whu.edu.cn, stating (1) Huggingface account name; (2) institute/company name; (3) the purpose of using this dataset.
## PwC4KPG dataset
we extract the **fields, tasks, methods, datasets, metrics, titles and abstracts** from the raw corpus of PwC, provided that the paper has a full title and abstract.
A total of 6,012 papers were extracted, of which 2,119 included all five categories of “keyphrases”, and the remaining 3,839 contained only some of them.
Note that PwC does not contain the research fields as we define them, so we used the “main_collection” of methods as an alternative.
**Train: 5,012 / Dev: 500 / Test: 500**
We randomly select 1,000 papers with full information,half of which are used for testing and the other half for validation. The remaining 5,012 served as the training set.
**Paper: JASIST 2023, Generating keyphrases for readers: A controllable keyphrase generation framework.**

```
@inproceedings{Jiang2023PwC4KPG,
title={ Generating keyphrases for readers: A controllable keyphrase generation framework},
author={Jiang, Yi and Meng, Rui and Huang, Yong and Lu, Wei and Liu, Jiawei},
booktitle={Journal of the Association for Information Science and Technology},
year={2023},
volume={74},
issue={7},
pages={759--774},
}
``` | [
0.04881129413843155,
-0.3353819251060486,
0.31537899374961853,
0.3300970792770386,
-0.3597971498966217,
0.18924635648727417,
0.10758855938911438,
-0.3191681504249573,
0.10450579226016998,
0.5645507574081421,
-0.39261239767074585,
-0.7290017008781433,
-0.6588349342346191,
0.5418091416358948... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Rdgaudio/TreinamentoRVC | Rdgaudio | 2023-11-13T16:11:36Z | 0 | 0 | null | [
"license:cc-by-nc-sa-4.0",
"region:us"
] | 2023-11-13T16:11:36Z | 2023-11-09T18:47:15.000Z | 2023-11-09T18:47:15 | ---
license: cc-by-nc-sa-4.0
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
arieg/bw_spec_cls_80_36 | arieg | 2023-11-09T18:51:31Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T18:51:31Z | 2023-11-09T18:51:14.000Z | 2023-11-09T18:51:14 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': '82630'
'1': '82631'
'2': '82881'
'3': '82886'
'4': '82890'
'5': '82892'
'6': '82893'
'7': '82914'
'8': '82915'
'9': '82916'
'10': '82917'
'11': '82918'
'12': '82919'
'13': '82920'
'14': '82921'
'15': '82928'
'16': '82929'
'17': '82930'
'18': '82931'
'19': '82932'
'20': '83600'
'21': '83612'
'22': '83613'
'23': '83715'
'24': '83717'
'25': '83718'
'26': '83719'
'27': '83789'
'28': '83790'
'29': '83791'
'30': '83903'
'31': '83911'
'32': '83913'
'33': '83954'
'34': '83960'
'35': '83969'
'36': '84009'
'37': '84055'
'38': '84056'
'39': '84058'
'40': '84095'
'41': '84096'
'42': '84097'
'43': '84111'
'44': '84135'
'45': '84136'
'46': '84139'
'47': '84141'
'48': '84142'
'49': '84144'
'50': '84154'
'51': '84155'
'52': '84156'
'53': '84157'
'54': '84158'
'55': '84159'
'56': '84195'
'57': '84198'
'58': '84200'
'59': '84201'
'60': '84202'
'61': '84264'
'62': '84290'
'63': '84291'
'64': '84405'
'65': '84417'
'66': '84423'
'67': '84483'
'68': '84484'
'69': '84485'
'70': '84486'
'71': '84605'
'72': '84743'
'73': '84757'
'74': '84768'
'75': '84788'
'76': '84817'
'77': '85027'
'78': '85038'
'79': '85039'
splits:
- name: train
num_bytes: 86231214.4
num_examples: 1600
- name: test
num_bytes: 21669535.0
num_examples: 400
download_size: 107649160
dataset_size: 107900749.4
---
# Dataset Card for "bw_spec_cls_80_36"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.7673241496086121,
-0.11210840195417404,
0.22960436344146729,
0.32208940386772156,
-0.2616190016269684,
-0.07895063608884811,
0.00258430652320385,
-0.24152587354183197,
0.5449079871177673,
0.49230074882507324,
-0.8448283672332764,
-0.7501130104064941,
-0.5117905735969543,
-0.218759030103... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
papanton/antonios | papanton | 2023-11-09T19:05:09Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T19:05:09Z | 2023-11-09T19:00:33.000Z | 2023-11-09T19:00:33 | ---
# For reference on dataset card metadata, see the spec: https://github.com/huggingface/hub-docs/blob/main/datasetcard.md?plain=1
# Doc / guide: https://huggingface.co/docs/hub/datasets-cards
{}
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
This dataset card aims to be a base template for new datasets. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/datasetcard_template.md?plain=1).
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.5322356224060059,
-0.5534716844558716,
0.1290130317211151,
0.23470577597618103,
-0.39626216888427734,
-0.11762470006942749,
-0.03545305132865906,
-0.6389272212982178,
0.5699822306632996,
0.7838326692581177,
-0.7834625840187073,
-0.9173274040222168,
-0.55633145570755,
0.13078093528747559... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ngcgarcia/foo | ngcgarcia | 2023-11-09T19:04:48Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T19:04:48Z | 2023-11-09T19:04:39.000Z | 2023-11-09T19:04:39 | asfdf | [
-0.3895011842250824,
-0.442147821187973,
0.10962357372045517,
1.1951137781143188,
-0.4632624685764313,
0.30276423692703247,
0.3302156627178192,
-0.45005711913108826,
0.1938348412513733,
1.0593156814575195,
-0.9975101351737976,
-0.2612268030643463,
-0.8550177812576294,
0.5819080471992493,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
sankovic/shirimdatasett | sankovic | 2023-11-09T19:05:44Z | 0 | 0 | null | [
"region:us"
] | 2023-11-09T19:05:44Z | 2023-11-09T19:04:53.000Z | 2023-11-09T19:04:53 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.