id stringlengths 2 115 | lastModified stringlengths 24 24 | tags list | author stringlengths 2 42 ⌀ | description stringlengths 0 68.7k ⌀ | citation stringlengths 0 10.7k ⌀ | cardData null | likes int64 0 3.55k | downloads int64 0 10.1M | card stringlengths 0 1.01M |
|---|---|---|---|---|---|---|---|---|---|
togethercomputer/RedPajama-Data-1T | 2023-06-30T22:06:10.000Z | [
"task_categories:text-generation",
"language:en",
"region:us"
] | togethercomputer | RedPajama is a clean-room, fully open-source implementation of the LLaMa dataset. | null | null | 878 | 17,226 | ---
task_categories:
- text-generation
language:
- en
pretty_name: Red Pajama 1T
---
### Getting Started
The dataset consists of 2084 jsonl files.
You can download the dataset using HuggingFace:
```python
from datasets import load_dataset
ds = load_dataset("togethercomputer/RedPajama-Data-1T")
```
Or you can directly download the files using the following command:
```
wget 'https://data.together.xyz/redpajama-data-1T/v1.0.0/urls.txt'
while read line; do
dload_loc=${line#https://data.together.xyz/redpajama-data-1T/v1.0.0/}
mkdir -p $(dirname $dload_loc)
wget "$line" -O "$dload_loc"
done < urls.txt
```
After downloading the files, you can load the dataset from disk by setting the `RED_PAJAMA_DATA_DIR` environment variable to the directory containing the files:
```python
import os
from datasets import load_dataset
os.environ["RED_PAJAMA_DATA_DIR"] = "/path/to/download"
ds = load_dataset("togethercomputer/RedPajama-Data-1T")
```
A smaller 1B-token sample of the dataset can be found [here](https://huggingface.co/datasets/togethercomputer/RedPajama-Data-1T-Sample).
A full set of scripts to recreate the dataset from scratch can be found [here](https://github.com/togethercomputer/RedPajama-Data).
### Dataset Summary
RedPajama is a clean-room, fully open-source implementation of the LLaMa dataset.
| Dataset | Token Count |
|---------------|-------------|
| Commoncrawl | 878 Billion |
| C4 | 175 Billion |
| GitHub | 59 Billion |
| Books | 26 Billion |
| ArXiv | 28 Billion |
| Wikipedia | 24 Billion |
| StackExchange | 20 Billion |
| Total | 1.2 Trillion |
### Languages
Primarily English, though the Wikipedia slice contains multiple languages.
## Dataset Structure
The dataset structure is as follows:
```json
{
"text": ...,
"meta": {"url": "...", "timestamp": "...", "source": "...", "language": "...", ...},
"red_pajama_subset": "common_crawl" | "c4" | "github" | "books" | "arxiv" | "wikipedia" | "stackexchange"
}
```
## Dataset Creation
This dataset was created to follow the LLaMa paper as closely as possible to try to reproduce its recipe.
### Source Data
#### Commoncrawl
We download five dumps from Commoncrawl, and run the dumps through the official `cc_net` pipeline.
We then deduplicate on the paragraph level, and filter out low quality text using a linear classifier trained to
classify paragraphs as Wikipedia references or random Commoncrawl samples.
#### C4
C4 is downloaded from Huggingface. The only preprocessing step is to bring the data into our own format.
#### GitHub
The raw GitHub data is downloaded from Google BigQuery. We deduplicate on the file level and filter out low quality
files and only keep projects that are distributed under the MIT, BSD, or Apache license.
#### Wikipedia
We use the Wikipedia dataset available on Huggingface, which is based on the Wikipedia dump from 2023-03-20 and contains
text in 20 different languages. The dataset comes in preprocessed format, so that hyperlinks, comments and other
formatting boilerplate has been removed.
#### Gutenberg and Books3
The PG19 subset of the Gutenberg Project and Books3 datasets are downloaded from Huggingface. After downloading, we use
simhash to remove near duplicates.
#### ArXiv
ArXiv data is downloaded from Amazon S3 in the `arxiv` requester pays bucket. We only keep latex source files and
remove preambles, comments, macros and bibliographies.
#### Stackexchange
The Stack Exchange split of the dataset is download from the
[Internet Archive](https://archive.org/download/stackexchange). Here we only keep the posts from the 28 largest sites,
remove html tags, group the posts into question-answer pairs, and order answers by their score.
### SHA256 Checksums
SHA256 checksums for the dataset files for each data source are available here:
```
https://data.together.xyz/redpajama-data-1T/v1.0.0/sha256/arxiv_SHA256SUMS.txt
https://data.together.xyz/redpajama-data-1T/v1.0.0/sha256/book_SHA256SUMS.txt
https://data.together.xyz/redpajama-data-1T/v1.0.0/sha256/c4_SHA256SUMS.txt
https://data.together.xyz/redpajama-data-1T/v1.0.0/sha256/common_crawl_SHA256SUMS.txt
https://data.together.xyz/redpajama-data-1T/v1.0.0/sha256/github_SHA256SUMS.txt
https://data.together.xyz/redpajama-data-1T/v1.0.0/sha256/stackexchange_SHA256SUMS.txt
https://data.together.xyz/redpajama-data-1T/v1.0.0/sha256/wikipedia_SHA256SUMS.txt
```
To cite RedPajama, please use:
```
@software{together2023redpajama,
author = {Together Computer},
title = {RedPajama: An Open Source Recipe to Reproduce LLaMA training dataset},
month = April,
year = 2023,
url = {https://github.com/togethercomputer/RedPajama-Data}
}
```
### License
Please refer to the licenses of the data subsets you use.
* [Common Crawl Foundation Terms of Use](https://commoncrawl.org/terms-of-use/full/)
* [C4 license](https://huggingface.co/datasets/allenai/c4#license)
* GitHub was limited to MIT, BSD, or Apache licenses only
* Books: [the_pile_books3 license](https://huggingface.co/datasets/the_pile_books3#licensing-information) and [pg19 license](https://huggingface.co/datasets/pg19#licensing-information)
* [ArXiv Terms of Use](https://info.arxiv.org/help/api/tou.html)
* [Wikipedia License](https://huggingface.co/datasets/wikipedia#licensing-information)
* [StackExchange license on the Internet Archive](https://archive.org/details/stackexchange)
<!--
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed]
--> |
yelp_review_full | 2023-01-25T15:03:32.000Z | [
"task_categories:text-classification",
"task_ids:sentiment-classification",
"annotations_creators:crowdsourced",
"language_creators:crowdsourced",
"multilinguality:monolingual",
"size_categories:100K<n<1M",
"source_datasets:original",
"language:en",
"license:other",
"arxiv:1509.01626",
"region:us"
] | null | The Yelp reviews dataset consists of reviews from Yelp. It is extracted from the Yelp Dataset Challenge 2015 data.
The Yelp reviews full star dataset is constructed by Xiang Zhang (xiang.zhang@nyu.edu) from the above dataset.
It is first used as a text classification benchmark in the following paper: Xiang Zhang, Junbo Zhao, Yann LeCun.
Character-level Convolutional Networks for Text Classification. Advances in Neural Information Processing Systems 28 (NIPS 2015). | @inproceedings{zhang2015character,
title={Character-level convolutional networks for text classification},
author={Zhang, Xiang and Zhao, Junbo and LeCun, Yann},
booktitle={Advances in neural information processing systems},
pages={649--657},
year={2015}
} | null | 33 | 16,117 | ---
annotations_creators:
- crowdsourced
language_creators:
- crowdsourced
language:
- en
license:
- other
multilinguality:
- monolingual
size_categories:
- 100K<n<1M
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- sentiment-classification
pretty_name: YelpReviewFull
license_details: yelp-licence
dataset_info:
features:
- name: label
dtype:
class_label:
names:
'0': 1 star
'1': 2 star
'2': 3 stars
'3': 4 stars
'4': 5 stars
- name: text
dtype: string
config_name: yelp_review_full
splits:
- name: train
num_bytes: 483811554
num_examples: 650000
- name: test
num_bytes: 37271188
num_examples: 50000
download_size: 196146755
dataset_size: 521082742
train-eval-index:
- config: yelp_review_full
task: text-classification
task_id: multi_class_classification
splits:
train_split: train
eval_split: test
col_mapping:
text: text
label: target
metrics:
- type: accuracy
name: Accuracy
- type: f1
name: F1 macro
args:
average: macro
- type: f1
name: F1 micro
args:
average: micro
- type: f1
name: F1 weighted
args:
average: weighted
- type: precision
name: Precision macro
args:
average: macro
- type: precision
name: Precision micro
args:
average: micro
- type: precision
name: Precision weighted
args:
average: weighted
- type: recall
name: Recall macro
args:
average: macro
- type: recall
name: Recall micro
args:
average: micro
- type: recall
name: Recall weighted
args:
average: weighted
---
---
# Dataset Card for YelpReviewFull
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [Yelp](https://www.yelp.com/dataset)
- **Repository:** [Crepe](https://github.com/zhangxiangxiao/Crepe)
- **Paper:** [Character-level Convolutional Networks for Text Classification](https://arxiv.org/abs/1509.01626)
- **Point of Contact:** [Xiang Zhang](mailto:xiang.zhang@nyu.edu)
### Dataset Summary
The Yelp reviews dataset consists of reviews from Yelp.
It is extracted from the Yelp Dataset Challenge 2015 data.
### Supported Tasks and Leaderboards
- `text-classification`, `sentiment-classification`: The dataset is mainly used for text classification: given the text, predict the sentiment.
### Languages
The reviews were mainly written in english.
## Dataset Structure
### Data Instances
A typical data point, comprises of a text and the corresponding label.
An example from the YelpReviewFull test set looks as follows:
```
{
'label': 0,
'text': 'I got \'new\' tires from them and within two weeks got a flat. I took my car to a local mechanic to see if i could get the hole patched, but they said the reason I had a flat was because the previous patch had blown - WAIT, WHAT? I just got the tire and never needed to have it patched? This was supposed to be a new tire. \\nI took the tire over to Flynn\'s and they told me that someone punctured my tire, then tried to patch it. So there are resentful tire slashers? I find that very unlikely. After arguing with the guy and telling him that his logic was far fetched he said he\'d give me a new tire \\"this time\\". \\nI will never go back to Flynn\'s b/c of the way this guy treated me and the simple fact that they gave me a used tire!'
}
```
### Data Fields
- 'text': The review texts are escaped using double quotes ("), and any internal double quote is escaped by 2 double quotes (""). New lines are escaped by a backslash followed with an "n" character, that is "\n".
- 'label': Corresponds to the score associated with the review (between 1 and 5).
### Data Splits
The Yelp reviews full star dataset is constructed by randomly taking 130,000 training samples and 10,000 testing samples for each review star from 1 to 5.
In total there are 650,000 trainig samples and 50,000 testing samples.
## Dataset Creation
### Curation Rationale
The Yelp reviews full star dataset is constructed by Xiang Zhang (xiang.zhang@nyu.edu) from the Yelp Dataset Challenge 2015. It is first used as a text classification benchmark in the following paper: Xiang Zhang, Junbo Zhao, Yann LeCun. Character-level Convolutional Networks for Text Classification. Advances in Neural Information Processing Systems 28 (NIPS 2015).
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
You can check the official [yelp-dataset-agreement](https://s3-media3.fl.yelpcdn.com/assets/srv0/engineering_pages/bea5c1e92bf3/assets/vendor/yelp-dataset-agreement.pdf).
### Citation Information
Xiang Zhang, Junbo Zhao, Yann LeCun. Character-level Convolutional Networks for Text Classification. Advances in Neural Information Processing Systems 28 (NIPS 2015).
### Contributions
Thanks to [@hfawaz](https://github.com/hfawaz) for adding this dataset. |
fashion_mnist | 2023-04-17T14:02:05.000Z | [
"task_categories:image-classification",
"task_ids:multi-class-image-classification",
"annotations_creators:expert-generated",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:mit",
"arxiv:1708.07747",
"region:us"
] | null | Fashion-MNIST is a dataset of Zalando's article images—consisting of a training set of
60,000 examples and a test set of 10,000 examples. Each example is a 28x28 grayscale image,
associated with a label from 10 classes. We intend Fashion-MNIST to serve as a direct drop-in
replacement for the original MNIST dataset for benchmarking machine learning algorithms.
It shares the same image size and structure of training and testing splits. | @article{DBLP:journals/corr/abs-1708-07747,
author = {Han Xiao and
Kashif Rasul and
Roland Vollgraf},
title = {Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning
Algorithms},
journal = {CoRR},
volume = {abs/1708.07747},
year = {2017},
url = {http://arxiv.org/abs/1708.07747},
archivePrefix = {arXiv},
eprint = {1708.07747},
timestamp = {Mon, 13 Aug 2018 16:47:27 +0200},
biburl = {https://dblp.org/rec/bib/journals/corr/abs-1708-07747},
bibsource = {dblp computer science bibliography, https://dblp.org}
} | null | 28 | 16,091 | ---
annotations_creators:
- expert-generated
language_creators:
- found
language:
- en
license:
- mit
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- image-classification
task_ids:
- multi-class-image-classification
paperswithcode_id: fashion-mnist
pretty_name: FashionMNIST
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': T - shirt / top
'1': Trouser
'2': Pullover
'3': Dress
'4': Coat
'5': Sandal
'6': Shirt
'7': Sneaker
'8': Bag
'9': Ankle boot
config_name: fashion_mnist
splits:
- name: train
num_bytes: 31296655
num_examples: 60000
- name: test
num_bytes: 5233818
num_examples: 10000
download_size: 30878645
dataset_size: 36530473
---
# Dataset Card for FashionMNIST
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [GitHub](https://github.com/zalandoresearch/fashion-mnist)
- **Repository:** [GitHub](https://github.com/zalandoresearch/fashion-mnist)
- **Paper:** [arXiv](https://arxiv.org/pdf/1708.07747.pdf)
- **Leaderboard:**
- **Point of Contact:**
### Dataset Summary
Fashion-MNIST is a dataset of Zalando's article images—consisting of a training set of 60,000 examples and a test set of 10,000 examples. Each example is a 28x28 grayscale image, associated with a label from 10 classes. We intend Fashion-MNIST to serve as a direct drop-in replacement for the original MNIST dataset for benchmarking machine learning algorithms. It shares the same image size and structure of training and testing splits.
### Supported Tasks and Leaderboards
- `image-classification`: The goal of this task is to classify a given image of Zalando's article into one of 10 classes. The leaderboard is available [here](https://paperswithcode.com/sota/image-classification-on-fashion-mnist).
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
A data point comprises an image and its label.
```
{
'image': <PIL.PngImagePlugin.PngImageFile image mode=L size=28x28 at 0x27601169DD8>,
'label': 9
}
```
### Data Fields
- `image`: A `PIL.Image.Image` object containing the 28x28 image. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]`.
- `label`: an integer between 0 and 9 representing the classes with the following mapping:
| Label | Description |
| --- | --- |
| 0 | T-shirt/top |
| 1 | Trouser |
| 2 | Pullover |
| 3 | Dress |
| 4 | Coat |
| 5 | Sandal |
| 6 | Shirt |
| 7 | Sneaker |
| 8 | Bag |
| 9 | Ankle boot |
### Data Splits
The data is split into training and test set. The training set contains 60,000 images and the test set 10,000 images.
## Dataset Creation
### Curation Rationale
**From the arXiv paper:**
The original MNIST dataset contains a lot of handwritten digits. Members of the AI/ML/Data Science community love this dataset and use it as a benchmark to validate their algorithms. In fact, MNIST is often the first dataset researchers try. "If it doesn't work on MNIST, it won't work at all", they said. "Well, if it does work on MNIST, it may still fail on others."
Here are some good reasons:
- MNIST is too easy. Convolutional nets can achieve 99.7% on MNIST. Classic machine learning algorithms can also achieve 97% easily. Check out our side-by-side benchmark for Fashion-MNIST vs. MNIST, and read "Most pairs of MNIST digits can be distinguished pretty well by just one pixel."
- MNIST is overused. In this April 2017 Twitter thread, Google Brain research scientist and deep learning expert Ian Goodfellow calls for people to move away from MNIST.
- MNIST can not represent modern CV tasks, as noted in this April 2017 Twitter thread, deep learning expert/Keras author François Chollet.
### Source Data
#### Initial Data Collection and Normalization
**From the arXiv paper:**
Fashion-MNIST is based on the assortment on Zalando’s website. Every fashion product on Zalando has a set of pictures shot by professional photographers, demonstrating different aspects of the product, i.e. front and back looks, details, looks with model and in an outfit. The original picture has a light-gray background (hexadecimal color: #fdfdfd) and stored in 762 × 1000 JPEG format. For efficiently serving different frontend components, the original picture is resampled with multiple resolutions, e.g. large, medium, small, thumbnail and tiny.
We use the front look thumbnail images of 70,000 unique products to build Fashion-MNIST. Those products come from different gender groups: men, women, kids and neutral. In particular, whitecolor products are not included in the dataset as they have low contrast to the background. The thumbnails (51 × 73) are then fed into the following conversion pipeline:
1. Converting the input to a PNG image.
2. Trimming any edges that are close to the color of the corner pixels. The “closeness” is defined by the distance within 5% of the maximum possible intensity in RGB space.
3. Resizing the longest edge of the image to 28 by subsampling the pixels, i.e. some rows and columns are skipped over.
4. Sharpening pixels using a Gaussian operator of the radius and standard deviation of 1.0, with increasing effect near outlines.
5. Extending the shortest edge to 28 and put the image to the center of the canvas.
6. Negating the intensities of the image.
7. Converting the image to 8-bit grayscale pixels.
#### Who are the source language producers?
**From the arXiv paper:**
Every fashion product on Zalando has a set of pictures shot by professional photographers, demonstrating different aspects of the product, i.e. front and back looks, details, looks with model and in an outfit.
### Annotations
#### Annotation process
**From the arXiv paper:**
For the class labels, they use the silhouette code of the product. The silhouette code is manually labeled by the in-house fashion experts and reviewed by a separate team at Zalando. Each product Zalando is the Europe’s largest online fashion platform. Each product contains only one silhouette code.
#### Who are the annotators?
**From the arXiv paper:**
The silhouette code is manually labeled by the in-house fashion experts and reviewed by a separate team at Zalando.
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
Han Xiao and Kashif Rasul and Roland Vollgraf
### Licensing Information
MIT Licence
### Citation Information
```
@article{DBLP:journals/corr/abs-1708-07747,
author = {Han Xiao and
Kashif Rasul and
Roland Vollgraf},
title = {Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning
Algorithms},
journal = {CoRR},
volume = {abs/1708.07747},
year = {2017},
url = {http://arxiv.org/abs/1708.07747},
archivePrefix = {arXiv},
eprint = {1708.07747},
timestamp = {Mon, 13 Aug 2018 16:47:27 +0200},
biburl = {https://dblp.org/rec/bib/journals/corr/abs-1708-07747},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
```
### Contributions
Thanks to [@gchhablani](https://github.com/gchablani) for adding this dataset. |
commonsense_qa | 2023-04-05T10:02:16.000Z | [
"task_categories:question-answering",
"task_ids:open-domain-qa",
"annotations_creators:crowdsourced",
"language_creators:crowdsourced",
"multilinguality:monolingual",
"size_categories:1K<n<10K",
"source_datasets:original",
"language:en",
"license:mit",
"arxiv:1811.00937",
"region:us"
] | null | CommonsenseQA is a new multiple-choice question answering dataset that requires different types of commonsense knowledge
to predict the correct answers . It contains 12,102 questions with one correct answer and four distractor answers.
The dataset is provided in two major training/validation/testing set splits: "Random split" which is the main evaluation
split, and "Question token split", see paper for details. | @inproceedings{talmor-etal-2019-commonsenseqa,
title = "{C}ommonsense{QA}: A Question Answering Challenge Targeting Commonsense Knowledge",
author = "Talmor, Alon and
Herzig, Jonathan and
Lourie, Nicholas and
Berant, Jonathan",
booktitle = "Proceedings of the 2019 Conference of the North {A}merican Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)",
month = jun,
year = "2019",
address = "Minneapolis, Minnesota",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/N19-1421",
doi = "10.18653/v1/N19-1421",
pages = "4149--4158",
archivePrefix = "arXiv",
eprint = "1811.00937",
primaryClass = "cs",
} | null | 23 | 16,048 | ---
annotations_creators:
- crowdsourced
language_creators:
- crowdsourced
language:
- en
license:
- mit
multilinguality:
- monolingual
pretty_name: CommonsenseQA
size_categories:
- 1K<n<10K
source_datasets:
- original
task_categories:
- question-answering
task_ids:
- open-domain-qa
paperswithcode_id: commonsenseqa
dataset_info:
features:
- name: id
dtype: string
- name: question
dtype: string
- name: question_concept
dtype: string
- name: choices
sequence:
- name: label
dtype: string
- name: text
dtype: string
- name: answerKey
dtype: string
splits:
- name: train
num_bytes: 2209044
num_examples: 9741
- name: validation
num_bytes: 274033
num_examples: 1221
- name: test
num_bytes: 258017
num_examples: 1140
download_size: 4680691
dataset_size: 2741094
---
# Dataset Card for "commonsense_qa"
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** https://www.tau-nlp.org/commonsenseqa
- **Repository:** https://github.com/jonathanherzig/commonsenseqa
- **Paper:** https://arxiv.org/abs/1811.00937
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 4.68 MB
- **Size of the generated dataset:** 2.18 MB
- **Total amount of disk used:** 6.86 MB
### Dataset Summary
CommonsenseQA is a new multiple-choice question answering dataset that requires different types of commonsense knowledge
to predict the correct answers . It contains 12,102 questions with one correct answer and four distractor answers.
The dataset is provided in two major training/validation/testing set splits: "Random split" which is the main evaluation
split, and "Question token split", see paper for details.
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
The dataset is in English (`en`).
## Dataset Structure
### Data Instances
#### default
- **Size of downloaded dataset files:** 4.68 MB
- **Size of the generated dataset:** 2.18 MB
- **Total amount of disk used:** 6.86 MB
An example of 'train' looks as follows:
```
{'id': '075e483d21c29a511267ef62bedc0461',
'question': 'The sanctions against the school were a punishing blow, and they seemed to what the efforts the school had made to change?',
'question_concept': 'punishing',
'choices': {'label': ['A', 'B', 'C', 'D', 'E'],
'text': ['ignore', 'enforce', 'authoritarian', 'yell at', 'avoid']},
'answerKey': 'A'}
```
### Data Fields
The data fields are the same among all splits.
#### default
- `id` (`str`): Unique ID.
- `question`: a `string` feature.
- `question_concept` (`str`): ConceptNet concept associated to the question.
- `choices`: a dictionary feature containing:
- `label`: a `string` feature.
- `text`: a `string` feature.
- `answerKey`: a `string` feature.
### Data Splits
| name | train | validation | test |
|---------|------:|-----------:|-----:|
| default | 9741 | 1221 | 1140 |
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
The dataset is licensed under the MIT License.
See: https://github.com/jonathanherzig/commonsenseqa/issues/5
### Citation Information
```
@inproceedings{talmor-etal-2019-commonsenseqa,
title = "{C}ommonsense{QA}: A Question Answering Challenge Targeting Commonsense Knowledge",
author = "Talmor, Alon and
Herzig, Jonathan and
Lourie, Nicholas and
Berant, Jonathan",
booktitle = "Proceedings of the 2019 Conference of the North {A}merican Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)",
month = jun,
year = "2019",
address = "Minneapolis, Minnesota",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/N19-1421",
doi = "10.18653/v1/N19-1421",
pages = "4149--4158",
archivePrefix = "arXiv",
eprint = "1811.00937",
primaryClass = "cs",
}
```
### Contributions
Thanks to [@thomwolf](https://github.com/thomwolf), [@lewtun](https://github.com/lewtun), [@albertvillanova](https://github.com/albertvillanova), [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset. |
mosaicml/dolly_hhrlhf | 2023-10-02T15:48:48.000Z | [
"task_categories:text-generation",
"language:en",
"license:cc-by-sa-3.0",
"region:us"
] | mosaicml | null | null | null | 87 | 15,344 | ---
dataset_info:
features:
- name: prompt
dtype: string
- name: response
dtype: string
splits:
- name: train
num_bytes: 43781455.002688624
num_examples: 59310
- name: test
num_bytes: 4479286.805304853
num_examples: 5129
download_size: 24882010
dataset_size: 48260741.80799348
license: cc-by-sa-3.0
task_categories:
- text-generation
language:
- en
pretty_name: Dolly HH-RLHF
---
# Dataset Card for "dolly_hhrlhf"
This dataset is a combination of [Databrick's dolly-15k](https://huggingface.co/datasets/databricks/databricks-dolly-15k) dataset and a filtered subset of [Anthropic's HH-RLHF](https://huggingface.co/datasets/Anthropic/hh-rlhf). It also includes a test split, which was missing in the original `dolly` set. That test set is composed of 200 randomly selected samples from `dolly` + 4,929 of the test set samples from HH-RLHF which made it through the filtering process. The train set contains 59,310 samples; `15,014 - 200 = 14,814` from Dolly, and the remaining 44,496 from HH-RLHF.
It is slightly larger than Alpaca, and in our experience of slightly higher quality, but is usable for commercial purposes so long as you follow the terms of the license.
## Filtering process
As mentioned, the HH-RLHF data in this dataset is filtered. Specifically, we take the first turn of the convesation, then remove any samples where the assistant:
- uses the word "human", "thank", or "sorry"
- asks a question
- uses a first person pronoun
This leaves samples which look like instruction-following, as opposed to conversation.
## License/Attribution
<!--
**Copyright (2023) MosaicML, Inc.**
-->
This dataset was developed at MosaicML (https://www.mosaicml.com) and its use is subject to the CC BY-SA 3.0 license.
Certain categories of material in the dataset include materials from the following sources, licensed under the CC BY-SA 3.0 license:
Wikipedia (various pages) - https://www.wikipedia.org/
Copyright © Wikipedia editors and contributors.
Databricks (https://www.databricks.com)
Copyright © Databricks
When citing this dataset, please use the following:
```
@misc{mosaicml2023dolly_hhrlhf,
author = {MosaicML},
title = {Dolly-HHRLHF Dataset},
year = {2023},
publisher = {HuggingFace Datasets},
howpublished = {https://huggingface.co/datasets/mosaicml/dolly_hhrlhf},
}
``` |
Open-Orca/OpenOrca | 2023-10-02T19:01:36.000Z | [
"task_categories:conversational",
"task_categories:text-classification",
"task_categories:token-classification",
"task_categories:table-question-answering",
"task_categories:question-answering",
"task_categories:zero-shot-classification",
"task_categories:summarization",
"task_categories:feature-extraction",
"task_categories:text-generation",
"task_categories:text2text-generation",
"size_categories:10M<n<100M",
"language:en",
"license:mit",
"arxiv:2306.02707",
"arxiv:2301.13688",
"region:us"
] | Open-Orca | null | null | null | 767 | 15,297 | ---
language:
- en
license: mit
task_categories:
- conversational
- text-classification
- token-classification
- table-question-answering
- question-answering
- zero-shot-classification
- summarization
- feature-extraction
- text-generation
- text2text-generation
pretty_name: OpenOrca
size_categories:
- 10M<n<100M
---
## Table of Contents
- [Dataset Summary](#dataset-summary)
- [Dataset Attribution](#dataset-attribution)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Dataset Use](#dataset-use)
- [Use Cases](#use-cases)
- [Usage Caveats](#usage-caveats)
- [Getting Started](#getting-started)
<p><h1>🐋 The OpenOrca Dataset! 🐋</h1></p>

<a name="dataset-announcement"></a>
We are thrilled to announce the release of the OpenOrca dataset!
This rich collection of augmented FLAN data aligns, as best as possible, with the distributions outlined in the [Orca paper](https://arxiv.org/abs/2306.02707).
It has been instrumental in generating high-performing model checkpoints and serves as a valuable resource for all NLP researchers and developers!
# Official Models
## Mistral-7B-OpenOrca
Our [latest model](https://huggingface.co/spaces/Open-Orca/Mistral-7B-OpenOrca), the first 7B to score better overall than all previous models below 30B.
98% of Llama2-70b-chat's performance, in a completely open 7B!
## OpenOrca-Platypus2-13B
Our [third model](https://huggingface.co/Open-Orca/OpenOrca-Platypus2-13B), the first 13B model to score higher than LLaMA1-65B on the HuggingFace Leaderboard!
Released in partnership with Platypus.
## LlongOrca 7B & 13B
* Our [first 7B release](https://huggingface.co/Open-Orca/LlongOrca-7B-16k), trained on top of LLongMA2 to achieve 16,000 tokens context. #1 long context 7B model at release time, with >99% of the overall #1 model's performance.
* [LlongOrca-13B-16k](https://huggingface.co/Open-Orca/LlongOrca-13B-16k), trained on top of LLongMA2. #1 long context 13B model at release time, with >97% of the overall #1 model's performance.
## OpenOrcaxOpenChat-Preview2-13B
Our [second model](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B), highlighting that we've surpassed the performance reported in the Orca paper.
Was #1 at release time, now surpassed by our own OpenOrca-Platypus2-13B.
Released in partnership with OpenChat.
## OpenOrca-Preview1-13B
[OpenOrca-Preview1-13B](https://huggingface.co/Open-Orca/OpenOrca-Preview1-13B)
This model was trained in less than a day, for <$200, with <10% of our data.
At release, it beat the current state of the art models on BigBench-Hard and AGIEval. Achieves ~60% of the improvements reported in the Orca paper.
<a name="dataset-summary"></a>
# Dataset Summary
The OpenOrca dataset is a collection of augmented [FLAN Collection data](https://arxiv.org/abs/2301.13688).
Currently ~1M GPT-4 completions, and ~3.2M GPT-3.5 completions.
It is tabularized in alignment with the distributions presented in the ORCA paper and currently represents a partial completion of the full intended dataset, with ongoing generation to expand its scope.
The data is primarily used for training and evaluation in the field of natural language processing.
<a name="dataset-attribution"></a>
# Dataset Attribution
We would like to give special recognition to the following contributors for their significant efforts and dedication:
Teknium
WingLian/Caseus
Eric Hartford
NanoBit
Pankaj
Winddude
Rohan
http://AlignmentLab.ai:
Autometa
Entropi
AtlasUnified
NeverendingToast
NanoBit
WingLian/Caseus
Also of course, as always, TheBloke, for being the backbone of the whole community.
Many thanks to NanoBit and Caseus, makers of [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl), for lending us their expertise on the platform that developed and trained manticore, minotaur, and many others!
We are welcoming sponsors or collaborators to help us build these models to the scale they deserve. Please reach out via our socials:
http://Alignmentlab.ai https://discord.gg/n9hXaBPWxx
Want to visualize our full dataset? Check out our [Nomic Atlas Map](https://atlas.nomic.ai/map/c1b88b47-2d9b-47e0-9002-b80766792582/2560fd25-52fe-42f1-a58f-ff5eccc890d2).
[<img src="https://huggingface.co/Open-Orca/OpenOrca-Preview1-13B/resolve/main/OpenOrca%20Nomic%20Atlas.png" alt="Atlas Nomic Dataset Map" width="400" height="400" />](https://atlas.nomic.ai/map/c1b88b47-2d9b-47e0-9002-b80766792582/2560fd25-52fe-42f1-a58f-ff5eccc890d2)
<a name="supported-tasks-and-leaderboards"></a>
# Supported Tasks and Leaderboards
This dataset supports a range of tasks including language modeling, text generation, and text augmentation.
It has been instrumental in the generation of multiple high-performing model checkpoints which have exhibited exceptional performance in our unit testing.
Further information on leaderboards will be updated as they become available.
<a name="languages"></a>
# Languages
The language of the data is primarily English.
<a name="dataset-structure"></a>
# Dataset Structure
<a name="data-instances"></a>
## Data Instances
A data instance in this dataset represents entries from the FLAN collection which have been augmented by submitting the listed question to either GPT-4 or GPT-3.5.
The response is then entered into the response field.
<a name="data-fields"></a>
## Data Fields
The fields are:
1) 'id', a unique numbered identifier which includes one of 'niv', 't0', 'cot', or 'flan' to represent which source FLAN Collection submix the 'question' is sourced from.
2) 'system_prompt', representing the System Prompt presented to the GPT-3.5 or GPT-4 API for the datapoint
3) 'question', representing a question entry as provided by the FLAN Collection
4) 'response', a response to that question received from a query to either GPT-3.5 or GPT-4.
<a name="data-splits"></a>
## Data Splits
The data is unsplit.
<a name="dataset-creation"></a>
# Dataset Creation
<a name="curation-rationale"></a>
## Curation Rationale
The dataset was created to provide a source of augmented text data for researchers and developers.
The datapoints are intended primarily to provide an enhancement of the core FLAN Collection data which relies upon the detailed step by step reasoning capabilities of GPT-3.5 and GPT-4.
This "reasoning trace" augmentation has demonstrated exceptional results, allowing a LLaMA-13B model trained with this data to rival or beat GPT-3.5 on broad sets of hard reasoning tasks which all models below 100B parameters had previously performed dramatically worse on.
<a name="source-data"></a>
## Source Data
The data is generated using techniques in alignment with the distributions outlined in the Orca paper, except as noted below:
1) There is not enough CoT data in the FLAN Collection to generate 150K zero-shot entries, as the paper purports to use.
We suspect this portion was either undocumented or misrepresented. We have used the ~75K points available.
2) We used the pre-generated FLAN Collection datasets hosted on HuggingFace under conceptofmind, e.g. [conceptofmind/flan2021](https://huggingface.co/datasets/conceptofmind/flan2021_submix_original).
These are referenced by the [official FLAN Collection repo](https://github.com/google-research/FLAN/tree/main/flan/v2) as the preferred data source.
However, these are a subset of the full FLAN Collection data, and have less than the required entries for the flan2021 and t0 submixes, by ~1.25M and 200k respectively.
Combined, this gave us ~1.5M fewer datapoints than in the original Orca paper. Completing the set is an ongoing work.
<a name="dataset-use"></a>
# Dataset Use
<a name="use-cases"></a>
## Use Cases
The dataset can be used for tasks related to language understanding, natural language processing, machine learning model training, and model performance evaluation.
<a name="usage-caveats"></a>
## Usage Caveats
Given that this is a work-in-progress dataset, it is recommended to regularly check for updates and improvements.
Further, the data should be used in accordance with the guidelines and recommendations outlined in the Orca paper.
<a name="getting-started"></a>
## Getting Started
This dataset is organized such that it can be naively loaded via Hugging Face datasets library.
We recommend using streaming due to the large size of the files.
Regular updates and data generation progress can be monitored through the OpenOrca repository on Hugging Face.
# Citation
```bibtex
@misc{OpenOrca,
title = {OpenOrca: An Open Dataset of GPT Augmented FLAN Reasoning Traces},
author = {Wing Lian and Bleys Goodson and Eugene Pentland and Austin Cook and Chanvichet Vong and "Teknium"},
year = {2023},
publisher = {HuggingFace},
journal = {HuggingFace repository},
howpublished = {\url{https://https://huggingface.co/Open-Orca/OpenOrca},
}
```
```bibtex
@misc{mukherjee2023orca,
title={Orca: Progressive Learning from Complex Explanation Traces of GPT-4},
author={Subhabrata Mukherjee and Arindam Mitra and Ganesh Jawahar and Sahaj Agarwal and Hamid Palangi and Ahmed Awadallah},
year={2023},
eprint={2306.02707},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
```bibtex
@misc{longpre2023flan,
title={The Flan Collection: Designing Data and Methods for Effective Instruction Tuning},
author={Shayne Longpre and Le Hou and Tu Vu and Albert Webson and Hyung Won Chung and Yi Tay and Denny Zhou and Quoc V. Le and Barret Zoph and Jason Wei and Adam Roberts},
year={2023},
eprint={2301.13688},
archivePrefix={arXiv},
primaryClass={cs.AI}
}
```
```bibtex
@misc{touvron2023llama,
title={Llama 2: Open Foundation and Fine-Tuned Chat Models},
author={Hugo Touvron and Louis Martin and Kevin Stone and Peter Albert and Amjad Almahairi and Yasmine Babaei and Nikolay Bashlykov and Soumya Batra and Prajjwal Bhargava and Shruti Bhosale and Dan Bikel and Lukas Blecher and Cristian Canton Ferrer and Moya Chen and Guillem Cucurull and David Esiobu and Jude Fernandes and Jeremy Fu and Wenyin Fu and Brian Fuller and Cynthia Gao and Vedanuj Goswami and Naman Goyal and Anthony Hartshorn and Saghar Hosseini and Rui Hou and Hakan Inan and Marcin Kardas and Viktor Kerkez and Madian Khabsa and Isabel Kloumann and Artem Korenev and Punit Singh Koura and Marie-Anne Lachaux and Thibaut Lavril and Jenya Lee and Diana Liskovich and Yinghai Lu and Yuning Mao and Xavier Martinet and Todor Mihaylov and Pushkar Mishra and Igor Molybog and Yixin Nie and Andrew Poulton and Jeremy Reizenstein and Rashi Rungta and Kalyan Saladi and Alan Schelten and Ruan Silva and Eric Michael Smith and Ranjan Subramanian and Xiaoqing Ellen Tan and Binh Tang and Ross Taylor and Adina Williams and Jian Xiang Kuan and Puxin Xu and Zheng Yan and Iliyan Zarov and Yuchen Zhang and Angela Fan and Melanie Kambadur and Sharan Narang and Aurelien Rodriguez and Robert Stojnic and Sergey Edunov and Thomas Scialom},
year={2023},
eprint= arXiv 2307.09288
}
@software{touvron2023llama,
title={LLaMA: Open and Efficient Foundation Language Models},
author={Touvron, Hugo and Lavril, Thibaut and Izacard, Gautier and Martinet, Xavier and Lachaux, Marie-Anne and Lacroix, Timoth{\'e}e and Rozi{\`e}re, Baptiste and Goyal, Naman and Hambro, Eric and Azhar, Faisal and Rodriguez, Aurelien and Joulin, Armand and Grave, Edouard and Lample, Guillaume},
journal={arXiv preprint arXiv:2302.13971},
year={2023}
}
``` |
hf-internal-testing/dummy_image_text_data | 2023-02-08T10:34:38.000Z | [
"region:us"
] | hf-internal-testing | null | null | null | 0 | 14,954 | ---
dataset_info:
features:
- name: image
dtype: image
- name: text
dtype: string
splits:
- name: train
num_bytes: 1944983.0
num_examples: 20
download_size: 1690123
dataset_size: 1944983.0
---
# Dataset Card for "dummy_image_text_data"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
skt/kobest_v1 | 2022-08-22T09:00:17.000Z | [
"annotations_creators:expert-generated",
"language_creators:expert-generated",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:ko",
"license:cc-by-sa-4.0",
"arxiv:2204.04541",
"region:us"
] | skt | The dataset contains data for KoBEST dataset | null | null | 17 | 14,875 | ---
pretty_name: KoBEST
annotations_creators:
- expert-generated
language_creators:
- expert-generated
language:
- ko
license:
- cc-by-sa-4.0
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- original
---
# Dataset Card for KoBEST
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Repository:** https://github.com/SKT-LSL/KoBEST_datarepo
- **Paper:**
- **Point of Contact:** https://github.com/SKT-LSL/KoBEST_datarepo/issues
### Dataset Summary
KoBEST is a Korean benchmark suite consists of 5 natural language understanding tasks that requires advanced knowledge in Korean.
### Supported Tasks and Leaderboards
Boolean Question Answering, Choice of Plausible Alternatives, Words-in-Context, HellaSwag, Sentiment Negation Recognition
### Languages
`ko-KR`
## Dataset Structure
### Data Instances
#### KB-BoolQ
An example of a data point looks as follows.
```
{'paragraph': '두아 리파(Dua Lipa, 1995년 8월 22일 ~ )는 잉글랜드의 싱어송라이터, 모델이다. BBC 사운드 오브 2016 명단에 노미닛되었다. 싱글 "Be the One"가 영국 싱글 차트 9위까지 오르는 등 성과를 보여주었다.',
'question': '두아 리파는 영국인인가?',
'label': 1}
```
#### KB-COPA
An example of a data point looks as follows.
```
{'premise': '물을 오래 끓였다.',
'question': '결과',
'alternative_1': '물의 양이 늘어났다.',
'alternative_2': '물의 양이 줄어들었다.',
'label': 1}
```
#### KB-WiC
An example of a data point looks as follows.
```
{'word': '양분',
'context_1': '토양에 [양분]이 풍부하여 나무가 잘 자란다. ',
'context_2': '태아는 모체로부터 [양분]과 산소를 공급받게 된다.',
'label': 1}
```
#### KB-HellaSwag
An example of a data point looks as follows.
```
{'context': '모자를 쓴 투수가 타자에게 온 힘을 다해 공을 던진다. 공이 타자에게 빠른 속도로 다가온다. 타자가 공을 배트로 친다. 배트에서 깡 소리가 난다. 공이 하늘 위로 날아간다.',
'ending_1': '외야수가 떨어지는 공을 글러브로 잡는다.',
'ending_2': '외야수가 공이 떨어질 위치에 자리를 잡는다.',
'ending_3': '심판이 아웃을 외친다.',
'ending_4': '외야수가 공을 따라 뛰기 시작한다.',
'label': 3}
```
#### KB-SentiNeg
An example of a data point looks as follows.
```
{'sentence': '택배사 정말 마음에 듬',
'label': 1}
```
### Data Fields
### KB-BoolQ
+ `paragraph`: a `string` feature
+ `question`: a `string` feature
+ `label`: a classification label, with possible values `False`(0) and `True`(1)
### KB-COPA
+ `premise`: a `string` feature
+ `question`: a `string` feature
+ `alternative_1`: a `string` feature
+ `alternative_2`: a `string` feature
+ `label`: an answer candidate label, with possible values `alternative_1`(0) and `alternative_2`(1)
### KB-WiC
+ `target_word`: a `string` feature
+ `context_1`: a `string` feature
+ `context_2`: a `string` feature
+ `label`: a classification label, with possible values `False`(0) and `True`(1)
### KB-HellaSwag
+ `target_word`: a `string` feature
+ `context_1`: a `string` feature
+ `context_2`: a `string` feature
+ `label`: a classification label, with possible values `False`(0) and `True`(1)
### KB-SentiNeg
+ `sentence`: a `string` feature
+ `label`: a classification label, with possible values `Negative`(0) and `Positive`(1)
### Data Splits
#### KB-BoolQ
+ train: 3,665
+ dev: 700
+ test: 1,404
#### KB-COPA
+ train: 3,076
+ dev: 1,000
+ test: 1,000
#### KB-WiC
+ train: 3,318
+ dev: 1,260
+ test: 1,260
#### KB-HellaSwag
+ train: 3,665
+ dev: 700
+ test: 1,404
#### KB-SentiNeg
+ train: 3,649
+ dev: 400
+ test: 397
+ test_originated: 397 (Corresponding training data where the test set is originated from.)
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
```
@misc{https://doi.org/10.48550/arxiv.2204.04541,
doi = {10.48550/ARXIV.2204.04541},
url = {https://arxiv.org/abs/2204.04541},
author = {Kim, Dohyeong and Jang, Myeongjun and Kwon, Deuk Sin and Davis, Eric},
title = {KOBEST: Korean Balanced Evaluation of Significant Tasks},
publisher = {arXiv},
year = {2022},
}
```
[More Information Needed]
### Contributions
Thanks to [@MJ-Jang](https://github.com/MJ-Jang) for adding this dataset. |
hate_speech_offensive | 2023-01-25T14:31:41.000Z | [
"task_categories:text-classification",
"annotations_creators:expert-generated",
"annotations_creators:crowdsourced",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:unknown",
"hate-speech-detection",
"arxiv:1703.04009",
"region:us"
] | null | An annotated dataset for hate speech and offensive language detection on tweets. | @inproceedings{hateoffensive,
title = {Automated Hate Speech Detection and the Problem of Offensive Language},
author = {Davidson, Thomas and Warmsley, Dana and Macy, Michael and Weber, Ingmar},
booktitle = {Proceedings of the 11th International AAAI Conference on Web and Social Media},
series = {ICWSM '17},
year = {2017},
location = {Montreal, Canada},
pages = {512-515}
} | null | 6 | 14,833 | ---
annotations_creators:
- expert-generated
- crowdsourced
language_creators:
- found
language:
- en
license:
- unknown
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- text-classification
task_ids: []
paperswithcode_id: hate-speech-and-offensive-language
pretty_name: Hate Speech and Offensive Language
tags:
- hate-speech-detection
dataset_info:
features:
- name: count
dtype: int64
- name: hate_speech_count
dtype: int64
- name: offensive_language_count
dtype: int64
- name: neither_count
dtype: int64
- name: class
dtype:
class_label:
names:
'0': hate speech
'1': offensive language
'2': neither
- name: tweet
dtype: string
splits:
- name: train
num_bytes: 3207826
num_examples: 24783
download_size: 2546446
dataset_size: 3207826
train-eval-index:
- config: default
task: text-classification
task_id: multi_class_classification
splits:
train_split: train
col_mapping:
tweet: text
label: target
metrics:
- type: accuracy
name: Accuracy
- type: f1
name: F1 macro
args:
average: macro
- type: f1
name: F1 micro
args:
average: micro
- type: f1
name: F1 weighted
args:
average: weighted
- type: precision
name: Precision macro
args:
average: macro
- type: precision
name: Precision micro
args:
average: micro
- type: precision
name: Precision weighted
args:
average: weighted
- type: recall
name: Recall macro
args:
average: macro
- type: recall
name: Recall micro
args:
average: micro
- type: recall
name: Recall weighted
args:
average: weighted
---
# Dataset Card for [Dataset Name]
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** https://github.com/t-davidson/hate-speech-and-offensive-language
- **Repository:** https://github.com/t-davidson/hate-speech-and-offensive-language
- **Paper:** https://arxiv.org/abs/1703.04009
- **Leaderboard:**
- **Point of Contact:** https://docs.google.com/forms/d/e/1FAIpQLSdrPNlfVBlqxun2tivzAtsZaOoPC5YYMocn-xscCgeRakLXHg/viewform?usp=pp_url&entry.1506871634&entry.147453066&entry.1390333885&entry.516829772
### Dataset Summary
An annotated dataset for hate speech and offensive language detection on tweets.
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
English (`en`)
## Dataset Structure
### Data Instances
```
{
"count": 3,
"hate_speech_annotation": 0,
"offensive_language_annotation": 0,
"neither_annotation": 3,
"label": 2, # "neither"
"tweet": "!!! RT @mayasolovely: As a woman you shouldn't complain about cleaning up your house. & as a man you should always take the trash out...")
}
```
### Data Fields
```
count: (Integer) number of users who coded each tweet (min is 3, sometimes more users coded a tweet when judgments were determined to be unreliable,
hate_speech_annotation: (Integer) number of users who judged the tweet to be hate speech,
offensive_language_annotation: (Integer) number of users who judged the tweet to be offensive,
neither_annotation: (Integer) number of users who judged the tweet to be neither offensive nor non-offensive,
label: (Class Label) class label for majority of CF users (0: 'hate-speech', 1: 'offensive-language' or 2: 'neither'),
tweet: (string)
```
### Data Splits
This dataset is not splitted, only the train split is available.
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
Usernames are not anonymized in the dataset.
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
MIT License
### Citation Information
@inproceedings{hateoffensive,
title = {Automated Hate Speech Detection and the Problem of Offensive Language},
author = {Davidson, Thomas and Warmsley, Dana and Macy, Michael and Weber, Ingmar},
booktitle = {Proceedings of the 11th International AAAI Conference on Web and Social Media},
series = {ICWSM '17},
year = {2017},
location = {Montreal, Canada},
pages = {512-515}
}
### Contributions
Thanks to [@hugoabonizio](https://github.com/hugoabonizio) for adding this dataset. |
HuggingFaceM4/tmp-pmd-synthetic-testing | 2022-10-05T17:16:27.000Z | [
"region:us"
] | HuggingFaceM4 | null | null | null | 1 | 14,799 | Entry not found |
wiki_qa | 2023-04-05T13:43:16.000Z | [
"task_categories:question-answering",
"task_ids:open-domain-qa",
"annotations_creators:crowdsourced",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:other",
"region:us"
] | null | Wiki Question Answering corpus from Microsoft | @InProceedings{YangYihMeek:EMNLP2015:WikiQA,
author = {{Yi}, Yang and {Wen-tau}, Yih and {Christopher} Meek},
title = "{WikiQA: A Challenge Dataset for Open-Domain Question Answering}",
journal = {Association for Computational Linguistics},
year = 2015,
doi = {10.18653/v1/D15-1237},
pages = {2013–2018},
} | null | 16 | 14,642 | ---
annotations_creators:
- crowdsourced
language_creators:
- found
language:
- en
license:
- other
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- question-answering
task_ids:
- open-domain-qa
paperswithcode_id: wikiqa
pretty_name: WikiQA
dataset_info:
features:
- name: question_id
dtype: string
- name: question
dtype: string
- name: document_title
dtype: string
- name: answer
dtype: string
- name: label
dtype:
class_label:
names:
'0': '0'
'1': '1'
splits:
- name: test
num_bytes: 1337903
num_examples: 6165
- name: train
num_bytes: 4469148
num_examples: 20360
- name: validation
num_bytes: 591833
num_examples: 2733
download_size: 7094233
dataset_size: 6398884
---
# Dataset Card for "wiki_qa"
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://www.microsoft.com/en-us/download/details.aspx?id=52419](https://www.microsoft.com/en-us/download/details.aspx?id=52419)
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Paper:** [WikiQA: A Challenge Dataset for Open-Domain Question Answering](https://aclanthology.org/D15-1237/)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 7.10 MB
- **Size of the generated dataset:** 6.40 MB
- **Total amount of disk used:** 13.50 MB
### Dataset Summary
Wiki Question Answering corpus from Microsoft.
The WikiQA corpus is a publicly available set of question and sentence pairs, collected and annotated for research on open-domain question answering.
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Dataset Structure
### Data Instances
#### default
- **Size of downloaded dataset files:** 7.10 MB
- **Size of the generated dataset:** 6.40 MB
- **Total amount of disk used:** 13.50 MB
An example of 'train' looks as follows.
```
{
"answer": "Glacier caves are often called ice caves , but this term is properly used to describe bedrock caves that contain year-round ice.",
"document_title": "Glacier cave",
"label": 0,
"question": "how are glacier caves formed?",
"question_id": "Q1"
}
```
### Data Fields
The data fields are the same among all splits.
#### default
- `question_id`: a `string` feature.
- `question`: a `string` feature.
- `document_title`: a `string` feature.
- `answer`: a `string` feature.
- `label`: a classification label, with possible values including `0` (0), `1` (1).
### Data Splits
| name |train|validation|test|
|-------|----:|---------:|---:|
|default|20360| 2733|6165|
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
MICROSOFT RESEARCH DATA LICENSE AGREEMENT
FOR
MICROSOFT RESEARCH WIKIQA CORPUS
These license terms are an agreement between Microsoft Corporation (or based on where you live, one of its
affiliates) and you. Please read them. They apply to the data associated with this license above, which includes
the media on which you received it, if any. The terms also apply to any Microsoft:
- updates,
- supplements,
- Internet-based services, and
- support services
for this data, unless other terms accompany those items. If so, those terms apply.
BY USING THE DATA, YOU ACCEPT THESE TERMS. IF YOU DO NOT ACCEPT THEM, DO NOT USE THE DATA.
If you comply with these license terms, you have the rights below.
1. SCOPE OF LICENSE.
a. You may use, copy, modify, create derivative works, and distribute the Dataset:
i. for research and technology development purposes only. Examples of research and technology
development uses are teaching, academic research, public demonstrations and experimentation ;
and
ii. to publish (or present papers or articles) on your results from using such Dataset.
b. The data is licensed, not sold. This agreement only gives you some rights to use the data. Microsoft reserves
all other rights. Unless applicable law gives you more rights despite this limitation, you may use the data only
as expressly permitted in this agreement. In doing so, you must comply with any technical limitations in the
data that only allow you to use it in certain ways.
You may not
- work around any technical limitations in the data;
- reverse engineer, decompile or disassemble the data, except and only to the extent that applicable law
expressly permits, despite this limitation;
- rent, lease or lend the data;
- transfer the data or this agreement to any third party; or
- use the data directly in a commercial product without Microsoft’s permission.
2. DISTRIBUTION REQUIREMENTS:
a. If you distribute the Dataset or any derivative works of the Dataset, you will distribute them under the
same terms and conditions as in this Agreement, and you will not grant other rights to the Dataset or
derivative works that are different from those provided by this Agreement.
b. If you have created derivative works of the Dataset, and distribute such derivative works, you will
cause the modified files to carry prominent notices so that recipients know that they are not receiving
Page 1 of 3the original Dataset. Such notices must state: (i) that you have changed the Dataset; and (ii) the date
of any changes.
3. DISTRIBUTION RESTRICTIONS. You may not: (a) alter any copyright, trademark or patent notice in the
Dataset; (b) use Microsoft’s trademarks in a way that suggests your derivative works or modifications come from
or are endorsed by Microsoft; (c) include the Dataset in malicious, deceptive or unlawful programs.
4. OWNERSHIP. Microsoft retains all right, title, and interest in and to any Dataset provided to you under this
Agreement. You acquire no interest in the Dataset you may receive under the terms of this Agreement.
5. LICENSE TO MICROSOFT. Microsoft is granted back, without any restrictions or limitations, a non-exclusive,
perpetual, irrevocable, royalty-free, assignable and sub-licensable license, to reproduce, publicly perform or
display, use, modify, post, distribute, make and have made, sell and transfer your modifications to and/or
derivative works of the Dataset, for any purpose.
6. FEEDBACK. If you give feedback about the Dataset to Microsoft, you give to Microsoft, without charge, the right
to use, share and commercialize your feedback in any way and for any purpose. You also give to third parties,
without charge, any patent rights needed for their products, technologies and services to use or interface with
any specific parts of a Microsoft dataset or service that includes the feedback. You will not give feedback that is
subject to a license that requires Microsoft to license its Dataset or documentation to third parties because we
include your feedback in them. These rights survive this Agreement.
7. EXPORT RESTRICTIONS. The Dataset is subject to United States export laws and regulations. You must
comply with all domestic and international export laws and regulations that apply to the Dataset. These laws
include restrictions on destinations, end users and end use. For additional information, see
www.microsoft.com/exporting.
8. ENTIRE AGREEMENT. This Agreement, and the terms for supplements, updates, Internet-based services and
support services that you use, are the entire agreement for the Dataset.
9. SUPPORT SERVICES. Because this data is “as is,” we may not provide support services for it.
10. APPLICABLE LAW.
a. United States. If you acquired the software in the United States, Washington state law governs the
interpretation of this agreement and applies to claims for breach of it, regardless of conflict of laws principles.
The laws of the state where you live govern all other claims, including claims under state consumer protection
laws, unfair competition laws, and in tort.
b. Outside the United States. If you acquired the software in any other country, the laws of that country
apply.
11. LEGAL EFFECT. This Agreement describes certain legal rights. You may have other rights under the laws of your
country. You may also have rights with respect to the party from whom you acquired the Dataset. This
Agreement does not change your rights under the laws of your country if the laws of your country do not permit
it to do so.
12. DISCLAIMER OF WARRANTY. The Dataset is licensed “as-is.” You bear the risk of using it. Microsoft gives no
express warranties, guarantees or conditions. You may have additional consumer rights or statutory guarantees
under your local laws which this agreement cannot change. To the extent permitted under your local laws,
Microsoft excludes the implied warranties of merchantability, fitness for a particular purpose and non-
infringement.
13. LIMITATION ON AND EXCLUSION OF REMEDIES AND DAMAGES. YOU CAN RECOVER FROM
MICROSOFT AND ITS SUPPLIERS ONLY DIRECT DAMAGES UP TO U.S. $5.00. YOU CANNOT RECOVER ANY
OTHER DAMAGES, INCLUDING CONSEQUENTIAL, LOST PROFITS, SPECIAL, INDIRECT OR INCIDENTAL
DAMAGES.
This limitation applies to
- anything related to the software, services, content (including code) on third party Internet sites, or third party
programs; and Page 2 of 3
- claims for breach of contract, breach of warranty, guarantee or condition, strict liability, negligence, or other
tort to the extent permitted by applicable law.
It also applies even if Microsoft knew or should have known about the possibility of the damages. The above
limitation or exclusion may not apply to you because your country may not allow the exclusion or limitation of
incidental, consequential or other damages.
### Citation Information
```
@inproceedings{yang-etal-2015-wikiqa,
title = "{W}iki{QA}: A Challenge Dataset for Open-Domain Question Answering",
author = "Yang, Yi and
Yih, Wen-tau and
Meek, Christopher",
booktitle = "Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing",
month = sep,
year = "2015",
address = "Lisbon, Portugal",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/D15-1237",
doi = "10.18653/v1/D15-1237",
pages = "2013--2018",
}
```
### Contributions
Thanks to [@patrickvonplaten](https://github.com/patrickvonplaten), [@mariamabarham](https://github.com/mariamabarham), [@lewtun](https://github.com/lewtun), [@thomwolf](https://github.com/thomwolf) for adding this dataset. |
bookcorpus | 2023-04-05T09:41:56.000Z | [
"task_categories:text-generation",
"task_categories:fill-mask",
"task_ids:language-modeling",
"task_ids:masked-language-modeling",
"annotations_creators:no-annotation",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10M<n<100M",
"source_datasets:original",
"language:en",
"license:unknown",
"arxiv:2105.05241",
"region:us"
] | null | Books are a rich source of both fine-grained information, how a character, an object or a scene looks like, as well as high-level semantics, what someone is thinking, feeling and how these states evolve through a story.This work aims to align books to their movie releases in order to providerich descriptive explanations for visual content that go semantically farbeyond the captions available in current datasets. \ | @InProceedings{Zhu_2015_ICCV,
title = {Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books},
author = {Zhu, Yukun and Kiros, Ryan and Zemel, Rich and Salakhutdinov, Ruslan and Urtasun, Raquel and Torralba, Antonio and Fidler, Sanja},
booktitle = {The IEEE International Conference on Computer Vision (ICCV)},
month = {December},
year = {2015}
} | null | 146 | 14,515 | ---
annotations_creators:
- no-annotation
language_creators:
- found
language:
- en
license:
- unknown
multilinguality:
- monolingual
pretty_name: BookCorpus
size_categories:
- 10M<n<100M
source_datasets:
- original
task_categories:
- text-generation
- fill-mask
task_ids:
- language-modeling
- masked-language-modeling
paperswithcode_id: bookcorpus
dataset_info:
features:
- name: text
dtype: string
config_name: plain_text
splits:
- name: train
num_bytes: 4853859824
num_examples: 74004228
download_size: 1179510242
dataset_size: 4853859824
---
# Dataset Card for BookCorpus
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://yknzhu.wixsite.com/mbweb](https://yknzhu.wixsite.com/mbweb)
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 1.18 GB
- **Size of the generated dataset:** 4.85 GB
- **Total amount of disk used:** 6.03 GB
### Dataset Summary
Books are a rich source of both fine-grained information, how a character, an object or a scene looks like, as well as high-level semantics, what someone is thinking, feeling and how these states evolve through a story.This work aims to align books to their movie releases in order to providerich descriptive explanations for visual content that go semantically farbeyond the captions available in current datasets.
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Dataset Structure
### Data Instances
#### plain_text
- **Size of downloaded dataset files:** 1.18 GB
- **Size of the generated dataset:** 4.85 GB
- **Total amount of disk used:** 6.03 GB
An example of 'train' looks as follows.
```
{
"text": "But I traded all my life for some lovin' and some gold"
}
```
### Data Fields
The data fields are the same among all splits.
#### plain_text
- `text`: a `string` feature.
### Data Splits
| name | train |
|----------|-------:|
|plain_text|74004228|
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
The books have been crawled from https://www.smashwords.com, see their [terms of service](https://www.smashwords.com/about/tos) for more information.
A data sheet for this dataset has also been created and published in [Addressing "Documentation Debt" in Machine Learning Research: A Retrospective Datasheet for BookCorpus](https://arxiv.org/abs/2105.05241).
### Citation Information
```
@InProceedings{Zhu_2015_ICCV,
title = {Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books},
author = {Zhu, Yukun and Kiros, Ryan and Zemel, Rich and Salakhutdinov, Ruslan and Urtasun, Raquel and Torralba, Antonio and Fidler, Sanja},
booktitle = {The IEEE International Conference on Computer Vision (ICCV)},
month = {December},
year = {2015}
}
```
### Contributions
Thanks to [@lewtun](https://github.com/lewtun), [@richarddwang](https://github.com/richarddwang), [@lhoestq](https://github.com/lhoestq), [@thomwolf](https://github.com/thomwolf) for adding this dataset. |
knkarthick/dialogsum | 2023-10-03T10:56:21.000Z | [
"task_categories:summarization",
"task_categories:text2text-generation",
"task_categories:text-generation",
"annotations_creators:expert-generated",
"language_creators:expert-generated",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:cc-by-nc-sa-4.0",
"dialogue-summary",
"one-liner-summary",
"meeting-title",
"email-subject",
"region:us"
] | knkarthick | null | null | null | 74 | 14,230 | ---
annotations_creators:
- expert-generated
language_creators:
- expert-generated
language:
- en
license: cc-by-nc-sa-4.0
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- summarization
- text2text-generation
- text-generation
task_ids: []
pretty_name: DIALOGSum Corpus
tags:
- dialogue-summary
- one-liner-summary
- meeting-title
- email-subject
---
# Dataset Card for DIALOGSum Corpus
## Dataset Description
### Links
- **Homepage:** https://aclanthology.org/2021.findings-acl.449
- **Repository:** https://github.com/cylnlp/dialogsum
- **Paper:** https://aclanthology.org/2021.findings-acl.449
- **Point of Contact:** https://huggingface.co/knkarthick
### Dataset Summary
DialogSum is a large-scale dialogue summarization dataset, consisting of 13,460 (Plus 100 holdout data for topic generation) dialogues with corresponding manually labeled summaries and topics.
### Languages
English
## Dataset Structure
### Data Instances
DialogSum is a large-scale dialogue summarization dataset, consisting of 13,460 dialogues (+1000 tests) split into train, test and validation.
The first instance in the training set:
{'id': 'train_0', 'summary': "Mr. Smith's getting a check-up, and Doctor Hawkins advises him to have one every year. Hawkins'll give some information about their classes and medications to help Mr. Smith quit smoking.", 'dialogue': "#Person1#: Hi, Mr. Smith. I'm Doctor Hawkins. Why are you here today?\n#Person2#: I found it would be a good idea to get a check-up.\n#Person1#: Yes, well, you haven't had one for 5 years. You should have one every year.\n#Person2#: I know. I figure as long as there is nothing wrong, why go see the doctor?\n#Person1#: Well, the best way to avoid serious illnesses is to find out about them early. So try to come at least once a year for your own good.\n#Person2#: Ok.\n#Person1#: Let me see here. Your eyes and ears look fine. Take a deep breath, please. Do you smoke, Mr. Smith?\n#Person2#: Yes.\n#Person1#: Smoking is the leading cause of lung cancer and heart disease, you know. You really should quit.\n#Person2#: I've tried hundreds of times, but I just can't seem to kick the habit.\n#Person1#: Well, we have classes and some medications that might help. I'll give you more information before you leave.\n#Person2#: Ok, thanks doctor.", 'topic': "get a check-up}
### Data Fields
- dialogue: text of dialogue.
- summary: human written summary of the dialogue.
- topic: human written topic/one liner of the dialogue.
- id: unique file id of an example.
### Data Splits
- train: 12460
- val: 500
- test: 1500
- holdout: 100 [Only 3 features: id, dialogue, topic]
## Dataset Creation
### Curation Rationale
In paper:
We collect dialogue data for DialogSum from three public dialogue corpora, namely Dailydialog (Li et al., 2017), DREAM (Sun et al., 2019) and MuTual (Cui et al., 2019), as well as an English speaking practice website. These datasets contain face-to-face spoken dialogues that cover a wide range of daily-life topics, including schooling, work, medication, shopping, leisure, travel. Most conversations take place between friends, colleagues, and between service providers and customers.
Compared with previous datasets, dialogues from DialogSum have distinct characteristics:
Under rich real-life scenarios, including more diverse task-oriented scenarios;
Have clear communication patterns and intents, which is valuable to serve as summarization sources;
Have a reasonable length, which comforts the purpose of automatic summarization.
We ask annotators to summarize each dialogue based on the following criteria:
Convey the most salient information;
Be brief;
Preserve important named entities within the conversation;
Be written from an observer perspective;
Be written in formal language.
### Who are the source language producers?
linguists
### Who are the annotators?
language experts
## Licensing Information
CC BY-NC-SA 4.0
## Citation Information
```
@inproceedings{chen-etal-2021-dialogsum,
title = "{D}ialog{S}um: {A} Real-Life Scenario Dialogue Summarization Dataset",
author = "Chen, Yulong and
Liu, Yang and
Chen, Liang and
Zhang, Yue",
booktitle = "Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021",
month = aug,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.findings-acl.449",
doi = "10.18653/v1/2021.findings-acl.449",
pages = "5062--5074",
```
## Contributions
Thanks to [@cylnlp](https://github.com/cylnlp) for adding this dataset. |
mozilla-foundation/common_voice_13_0 | 2023-06-26T15:23:12.000Z | [
"task_categories:automatic-speech-recognition",
"annotations_creators:crowdsourced",
"language_creators:crowdsourced",
"multilinguality:multilingual",
"source_datasets:extended|common_voice",
"license:cc0-1.0",
"arxiv:1912.06670",
"region:us"
] | mozilla-foundation | null | @inproceedings{commonvoice:2020,
author = {Ardila, R. and Branson, M. and Davis, K. and Henretty, M. and Kohler, M. and Meyer, J. and Morais, R. and Saunders, L. and Tyers, F. M. and Weber, G.},
title = {Common Voice: A Massively-Multilingual Speech Corpus},
booktitle = {Proceedings of the 12th Conference on Language Resources and Evaluation (LREC 2020)},
pages = {4211--4215},
year = 2020
} | null | 74 | 14,203 | ---
pretty_name: Common Voice Corpus 13.0
annotations_creators:
- crowdsourced
language_creators:
- crowdsourced
language_bcp47:
- ab
- ar
- as
- ast
- az
- ba
- bas
- be
- bg
- bn
- br
- ca
- ckb
- cnh
- cs
- cv
- cy
- da
- de
- dv
- dyu
- el
- en
- eo
- es
- et
- eu
- fa
- fi
- fr
- fy-NL
- ga-IE
- gl
- gn
- ha
- hi
- hsb
- hu
- hy-AM
- ia
- id
- ig
- is
- it
- ja
- ka
- kab
- kk
- kmr
- ko
- ky
- lg
- lo
- lt
- lv
- mdf
- mhr
- mk
- ml
- mn
- mr
- mrj
- mt
- myv
- nan-tw
- ne-NP
- nl
- nn-NO
- oc
- or
- pa-IN
- pl
- pt
- quy
- rm-sursilv
- rm-vallader
- ro
- ru
- rw
- sah
- sat
- sc
- sk
- skr
- sl
- sr
- sv-SE
- sw
- ta
- th
- ti
- tig
- tk
- tok
- tr
- tt
- tw
- ug
- uk
- ur
- uz
- vi
- vot
- yo
- yue
- zh-CN
- zh-HK
- zh-TW
license:
- cc0-1.0
multilinguality:
- multilingual
size_categories:
ab:
- 10K<n<100K
ar:
- 100K<n<1M
as:
- 1K<n<10K
ast:
- 1K<n<10K
az:
- n<1K
ba:
- 100K<n<1M
bas:
- 1K<n<10K
be:
- 1M<n<10M
bg:
- 10K<n<100K
bn:
- 1M<n<10M
br:
- 10K<n<100K
ca:
- 1M<n<10M
ckb:
- 100K<n<1M
cnh:
- 1K<n<10K
cs:
- 100K<n<1M
cv:
- 10K<n<100K
cy:
- 100K<n<1M
da:
- 10K<n<100K
de:
- 100K<n<1M
dv:
- 10K<n<100K
dyu:
- n<1K
el:
- 10K<n<100K
en:
- 1M<n<10M
eo:
- 1M<n<10M
es:
- 1M<n<10M
et:
- 10K<n<100K
eu:
- 100K<n<1M
fa:
- 100K<n<1M
fi:
- 10K<n<100K
fr:
- 100K<n<1M
fy-NL:
- 100K<n<1M
ga-IE:
- 10K<n<100K
gl:
- 10K<n<100K
gn:
- 1K<n<10K
ha:
- 10K<n<100K
hi:
- 10K<n<100K
hsb:
- 1K<n<10K
hu:
- 10K<n<100K
hy-AM:
- 1K<n<10K
ia:
- 10K<n<100K
id:
- 10K<n<100K
ig:
- 1K<n<10K
is:
- n<1K
it:
- 100K<n<1M
ja:
- 100K<n<1M
ka:
- 10K<n<100K
kab:
- 100K<n<1M
kk:
- 1K<n<10K
kmr:
- 10K<n<100K
ko:
- 1K<n<10K
ky:
- 10K<n<100K
lg:
- 100K<n<1M
lo:
- n<1K
lt:
- 10K<n<100K
lv:
- 10K<n<100K
mdf:
- n<1K
mhr:
- 100K<n<1M
mk:
- n<1K
ml:
- 1K<n<10K
mn:
- 10K<n<100K
mr:
- 10K<n<100K
mrj:
- 10K<n<100K
mt:
- 10K<n<100K
myv:
- 1K<n<10K
nan-tw:
- 10K<n<100K
ne-NP:
- n<1K
nl:
- 10K<n<100K
nn-NO:
- n<1K
oc:
- 1K<n<10K
or:
- 1K<n<10K
pa-IN:
- 1K<n<10K
pl:
- 100K<n<1M
pt:
- 100K<n<1M
quy:
- n<1K
rm-sursilv:
- 1K<n<10K
rm-vallader:
- 1K<n<10K
ro:
- 10K<n<100K
ru:
- 100K<n<1M
rw:
- 1M<n<10M
sah:
- 1K<n<10K
sat:
- n<1K
sc:
- 1K<n<10K
sk:
- 10K<n<100K
skr:
- 1K<n<10K
sl:
- 10K<n<100K
sr:
- 1K<n<10K
sv-SE:
- 10K<n<100K
sw:
- 100K<n<1M
ta:
- 100K<n<1M
th:
- 100K<n<1M
ti:
- n<1K
tig:
- n<1K
tk:
- 1K<n<10K
tok:
- 10K<n<100K
tr:
- 10K<n<100K
tt:
- 10K<n<100K
tw:
- n<1K
ug:
- 10K<n<100K
uk:
- 10K<n<100K
ur:
- 100K<n<1M
uz:
- 100K<n<1M
vi:
- 10K<n<100K
vot:
- n<1K
yo:
- 1K<n<10K
yue:
- 10K<n<100K
zh-CN:
- 100K<n<1M
zh-HK:
- 100K<n<1M
zh-TW:
- 100K<n<1M
source_datasets:
- extended|common_voice
task_categories:
- automatic-speech-recognition
paperswithcode_id: common-voice
extra_gated_prompt: "By clicking on “Access repository” below, you also agree to not attempt to determine the identity of speakers in the Common Voice dataset."
---
# Dataset Card for Common Voice Corpus 13.0
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [How to use](#how-to-use)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** https://commonvoice.mozilla.org/en/datasets
- **Repository:** https://github.com/common-voice/common-voice
- **Paper:** https://arxiv.org/abs/1912.06670
- **Leaderboard:** https://paperswithcode.com/dataset/common-voice
- **Point of Contact:** [Vaibhav Srivastav](mailto:vaibhav@huggingface.co)
### Dataset Summary
The Common Voice dataset consists of a unique MP3 and corresponding text file.
Many of the 27141 recorded hours in the dataset also include demographic metadata like age, sex, and accent
that can help improve the accuracy of speech recognition engines.
The dataset currently consists of 17689 validated hours in 108 languages, but more voices and languages are always added.
Take a look at the [Languages](https://commonvoice.mozilla.org/en/languages) page to request a language or start contributing.
### Supported Tasks and Leaderboards
The results for models trained on the Common Voice datasets are available via the
[🤗 Autoevaluate Leaderboard](https://huggingface.co/spaces/autoevaluate/leaderboards?dataset=mozilla-foundation%2Fcommon_voice_11_0&only_verified=0&task=automatic-speech-recognition&config=ar&split=test&metric=wer)
### Languages
```
Abkhaz, Arabic, Armenian, Assamese, Asturian, Azerbaijani, Basaa, Bashkir, Basque, Belarusian, Bengali, Breton, Bulgarian, Cantonese, Catalan, Central Kurdish, Chinese (China), Chinese (Hong Kong), Chinese (Taiwan), Chuvash, Czech, Danish, Dhivehi, Dioula, Dutch, English, Erzya, Esperanto, Estonian, Finnish, French, Frisian, Galician, Georgian, German, Greek, Guarani, Hakha Chin, Hausa, Hill Mari, Hindi, Hungarian, Icelandic, Igbo, Indonesian, Interlingua, Irish, Italian, Japanese, Kabyle, Kazakh, Kinyarwanda, Korean, Kurmanji Kurdish, Kyrgyz, Lao, Latvian, Lithuanian, Luganda, Macedonian, Malayalam, Maltese, Marathi, Meadow Mari, Moksha, Mongolian, Nepali, Norwegian Nynorsk, Occitan, Odia, Persian, Polish, Portuguese, Punjabi, Quechua Chanka, Romanian, Romansh Sursilvan, Romansh Vallader, Russian, Sakha, Santali (Ol Chiki), Saraiki, Sardinian, Serbian, Slovak, Slovenian, Sorbian, Upper, Spanish, Swahili, Swedish, Taiwanese (Minnan), Tamil, Tatar, Thai, Tigre, Tigrinya, Toki Pona, Turkish, Turkmen, Twi, Ukrainian, Urdu, Uyghur, Uzbek, Vietnamese, Votic, Welsh, Yoruba
```
## How to use
The `datasets` library allows you to load and pre-process your dataset in pure Python, at scale. The dataset can be downloaded and prepared in one call to your local drive by using the `load_dataset` function.
For example, to download the Hindi config, simply specify the corresponding language config name (i.e., "hi" for Hindi):
```python
from datasets import load_dataset
cv_13 = load_dataset("mozilla-foundation/common_voice_13_0", "hi", split="train")
```
Using the datasets library, you can also stream the dataset on-the-fly by adding a `streaming=True` argument to the `load_dataset` function call. Loading a dataset in streaming mode loads individual samples of the dataset at a time, rather than downloading the entire dataset to disk.
```python
from datasets import load_dataset
cv_13 = load_dataset("mozilla-foundation/common_voice_13_0", "hi", split="train", streaming=True)
print(next(iter(cv_13)))
```
*Bonus*: create a [PyTorch dataloader](https://huggingface.co/docs/datasets/use_with_pytorch) directly with your own datasets (local/streamed).
### Local
```python
from datasets import load_dataset
from torch.utils.data.sampler import BatchSampler, RandomSampler
cv_13 = load_dataset("mozilla-foundation/common_voice_13_0", "hi", split="train")
batch_sampler = BatchSampler(RandomSampler(cv_13), batch_size=32, drop_last=False)
dataloader = DataLoader(cv_13, batch_sampler=batch_sampler)
```
### Streaming
```python
from datasets import load_dataset
from torch.utils.data import DataLoader
cv_13 = load_dataset("mozilla-foundation/common_voice_13_0", "hi", split="train")
dataloader = DataLoader(cv_13, batch_size=32)
```
To find out more about loading and preparing audio datasets, head over to [hf.co/blog/audio-datasets](https://huggingface.co/blog/audio-datasets).
### Example scripts
Train your own CTC or Seq2Seq Automatic Speech Recognition models on Common Voice 13 with `transformers` - [here](https://github.com/huggingface/transformers/tree/main/examples/pytorch/speech-recognition).
## Dataset Structure
### Data Instances
A typical data point comprises the `path` to the audio file and its `sentence`.
Additional fields include `accent`, `age`, `client_id`, `up_votes`, `down_votes`, `gender`, `locale` and `segment`.
```python
{
'client_id': 'd59478fbc1ee646a28a3c652a119379939123784d99131b865a89f8b21c81f69276c48bd574b81267d9d1a77b83b43e6d475a6cfc79c232ddbca946ae9c7afc5',
'path': 'et/clips/common_voice_et_18318995.mp3',
'audio': {
'path': 'et/clips/common_voice_et_18318995.mp3',
'array': array([-0.00048828, -0.00018311, -0.00137329, ..., 0.00079346, 0.00091553, 0.00085449], dtype=float32),
'sampling_rate': 48000
},
'sentence': 'Tasub kokku saada inimestega, keda tunned juba ammust ajast saati.',
'up_votes': 2,
'down_votes': 0,
'age': 'twenties',
'gender': 'male',
'accent': '',
'locale': 'et',
'segment': ''
}
```
### Data Fields
`client_id` (`string`): An id for which client (voice) made the recording
`path` (`string`): The path to the audio file
`audio` (`dict`): A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: `dataset[0]["audio"]` the audio file is automatically decoded and resampled to `dataset.features["audio"].sampling_rate`. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the `"audio"` column, *i.e.* `dataset[0]["audio"]` should **always** be preferred over `dataset["audio"][0]`.
`sentence` (`string`): The sentence the user was prompted to speak
`up_votes` (`int64`): How many upvotes the audio file has received from reviewers
`down_votes` (`int64`): How many downvotes the audio file has received from reviewers
`age` (`string`): The age of the speaker (e.g. `teens`, `twenties`, `fifties`)
`gender` (`string`): The gender of the speaker
`accent` (`string`): Accent of the speaker
`locale` (`string`): The locale of the speaker
`segment` (`string`): Usually an empty field
### Data Splits
The speech material has been subdivided into portions for dev, train, test, validated, invalidated, reported and other.
The validated data is data that has been validated with reviewers and received upvotes that the data is of high quality.
The invalidated data is data has been invalidated by reviewers
and received downvotes indicating that the data is of low quality.
The reported data is data that has been reported, for different reasons.
The other data is data that has not yet been reviewed.
The dev, test, train are all data that has been reviewed, deemed of high quality and split into dev, test and train.
## Data Preprocessing Recommended by Hugging Face
The following are data preprocessing steps advised by the Hugging Face team. They are accompanied by an example code snippet that shows how to put them to practice.
Many examples in this dataset have trailing quotations marks, e.g _“the cat sat on the mat.“_. These trailing quotation marks do not change the actual meaning of the sentence, and it is near impossible to infer whether a sentence is a quotation or not a quotation from audio data alone. In these cases, it is advised to strip the quotation marks, leaving: _the cat sat on the mat_.
In addition, the majority of training sentences end in punctuation ( . or ? or ! ), whereas just a small proportion do not. In the dev set, **almost all** sentences end in punctuation. Thus, it is recommended to append a full-stop ( . ) to the end of the small number of training examples that do not end in punctuation.
```python
from datasets import load_dataset
ds = load_dataset("mozilla-foundation/common_voice_13_0", "en", use_auth_token=True)
def prepare_dataset(batch):
"""Function to preprocess the dataset with the .map method"""
transcription = batch["sentence"]
if transcription.startswith('"') and transcription.endswith('"'):
# we can remove trailing quotation marks as they do not affect the transcription
transcription = transcription[1:-1]
if transcription[-1] not in [".", "?", "!"]:
# append a full-stop to sentences that do not end in punctuation
transcription = transcription + "."
batch["sentence"] = transcription
return batch
ds = ds.map(prepare_dataset, desc="preprocess dataset")
```
## Dataset Creation
### Curation Rationale
[Needs More Information]
### Source Data
#### Initial Data Collection and Normalization
[Needs More Information]
#### Who are the source language producers?
[Needs More Information]
### Annotations
#### Annotation process
[Needs More Information]
#### Who are the annotators?
[Needs More Information]
### Personal and Sensitive Information
The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in the Common Voice dataset.
## Considerations for Using the Data
### Social Impact of Dataset
The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in the Common Voice dataset.
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
Public Domain, [CC-0](https://creativecommons.org/share-your-work/public-domain/cc0/)
### Citation Information
```
@inproceedings{commonvoice:2020,
author = {Ardila, R. and Branson, M. and Davis, K. and Henretty, M. and Kohler, M. and Meyer, J. and Morais, R. and Saunders, L. and Tyers, F. M. and Weber, G.},
title = {Common Voice: A Massively-Multilingual Speech Corpus},
booktitle = {Proceedings of the 12th Conference on Language Resources and Evaluation (LREC 2020)},
pages = {4211--4215},
year = 2020
}
``` |
mlabonne/guanaco-llama2-1k | 2023-08-25T16:49:41.000Z | [
"region:us"
] | mlabonne | null | null | null | 50 | 14,161 | ---
dataset_info:
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 1654448
num_examples: 1000
download_size: 966693
dataset_size: 1654448
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Guanaco-1k: Lazy Llama 2 Formatting
This is a subset (1000 samples) of the excellent [`timdettmers/openassistant-guanaco`](https://huggingface.co/datasets/timdettmers/openassistant-guanaco) dataset, processed to match Llama 2's prompt format as described [in this article](https://huggingface.co/blog/llama2#how-to-prompt-llama-2). It was created using the following [colab notebook](https://colab.research.google.com/drive/1Ad7a9zMmkxuXTOh1Z7-rNSICA4dybpM2?usp=sharing).
Useful if you don't want to reformat it by yourself (e.g., using a script). It was designed for [this article](https://mlabonne.github.io/blog/posts/Fine_Tune_Your_Own_Llama_2_Model_in_a_Colab_Notebook.html) about fine-tuning a Llama 2 (chat) model in a Google Colab.
|
tatsu-lab/alpaca_eval | 2023-06-09T11:58:42.000Z | [
"license:cc-by-nc-4.0",
"region:us"
] | tatsu-lab | Data for alpaca_eval, which aims to help automatic evaluation of instruction-following models | @misc{alpaca_eval,
author = {Xuechen Li and Tianyi Zhang and Yann Dubois and Rohan Taori and Ishaan Gulrajani and Carlos Guestrin and Percy Liang and Tatsunori B. Hashimoto },
title = {AlpacaEval: An Automatic Evaluator of Instruction-following Models},
year = {2023},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\\url{https://github.com/tatsu-lab/alpaca_eval}}
} | null | 16 | 13,828 | ---
license: cc-by-nc-4.0
---
|
iohadrubin/c5 | 2023-10-07T06:13:07.000Z | [
"region:us"
] | iohadrubin | A colossal, cleaned version of Common Crawl's web crawl corpus.
Based on Common Crawl dataset: "https://commoncrawl.org".
This is the processed version of Google's C5 dataset by AllenAI. | @article{2019t5,
author = {Colin Raffel and Noam Shazeer and Adam Roberts and Katherine Lee and Sharan Narang and Michael Matena and Yanqi Zhou and Wei Li and Peter J. Liu},
title = {Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer},
journal = {arXiv e-prints},
year = {2019},
archivePrefix = {arXiv},
eprint = {1910.10683},
} | null | 0 | 13,792 | Entry not found |
librispeech_asr | 2022-11-18T20:18:42.000Z | [
"task_categories:automatic-speech-recognition",
"task_categories:audio-classification",
"task_ids:speaker-identification",
"annotations_creators:expert-generated",
"language_creators:crowdsourced",
"language_creators:expert-generated",
"multilinguality:monolingual",
"size_categories:100K<n<1M",
"source_datasets:original",
"language:en",
"license:cc-by-4.0",
"region:us"
] | null | LibriSpeech is a corpus of approximately 1000 hours of read English speech with sampling rate of 16 kHz,
prepared by Vassil Panayotov with the assistance of Daniel Povey. The data is derived from read
audiobooks from the LibriVox project, and has been carefully segmented and aligned.87 | @inproceedings{panayotov2015librispeech,
title={Librispeech: an ASR corpus based on public domain audio books},
author={Panayotov, Vassil and Chen, Guoguo and Povey, Daniel and Khudanpur, Sanjeev},
booktitle={Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on},
pages={5206--5210},
year={2015},
organization={IEEE}
} | null | 57 | 13,218 | ---
pretty_name: LibriSpeech
annotations_creators:
- expert-generated
language_creators:
- crowdsourced
- expert-generated
language:
- en
license:
- cc-by-4.0
multilinguality:
- monolingual
paperswithcode_id: librispeech-1
size_categories:
- 100K<n<1M
source_datasets:
- original
task_categories:
- automatic-speech-recognition
- audio-classification
task_ids:
- speaker-identification
dataset_info:
- config_name: clean
features:
- name: file
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 16000
- name: text
dtype: string
- name: speaker_id
dtype: int64
- name: chapter_id
dtype: int64
- name: id
dtype: string
splits:
- name: train.100
num_bytes: 6619683041
num_examples: 28539
- name: train.360
num_bytes: 23898214592
num_examples: 104014
- name: validation
num_bytes: 359572231
num_examples: 2703
- name: test
num_bytes: 367705423
num_examples: 2620
download_size: 30121377654
dataset_size: 31245175287
- config_name: other
features:
- name: file
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 16000
- name: text
dtype: string
- name: speaker_id
dtype: int64
- name: chapter_id
dtype: int64
- name: id
dtype: string
splits:
- name: train.500
num_bytes: 31810256902
num_examples: 148688
- name: validation
num_bytes: 337283304
num_examples: 2864
- name: test
num_bytes: 352396474
num_examples: 2939
download_size: 31236565377
dataset_size: 32499936680
- config_name: all
features:
- name: file
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 16000
- name: text
dtype: string
- name: speaker_id
dtype: int64
- name: chapter_id
dtype: int64
- name: id
dtype: string
splits:
- name: train.clean.100
num_bytes: 6627791685
num_examples: 28539
- name: train.clean.360
num_bytes: 23927767570
num_examples: 104014
- name: train.other.500
num_bytes: 31852502880
num_examples: 148688
- name: validation.clean
num_bytes: 359505691
num_examples: 2703
- name: validation.other
num_bytes: 337213112
num_examples: 2864
- name: test.clean
num_bytes: 368449831
num_examples: 2620
- name: test.other
num_bytes: 353231518
num_examples: 2939
download_size: 61357943031
dataset_size: 63826462287
---
# Dataset Card for librispeech_asr
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [LibriSpeech ASR corpus](http://www.openslr.org/12)
- **Repository:** [Needs More Information]
- **Paper:** [LibriSpeech: An ASR Corpus Based On Public Domain Audio Books](https://www.danielpovey.com/files/2015_icassp_librispeech.pdf)
- **Leaderboard:** [The 🤗 Speech Bench](https://huggingface.co/spaces/huggingface/hf-speech-bench)
- **Point of Contact:** [Daniel Povey](mailto:dpovey@gmail.com)
### Dataset Summary
LibriSpeech is a corpus of approximately 1000 hours of 16kHz read English speech, prepared by Vassil Panayotov with the assistance of Daniel Povey. The data is derived from read audiobooks from the LibriVox project, and has been carefully segmented and aligned.
### Supported Tasks and Leaderboards
- `automatic-speech-recognition`, `audio-speaker-identification`: The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). The task has an active Hugging Face leaderboard which can be found at https://huggingface.co/spaces/huggingface/hf-speech-bench. The leaderboard ranks models uploaded to the Hub based on their WER. An external leaderboard at https://paperswithcode.com/sota/speech-recognition-on-librispeech-test-clean ranks the latest models from research and academia.
### Languages
The audio is in English. There are two configurations: `clean` and `other`.
The speakers in the corpus were ranked according to the WER of the transcripts of a model trained on
a different dataset, and were divided roughly in the middle,
with the lower-WER speakers designated as "clean" and the higher WER speakers designated as "other".
## Dataset Structure
### Data Instances
A typical data point comprises the path to the audio file, usually called `file` and its transcription, called `text`. Some additional information about the speaker and the passage which contains the transcription is provided.
```
{'chapter_id': 141231,
'file': '/home/patrick/.cache/huggingface/datasets/downloads/extracted/b7ded9969e09942ab65313e691e6fc2e12066192ee8527e21d634aca128afbe2/dev_clean/1272/141231/1272-141231-0000.flac',
'audio': {'path': '/home/patrick/.cache/huggingface/datasets/downloads/extracted/b7ded9969e09942ab65313e691e6fc2e12066192ee8527e21d634aca128afbe2/dev_clean/1272/141231/1272-141231-0000.flac',
'array': array([-0.00048828, -0.00018311, -0.00137329, ..., 0.00079346,
0.00091553, 0.00085449], dtype=float32),
'sampling_rate': 16000},
'id': '1272-141231-0000',
'speaker_id': 1272,
'text': 'A MAN SAID TO THE UNIVERSE SIR I EXIST'}
```
### Data Fields
- file: A path to the downloaded audio file in .flac format.
- audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: `dataset[0]["audio"]` the audio file is automatically decoded and resampled to `dataset.features["audio"].sampling_rate`. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the `"audio"` column, *i.e.* `dataset[0]["audio"]` should **always** be preferred over `dataset["audio"][0]`.
- text: the transcription of the audio file.
- id: unique id of the data sample.
- speaker_id: unique id of the speaker. The same speaker id can be found for multiple data samples.
- chapter_id: id of the audiobook chapter which includes the transcription.
### Data Splits
The size of the corpus makes it impractical, or at least inconvenient
for some users, to distribute it as a single large archive. Thus the
training portion of the corpus is split into three subsets, with approximate size 100, 360 and 500 hours respectively.
A simple automatic
procedure was used to select the audio in the first two sets to be, on
average, of higher recording quality and with accents closer to US
English. An acoustic model was trained on WSJ’s si-84 data subset
and was used to recognize the audio in the corpus, using a bigram
LM estimated on the text of the respective books. We computed the
Word Error Rate (WER) of this automatic transcript relative to our
reference transcripts obtained from the book texts.
The speakers in the corpus were ranked according to the WER of
the WSJ model’s transcripts, and were divided roughly in the middle,
with the lower-WER speakers designated as "clean" and the higher-WER speakers designated as "other".
For "clean", the data is split into train, validation, and test set. The train set is further split into train.100 and train.360
respectively accounting for 100h and 360h of the training data.
For "other", the data is split into train, validation, and test set. The train set contains approximately 500h of recorded speech.
| | Train.500 | Train.360 | Train.100 | Valid | Test |
| ----- | ------ | ----- | ---- | ---- | ---- |
| clean | - | 104014 | 28539 | 2703 | 2620|
| other | 148688 | - | - | 2864 | 2939 |
## Dataset Creation
### Curation Rationale
[Needs More Information]
### Source Data
#### Initial Data Collection and Normalization
[Needs More Information]
#### Who are the source language producers?
[Needs More Information]
### Annotations
#### Annotation process
[Needs More Information]
#### Who are the annotators?
[Needs More Information]
### Personal and Sensitive Information
The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in this dataset.
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[Needs More Information]
## Additional Information
### Dataset Curators
The dataset was initially created by Vassil Panayotov, Guoguo Chen, Daniel Povey, and Sanjeev Khudanpur.
### Licensing Information
[CC BY 4.0](https://creativecommons.org/licenses/by/4.0/)
### Citation Information
```
@inproceedings{panayotov2015librispeech,
title={Librispeech: an ASR corpus based on public domain audio books},
author={Panayotov, Vassil and Chen, Guoguo and Povey, Daniel and Khudanpur, Sanjeev},
booktitle={Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on},
pages={5206--5210},
year={2015},
organization={IEEE}
}
```
### Contributions
Thanks to [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset. |
cifar100 | 2023-01-25T14:27:57.000Z | [
"task_categories:image-classification",
"annotations_creators:crowdsourced",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:extended|other-80-Million-Tiny-Images",
"language:en",
"license:unknown",
"region:us"
] | null | The CIFAR-100 dataset consists of 60000 32x32 colour images in 100 classes, with 600 images
per class. There are 500 training images and 100 testing images per class. There are 50000 training images and 10000 test images. The 100 classes are grouped into 20 superclasses.
There are two labels per image - fine label (actual class) and coarse label (superclass). | @TECHREPORT{Krizhevsky09learningmultiple,
author = {Alex Krizhevsky},
title = {Learning multiple layers of features from tiny images},
institution = {},
year = {2009}
} | null | 14 | 13,190 | ---
annotations_creators:
- crowdsourced
language_creators:
- found
language:
- en
license:
- unknown
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- extended|other-80-Million-Tiny-Images
task_categories:
- image-classification
task_ids: []
paperswithcode_id: cifar-100
pretty_name: Cifar100
dataset_info:
features:
- name: img
dtype: image
- name: fine_label
dtype:
class_label:
names:
'0': apple
'1': aquarium_fish
'2': baby
'3': bear
'4': beaver
'5': bed
'6': bee
'7': beetle
'8': bicycle
'9': bottle
'10': bowl
'11': boy
'12': bridge
'13': bus
'14': butterfly
'15': camel
'16': can
'17': castle
'18': caterpillar
'19': cattle
'20': chair
'21': chimpanzee
'22': clock
'23': cloud
'24': cockroach
'25': couch
'26': cra
'27': crocodile
'28': cup
'29': dinosaur
'30': dolphin
'31': elephant
'32': flatfish
'33': forest
'34': fox
'35': girl
'36': hamster
'37': house
'38': kangaroo
'39': keyboard
'40': lamp
'41': lawn_mower
'42': leopard
'43': lion
'44': lizard
'45': lobster
'46': man
'47': maple_tree
'48': motorcycle
'49': mountain
'50': mouse
'51': mushroom
'52': oak_tree
'53': orange
'54': orchid
'55': otter
'56': palm_tree
'57': pear
'58': pickup_truck
'59': pine_tree
'60': plain
'61': plate
'62': poppy
'63': porcupine
'64': possum
'65': rabbit
'66': raccoon
'67': ray
'68': road
'69': rocket
'70': rose
'71': sea
'72': seal
'73': shark
'74': shrew
'75': skunk
'76': skyscraper
'77': snail
'78': snake
'79': spider
'80': squirrel
'81': streetcar
'82': sunflower
'83': sweet_pepper
'84': table
'85': tank
'86': telephone
'87': television
'88': tiger
'89': tractor
'90': train
'91': trout
'92': tulip
'93': turtle
'94': wardrobe
'95': whale
'96': willow_tree
'97': wolf
'98': woman
'99': worm
- name: coarse_label
dtype:
class_label:
names:
'0': aquatic_mammals
'1': fish
'2': flowers
'3': food_containers
'4': fruit_and_vegetables
'5': household_electrical_devices
'6': household_furniture
'7': insects
'8': large_carnivores
'9': large_man-made_outdoor_things
'10': large_natural_outdoor_scenes
'11': large_omnivores_and_herbivores
'12': medium_mammals
'13': non-insect_invertebrates
'14': people
'15': reptiles
'16': small_mammals
'17': trees
'18': vehicles_1
'19': vehicles_2
config_name: cifar100
splits:
- name: train
num_bytes: 112751396
num_examples: 50000
- name: test
num_bytes: 22605519
num_examples: 10000
download_size: 169001437
dataset_size: 135356915
---
# Dataset Card for CIFAR-100
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [CIFAR Datasets](https://www.cs.toronto.edu/~kriz/cifar.html)
- **Repository:**
- **Paper:** [Paper](https://www.cs.toronto.edu/~kriz/learning-features-2009-TR.pdf)
- **Leaderboard:**
- **Point of Contact:**
### Dataset Summary
The CIFAR-100 dataset consists of 60000 32x32 colour images in 100 classes, with 600 images
per class. There are 500 training images and 100 testing images per class. There are 50000 training images and 10000 test images. The 100 classes are grouped into 20 superclasses.
There are two labels per image - fine label (actual class) and coarse label (superclass).
### Supported Tasks and Leaderboards
- `image-classification`: The goal of this task is to classify a given image into one of 100 classes. The leaderboard is available [here](https://paperswithcode.com/sota/image-classification-on-cifar-100).
### Languages
English
## Dataset Structure
### Data Instances
A sample from the training set is provided below:
```
{
'img': <PIL.PngImagePlugin.PngImageFile image mode=RGB size=32x32 at 0x2767F58E080>, 'fine_label': 19,
'coarse_label': 11
}
```
### Data Fields
- `img`: A `PIL.Image.Image` object containing the 32x32 image. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]`
- `fine_label`: an `int` classification label with the following mapping:
`0`: apple
`1`: aquarium_fish
`2`: baby
`3`: bear
`4`: beaver
`5`: bed
`6`: bee
`7`: beetle
`8`: bicycle
`9`: bottle
`10`: bowl
`11`: boy
`12`: bridge
`13`: bus
`14`: butterfly
`15`: camel
`16`: can
`17`: castle
`18`: caterpillar
`19`: cattle
`20`: chair
`21`: chimpanzee
`22`: clock
`23`: cloud
`24`: cockroach
`25`: couch
`26`: cra
`27`: crocodile
`28`: cup
`29`: dinosaur
`30`: dolphin
`31`: elephant
`32`: flatfish
`33`: forest
`34`: fox
`35`: girl
`36`: hamster
`37`: house
`38`: kangaroo
`39`: keyboard
`40`: lamp
`41`: lawn_mower
`42`: leopard
`43`: lion
`44`: lizard
`45`: lobster
`46`: man
`47`: maple_tree
`48`: motorcycle
`49`: mountain
`50`: mouse
`51`: mushroom
`52`: oak_tree
`53`: orange
`54`: orchid
`55`: otter
`56`: palm_tree
`57`: pear
`58`: pickup_truck
`59`: pine_tree
`60`: plain
`61`: plate
`62`: poppy
`63`: porcupine
`64`: possum
`65`: rabbit
`66`: raccoon
`67`: ray
`68`: road
`69`: rocket
`70`: rose
`71`: sea
`72`: seal
`73`: shark
`74`: shrew
`75`: skunk
`76`: skyscraper
`77`: snail
`78`: snake
`79`: spider
`80`: squirrel
`81`: streetcar
`82`: sunflower
`83`: sweet_pepper
`84`: table
`85`: tank
`86`: telephone
`87`: television
`88`: tiger
`89`: tractor
`90`: train
`91`: trout
`92`: tulip
`93`: turtle
`94`: wardrobe
`95`: whale
`96`: willow_tree
`97`: wolf
`98`: woman
`99`: worm
- `coarse_label`: an `int` coarse classification label with following mapping:
`0`: aquatic_mammals
`1`: fish
`2`: flowers
`3`: food_containers
`4`: fruit_and_vegetables
`5`: household_electrical_devices
`6`: household_furniture
`7`: insects
`8`: large_carnivores
`9`: large_man-made_outdoor_things
`10`: large_natural_outdoor_scenes
`11`: large_omnivores_and_herbivores
`12`: medium_mammals
`13`: non-insect_invertebrates
`14`: people
`15`: reptiles
`16`: small_mammals
`17`: trees
`18`: vehicles_1
`19`: vehicles_2
### Data Splits
| name |train|test|
|----------|----:|---------:|
|cifar100|50000| 10000|
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
```
@TECHREPORT{Krizhevsky09learningmultiple,
author = {Alex Krizhevsky},
title = {Learning multiple layers of features from tiny images},
institution = {},
year = {2009}
}
```
### Contributions
Thanks to [@gchhablani](https://github.com/gchablani) for adding this dataset. |
mteb/sts22-crosslingual-sts | 2022-09-27T19:10:13.000Z | [
"language:ar",
"language:de",
"language:en",
"language:es",
"language:fr",
"language:it",
"language:pl",
"language:ru",
"language:tr",
"language:zh",
"region:us"
] | mteb | SemEval 2022 Task 8: Multilingual News Article Similarity | \ | null | 4 | 13,067 | ---
language:
- ar
- de
- en
- es
- fr
- it
- pl
- ru
- tr
- zh
---
Scores in this dataset have been inverted to be from least to most similar!
The scores in the original STS22 task were from most to least similar. |
boolq | 2023-04-05T09:42:01.000Z | [
"task_categories:text-classification",
"task_ids:natural-language-inference",
"annotations_creators:crowdsourced",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:cc-by-sa-3.0",
"region:us"
] | null | BoolQ is a question answering dataset for yes/no questions containing 15942 examples. These questions are naturally
occurring ---they are generated in unprompted and unconstrained settings.
Each example is a triplet of (question, passage, answer), with the title of the page as optional additional context.
The text-pair classification setup is similar to existing natural language inference tasks. | @inproceedings{clark2019boolq,
title = {BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions},
author = {Clark, Christopher and Lee, Kenton and Chang, Ming-Wei, and Kwiatkowski, Tom and Collins, Michael, and Toutanova, Kristina},
booktitle = {NAACL},
year = {2019},
} | null | 24 | 12,941 | ---
annotations_creators:
- crowdsourced
language_creators:
- found
language:
- en
license:
- cc-by-sa-3.0
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- natural-language-inference
paperswithcode_id: boolq
pretty_name: BoolQ
dataset_info:
features:
- name: question
dtype: string
- name: answer
dtype: bool
- name: passage
dtype: string
splits:
- name: train
num_bytes: 5829592
num_examples: 9427
- name: validation
num_bytes: 1998190
num_examples: 3270
download_size: 8764539
dataset_size: 7827782
---
# Dataset Card for Boolq
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://github.com/google-research-datasets/boolean-questions](https://github.com/google-research-datasets/boolean-questions)
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 8.77 MB
- **Size of the generated dataset:** 7.83 MB
- **Total amount of disk used:** 16.59 MB
### Dataset Summary
BoolQ is a question answering dataset for yes/no questions containing 15942 examples. These questions are naturally
occurring ---they are generated in unprompted and unconstrained settings.
Each example is a triplet of (question, passage, answer), with the title of the page as optional additional context.
The text-pair classification setup is similar to existing natural language inference tasks.
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Dataset Structure
### Data Instances
#### default
- **Size of downloaded dataset files:** 8.77 MB
- **Size of the generated dataset:** 7.83 MB
- **Total amount of disk used:** 16.59 MB
An example of 'validation' looks as follows.
```
This example was too long and was cropped:
{
"answer": false,
"passage": "\"All biomass goes through at least some of these steps: it needs to be grown, collected, dried, fermented, distilled, and burned...",
"question": "does ethanol take more energy make that produces"
}
```
### Data Fields
The data fields are the same among all splits.
#### default
- `question`: a `string` feature.
- `answer`: a `bool` feature.
- `passage`: a `string` feature.
### Data Splits
| name |train|validation|
|-------|----:|---------:|
|default| 9427| 3270|
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
BoolQ is released under the [Creative Commons Share-Alike 3.0](https://creativecommons.org/licenses/by-sa/3.0/) license.
### Citation Information
```
@inproceedings{clark2019boolq,
title = {BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions},
author = {Clark, Christopher and Lee, Kenton and Chang, Ming-Wei, and Kwiatkowski, Tom and Collins, Michael, and Toutanova, Kristina},
booktitle = {NAACL},
year = {2019},
}
```
### Contributions
Thanks to [@lewtun](https://github.com/lewtun), [@lhoestq](https://github.com/lhoestq), [@thomwolf](https://github.com/thomwolf), [@patrickvonplaten](https://github.com/patrickvonplaten), [@albertvillanova](https://github.com/albertvillanova) for adding this dataset. |
Open-Orca/FLAN | 2023-08-02T15:08:01.000Z | [
"size_categories:1B<n<10B",
"language:en",
"license:cc-by-4.0",
"arxiv:2301.13688",
"arxiv:2109.01652",
"arxiv:2110.08207",
"arxiv:2204.07705",
"region:us"
] | Open-Orca | null | null | null | 96 | 12,725 | ---
license: cc-by-4.0
language:
- en
library_name: transformers
pipeline_tag: text-generation
datasets:
- Open-Orca/OpenOrca
size_categories:
- 1B<n<10B
---
<p><h1>🍮 The WHOLE FLAN Collection! 🍮</h1></p>

# Overview
This repository includes the full dataset from the [FLAN Collection](https://ai.googleblog.com/2023/02/the-flan-collection-advancing-open.html), totalling ~300GB as parquets.
Generated using the official seqio templating from the [Google FLAN Collection GitHub repo](https://github.com/google-research/FLAN/tree/main/flan/v2).
The data is subject to all the same licensing of the component datasets.
To keep up with our continued work on OpenOrca and other exciting research, find our Discord here:
https://AlignmentLab.ai
# Motivation
This work was done as part of the requirements for the OpenOrca project.
There was not a large enough subset of FLAN Collection generated publicly to subsample from to complete the work.
So, we opted to process the entire collection ourselves.
Generating this requires an understanding of seqio and a Linux server with 512GB of CPU ram, as well as fast drives and custom limits for many parameters beyond what is default on Linux server distributions (e.g., requiring up to 45,000 threads running at once).
It takes downloading over 400GB of datasets, working around tfds bugs, and then processing the datasets over the course of several days.
We provide this repo as a resource to other ML researchers, as it saves these time consuming and laborious steps to getting the data into a more accessible format for further consumption.
# Data
## Organization
* JSON files at top level are used for subsampling in OpenOrca
* Parquets in subdirectories contain the entire FLAN collection in Dask-sharded folders by submix fractions
## Zero-Shot vs Few-Shot and Options vs No-Options
The core sub-collections of FLAN are `CoT`, `Dialog`, `NIv2`, `T0`, and `flan2021`.
Within those sub-collections are four "remixes" of the data that are templated differently:
* `Zero-Shot` and `Few-Shot`
* `Zero-Shot` provides a prompt, question, or challenge without any exemplaries prior
* `Few-Shot` provides exemplaries first
* `Options` and `No-Options`
* `Options` provides a question or challenge with multiple-choice (e.g. A/B/C/D) answer options provided to select from
* `No-Options` requires a free-form answer
For every sub-collection, only some of the "remixes" may officially be provided. All available have been generated in full without any redaction or sub-sampling.
An example: `t0_fsopt_data` folder contains the sub-collection `T0`'s Few-Shot (FS), Options (OPT) remix set.
Notably, this is the largest "remix" and the one that necessitates 512GB CPU ram to generate. The raw json output is nearly 200GB.
## Parquet Sizes
Each sub-collection's individual remixes are provided as [Parquet](https://huggingface.co/docs/datasets/loading#parquet) files which have been sharded by [Dask](https://huggingface.co/docs/datasets/main/en/filesystems#dask) into ~160MB chunks (starting from 256MB blocks of the source jsonl files).
The folder structure along with size sums is provided below.
```
$ du -h --max-depth=1 ./
9.1G ./niv2_fsopt_data
2.4G ./niv2_zsopt_data
59G ./flan_fsopt_data
984M ./dialog_zsopt_data
11G ./flan_zsopt_data
8.6G ./dialog_fsopt_data
16G ./t0_zsnoopt_data
149M ./cot_fsopt_data
20M ./cot_zsopt_data
17G ./t0_zsopt_data
11G ./flan_zsnoopt_data
101G ./t0_fsopt_data
25G ./flan_fsnoopt_data
39G ./t0_fsnoopt_data
296G ./
```
# Citations
```bibtex
@misc{goodson2023huggyflan
title={Fine FLAN: Seqio to Parquet So You Don't Have To},
author={Bleys Goodson},
year={2023},
publisher = {HuggingFace},
journal = {HuggingFace repository},
howpublished = {\url{https://https://huggingface.co/datasets/Open-Orca/FLAN},
}
```
```bibtex
@misc{longpre2023flan,
title={The Flan Collection: Designing Data and Methods for Effective Instruction Tuning},
author={Shayne Longpre and Le Hou and Tu Vu and Albert Webson and Hyung Won Chung and Yi Tay and Denny Zhou and Quoc V. Le and Barret Zoph and Jason Wei and Adam Roberts},
year={2023},
eprint={2301.13688},
archivePrefix={arXiv},
primaryClass={cs.AI}
}
```
```bibtex
@misc{wei2022finetuned,
title={Finetuned Language Models Are Zero-Shot Learners},
author={Jason Wei and Maarten Bosma and Vincent Y. Zhao and Kelvin Guu and Adams Wei Yu and Brian Lester and Nan Du and Andrew M. Dai and Quoc V. Le},
year={2022},
eprint={2109.01652},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
```bibtex
@misc{sanh2022multitask,
title={Multitask Prompted Training Enables Zero-Shot Task Generalization},
author={Victor Sanh and Albert Webson and Colin Raffel and Stephen H. Bach and Lintang Sutawika and Zaid Alyafeai and Antoine Chaffin and Arnaud Stiegler and Teven Le Scao and Arun Raja and Manan Dey and M Saiful Bari and Canwen Xu and Urmish Thakker and Shanya Sharma Sharma and Eliza Szczechla and Taewoon Kim and Gunjan Chhablani and Nihal Nayak and Debajyoti Datta and Jonathan Chang and Mike Tian-Jian Jiang and Han Wang and Matteo Manica and Sheng Shen and Zheng Xin Yong and Harshit Pandey and Rachel Bawden and Thomas Wang and Trishala Neeraj and Jos Rozen and Abheesht Sharma and Andrea Santilli and Thibault Fevry and Jason Alan Fries and Ryan Teehan and Tali Bers and Stella Biderman and Leo Gao and Thomas Wolf and Alexander M. Rush},
year={2022},
eprint={2110.08207},
archivePrefix={arXiv},
primaryClass={cs.LG}
}
```
```bibtex
@misc{wang2022supernaturalinstructions,
title={Super-NaturalInstructions: Generalization via Declarative Instructions on 1600+ NLP Tasks},
author={Yizhong Wang and Swaroop Mishra and Pegah Alipoormolabashi and Yeganeh Kordi and Amirreza Mirzaei and Anjana Arunkumar and Arjun Ashok and Arut Selvan Dhanasekaran and Atharva Naik and David Stap and Eshaan Pathak and Giannis Karamanolakis and Haizhi Gary Lai and Ishan Purohit and Ishani Mondal and Jacob Anderson and Kirby Kuznia and Krima Doshi and Maitreya Patel and Kuntal Kumar Pal and Mehrad Moradshahi and Mihir Parmar and Mirali Purohit and Neeraj Varshney and Phani Rohitha Kaza and Pulkit Verma and Ravsehaj Singh Puri and Rushang Karia and Shailaja Keyur Sampat and Savan Doshi and Siddhartha Mishra and Sujan Reddy and Sumanta Patro and Tanay Dixit and Xudong Shen and Chitta Baral and Yejin Choi and Noah A. Smith and Hannaneh Hajishirzi and Daniel Khashabi},
year={2022},
eprint={2204.07705},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
``` |
opentensor/openvalidators | 2023-09-25T14:03:34.000Z | [
"size_categories:1M<n<10M",
"license:mit",
"region:us"
] | opentensor | null | null | null | 6 | 12,135 | ---
license: mit
viewer: False
size_categories:
- 1M<n<10M
---
# Dataset Card for Openvalidators dataset
## Dataset Description
- **Repository:** https://github.com/opentensor/validators
- **Homepage:** https://bittensor.com/
### Dataset Summary
The OpenValidators dataset, created by the OpenTensor Foundation, is a continuously growing collection of data generated
by the [OpenValidators](https://github.com/opentensor/validators) project in [W&B](https://wandb.ai/opentensor-dev/openvalidators/table).
It contains millions of records and serves researchers, data scientists, and miners in the Bittensor network.
The dataset provides information on network performance, node behaviors, and wandb run details.
Researchers can gain insights and detect patterns, while data scientists can use it for training models and analysis.
Miners can use the generated data to fine-tune their models and enhance their incentives in the network.
The dataset's continuous updates support collaboration and innovation in decentralized computing.
### Version support and revisions
This dataset is in constant evolution, so in order to facilitate data management, each data schema is versioned in
a hugging face dataset branch, so legacy data can be easily retrieved.
The main branch (or default revision) will always be the latest version of the dataset, following the latest schema adopted
by the openvalidators.
The current state of data organization is as following:
- `v1.0`: All data collected from the first openvalidators schema, ranging from version `1.0.0` to `1.0.8`.
- `main`: Current state of the dataset, following the latest schema adopted by the openvalidators (>= `1.1.0`).
### How to use
The `datasets` library allows you to load and pre-process your dataset in pure Python, at scale.
The OpenValidators dataset gives you the granularity of extracting data by **run_id**, by **OpenValidators version** and
by **multiple OpenValidators versions.**
The dataset can be downloaded and prepared in one call to your local drive by using the `load_dataset` function.
**Downloading by run id**
For example, to download the data for a specific run, simply specify the corresponding **OpenValidators version** and the **wandb run id** in the format `version/raw_data/run_id.parquet`:
```python
from datasets import load_dataset
version = '1.1.0' # OpenValidators version
run_id = '0drg98iy' # WandB run id
run_id_dataset = load_dataset('opentensor/openvalidators', data_files=f'{version}/raw_data/{run_id}.parquet')
```
_Please note that only completed run_ids are included in the dataset. Runs that are still in progress will be ingested shortly after they finish._
**Downloading by OpenValidators version**
One can also leverage the `datasets` library to download all the runs within a determined **OpenValidators** version. That can be useful for researchers and data enthusiasts that are looking to do analysis in a specific **OpenValidators** version state.
```python
from datasets import load_dataset
version = '1.1.0' # Openvalidators version
version_dataset = load_dataset('opentensor/openvalidators', data_files=f'{version}/raw_data/*')
```
**Downloading by multiple OpenValidators version**
Utilizing the `datasets` library, users can efficiently download runs from multiple **OpenValidators** versions. By accessing data from various OpenValidators versions, users can undertake downstream tasks such as data fine-tuning for mining or to perform big data analysis.
```python
from datasets import load_dataset
versions = ['1.1.0', '1.1.1', ...] # Desired versions for extraction
data_files = [f'{version}/raw_data/*' for version in versions] # Set data files directories
dataset = load_dataset('opentensor/openvalidators', data_files={ 'test': data_files })
```
**Downloading legacy data using revisions**
```python
from datasets import load_dataset
version = '1.0.4' # OpenValidators version
run_id = '0plco3n0' # WandB run id
revision = 'v1.0' # Dataset revision
run_id_dataset = load_dataset('opentensor/openvalidators', data_files=f'{version}/raw_data/{run_id}.parquet', revision=revision)
```
> Note: You can interact with legacy data in all the ways mentioned above, as long as your data scope is within the same revision.
**Analyzing metadata**
All the state related to the details of the wandb data ingestion can be accessed easily using pandas and hugging face datasets structure. This data contains relevant information regarding the metadata of the run, including user information, config information and ingestion state.
```python
import pandas as pd
version = '1.1.0' # OpenValidators version for metadata analysis
df = pd.read_csv(f'hf://datasets/opentensor/openvalidators/{version}/metadata.csv')
```
## Dataset Structure
### Data Instances
**versioned raw_data**
The data is provided as-in the wandb logs, without further preprocessing or tokenization. This data is located at `version/raw_data` where each file is a wandb run.
**metadata**
This dataset defines the current state of the wandb data ingestion by **run id**.
### Data Fields
**Raw data**
The versioned raw_data collected from W&B follows the following schema:
- `rewards`: (float64) Reward vector for given step
- `completion_times`: (float64) List of completion times for a given prompt
- `completions`: (string) List of completions received for a given prompt
- `_runtime`: (float64) Runtime of the event
- `_timestamp`: (float64) Timestamp of the event
- `name`: (string) Prompt type, e.g. 'followup', 'answer', 'augment'
- `block`: (float64) Current block at given step
- `gating_loss`: (float64) Gating model loss for given step
- `rlhf_reward_model`: (float64) Output vector of the rlhf reward model
- `relevance_filter`: (float64) Output vector of the relevance scoring reward model
- `dahoas_reward_model`: (float64) Output vector of the dahoas reward model
- `blacklist_filter`:(float64) Output vector of the blacklist filter
- `nsfw_filter`:(float64) Output vector of the nsfw filter
- `prompt_reward_model`:(float64) Output vector of the prompt reward model
- `reciprocate_reward_model`:(float64) Output vector of the reciprocate reward model
- `diversity_reward_model`:(float64) Output vector of the diversity reward model
- `set_weights`: (float64) Output vector of the set weights
- `uids`:(int64) Queried uids
- `_step`: (int64) Step of the event
- `prompt`: (string) Prompt text string
- `step_length`: (float64) Elapsed time between the beginning of a run step to the end of a run step
- `best`: (string) Best completion for given prompt
**Metadata**
- `run_id`: (string) Wandb Run Id
- `completed`: (boolean) Flag indicating if the run_id is completed (finished, crashed or killed)
- `downloaded`: (boolean) Flag indicating if the run_id data has been downloaded
- `last_checkpoint`: (string) Last checkpoint of the run_id
- `hotkey`: (string) Hotkey associated with the run_id
- `openvalidators_version`: (string) Version of OpenValidators associated with the run_id
- `problematic`: (boolean) Flag indicating if the run_id data had problems to be ingested
- `problematic_reason`: (string) Reason for the run_id being problematic (Exception message)
- `wandb_json_config`: (string) JSON configuration associated with the run_id in Wandb
- `wandb_run_name`: (string) Name of the Wandb run
- `wandb_user_info`: (string) Username information associated with the Wandb run
- `wandb_tags`: (list) List of tags associated with the Wandb run
- `wandb_createdAt`: (string) Timestamp of the run creation in Wandb
## Dataset Creation
### Curation Rationale
This dataset was curated to provide a comprehensive and reliable collection of historical data obtained by the execution of different OpenValidators in the bittensor network.
The goal is to support researchers, data scientists and developers with data generated in the network, facilitating the discovery of new insights, network analysis, troubleshooting, and data extraction for downstream tasks like mining.
### Source Data
#### Initial Data Collection and Normalization
The initial data collection process for this dataset involves recurrent collection by a specialized worker responsible for extracting data from wandb and ingesting it into the Hugging Face datasets structure. The collected data is organized based on the OpenValidators version and run ID to facilitate efficient data management and granular access. Each run is collected based on its corresponding OpenValidators version tag and grouped into version-specific folders. Within each version folder, a `metadata.csv` file is included to manage the collection state, while the raw data of each run is saved in the `.parquet` format with the file name corresponding to the run ID (e.g., `run_id.parquet`). Please note that the code for this data collection process will be released for transparency and reproducibility.
#### Who are the source language producers?
The language producers for this dataset are all the openvalidators that are logging their data into wandb in conjunction of other nodes of the bittensor network. The main wandb page where the data is sent can be accessed at https://wandb.ai/opentensor-dev/openvalidators/table.
### Licensing Information
The dataset is licensed under the [MIT License](https://github.com/opentensor/validators/blob/main/LICENSE)
### Supported Tasks and Leaderboards
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] |
AmazonScience/massive | 2022-11-16T15:44:51.000Z | [
"task_categories:text-classification",
"task_ids:intent-classification",
"task_ids:multi-class-classification",
"annotations_creators:expert-generated",
"language_creators:found",
"multilinguality:af-ZA",
"multilinguality:am-ET",
"multilinguality:ar-SA",
"multilinguality:az-AZ",
"multilinguality:bn-BD",
"multilinguality:ca-ES",
"multilinguality:cy-GB",
"multilinguality:da-DK",
"multilinguality:de-DE",
"multilinguality:el-GR",
"multilinguality:en-US",
"multilinguality:es-ES",
"multilinguality:fa-IR",
"multilinguality:fi-FI",
"multilinguality:fr-FR",
"multilinguality:he-IL",
"multilinguality:hi-IN",
"multilinguality:hu-HU",
"multilinguality:hy-AM",
"multilinguality:id-ID",
"multilinguality:is-IS",
"multilinguality:it-IT",
"multilinguality:ja-JP",
"multilinguality:jv-ID",
"multilinguality:ka-GE",
"multilinguality:km-KH",
"multilinguality:kn-IN",
"multilinguality:ko-KR",
"multilinguality:lv-LV",
"multilinguality:ml-IN",
"multilinguality:mn-MN",
"multilinguality:ms-MY",
"multilinguality:my-MM",
"multilinguality:nb-NO",
"multilinguality:nl-NL",
"multilinguality:pl-PL",
"multilinguality:pt-PT",
"multilinguality:ro-RO",
"multilinguality:ru-RU",
"multilinguality:sl-SL",
"multilinguality:sq-AL",
"multilinguality:sv-SE",
"multilinguality:sw-KE",
"multilinguality:ta-IN",
"multilinguality:te-IN",
"multilinguality:th-TH",
"multilinguality:tl-PH",
"multilinguality:tr-TR",
"multilinguality:ur-PK",
"multilinguality:vi-VN",
"multilinguality:zh-CN",
"multilinguality:zh-TW",
"size_categories:100K<n<1M",
"source_datasets:original",
"license:cc-by-4.0",
"natural-language-understanding",
"arxiv:2204.08582",
"region:us"
] | AmazonScience | MASSIVE is a parallel dataset of > 1M utterances across 51 languages with annotations
for the Natural Language Understanding tasks of intent prediction and slot annotation.
Utterances span 60 intents and include 55 slot types. MASSIVE was created by localizing
the SLURP dataset, composed of general Intelligent Voice Assistant single-shot interactions. | @misc{fitzgerald2022massive,
title={MASSIVE: A 1M-Example Multilingual Natural Language Understanding Dataset with 51 Typologically-Diverse Languages},
author={Jack FitzGerald and Christopher Hench and Charith Peris and Scott Mackie and Kay Rottmann and Ana Sanchez and Aaron Nash and Liam Urbach and Vishesh Kakarala and Richa Singh and Swetha Ranganath and Laurie Crist and Misha Britan and Wouter Leeuwis and Gokhan Tur and Prem Natarajan},
year={2022},
eprint={2204.08582},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
@inproceedings{bastianelli-etal-2020-slurp,
title = "{SLURP}: A Spoken Language Understanding Resource Package",
author = "Bastianelli, Emanuele and
Vanzo, Andrea and
Swietojanski, Pawel and
Rieser, Verena",
booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)",
month = nov,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2020.emnlp-main.588",
doi = "10.18653/v1/2020.emnlp-main.588",
pages = "7252--7262",
abstract = "Spoken Language Understanding infers semantic meaning directly from audio data, and thus promises to reduce error propagation and misunderstandings in end-user applications. However, publicly available SLU resources are limited. In this paper, we release SLURP, a new SLU package containing the following: (1) A new challenging dataset in English spanning 18 domains, which is substantially bigger and linguistically more diverse than existing datasets; (2) Competitive baselines based on state-of-the-art NLU and ASR systems; (3) A new transparent metric for entity labelling which enables a detailed error analysis for identifying potential areas of improvement. SLURP is available at https://github.com/pswietojanski/slurp."
} | null | 37 | 12,119 | ---
annotations_creators:
- expert-generated
language_creators:
- found
license:
- cc-by-4.0
multilinguality:
- af-ZA
- am-ET
- ar-SA
- az-AZ
- bn-BD
- ca-ES
- cy-GB
- da-DK
- de-DE
- el-GR
- en-US
- es-ES
- fa-IR
- fi-FI
- fr-FR
- he-IL
- hi-IN
- hu-HU
- hy-AM
- id-ID
- is-IS
- it-IT
- ja-JP
- jv-ID
- ka-GE
- km-KH
- kn-IN
- ko-KR
- lv-LV
- ml-IN
- mn-MN
- ms-MY
- my-MM
- nb-NO
- nl-NL
- pl-PL
- pt-PT
- ro-RO
- ru-RU
- sl-SL
- sq-AL
- sv-SE
- sw-KE
- ta-IN
- te-IN
- th-TH
- tl-PH
- tr-TR
- ur-PK
- vi-VN
- zh-CN
- zh-TW
size_categories:
- 100K<n<1M
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- intent-classification
- multi-class-classification
paperswithcode_id: massive
pretty_name: MASSIVE
language_bcp47:
- af-ZA
- am-ET
- ar-SA
- az-AZ
- bn-BD
- ca-ES
- cy-GB
- da-DK
- de-DE
- el-GR
- en-US
- es-ES
- fa-IR
- fi-FI
- fr-FR
- he-IL
- hi-IN
- hu-HU
- hy-AM
- id-ID
- is-IS
- it-IT
- ja-JP
- jv-ID
- ka-GE
- km-KH
- kn-IN
- ko-KR
- lv-LV
- ml-IN
- mn-MN
- ms-MY
- my-MM
- nb-NO
- nl-NL
- pl-PL
- pt-PT
- ro-RO
- ru-RU
- sl-SL
- sq-AL
- sv-SE
- sw-KE
- ta-IN
- te-IN
- th-TH
- tl-PH
- tr-TR
- ur-PK
- vi-VN
- zh-CN
- zh-TW
tags:
- natural-language-understanding
---
# MASSIVE 1.1: A 1M-Example Multilingual Natural Language Understanding Dataset with 52 Typologically-Diverse Languages
## Table of Contents
- [Dataset Card for [Needs More Information]](#dataset-card-for-needs-more-information)
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Initial Data Collection and Normalization](#initial-data-collection-and-normalization)
- [Who are the source language producers?](#who-are-the-source-language-producers)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [No Warranty](#no-warranty)
- [Citation Information](#citation-information)
## Dataset Description
- **Homepage:** https://github.com/alexa/massive
- **Repository:** https://github.com/alexa/massive
- **Paper:** https://arxiv.org/abs/2204.08582
- **Leaderboard:** https://eval.ai/web/challenges/challenge-page/1697/overview
- **Point of Contact:** [GitHub](https://github.com/alexa/massive/issues)
### Dataset Summary
MASSIVE 1.1 is a parallel dataset of > 1M utterances across 52 languages with annotations for the Natural Language Understanding tasks of intent prediction and slot annotation. Utterances span 60 intents and include 55 slot types. MASSIVE was created by localizing the SLURP dataset, composed of general Intelligent Voice Assistant single-shot interactions.
| Name | Lang | Utt/Lang | Domains | Intents | Slots |
|:-------------------------------------------------------------------------------:|:-------:|:--------------:|:-------:|:--------:|:------:|
| MASSIVE 1.1 | 52 | 19,521 | 18 | 60 | 55 |
| SLURP (Bastianelli et al., 2020) | 1 | 16,521 | 18 | 60 | 55 |
| NLU Evaluation Data (Liu et al., 2019) | 1 | 25,716 | 18 | 54 | 56 |
| Airline Travel Information System (ATIS) (Price, 1990) | 1 | 5,871 | 1 | 26 | 129 |
| ATIS with Hindi and Turkish (Upadhyay et al., 2018) | 3 | 1,315-5,871 | 1 | 26 | 129 |
| MultiATIS++ (Xu et al., 2020) | 9 | 1,422-5,897 | 1 | 21-26 | 99-140 |
| Snips (Coucke et al., 2018) | 1 | 14,484 | - | 7 | 53 |
| Snips with French (Saade et al., 2019) | 2 | 4,818 | 2 | 14-15 | 11-12 |
| Task Oriented Parsing (TOP) (Gupta et al., 2018) | 1 | 44,873 | 2 | 25 | 36 |
| Multilingual Task-Oriented Semantic Parsing (MTOP) (Li et al., 2021) | 6 | 15,195-22,288 | 11 | 104-113 | 72-75 |
| Cross-Lingual Multilingual Task Oriented Dialog (Schuster et al., 2019) | 3 | 5,083-43,323 | 3 | 12 | 11 |
| Microsoft Dialog Challenge (Li et al., 2018) | 1 | 38,276 | 3 | 11 | 29 |
| Fluent Speech Commands (FSC) (Lugosch et al., 2019) | 1 | 30,043 | - | 31 | - |
| Chinese Audio-Textual Spoken Language Understanding (CATSLU) (Zhu et al., 2019) | 1 | 16,258 | 4 | - | 94 |
### Supported Tasks and Leaderboards
The dataset can be used to train a model for `natural-language-understanding` (NLU) :
- `intent-classification`
- `multi-class-classification`
- `natural-language-understanding`
### Languages
The MASSIVE 1.1 corpora consists of parallel sentences from 52 languages :
- `Afrikaans - South Africa (af-ZA)`
- `Amharic - Ethiopia (am-ET)`
- `Arabic - Saudi Arabia (ar-SA)`
- `Azeri - Azerbaijan (az-AZ)`
- `Bengali - Bangladesh (bn-BD)`
- `Catalan - Spain (ca-ES)`
- `Chinese - China (zh-CN)`
- `Chinese - Taiwan (zh-TW)`
- `Danish - Denmark (da-DK)`
- `German - Germany (de-DE)`
- `Greek - Greece (el-GR)`
- `English - United States (en-US)`
- `Spanish - Spain (es-ES)`
- `Farsi - Iran (fa-IR)`
- `Finnish - Finland (fi-FI)`
- `French - France (fr-FR)`
- `Hebrew - Israel (he-IL)`
- `Hungarian - Hungary (hu-HU)`
- `Armenian - Armenia (hy-AM)`
- `Indonesian - Indonesia (id-ID)`
- `Icelandic - Iceland (is-IS)`
- `Italian - Italy (it-IT)`
- `Japanese - Japan (ja-JP)`
- `Javanese - Indonesia (jv-ID)`
- `Georgian - Georgia (ka-GE)`
- `Khmer - Cambodia (km-KH)`
- `Korean - Korea (ko-KR)`
- `Latvian - Latvia (lv-LV)`
- `Mongolian - Mongolia (mn-MN)`
- `Malay - Malaysia (ms-MY)`
- `Burmese - Myanmar (my-MM)`
- `Norwegian - Norway (nb-NO)`
- `Dutch - Netherlands (nl-NL)`
- `Polish - Poland (pl-PL)`
- `Portuguese - Portugal (pt-PT)`
- `Romanian - Romania (ro-RO)`
- `Russian - Russia (ru-RU)`
- `Slovanian - Slovania (sl-SL)`
- `Albanian - Albania (sq-AL)`
- `Swedish - Sweden (sv-SE)`
- `Swahili - Kenya (sw-KE)`
- `Hindi - India (hi-IN)`
- `Kannada - India (kn-IN)`
- `Malayalam - India (ml-IN)`
- `Tamil - India (ta-IN)`
- `Telugu - India (te-IN)`
- `Thai - Thailand (th-TH)`
- `Tagalog - Philippines (tl-PH)`
- `Turkish - Turkey (tr-TR)`
- `Urdu - Pakistan (ur-PK)`
- `Vietnamese - Vietnam (vi-VN)`
- `Welsh - United Kingdom (cy-GB)`
## Load the dataset with HuggingFace
```python
from datasets import load_dataset
dataset = load_dataset("AmazonScience/massive", "en-US", split='train')
print(dataset[0])
```
## Dataset Structure
### Data Instances
```json
{
"id": "0",
"locale": "fr-FR",
"partition": "test",
"scenario": "alarm",
"intent": "alarm_set",
"utt": "réveille-moi à cinq heures du matin cette semaine",
"annot_utt": "réveille-moi à [time : cinq heures du matin] [date : cette semaine]",
"worker_id": "22",
"slot_method": [
{ "slot": "time", "method": "translation" },
{ "slot": "date", "method": "translation" }
],
"judgments": [
{
"worker_id": "22",
"intent_score": 1,
"slots_score": 1,
"grammar_score": 4,
"spelling_score": 2,
"language_identification": "target"
},
{
"worker_id": "8",
"intent_score": 1,
"slots_score": 1,
"grammar_score": 4,
"spelling_score": 2,
"language_identification": "target"
},
{
"worker_id": "0",
"intent_score": 1,
"slots_score": 1,
"grammar_score": 4,
"spelling_score": 2,
"language_identification": "target"
}
]
}
```
### Data Fields
`id`: maps to the original ID in the [SLURP](https://github.com/pswietojanski/slurp) collection. Mapping back to the SLURP en-US utterance, this utterance served as the basis for this localization.
`locale`: is the language and country code accoring to ISO-639-1 and ISO-3166.
`partition`: is either `train`, `dev`, or `test`, according to the original split in [SLURP](https://github.com/pswietojanski/slurp).
`scenario`: is the general domain, aka "scenario" in SLURP terminology, of an utterance
`intent`: is the specific intent of an utterance within a domain formatted as `{scenario}_{intent}`
`utt`: the raw utterance text without annotations
`annot_utt`: the text from `utt` with slot annotations formatted as `[{label} : {entity}]`
`worker_id`: The obfuscated worker ID from MTurk of the worker completing the localization of the utterance. Worker IDs are specific to a locale and do *not* map across locales.
`slot_method`: for each slot in the utterance, whether that slot was a `translation` (i.e., same expression just in the target language), `localization` (i.e., not the same expression but a different expression was chosen more suitable to the phrase in that locale), or `unchanged` (i.e., the original en-US slot value was copied over without modification).
`judgments`: Each judgment collected for the localized utterance has 6 keys. `worker_id` is the obfuscated worker ID from MTurk of the worker completing the judgment. Worker IDs are specific to a locale and do *not* map across locales, but *are* consistent across the localization tasks and the judgment tasks, e.g., judgment worker ID 32 in the example above may appear as the localization worker ID for the localization of a different de-DE utterance, in which case it would be the same worker.
```plain
intent_score : "Does the sentence match the intent?"
0: No
1: Yes
2: It is a reasonable interpretation of the goal
slots_score : "Do all these terms match the categories in square brackets?"
0: No
1: Yes
2: There are no words in square brackets (utterance without a slot)
grammar_score : "Read the sentence out loud. Ignore any spelling, punctuation, or capitalization errors. Does it sound natural?"
0: Completely unnatural (nonsensical, cannot be understood at all)
1: Severe errors (the meaning cannot be understood and doesn't sound natural in your language)
2: Some errors (the meaning can be understood but it doesn't sound natural in your language)
3: Good enough (easily understood and sounds almost natural in your language)
4: Perfect (sounds natural in your language)
spelling_score : "Are all words spelled correctly? Ignore any spelling variances that may be due to differences in dialect. Missing spaces should be marked as a spelling error."
0: There are more than 2 spelling errors
1: There are 1-2 spelling errors
2: All words are spelled correctly
language_identification : "The following sentence contains words in the following languages (check all that apply)"
1: target
2: english
3: other
4: target & english
5: target & other
6: english & other
7: target & english & other
```
### Data Splits
|Language|Train|Dev|Test|
|:---:|:---:|:---:|:---:|
|af-ZA|11514|2033|2974|
|am-ET|11514|2033|2974|
|ar-SA|11514|2033|2974|
|az-AZ|11514|2033|2974|
|bn-BD|11514|2033|2974|
|ca-ES|11514|2033|2974|
|cy-GB|11514|2033|2974|
|da-DK|11514|2033|2974|
|de-DE|11514|2033|2974|
|el-GR|11514|2033|2974|
|en-US|11514|2033|2974|
|es-ES|11514|2033|2974|
|fa-IR|11514|2033|2974|
|fi-FI|11514|2033|2974|
|fr-FR|11514|2033|2974|
|he-IL|11514|2033|2974|
|hi-IN|11514|2033|2974|
|hu-HU|11514|2033|2974|
|hy-AM|11514|2033|2974|
|id-ID|11514|2033|2974|
|is-IS|11514|2033|2974|
|it-IT|11514|2033|2974|
|ja-JP|11514|2033|2974|
|jv-ID|11514|2033|2974|
|ka-GE|11514|2033|2974|
|km-KH|11514|2033|2974|
|kn-IN|11514|2033|2974|
|ko-KR|11514|2033|2974|
|lv-LV|11514|2033|2974|
|ml-IN|11514|2033|2974|
|mn-MN|11514|2033|2974|
|ms-MY|11514|2033|2974|
|my-MM|11514|2033|2974|
|nb-NO|11514|2033|2974|
|nl-NL|11514|2033|2974|
|pl-PL|11514|2033|2974|
|pt-PT|11514|2033|2974|
|ro-RO|11514|2033|2974|
|ru-RU|11514|2033|2974|
|sl-SL|11514|2033|2974|
|sq-AL|11514|2033|2974|
|sv-SE|11514|2033|2974|
|sw-KE|11514|2033|2974|
|ta-IN|11514|2033|2974|
|te-IN|11514|2033|2974|
|th-TH|11514|2033|2974|
|tl-PH|11514|2033|2974|
|tr-TR|11514|2033|2974|
|ur-PK|11514|2033|2974|
|vi-VN|11514|2033|2974|
|zh-CN|11514|2033|2974|
|zh-TW|11514|2033|2974|
### Personal and Sensitive Information
The corpora is free of personal or sensitive information.
## Additional Information
### Dataset Curators
__MASSIVE__: Jack FitzGerald and Christopher Hench and Charith Peris and Scott Mackie and Kay Rottmann and Ana Sanchez and Aaron Nash and Liam Urbach and Vishesh Kakarala and Richa Singh and Swetha Ranganath and Laurie Crist and Misha Britan and Wouter Leeuwis and Gokhan Tur and Prem Natarajan.
__SLURP__: Bastianelli, Emanuele and Vanzo, Andrea and Swietojanski, Pawel and Rieser, Verena.
__Hugging Face Upload and Integration__: Labrak Yanis (Not affiliated with the original corpus)
### Licensing Information
```plain
Copyright Amazon.com Inc. or its affiliates.
Attribution 4.0 International
=======================================================================
Creative Commons Corporation ("Creative Commons") is not a law firm and
does not provide legal services or legal advice. Distribution of
Creative Commons public licenses does not create a lawyer-client or
other relationship. Creative Commons makes its licenses and related
information available on an "as-is" basis. Creative Commons gives no
warranties regarding its licenses, any material licensed under their
terms and conditions, or any related information. Creative Commons
disclaims all liability for damages resulting from their use to the
fullest extent possible.
Using Creative Commons Public Licenses
Creative Commons public licenses provide a standard set of terms and
conditions that creators and other rights holders may use to share
original works of authorship and other material subject to copyright
and certain other rights specified in the public license below. The
following considerations are for informational purposes only, are not
exhaustive, and do not form part of our licenses.
Considerations for licensors: Our public licenses are
intended for use by those authorized to give the public
permission to use material in ways otherwise restricted by
copyright and certain other rights. Our licenses are
irrevocable. Licensors should read and understand the terms
and conditions of the license they choose before applying it.
Licensors should also secure all rights necessary before
applying our licenses so that the public can reuse the
material as expected. Licensors should clearly mark any
material not subject to the license. This includes other CC-
licensed material, or material used under an exception or
limitation to copyright. More considerations for licensors:
wiki.creativecommons.org/Considerations_for_licensors
Considerations for the public: By using one of our public
licenses, a licensor grants the public permission to use the
licensed material under specified terms and conditions. If
the licensor's permission is not necessary for any reason--for
example, because of any applicable exception or limitation to
copyright--then that use is not regulated by the license. Our
licenses grant only permissions under copyright and certain
other rights that a licensor has authority to grant. Use of
the licensed material may still be restricted for other
reasons, including because others have copyright or other
rights in the material. A licensor may make special requests,
such as asking that all changes be marked or described.
Although not required by our licenses, you are encouraged to
respect those requests where reasonable. More considerations
for the public:
wiki.creativecommons.org/Considerations_for_licensees
=======================================================================
Creative Commons Attribution 4.0 International Public License
By exercising the Licensed Rights (defined below), You accept and agree
to be bound by the terms and conditions of this Creative Commons
Attribution 4.0 International Public License ("Public License"). To the
extent this Public License may be interpreted as a contract, You are
granted the Licensed Rights in consideration of Your acceptance of
these terms and conditions, and the Licensor grants You such rights in
consideration of benefits the Licensor receives from making the
Licensed Material available under these terms and conditions.
Section 1 -- Definitions.
a. Adapted Material means material subject to Copyright and Similar
Rights that is derived from or based upon the Licensed Material
and in which the Licensed Material is translated, altered,
arranged, transformed, or otherwise modified in a manner requiring
permission under the Copyright and Similar Rights held by the
Licensor. For purposes of this Public License, where the Licensed
Material is a musical work, performance, or sound recording,
Adapted Material is always produced where the Licensed Material is
synched in timed relation with a moving image.
b. Adapter's License means the license You apply to Your Copyright
and Similar Rights in Your contributions to Adapted Material in
accordance with the terms and conditions of this Public License.
c. Copyright and Similar Rights means copyright and/or similar rights
closely related to copyright including, without limitation,
performance, broadcast, sound recording, and Sui Generis Database
Rights, without regard to how the rights are labeled or
categorized. For purposes of this Public License, the rights
specified in Section 2(b)(1)-(2) are not Copyright and Similar
Rights.
d. Effective Technological Measures means those measures that, in the
absence of proper authority, may not be circumvented under laws
fulfilling obligations under Article 11 of the WIPO Copyright
Treaty adopted on December 20, 1996, and/or similar international
agreements.
e. Exceptions and Limitations means fair use, fair dealing, and/or
any other exception or limitation to Copyright and Similar Rights
that applies to Your use of the Licensed Material.
f. Licensed Material means the artistic or literary work, database,
or other material to which the Licensor applied this Public
License.
g. Licensed Rights means the rights granted to You subject to the
terms and conditions of this Public License, which are limited to
all Copyright and Similar Rights that apply to Your use of the
Licensed Material and that the Licensor has authority to license.
h. Licensor means the individual(s) or entity(ies) granting rights
under this Public License.
i. Share means to provide material to the public by any means or
process that requires permission under the Licensed Rights, such
as reproduction, public display, public performance, distribution,
dissemination, communication, or importation, and to make material
available to the public including in ways that members of the
public may access the material from a place and at a time
individually chosen by them.
j. Sui Generis Database Rights means rights other than copyright
resulting from Directive 96/9/EC of the European Parliament and of
the Council of 11 March 1996 on the legal protection of databases,
as amended and/or succeeded, as well as other essentially
equivalent rights anywhere in the world.
k. You means the individual or entity exercising the Licensed Rights
under this Public License. Your has a corresponding meaning.
Section 2 -- Scope.
a. License grant.
1. Subject to the terms and conditions of this Public License,
the Licensor hereby grants You a worldwide, royalty-free,
non-sublicensable, non-exclusive, irrevocable license to
exercise the Licensed Rights in the Licensed Material to:
a. reproduce and Share the Licensed Material, in whole or
in part; and
b. produce, reproduce, and Share Adapted Material.
2. Exceptions and Limitations. For the avoidance of doubt, where
Exceptions and Limitations apply to Your use, this Public
License does not apply, and You do not need to comply with
its terms and conditions.
3. Term. The term of this Public License is specified in Section
6(a).
4. Media and formats; technical modifications allowed. The
Licensor authorizes You to exercise the Licensed Rights in
all media and formats whether now known or hereafter created,
and to make technical modifications necessary to do so. The
Licensor waives and/or agrees not to assert any right or
authority to forbid You from making technical modifications
necessary to exercise the Licensed Rights, including
technical modifications necessary to circumvent Effective
Technological Measures. For purposes of this Public License,
simply making modifications authorized by this Section 2(a)
(4) never produces Adapted Material.
5. Downstream recipients.
a. Offer from the Licensor -- Licensed Material. Every
recipient of the Licensed Material automatically
receives an offer from the Licensor to exercise the
Licensed Rights under the terms and conditions of this
Public License.
b. No downstream restrictions. You may not offer or impose
any additional or different terms or conditions on, or
apply any Effective Technological Measures to, the
Licensed Material if doing so restricts exercise of the
Licensed Rights by any recipient of the Licensed
Material.
6. No endorsement. Nothing in this Public License constitutes or
may be construed as permission to assert or imply that You
are, or that Your use of the Licensed Material is, connected
with, or sponsored, endorsed, or granted official status by,
the Licensor or others designated to receive attribution as
provided in Section 3(a)(1)(A)(i).
b. Other rights.
1. Moral rights, such as the right of integrity, are not
licensed under this Public License, nor are publicity,
privacy, and/or other similar personality rights; however, to
the extent possible, the Licensor waives and/or agrees not to
assert any such rights held by the Licensor to the limited
extent necessary to allow You to exercise the Licensed
Rights, but not otherwise.
2. Patent and trademark rights are not licensed under this
Public License.
3. To the extent possible, the Licensor waives any right to
collect royalties from You for the exercise of the Licensed
Rights, whether directly or through a collecting society
under any voluntary or waivable statutory or compulsory
licensing scheme. In all other cases the Licensor expressly
reserves any right to collect such royalties.
Section 3 -- License Conditions.
Your exercise of the Licensed Rights is expressly made subject to the
following conditions.
a. Attribution.
1. If You Share the Licensed Material (including in modified
form), You must:
a. retain the following if it is supplied by the Licensor
with the Licensed Material:
i. identification of the creator(s) of the Licensed
Material and any others designated to receive
attribution, in any reasonable manner requested by
the Licensor (including by pseudonym if
designated);
ii. a copyright notice;
iii. a notice that refers to this Public License;
iv. a notice that refers to the disclaimer of
warranties;
v. a URI or hyperlink to the Licensed Material to the
extent reasonably practicable;
b. indicate if You modified the Licensed Material and
retain an indication of any previous modifications; and
c. indicate the Licensed Material is licensed under this
Public License, and include the text of, or the URI or
hyperlink to, this Public License.
2. You may satisfy the conditions in Section 3(a)(1) in any
reasonable manner based on the medium, means, and context in
which You Share the Licensed Material. For example, it may be
reasonable to satisfy the conditions by providing a URI or
hyperlink to a resource that includes the required
information.
3. If requested by the Licensor, You must remove any of the
information required by Section 3(a)(1)(A) to the extent
reasonably practicable.
4. If You Share Adapted Material You produce, the Adapter's
License You apply must not prevent recipients of the Adapted
Material from complying with this Public License.
Section 4 -- Sui Generis Database Rights.
Where the Licensed Rights include Sui Generis Database Rights that
apply to Your use of the Licensed Material:
a. for the avoidance of doubt, Section 2(a)(1) grants You the right
to extract, reuse, reproduce, and Share all or a substantial
portion of the contents of the database;
b. if You include all or a substantial portion of the database
contents in a database in which You have Sui Generis Database
Rights, then the database in which You have Sui Generis Database
Rights (but not its individual contents) is Adapted Material; and
c. You must comply with the conditions in Section 3(a) if You Share
all or a substantial portion of the contents of the database.
For the avoidance of doubt, this Section 4 supplements and does not
replace Your obligations under this Public License where the Licensed
Rights include other Copyright and Similar Rights.
Section 5 -- Disclaimer of Warranties and Limitation of Liability.
a. UNLESS OTHERWISE SEPARATELY UNDERTAKEN BY THE LICENSOR, TO THE
EXTENT POSSIBLE, THE LICENSOR OFFERS THE LICENSED MATERIAL AS-IS
AND AS-AVAILABLE, AND MAKES NO REPRESENTATIONS OR WARRANTIES OF
ANY KIND CONCERNING THE LICENSED MATERIAL, WHETHER EXPRESS,
IMPLIED, STATUTORY, OR OTHER. THIS INCLUDES, WITHOUT LIMITATION,
WARRANTIES OF TITLE, MERCHANTABILITY, FITNESS FOR A PARTICULAR
PURPOSE, NON-INFRINGEMENT, ABSENCE OF LATENT OR OTHER DEFECTS,
ACCURACY, OR THE PRESENCE OR ABSENCE OF ERRORS, WHETHER OR NOT
KNOWN OR DISCOVERABLE. WHERE DISCLAIMERS OF WARRANTIES ARE NOT
ALLOWED IN FULL OR IN PART, THIS DISCLAIMER MAY NOT APPLY TO YOU.
b. TO THE EXTENT POSSIBLE, IN NO EVENT WILL THE LICENSOR BE LIABLE
TO YOU ON ANY LEGAL THEORY (INCLUDING, WITHOUT LIMITATION,
NEGLIGENCE) OR OTHERWISE FOR ANY DIRECT, SPECIAL, INDIRECT,
INCIDENTAL, CONSEQUENTIAL, PUNITIVE, EXEMPLARY, OR OTHER LOSSES,
COSTS, EXPENSES, OR DAMAGES ARISING OUT OF THIS PUBLIC LICENSE OR
USE OF THE LICENSED MATERIAL, EVEN IF THE LICENSOR HAS BEEN
ADVISED OF THE POSSIBILITY OF SUCH LOSSES, COSTS, EXPENSES, OR
DAMAGES. WHERE A LIMITATION OF LIABILITY IS NOT ALLOWED IN FULL OR
IN PART, THIS LIMITATION MAY NOT APPLY TO YOU.
c. The disclaimer of warranties and limitation of liability provided
above shall be interpreted in a manner that, to the extent
possible, most closely approximates an absolute disclaimer and
waiver of all liability.
Section 6 -- Term and Termination.
a. This Public License applies for the term of the Copyright and
Similar Rights licensed here. However, if You fail to comply with
this Public License, then Your rights under this Public License
terminate automatically.
b. Where Your right to use the Licensed Material has terminated under
Section 6(a), it reinstates:
1. automatically as of the date the violation is cured, provided
it is cured within 30 days of Your discovery of the
violation; or
2. upon express reinstatement by the Licensor.
For the avoidance of doubt, this Section 6(b) does not affect any
right the Licensor may have to seek remedies for Your violations
of this Public License.
c. For the avoidance of doubt, the Licensor may also offer the
Licensed Material under separate terms or conditions or stop
distributing the Licensed Material at any time; however, doing so
will not terminate this Public License.
d. Sections 1, 5, 6, 7, and 8 survive termination of this Public
License.
Section 7 -- Other Terms and Conditions.
a. The Licensor shall not be bound by any additional or different
terms or conditions communicated by You unless expressly agreed.
b. Any arrangements, understandings, or agreements regarding the
Licensed Material not stated herein are separate from and
independent of the terms and conditions of this Public License.
Section 8 -- Interpretation.
a. For the avoidance of doubt, this Public License does not, and
shall not be interpreted to, reduce, limit, restrict, or impose
conditions on any use of the Licensed Material that could lawfully
be made without permission under this Public License.
b. To the extent possible, if any provision of this Public License is
deemed unenforceable, it shall be automatically reformed to the
minimum extent necessary to make it enforceable. If the provision
cannot be reformed, it shall be severed from this Public License
without affecting the enforceability of the remaining terms and
conditions.
c. No term or condition of this Public License will be waived and no
failure to comply consented to unless expressly agreed to by the
Licensor.
d. Nothing in this Public License constitutes or may be interpreted
as a limitation upon, or waiver of, any privileges and immunities
that apply to the Licensor or You, including from the legal
processes of any jurisdiction or authority.
=======================================================================
Creative Commons is not a party to its public licenses.
Notwithstanding, Creative Commons may elect to apply one of its public
licenses to material it publishes and in those instances will be
considered the “Licensor.” The text of the Creative Commons public
licenses is dedicated to the public domain under the CC0 Public Domain
Dedication. Except for the limited purpose of indicating that material
is shared under a Creative Commons public license or as otherwise
permitted by the Creative Commons policies published at
creativecommons.org/policies, Creative Commons does not authorize the
use of the trademark "Creative Commons" or any other trademark or logo
of Creative Commons without its prior written consent including,
without limitation, in connection with any unauthorized modifications
to any of its public licenses or any other arrangements,
understandings, or agreements concerning use of licensed material. For
the avoidance of doubt, this paragraph does not form part of the public
licenses.
Creative Commons may be contacted at creativecommons.org.
```
### Citation Information
Please cite the following papers when using this dataset.
```latex
@misc{fitzgerald2022massive,
title={MASSIVE: A 1M-Example Multilingual Natural Language Understanding Dataset with 51 Typologically-Diverse Languages},
author={Jack FitzGerald and Christopher Hench and Charith Peris and Scott Mackie and Kay Rottmann and Ana Sanchez and Aaron Nash and Liam Urbach and Vishesh Kakarala and Richa Singh and Swetha Ranganath and Laurie Crist and Misha Britan and Wouter Leeuwis and Gokhan Tur and Prem Natarajan},
year={2022},
eprint={2204.08582},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
@inproceedings{bastianelli-etal-2020-slurp,
title = "{SLURP}: A Spoken Language Understanding Resource Package",
author = "Bastianelli, Emanuele and
Vanzo, Andrea and
Swietojanski, Pawel and
Rieser, Verena",
booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)",
month = nov,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2020.emnlp-main.588",
doi = "10.18653/v1/2020.emnlp-main.588",
pages = "7252--7262",
abstract = "Spoken Language Understanding infers semantic meaning directly from audio data, and thus promises to reduce error propagation and misunderstandings in end-user applications. However, publicly available SLU resources are limited. In this paper, we release SLURP, a new SLU package containing the following: (1) A new challenging dataset in English spanning 18 domains, which is substantially bigger and linguistically more diverse than existing datasets; (2) Competitive baselines based on state-of-the-art NLU and ASR systems; (3) A new transparent metric for entity labelling which enables a detailed error analysis for identifying potential areas of improvement. SLURP is available at https://github.com/pswietojanski/slurp."
}
```
|
bigscience/P3 | 2023-02-01T13:38:41.000Z | [
"task_categories:other",
"annotations_creators:crowdsourced",
"annotations_creators:expert-generated",
"multilinguality:monolingual",
"size_categories:100M<n<1B",
"language:en",
"license:apache-2.0",
"arxiv:2110.08207",
"region:us"
] | bigscience | P3 (Public Pool of Prompts) is a collection of prompted English datasets covering a diverse set of NLP tasks. A prompt is the combination of an input template and a target template. The templates are functions mapping a data example into natural language for the input and target sequences. For example, in the case of an NLI dataset, the data example would include fields for *Premise, Hypothesis, Label*. An input template would be *If {Premise} is true, is it also true that {Hypothesis}?*, whereas a target template can be defined with the label choices *Choices[label]*. Here *Choices* is prompt-specific metadata that consists of the options *yes, maybe, no* corresponding to *label* being entailment (0), neutral (1) or contradiction (2).
Prompts are collected using [Promptsource](https://github.com/bigscience-workshop/promptsource), an interface to interactively write prompts on datasets, and collect prompt-specific metadata such as evaluation metrics. As of October 13th, there are 2'000 prompts collected for 270+ data(sub)sets. The collection of prompts of P3 is publicly available on [Promptsource](https://github.com/bigscience-workshop/promptsource).
To train [T0*](https://huggingface.co/bigscience/T0pp), we used a subset of the prompts available in Promptsource (see details [here](https://huggingface.co/bigscience/T0pp#training-data)). However, some of the prompts use `random.choice`, a method that selects uniformly at random an option in a list of valid possibilities. For reproducibility purposes, we release the collection of prompted examples used to train T0*. **The data available here are the materialized version of the prompted datasets used in [Multitask Prompted Training Enables Zero-Shot Task Generalization](https://arxiv.org/abs/2110.08207) which represent only a subset of the datasets for which there is at least one prompt in Promptsource.** | @misc{sanh2021multitask,
title={Multitask Prompted Training Enables Zero-Shot Task Generalization},
author={Victor Sanh and Albert Webson and Colin Raffel and Stephen H. Bach and Lintang Sutawika and Zaid Alyafeai and Antoine Chaffin and Arnaud Stiegler and Teven Le Scao and Arun Raja and Manan Dey and M Saiful Bari and Canwen Xu and Urmish Thakker and Shanya Sharma Sharma and Eliza Szczechla and Taewoon Kim and Gunjan Chhablani and Nihal Nayak and Debajyoti Datta and Jonathan Chang and Mike Tian-Jian Jiang and Han Wang and Matteo Manica and Sheng Shen and Zheng Xin Yong and Harshit Pandey and Rachel Bawden and Thomas Wang and Trishala Neeraj and Jos Rozen and Abheesht Sharma and Andrea Santilli and Thibault Fevry and Jason Alan Fries and Ryan Teehan and Stella Biderman and Leo Gao and Tali Bers and Thomas Wolf and Alexander M. Rush},
year={2021},
eprint={2110.08207},
archivePrefix={arXiv},
primaryClass={cs.LG}
} | null | 157 | 12,012 | ---
annotations_creators:
- crowdsourced
- expert-generated
language:
- en
license:
- apache-2.0
multilinguality:
- monolingual
pretty_name: P3
size_categories:
- 100M<n<1B
task_categories:
- other
---
# Dataset Card for P3
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Additional Information](#additional-information)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** https://bigscience.huggingface.co/promptsource
- **Repository:** https://github.com/bigscience-workshop/promptsource/
- **Paper:** [Multitask Prompted Training Enables Zero-Shot Task Generalization](https://arxiv.org/abs/2110.08207)
- **Point of Contact:** [Victor Sanh](mailto:victor@huggingface.co)
### Dataset Summary
P3 (Public Pool of Prompts) is a collection of prompted English datasets covering a diverse set of NLP tasks. A prompt is the combination of an input template and a target template. The templates are functions mapping a data example into natural language for the input and target sequences. For example, in the case of an NLI dataset, the data example would include fields for *Premise, Hypothesis, Label*. An input template would be *If {Premise} is true, is it also true that {Hypothesis}?*, whereas a target template can be defined with the label choices *Choices[label]*. Here *Choices* is prompt-specific metadata that consists of the options *yes, maybe, no* corresponding to *label* being entailment (0), neutral (1) or contradiction (2).
Prompts are collected using [Promptsource](https://github.com/bigscience-workshop/promptsource), an interface to interactively write prompts on datasets, and collect prompt-specific metadata such as evaluation metrics. As of October 13th, there are 2'000 prompts collected for 270+ data(sub)sets. The collection of prompts of P3 is publicly available on [Promptsource](https://github.com/bigscience-workshop/promptsource).
To train [T0*](https://huggingface.co/bigscience/T0pp), we used a subset of the prompts available in Promptsource (see details [here](https://huggingface.co/bigscience/T0pp#training-data)). However, some of the prompts use `random.choice`, a method that selects uniformly at random an option in a list of valid possibilities. For reproducibility purposes, we release the collection of prompted examples used to train T0*. **The data available here are the materialized version of the prompted datasets used in [Multitask Prompted Training Enables Zero-Shot Task Generalization](https://arxiv.org/abs/2110.08207) which represent only a subset of the datasets for which there is at least one prompt in Promptsource.**
### Supported Tasks and Leaderboards
The tasks represented in P3 cover a diverse set of NLP tasks including multiple-choice QA, sentiment analysis or natural language inference. We detail the full list of datasets in [Source Data](#source-data).
### Languages
The data in P3 are in English (BCP-47 `en`).
## Dataset Structure
### Data Instances
An example of "train" looks as follows:
```bash
{
'answer_choices': ['safe', 'trolley'],
'inputs': [86, 8, 7142, 666, 6, 405, 8, 3, 834, 1518, 21, 1346, 42, 31682, 58, 37, 3, 929, 9, 3042, 63, 2765, 808, 8, 2045, 6448, 326, 13, 8, 31682, 11, 3, 24052, 135, 16, 8, 1346, 552, 8, 3, 834, 47, 6364, 5], 'inputs_pretokenized': 'In the sentence below, does the _ stand for safe or trolley?\nThe treasury workers took the gold bars off of the trolley and stacked them in the safe until the _ was empty.',
'targets': [31682, 1],
'targets_pretokenized': '\ntrolley'
}
```
In the case of rank classification (letting the model select its the prediction the option with the highest log-likelihood), an example looks as follows:
```bash
{
'idx': [5, 0],
'inputs': [86, 8, 7142, 666, 6, 405, 8, 3, 834, 1518, 21, 19454, 42, 22227, 58, 19454, 744, 31, 17, 2112, 4553, 17742, 7, 12, 1953, 6, 298, 22227, 966, 373, 405, 5, 3, 834, 19, 72, 952, 12, 619, 16, 3, 9, 17742, 3298, 5],
'inputs_pretokenized': "In the sentence below, does the _ stand for Kyle or Logan?\nKyle doesn't wear leg warmers to bed, while Logan almost always does. _ is more likely to live in a warmer climate.",
'is_correct': True,
'targets': [19454, 1],
'targets_pretokenized': 'Kyle',
'weight': 1.0
}
```
To check all the prompted examples, you can use the [Promptsource hosted tool](http://bigscience.huggingface.co/promptsource) and choose the `Prompted dataset viewer` mode in the left panel.
### Data Fields
The data fields are the same among all splits:
- `answer_choices`: the choices (in natural language) available to the model
- `inputs_pretokenized`: the natural language input fed to the model
- `targets_pretokenized`: the natural language target that the model has to generate
- `inputs`: the tokenized input with [T5](https://huggingface.co/google/t5-v1_1-base)'s tokenizer
- `targets`: the tokenized target with [T5](https://huggingface.co/google/t5-v1_1-base)'s tokenizer
- `idx`: identifier of the (example, answer_option_id) in the case of rank classification
- `weight`: a weight for the example produced by seqio (always set to 1.0 in practise)
- `is_correct`: whether the (example, answer_option_id) is the correct one
### Data Splits
The list of data splits and their respective sizes is very long. You'll find the whole list in this [file](https://huggingface.co/datasets/bigscience/P3/blob/main/tasks_splits_and_features.py).
## Dataset Creation
### Curation Rationale
The Public Pool of Prompts relies on the Hugging Face Dataset library. Any public dataset in the Datasets library can be prompted. We select the datasets that have at least one subset in English and excluded datasets containing (predominantly) non-natural language examples.
We conservatively decided not to prompt datasets that contain potentially harmful content (for instance, datasets built on social media content). However, we sometimes prompt datasets that are purposefully built to measure bias and fairness of trained models, and reserve these prompted datasets (the validation or test sets) for evaluation purposes.
### Source Data
Here's the full list of the datasets present in the materialized version of P3:
- Multiple-Choice QA
- CommonsenseQA
- DREAM
- QUAIL
- QuaRTz
- Social IQA
- WiQA
- Cosmos
- QASC
- Quarel
- SciQ
- Wiki Hop
- ARC
- OpenBookQA
- MultiRC
- PIQA
- RACE
- HellaSwag
- BoolQ
- Extractive QA
- Adversarial QA
- Quoref
- DuoRC
- ROPES
- SQuAD v2
- ReCoRD
- Close-book QA
- Hotpot QA
- Wiki QA
- Trivia QA
- Web Questions
- Structure-to-text
- Common Gen
- Wiki Bio
- Sentiment
- Amazon
- App Reviews
- IMDB
- Rotten Tomatoes
- Yelp
- Summarization
- CNN Daily Mail
- Gigaword
- MultiNews
- SamSum
- XSum
- Topic Classification
- AG News
- DBPedia
- TREC
- Paraphrase Identification
- MRPC
- PAWS
- QQP
- Natural Language Inference
- ANLI
- CB
- RTE
- Coreference Resolution
- WSC
- Winogrande
- Word Sense disambiguation
- WiC
- Sentence Completion
- COPA
- HellaSwag
- Story Cloze
### Annotations
The prompts available in Promptsource are collected as part of BigScience, one-year long research workshop on large multilingual models and datasets. 36 contributors affiliated with 24 institutions in 8 countries participated to the prompt collection. Contributors are in majority machine learning researchers or machine learning engineers.
The main annotation guideline was that prompts needed to be grammatical and understandable by a native English speaker with no prior experience of the tasks. Additionally, prompts that required explicit counting or numerical indexing were removed in favor of natural language variants, e.g., instead of predicting indices of a span to extract (e.g. in extractive question answering), the model was expected to copy the span's text instead. With these minimal constraints, prompt writers were encouraged to use both formal and creative prompts and various orderings of the data. Most of the prompts correspond directly to a version of the original proposed task, although we also allowed prompts that permuted the original task (for instance, generating a document from its summary) or allowed for ambiguous output (for instance, not indicating a list of available choices).
The full annotation given to the contributors can be found [here](https://github.com/bigscience-workshop/promptsource/blob/main/CONTRIBUTING.md). *Note to self: the link is currently being updated with the)
## Additional Information
### Licensing Information
The dataset is released under Apache 2.0.
### Citation Information
```bibtex
@misc{sanh2021multitask,
title={Multitask Prompted Training Enables Zero-Shot Task Generalization},
author={Victor Sanh and Albert Webson and Colin Raffel and Stephen H. Bach and Lintang Sutawika and Zaid Alyafeai and Antoine Chaffin and Arnaud Stiegler and Teven Le Scao and Arun Raja and Manan Dey and M Saiful Bari and Canwen Xu and Urmish Thakker and Shanya Sharma Sharma and Eliza Szczechla and Taewoon Kim and Gunjan Chhablani and Nihal Nayak and Debajyoti Datta and Jonathan Chang and Mike Tian-Jian Jiang and Han Wang and Matteo Manica and Sheng Shen and Zheng Xin Yong and Harshit Pandey and Rachel Bawden and Thomas Wang and Trishala Neeraj and Jos Rozen and Abheesht Sharma and Andrea Santilli and Thibault Fevry and Jason Alan Fries and Ryan Teehan and Stella Biderman and Leo Gao and Tali Bers and Thomas Wolf and Alexander M. Rush},
year={2021},
eprint={2110.08207},
archivePrefix={arXiv},
primaryClass={cs.LG}
}
```
### Contributions
Thanks to the contributors of [promptsource](https://github.com/bigscience-workshop/promptsource/graphs/contributors) for adding this dataset.
|
amazon_polarity | 2023-01-25T14:26:12.000Z | [
"task_categories:text-classification",
"task_ids:sentiment-classification",
"annotations_creators:crowdsourced",
"language_creators:crowdsourced",
"multilinguality:monolingual",
"size_categories:1M<n<10M",
"source_datasets:original",
"language:en",
"license:apache-2.0",
"arxiv:1509.01626",
"region:us"
] | null | The Amazon reviews dataset consists of reviews from amazon.
The data span a period of 18 years, including ~35 million reviews up to March 2013.
Reviews include product and user information, ratings, and a plaintext review. | @inproceedings{mcauley2013hidden,
title={Hidden factors and hidden topics: understanding rating dimensions with review text},
author={McAuley, Julian and Leskovec, Jure},
booktitle={Proceedings of the 7th ACM conference on Recommender systems},
pages={165--172},
year={2013}
} | null | 27 | 11,958 | ---
annotations_creators:
- crowdsourced
language_creators:
- crowdsourced
language:
- en
license:
- apache-2.0
multilinguality:
- monolingual
size_categories:
- 1M<n<10M
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- sentiment-classification
pretty_name: Amazon Review Polarity
dataset_info:
features:
- name: label
dtype:
class_label:
names:
'0': negative
'1': positive
- name: title
dtype: string
- name: content
dtype: string
config_name: amazon_polarity
splits:
- name: train
num_bytes: 1604364432
num_examples: 3600000
- name: test
num_bytes: 178176193
num_examples: 400000
download_size: 688339454
dataset_size: 1782540625
train-eval-index:
- config: amazon_polarity
task: text-classification
task_id: binary_classification
splits:
train_split: train
eval_split: test
col_mapping:
content: text
label: target
metrics:
- type: accuracy
name: Accuracy
- type: f1
name: F1 macro
args:
average: macro
- type: f1
name: F1 micro
args:
average: micro
- type: f1
name: F1 weighted
args:
average: weighted
- type: precision
name: Precision macro
args:
average: macro
- type: precision
name: Precision micro
args:
average: micro
- type: precision
name: Precision weighted
args:
average: weighted
- type: recall
name: Recall macro
args:
average: macro
- type: recall
name: Recall micro
args:
average: micro
- type: recall
name: Recall weighted
args:
average: weighted
---
# Dataset Card for Amazon Review Polarity
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** https://registry.opendata.aws/
- **Repository:** https://github.com/zhangxiangxiao/Crepe
- **Paper:** https://arxiv.org/abs/1509.01626
- **Leaderboard:** [Needs More Information]
- **Point of Contact:** [Xiang Zhang](mailto:xiang.zhang@nyu.edu)
### Dataset Summary
The Amazon reviews dataset consists of reviews from amazon.
The data span a period of 18 years, including ~35 million reviews up to March 2013.
Reviews include product and user information, ratings, and a plaintext review.
### Supported Tasks and Leaderboards
- `text-classification`, `sentiment-classification`: The dataset is mainly used for text classification: given the content and the title, predict the correct star rating.
### Languages
Mainly English.
## Dataset Structure
### Data Instances
A typical data point, comprises of a title, a content and the corresponding label.
An example from the AmazonPolarity test set looks as follows:
```
{
'title':'Great CD',
'content':"My lovely Pat has one of the GREAT voices of her generation. I have listened to this CD for YEARS and I still LOVE IT. When I'm in a good mood it makes me feel better. A bad mood just evaporates like sugar in the rain. This CD just oozes LIFE. Vocals are jusat STUUNNING and lyrics just kill. One of life's hidden gems. This is a desert isle CD in my book. Why she never made it big is just beyond me. Everytime I play this, no matter black, white, young, old, male, female EVERYBODY says one thing ""Who was that singing ?""",
'label':1
}
```
### Data Fields
- 'title': a string containing the title of the review - escaped using double quotes (") and any internal double quote is escaped by 2 double quotes (""). New lines are escaped by a backslash followed with an "n" character, that is "\n".
- 'content': a string containing the body of the document - escaped using double quotes (") and any internal double quote is escaped by 2 double quotes (""). New lines are escaped by a backslash followed with an "n" character, that is "\n".
- 'label': either 1 (positive) or 0 (negative) rating.
### Data Splits
The Amazon reviews polarity dataset is constructed by taking review score 1 and 2 as negative, and 4 and 5 as positive. Samples of score 3 is ignored. Each class has 1,800,000 training samples and 200,000 testing samples.
## Dataset Creation
### Curation Rationale
The Amazon reviews polarity dataset is constructed by Xiang Zhang (xiang.zhang@nyu.edu). It is used as a text classification benchmark in the following paper: Xiang Zhang, Junbo Zhao, Yann LeCun. Character-level Convolutional Networks for Text Classification. Advances in Neural Information Processing Systems 28 (NIPS 2015).
### Source Data
#### Initial Data Collection and Normalization
[Needs More Information]
#### Who are the source language producers?
[Needs More Information]
### Annotations
#### Annotation process
[Needs More Information]
#### Who are the annotators?
[Needs More Information]
### Personal and Sensitive Information
[Needs More Information]
## Considerations for Using the Data
### Social Impact of Dataset
[Needs More Information]
### Discussion of Biases
[Needs More Information]
### Other Known Limitations
[Needs More Information]
## Additional Information
### Dataset Curators
[Needs More Information]
### Licensing Information
Apache License 2.0
### Citation Information
McAuley, Julian, and Jure Leskovec. "Hidden factors and hidden topics: understanding rating dimensions with review text." In Proceedings of the 7th ACM conference on Recommender systems, pp. 165-172. 2013.
Xiang Zhang, Junbo Zhao, Yann LeCun. Character-level Convolutional Networks for Text Classification. Advances in Neural Information Processing Systems 28 (NIPS 2015)
### Contributions
Thanks to [@hfawaz](https://github.com/hfawaz) for adding this dataset. |
Dahoas/full-hh-rlhf | 2023-02-23T17:29:46.000Z | [
"region:us"
] | Dahoas | null | null | null | 50 | 11,813 | ---
dataset_info:
features:
- name: prompt
dtype: string
- name: response
dtype: string
- name: chosen
dtype: string
- name: rejected
dtype: string
splits:
- name: train
num_bytes: 203150123
num_examples: 112052
- name: test
num_bytes: 22606646
num_examples: 12451
download_size: 136150742
dataset_size: 225756769
---
# Dataset Card for "full-hh-rlhf"
Anthropic's HH dataset reformatted into prompt, chosen, rejected samples. |
monology/pile-uncopyrighted | 2023-08-31T03:45:38.000Z | [
"license:other",
"arxiv:2101.00027",
"region:us"
] | monology | null | null | null | 10 | 11,689 | ---
license: other
---
# Pile Uncopyrighted
In response to [authors demanding that LLMs stop using their works](https://tcrn.ch/3rtpIDn), here's a copy of [The Pile](https://huggingface.co/datasets/monology/pile) with all copyrighted content removed.
Please consider using this dataset to train your future LLMs, to respect authors and abide by copyright law.
Creating an uncopyrighted version of a larger dataset (ie RedPajama) is planned, with no ETA.
**Methodology**
Cleaning was performed by removing everything from the Books3, BookCorpus2, OpenSubtitles, YTSubtitles, and OWT2 subsets.
Based on section 7.1 of [the original paper](https://arxiv.org/abs/2101.00027), these datasets are the only ones which are not explicitly allowed to be used in AI training. |
THUDM/LongBench | 2023-08-29T04:51:14.000Z | [
"task_categories:question-answering",
"task_categories:text-generation",
"task_categories:summarization",
"task_categories:conversational",
"task_categories:text-classification",
"size_categories:1K<n<10K",
"language:en",
"language:zh",
"Long Context",
"arxiv:2308.14508",
"arxiv:2108.00573",
"arxiv:1712.07040",
"arxiv:2105.03011",
"arxiv:2104.02112",
"arxiv:2104.05938",
"arxiv:2305.05280",
"arxiv:2303.09752",
"arxiv:1910.10683",
"arxiv:2306.14893",
"arxiv:2306.03091",
"region:us"
] | THUDM | LongBench is a comprehensive benchmark for multilingual and multi-task purposes, with the goal to fully measure and evaluate the ability of pre-trained language models to understand long text. This dataset consists of twenty different tasks, covering key long-text application scenarios such as multi-document QA, single-document QA, summarization, few-shot learning, synthetic tasks, and code completion. | null | null | 28 | 11,375 | ---
task_categories:
- question-answering
- text-generation
- summarization
- conversational
- text-classification
language:
- en
- zh
tags:
- Long Context
size_categories:
- 1K<n<10K
---
# Introduction
**LongBench** is the first benchmark for bilingual, multitask, and comprehensive assessment of **long context understanding** capabilities of large language models. LongBench includes different languages (Chinese and English) to provide a more comprehensive evaluation of the large models' multilingual capabilities on long contexts. In addition, LongBench is composed of six major categories and twenty one different tasks, covering key long-text application scenarios such as single-document QA, multi-document QA, summarization, few-shot learning, synthetic tasks and code completion.
We are fully aware of the potentially high costs involved in the model evaluation process, especially in the context of long context scenarios (such as manual annotation costs or API call costs). Therefore, we adopt a fully automated evaluation method, aimed at measuring and evaluating the model's ability to understand long contexts at the lowest cost.
LongBench includes 14 English tasks, 5 Chinese tasks, and 2 code tasks, with the average length of most tasks ranging from 5k to 15k, and a total of 4,750 test data. For detailed statistics and construction methods of LongBench tasks, please refer [here](task.md). In addition, we provide LongBench-E, a test set with a more uniform length distribution constructed by uniform sampling, with comparable amounts of data in the 0-4k, 4k-8k, and 8k+ length intervals to provide an analysis of the model's performance variations at different input lengths.
Github Repo for LongBench: https://github.com/THUDM/LongBench
Arxiv Paper for LongBench: https://arxiv.org/pdf/2308.14508.pdf
# How to use it?
#### Loading Data
```python
from datasets import load_dataset
datasets = ["narrativeqa", "qasper", "multifieldqa_en", "multifieldqa_zh", "hotpotqa", "2wikimqa", "musique", \
"dureader", "gov_report", "qmsum", "multi_news", "vcsum", "trec", "triviaqa", "samsum", "lsht", \
"passage_count", "passage_retrieval_en", "passage_retrieval_zh", "lcc", "repobench-p"]
for dataset in datasets:
data = load_dataset('THUDM/LongBench', dataset, split='test')
```
Similarly, you can load the **LongBench-E** data
```python
from datasets import load_dataset
datasets = ["qasper", "multifieldqa_en", "hotpotqa", "2wikimqa", "gov_report", "multi_news", "trec", \
"triviaqa", "samsum", "passage_count", "passage_retrieval_en", "lcc", "repobench-p"]
for dataset in datasets:
data = load_dataset('THUDM/LongBench', f"{dataset}_e", split='test')
```
Alternatively, you can download the folder from [this link](https://huggingface.co/datasets/THUDM/LongBench/resolve/main/data.zip) to load the data.
#### Data Format
All data in **LongBench** (LongBench-E) are standardized to the following format:
```json
{
"input": "The input/command for the task, usually short, such as questions in QA, queries in Few-shot tasks, etc",
"context": "The long context required for the task, such as documents, cross-file code, few-shot examples in Few-shot tasks",
"answers": "A List of all true answers",
"length": "Total length of the first three items (counted in characters for Chinese and words for English)",
"dataset": "The name of the dataset to which this piece of data belongs",
"language": "The language of this piece of data",
"all_classes": "All categories in classification tasks, null for non-classification tasks",
"_id": "Random id for each piece of data"
}
```
#### Evaluation
This repository provides data download for LongBench. If you wish to use this dataset for automated evaluation, please refer to our [github](https://github.com/THUDM/LongBench).
# Task statistics
| Task | Task Type | Eval metric | Avg len |Language | \#Sample |
| :-------- | :-----------:| :-----------: |:-------: | :-----------: |:--------: |
| HotpotQA | Multi-doc QA | F1 |9,151 |EN |200 |
| 2WikiMultihopQA| Multi-doc QA | F1 |4,887 |EN |200 |
| MuSiQue| Multi-doc QA | F1 |11,214 |EN |200 |
| DuReader| Multi-doc QA | Rouge-L |15,768 |ZH |200 |
| MultiFieldQA-en| Single-doc QA | F1 |4,559 |EN |150 |
| MultiFieldQA-zh| Single-doc QA | F1 |6,701 |ZH |200 |
| NarrativeQA| Single-doc QA | F1 |18,409 |EN |200 |
| Qasper| Single-doc QA | F1 |3,619 |EN |200 |
| GovReport| Summarization | Rouge-L |8,734 |EN |200 |
| QMSum| Summarization | Rouge-L |10,614 |EN |200 |
| MultiNews| Summarization | Rouge-L |2,113 |EN |200 |
| VCSUM| Summarization | Rouge-L |15,380 |ZH |200 |
| TriviaQA| Few shot | F1 |8,209 |EN |200 |
| SAMSum| Few shot | Rouge-L |6,258 |EN |200 |
| TREC| Few shot | Accuracy |5,177 |EN |200 |
| LSHT| Few shot | Accuracy |22,337 |ZH |200 |
| PassageRetrieval-en| Synthetic | Accuracy |9,289 |EN |200 |
| PassageCount| Synthetic | Accuracy |11,141 |EN |200 |
| PassageRetrieval-zh | Synthetic | Accuracy |6,745 |ZH |200 |
| LCC| Code | Edit Sim |1,235 |Python/C#/Java |500 |
| RepoBench-P| Code | Edit Sim |4,206 |Python/Java |500 |
> Note: In order to avoid discrepancies caused by different tokenizers, we use the word count (using Python's split function) to calculate the average length of English datasets and code datasets, and use the character count to calculate the average length of Chinese datasets.
# Task description
| Task | Task Description |
| :---------------- | :----------------------------------------------------------- |
| HotpotQA | Answer related questions based on multiple given documents |
| 2WikiMultihopQA | Answer related questions based on multiple given documents |
| MuSiQue | Answer related questions based on multiple given documents |
| DuReader | Answer related Chinese questions based on multiple retrieved documents |
| MultiFieldQA-en | Answer English questions based on a long article, which comes from a relatively diverse field |
| MultiFieldQA-zh | Answer Chinese questions based on a long article, which comes from a relatively diverse field |
| NarrativeQA | Answer questions based on stories or scripts, including understanding of important elements such as characters, plots, themes, etc. |
| Qasper | Answer questions based on a NLP research paper, questions proposed and answered by NLP practitioners |
| GovReport | A summarization task that requires summarizing government work reports |
| MultiNews | A multi-doc summarization that requires summarizing over multiple news |
| QMSum | A summarization task that requires summarizing meeting records based on user queries |
| VCSUM | A summarization task that requires summarizing Chinese meeting records |
| SAMSum | A dialogue summarization task, providing several few-shot examples |
| TriviaQA | Single document question answering task, providing several few-shot examples |
| NQ | Single document question answering task, providing several few-shot examples |
| TREC | A classification task that requires categorizing questions, includes 50 categories in total |
| LSHT | A Chinese classification task that requires categorizing news, includes 24 categories in total |
| PassageRetrieval-en | Given 30 English Wikipedia paragraphs, determine which paragraph the given summary corresponds to |
| PassageCount | Determine the total number of different paragraphs in a given repetitive article |
| PassageRetrieval-zh | Given several Chinese paragraphs from the C4 data set, determine which paragraph the given abstract corresponds to |
| LCC | Given a long piece of code, predict the next line of code |
| RepoBench-P | Given code in multiple files within a GitHub repository (including cross-file dependencies), predict the next line of code |
# Task construction
> Note: For all tasks constructed from existing datasets, we use data from the validation or test set of the existing dataset (except for VCSUM).
- The tasks of [HotpotQA](https://hotpotqa.github.io/), [2WikiMultihopQA](https://aclanthology.org/2020.coling-main.580/), [MuSiQue](https://arxiv.org/abs/2108.00573), and [DuReader](https://github.com/baidu/DuReader) are built based on the original datasets and processed to be suitable for long context evaluation. Specifically, for questions in the validation set, we select the evidence passage that contains the answer and several distracting articles. These articles together with the original question constitute the input of the tasks.
- The tasks of MultiFiedQA-zh and MultiFieldQA-en consist of long artical data from about 10 sources, including Latex papers, judicial documents, government work reports, and PDF documents indexed by Google. For each long artical, we invite several PhD and master students to annotate, i.e., to ask questions based on the long artical and give the correct answers. To better automate evaluation, we ask the annotators to propose questions with definitive answers as much as possible.
- The tasks of [NarrativeQA](https://arxiv.org/pdf/1712.07040.pdf), [Qasper](https://arxiv.org/pdf/2105.03011.pdf), [GovReport](https://arxiv.org/pdf/2104.02112.pdf), [QMSum](https://arxiv.org/pdf/2104.05938.pdf) and [MultiNews](https://aclanthology.org/P19-1102.pdf) directly use the data provided by the original papers. In the specific construction, we use the template provided by [ZeroSCROLLS](https://www.zero.scrolls-benchmark.com/) to convert the corresponding data into pure text input.
- The [VCSUM](https://arxiv.org/abs/2305.05280) task is built based on the original dataset, and we design a corresponding template to convert the corresponding data into pure text input.
- The [TriviaQA](https://nlp.cs.washington.edu/triviaqa/) task is constructed in the manner of [CoLT5](https://arxiv.org/abs/2303.09752), which provides several examples of question and answering based on documents, and requires the language model to answer related questions based on new documents.
- The tasks of [SAMSum](https://aclanthology.org/D19-5409.pdf), [TREC](https://aclanthology.org/C02-1150.pdf) and [LSHT](http://tcci.ccf.org.cn/conference/2014/dldoc/evatask6.pdf) are built based on the original datasets. For each question in the validation set, we sample several data from the training set to form few-shot examples. These examples together with the questions in the validation set constitute the input for this task.
- The PassageRetrieval-en task is constructed based on English Wikipedia. For each piece of data, we randomly sample 30 paragraphs from English Wikipedia and select one for summarization (using GPT-3.5-Turbo). This task requires the model to give the original paragraph name to which the summary corresponds.
- The PassageCount task is constructed based on the English wiki. For each piece of data, we randomly sample several passages from English Wikipedia, repeat each paragraph at random several times, and finally shuffle the paragraphs. This task requires the model to determine the total number of different paragraphs in the given context.
- The PasskeyRetrieval-zh task is constructed based on [C4](https://arxiv.org/abs/1910.10683). For each piece of data, we randomly sample several Chinese paragraphs from C4 and select one of them for summarization (using GPT-3.5-Turbo). This task requires the model to give the original paragraph name to which the summary corresponds.
- For the [LCC](https://arxiv.org/abs/2306.14893) task, we sample from the original code completion dataset. In the [RepoBench-P](https://arxiv.org/abs/2306.03091) task, we select the most challenging XF-F (Cross-File-First) setting from the original dataset and refer to the Oracle-Filled scenario in the paper. For each original piece of data, we randomly extract multiple cross-file code snippets, including the gold cross-file code snippet, and concatenate them as input, requiring the model to effectively use cross-file code for completion.
# LongBench-E statistics
| Task | Task Type | \#data in 0-4k | \#data in 4-8k | \#data in 8k+|
| :--------- | :-----------:| :-----------: |:---------: | :-------------: |
| HotpotQA | Multi-doc QA | 100 |100 |100 |
| 2WikiMultihopQA| Multi-doc QA | 100 |100 |100 |
| MultiFieldQA-en| Single-doc QA | 67 |70 |13 |
| Qasper| Single-doc QA | 100 |100 |24 |
| GovReport| Summarization | 100 |100 |100 |
| MultiNews| Summarization | 100 |100 |94 |
| TriviaQA| Few shot | 100 |100 |100 |
| SAMSum| Few shot | 100 |100 |100 |
| TREC| Few shot | 100 |100 |100 |
| PassageRetrieval-en| Synthetic | 100 |100 |100 |
| PassageCount| Synthetic | 100 |100 |100 |
| LCC| Code | 100 |100 |100 |
| RepoBench-P| Code | 100 |100 |100 |
# Citation
```
@misc{bai2023longbench,
title={LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding},
author={Yushi Bai and Xin Lv and Jiajie Zhang and Hongchang Lyu and Jiankai Tang and Zhidian Huang and Zhengxiao Du and Xiao Liu and Aohan Zeng and Lei Hou and Yuxiao Dong and Jie Tang and Juanzi Li},
year={2023},
eprint={2308.14508},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
``` |
open-llm-leaderboard/details_Aspik101__trurl-2-7b-pl-instruct_unload | 2023-09-22T23:43:57.000Z | [
"region:us"
] | open-llm-leaderboard | null | null | null | 0 | 11,208 | ---
pretty_name: Evaluation run of Aspik101/trurl-2-7b-pl-instruct_unload
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [Aspik101/trurl-2-7b-pl-instruct_unload](https://huggingface.co/Aspik101/trurl-2-7b-pl-instruct_unload)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 64 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 2 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the agregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_Aspik101__trurl-2-7b-pl-instruct_unload\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-09-22T23:43:45.355114](https://huggingface.co/datasets/open-llm-leaderboard/details_Aspik101__trurl-2-7b-pl-instruct_unload/blob/main/results_2023-09-22T23-43-45.355114.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"em\": 0.16096895973154363,\n\
\ \"em_stderr\": 0.0037635677120072437,\n \"f1\": 0.22060822147650985,\n\
\ \"f1_stderr\": 0.0038388662178584767,\n \"acc\": 0.3986331756197593,\n\
\ \"acc_stderr\": 0.009900867594093993\n },\n \"harness|drop|3\": {\n\
\ \"em\": 0.16096895973154363,\n \"em_stderr\": 0.0037635677120072437,\n\
\ \"f1\": 0.22060822147650985,\n \"f1_stderr\": 0.0038388662178584767\n\
\ },\n \"harness|gsm8k|5\": {\n \"acc\": 0.07429871114480667,\n \
\ \"acc_stderr\": 0.007223844172845576\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.7229676400947119,\n \"acc_stderr\": 0.012577891015342412\n\
\ }\n}\n```"
repo_url: https://huggingface.co/Aspik101/trurl-2-7b-pl-instruct_unload
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|arc:challenge|25_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_drop_3
data_files:
- split: 2023_09_22T23_43_45.355114
path:
- '**/details_harness|drop|3_2023-09-22T23-43-45.355114.parquet'
- split: latest
path:
- '**/details_harness|drop|3_2023-09-22T23-43-45.355114.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_09_22T23_43_45.355114
path:
- '**/details_harness|gsm8k|5_2023-09-22T23-43-45.355114.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-09-22T23-43-45.355114.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hellaswag|10_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-08-17T14:40:19.486608.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-management|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-virology|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- '**/details_harness|truthfulqa:mc|0_2023-08-17T14:40:19.486608.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2023-08-17T14:40:19.486608.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_09_22T23_43_45.355114
path:
- '**/details_harness|winogrande|5_2023-09-22T23-43-45.355114.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-09-22T23-43-45.355114.parquet'
- config_name: results
data_files:
- split: 2023_08_17T14_40_19.486608
path:
- results_2023-08-17T14:40:19.486608.parquet
- split: 2023_09_22T23_43_45.355114
path:
- results_2023-09-22T23-43-45.355114.parquet
- split: latest
path:
- results_2023-09-22T23-43-45.355114.parquet
---
# Dataset Card for Evaluation run of Aspik101/trurl-2-7b-pl-instruct_unload
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/Aspik101/trurl-2-7b-pl-instruct_unload
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [Aspik101/trurl-2-7b-pl-instruct_unload](https://huggingface.co/Aspik101/trurl-2-7b-pl-instruct_unload) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 64 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 2 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_Aspik101__trurl-2-7b-pl-instruct_unload",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-09-22T23:43:45.355114](https://huggingface.co/datasets/open-llm-leaderboard/details_Aspik101__trurl-2-7b-pl-instruct_unload/blob/main/results_2023-09-22T23-43-45.355114.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"em": 0.16096895973154363,
"em_stderr": 0.0037635677120072437,
"f1": 0.22060822147650985,
"f1_stderr": 0.0038388662178584767,
"acc": 0.3986331756197593,
"acc_stderr": 0.009900867594093993
},
"harness|drop|3": {
"em": 0.16096895973154363,
"em_stderr": 0.0037635677120072437,
"f1": 0.22060822147650985,
"f1_stderr": 0.0038388662178584767
},
"harness|gsm8k|5": {
"acc": 0.07429871114480667,
"acc_stderr": 0.007223844172845576
},
"harness|winogrande|5": {
"acc": 0.7229676400947119,
"acc_stderr": 0.012577891015342412
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] |
open-llm-leaderboard/details_togethercomputer__GPT-JT-6B-v0 | 2023-08-27T12:37:30.000Z | [
"region:us"
] | open-llm-leaderboard | null | null | null | 0 | 10,794 | ---
pretty_name: Evaluation run of togethercomputer/GPT-JT-6B-v0
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [togethercomputer/GPT-JT-6B-v0](https://huggingface.co/togethercomputer/GPT-JT-6B-v0)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 61 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the agregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_togethercomputer__GPT-JT-6B-v0\"\
,\n\t\"harness_truthfulqa_mc_0\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\
\nThese are the [latest results from run 2023-07-19T15:42:14.994932](https://huggingface.co/datasets/open-llm-leaderboard/details_togethercomputer__GPT-JT-6B-v0/blob/main/results_2023-07-19T15%3A42%3A14.994932.json)\
\ (note that their might be results for other tasks in the repos if successive evals\
\ didn't cover the same tasks. You find each in the results and the \"latest\" split\
\ for each eval):\n\n```python\n{\n \"all\": {\n \"acc\": 0.4917382409100623,\n\
\ \"acc_stderr\": 0.03534934330073521,\n \"acc_norm\": 0.4953627760622265,\n\
\ \"acc_norm_stderr\": 0.03534663453635268,\n \"mc1\": 0.23255813953488372,\n\
\ \"mc1_stderr\": 0.014789157531080515,\n \"mc2\": 0.3889085255663652,\n\
\ \"mc2_stderr\": 0.013877851892873926\n },\n \"harness|arc:challenge|25\"\
: {\n \"acc\": 0.38993174061433444,\n \"acc_stderr\": 0.01425295984889289,\n\
\ \"acc_norm\": 0.4206484641638225,\n \"acc_norm_stderr\": 0.014426211252508406\n\
\ },\n \"harness|hellaswag|10\": {\n \"acc\": 0.4965146385182235,\n\
\ \"acc_stderr\": 0.004989660180792163,\n \"acc_norm\": 0.6796454889464251,\n\
\ \"acc_norm_stderr\": 0.004656591678606739\n },\n \"harness|hendrycksTest-abstract_algebra|5\"\
: {\n \"acc\": 0.41,\n \"acc_stderr\": 0.049431107042371025,\n \
\ \"acc_norm\": 0.41,\n \"acc_norm_stderr\": 0.049431107042371025\n \
\ },\n \"harness|hendrycksTest-anatomy|5\": {\n \"acc\": 0.6592592592592592,\n\
\ \"acc_stderr\": 0.04094376269996793,\n \"acc_norm\": 0.6592592592592592,\n\
\ \"acc_norm_stderr\": 0.04094376269996793\n },\n \"harness|hendrycksTest-astronomy|5\"\
: {\n \"acc\": 0.625,\n \"acc_stderr\": 0.039397364351956274,\n \
\ \"acc_norm\": 0.625,\n \"acc_norm_stderr\": 0.039397364351956274\n\
\ },\n \"harness|hendrycksTest-business_ethics|5\": {\n \"acc\": 0.55,\n\
\ \"acc_stderr\": 0.049999999999999996,\n \"acc_norm\": 0.55,\n \
\ \"acc_norm_stderr\": 0.049999999999999996\n },\n \"harness|hendrycksTest-clinical_knowledge|5\"\
: {\n \"acc\": 0.569811320754717,\n \"acc_stderr\": 0.030471445867183238,\n\
\ \"acc_norm\": 0.569811320754717,\n \"acc_norm_stderr\": 0.030471445867183238\n\
\ },\n \"harness|hendrycksTest-college_biology|5\": {\n \"acc\": 0.5763888888888888,\n\
\ \"acc_stderr\": 0.04132125019723369,\n \"acc_norm\": 0.5763888888888888,\n\
\ \"acc_norm_stderr\": 0.04132125019723369\n },\n \"harness|hendrycksTest-college_chemistry|5\"\
: {\n \"acc\": 0.42,\n \"acc_stderr\": 0.049604496374885836,\n \
\ \"acc_norm\": 0.42,\n \"acc_norm_stderr\": 0.049604496374885836\n \
\ },\n \"harness|hendrycksTest-college_computer_science|5\": {\n \"\
acc\": 0.45,\n \"acc_stderr\": 0.05,\n \"acc_norm\": 0.45,\n \
\ \"acc_norm_stderr\": 0.05\n },\n \"harness|hendrycksTest-college_mathematics|5\"\
: {\n \"acc\": 0.44,\n \"acc_stderr\": 0.04988876515698589,\n \
\ \"acc_norm\": 0.44,\n \"acc_norm_stderr\": 0.04988876515698589\n \
\ },\n \"harness|hendrycksTest-college_medicine|5\": {\n \"acc\": 0.5375722543352601,\n\
\ \"acc_stderr\": 0.0380168510452446,\n \"acc_norm\": 0.5375722543352601,\n\
\ \"acc_norm_stderr\": 0.0380168510452446\n },\n \"harness|hendrycksTest-college_physics|5\"\
: {\n \"acc\": 0.3137254901960784,\n \"acc_stderr\": 0.04617034827006717,\n\
\ \"acc_norm\": 0.3137254901960784,\n \"acc_norm_stderr\": 0.04617034827006717\n\
\ },\n \"harness|hendrycksTest-computer_security|5\": {\n \"acc\":\
\ 0.69,\n \"acc_stderr\": 0.04648231987117316,\n \"acc_norm\": 0.69,\n\
\ \"acc_norm_stderr\": 0.04648231987117316\n },\n \"harness|hendrycksTest-conceptual_physics|5\"\
: {\n \"acc\": 0.44680851063829785,\n \"acc_stderr\": 0.0325005368436584,\n\
\ \"acc_norm\": 0.44680851063829785,\n \"acc_norm_stderr\": 0.0325005368436584\n\
\ },\n \"harness|hendrycksTest-econometrics|5\": {\n \"acc\": 0.3684210526315789,\n\
\ \"acc_stderr\": 0.04537815354939392,\n \"acc_norm\": 0.3684210526315789,\n\
\ \"acc_norm_stderr\": 0.04537815354939392\n },\n \"harness|hendrycksTest-electrical_engineering|5\"\
: {\n \"acc\": 0.5793103448275863,\n \"acc_stderr\": 0.0411391498118926,\n\
\ \"acc_norm\": 0.5793103448275863,\n \"acc_norm_stderr\": 0.0411391498118926\n\
\ },\n \"harness|hendrycksTest-elementary_mathematics|5\": {\n \"acc\"\
: 0.2671957671957672,\n \"acc_stderr\": 0.02278967314577657,\n \"\
acc_norm\": 0.2671957671957672,\n \"acc_norm_stderr\": 0.02278967314577657\n\
\ },\n \"harness|hendrycksTest-formal_logic|5\": {\n \"acc\": 0.3253968253968254,\n\
\ \"acc_stderr\": 0.04190596438871136,\n \"acc_norm\": 0.3253968253968254,\n\
\ \"acc_norm_stderr\": 0.04190596438871136\n },\n \"harness|hendrycksTest-global_facts|5\"\
: {\n \"acc\": 0.41,\n \"acc_stderr\": 0.049431107042371025,\n \
\ \"acc_norm\": 0.41,\n \"acc_norm_stderr\": 0.049431107042371025\n \
\ },\n \"harness|hendrycksTest-high_school_biology|5\": {\n \"acc\"\
: 0.4870967741935484,\n \"acc_stderr\": 0.028434533152681855,\n \"\
acc_norm\": 0.4870967741935484,\n \"acc_norm_stderr\": 0.028434533152681855\n\
\ },\n \"harness|hendrycksTest-high_school_chemistry|5\": {\n \"acc\"\
: 0.3448275862068966,\n \"acc_stderr\": 0.033442837442804574,\n \"\
acc_norm\": 0.3448275862068966,\n \"acc_norm_stderr\": 0.033442837442804574\n\
\ },\n \"harness|hendrycksTest-high_school_computer_science|5\": {\n \
\ \"acc\": 0.53,\n \"acc_stderr\": 0.050161355804659205,\n \"acc_norm\"\
: 0.53,\n \"acc_norm_stderr\": 0.050161355804659205\n },\n \"harness|hendrycksTest-high_school_european_history|5\"\
: {\n \"acc\": 0.42424242424242425,\n \"acc_stderr\": 0.03859268142070262,\n\
\ \"acc_norm\": 0.42424242424242425,\n \"acc_norm_stderr\": 0.03859268142070262\n\
\ },\n \"harness|hendrycksTest-high_school_geography|5\": {\n \"acc\"\
: 0.6313131313131313,\n \"acc_stderr\": 0.034373055019806184,\n \"\
acc_norm\": 0.6313131313131313,\n \"acc_norm_stderr\": 0.034373055019806184\n\
\ },\n \"harness|hendrycksTest-high_school_government_and_politics|5\": {\n\
\ \"acc\": 0.6217616580310881,\n \"acc_stderr\": 0.034998072761933396,\n\
\ \"acc_norm\": 0.6217616580310881,\n \"acc_norm_stderr\": 0.034998072761933396\n\
\ },\n \"harness|hendrycksTest-high_school_macroeconomics|5\": {\n \
\ \"acc\": 0.4153846153846154,\n \"acc_stderr\": 0.024985354923102325,\n\
\ \"acc_norm\": 0.4153846153846154,\n \"acc_norm_stderr\": 0.024985354923102325\n\
\ },\n \"harness|hendrycksTest-high_school_mathematics|5\": {\n \"\
acc\": 0.22592592592592592,\n \"acc_stderr\": 0.025497532639609553,\n \
\ \"acc_norm\": 0.22592592592592592,\n \"acc_norm_stderr\": 0.025497532639609553\n\
\ },\n \"harness|hendrycksTest-high_school_microeconomics|5\": {\n \
\ \"acc\": 0.42016806722689076,\n \"acc_stderr\": 0.03206183783236153,\n\
\ \"acc_norm\": 0.42016806722689076,\n \"acc_norm_stderr\": 0.03206183783236153\n\
\ },\n \"harness|hendrycksTest-high_school_physics|5\": {\n \"acc\"\
: 0.2980132450331126,\n \"acc_stderr\": 0.03734535676787198,\n \"\
acc_norm\": 0.2980132450331126,\n \"acc_norm_stderr\": 0.03734535676787198\n\
\ },\n \"harness|hendrycksTest-high_school_psychology|5\": {\n \"acc\"\
: 0.5504587155963303,\n \"acc_stderr\": 0.021327881417823373,\n \"\
acc_norm\": 0.5504587155963303,\n \"acc_norm_stderr\": 0.021327881417823373\n\
\ },\n \"harness|hendrycksTest-high_school_statistics|5\": {\n \"acc\"\
: 0.36574074074074076,\n \"acc_stderr\": 0.03284738857647206,\n \"\
acc_norm\": 0.36574074074074076,\n \"acc_norm_stderr\": 0.03284738857647206\n\
\ },\n \"harness|hendrycksTest-high_school_us_history|5\": {\n \"acc\"\
: 0.4362745098039216,\n \"acc_stderr\": 0.03480693138457039,\n \"\
acc_norm\": 0.4362745098039216,\n \"acc_norm_stderr\": 0.03480693138457039\n\
\ },\n \"harness|hendrycksTest-high_school_world_history|5\": {\n \"\
acc\": 0.48523206751054854,\n \"acc_stderr\": 0.032533028078777386,\n \
\ \"acc_norm\": 0.48523206751054854,\n \"acc_norm_stderr\": 0.032533028078777386\n\
\ },\n \"harness|hendrycksTest-human_aging|5\": {\n \"acc\": 0.547085201793722,\n\
\ \"acc_stderr\": 0.033408675019233246,\n \"acc_norm\": 0.547085201793722,\n\
\ \"acc_norm_stderr\": 0.033408675019233246\n },\n \"harness|hendrycksTest-human_sexuality|5\"\
: {\n \"acc\": 0.6412213740458015,\n \"acc_stderr\": 0.04206739313864908,\n\
\ \"acc_norm\": 0.6412213740458015,\n \"acc_norm_stderr\": 0.04206739313864908\n\
\ },\n \"harness|hendrycksTest-international_law|5\": {\n \"acc\":\
\ 0.5289256198347108,\n \"acc_stderr\": 0.04556710331269498,\n \"\
acc_norm\": 0.5289256198347108,\n \"acc_norm_stderr\": 0.04556710331269498\n\
\ },\n \"harness|hendrycksTest-jurisprudence|5\": {\n \"acc\": 0.6018518518518519,\n\
\ \"acc_stderr\": 0.04732332615978815,\n \"acc_norm\": 0.6018518518518519,\n\
\ \"acc_norm_stderr\": 0.04732332615978815\n },\n \"harness|hendrycksTest-logical_fallacies|5\"\
: {\n \"acc\": 0.6012269938650306,\n \"acc_stderr\": 0.038470214204560246,\n\
\ \"acc_norm\": 0.6012269938650306,\n \"acc_norm_stderr\": 0.038470214204560246\n\
\ },\n \"harness|hendrycksTest-machine_learning|5\": {\n \"acc\": 0.44642857142857145,\n\
\ \"acc_stderr\": 0.04718471485219588,\n \"acc_norm\": 0.44642857142857145,\n\
\ \"acc_norm_stderr\": 0.04718471485219588\n },\n \"harness|hendrycksTest-management|5\"\
: {\n \"acc\": 0.7281553398058253,\n \"acc_stderr\": 0.044052680241409216,\n\
\ \"acc_norm\": 0.7281553398058253,\n \"acc_norm_stderr\": 0.044052680241409216\n\
\ },\n \"harness|hendrycksTest-marketing|5\": {\n \"acc\": 0.6196581196581197,\n\
\ \"acc_stderr\": 0.03180425204384099,\n \"acc_norm\": 0.6196581196581197,\n\
\ \"acc_norm_stderr\": 0.03180425204384099\n },\n \"harness|hendrycksTest-medical_genetics|5\"\
: {\n \"acc\": 0.69,\n \"acc_stderr\": 0.04648231987117316,\n \
\ \"acc_norm\": 0.69,\n \"acc_norm_stderr\": 0.04648231987117316\n \
\ },\n \"harness|hendrycksTest-miscellaneous|5\": {\n \"acc\": 0.4904214559386973,\n\
\ \"acc_stderr\": 0.017876682275340866,\n \"acc_norm\": 0.4904214559386973,\n\
\ \"acc_norm_stderr\": 0.017876682275340866\n },\n \"harness|hendrycksTest-moral_disputes|5\"\
: {\n \"acc\": 0.47109826589595377,\n \"acc_stderr\": 0.026874085883518348,\n\
\ \"acc_norm\": 0.47109826589595377,\n \"acc_norm_stderr\": 0.026874085883518348\n\
\ },\n \"harness|hendrycksTest-moral_scenarios|5\": {\n \"acc\": 0.23910614525139665,\n\
\ \"acc_stderr\": 0.01426555419233114,\n \"acc_norm\": 0.23910614525139665,\n\
\ \"acc_norm_stderr\": 0.01426555419233114\n },\n \"harness|hendrycksTest-nutrition|5\"\
: {\n \"acc\": 0.49673202614379086,\n \"acc_stderr\": 0.02862930519400354,\n\
\ \"acc_norm\": 0.49673202614379086,\n \"acc_norm_stderr\": 0.02862930519400354\n\
\ },\n \"harness|hendrycksTest-philosophy|5\": {\n \"acc\": 0.45980707395498394,\n\
\ \"acc_stderr\": 0.028306190403305696,\n \"acc_norm\": 0.45980707395498394,\n\
\ \"acc_norm_stderr\": 0.028306190403305696\n },\n \"harness|hendrycksTest-prehistory|5\"\
: {\n \"acc\": 0.5246913580246914,\n \"acc_stderr\": 0.02778680093142745,\n\
\ \"acc_norm\": 0.5246913580246914,\n \"acc_norm_stderr\": 0.02778680093142745\n\
\ },\n \"harness|hendrycksTest-professional_accounting|5\": {\n \"\
acc\": 0.375886524822695,\n \"acc_stderr\": 0.02889395541211588,\n \
\ \"acc_norm\": 0.375886524822695,\n \"acc_norm_stderr\": 0.02889395541211588\n\
\ },\n \"harness|hendrycksTest-professional_law|5\": {\n \"acc\": 0.3428943937418514,\n\
\ \"acc_stderr\": 0.012123463271585892,\n \"acc_norm\": 0.3428943937418514,\n\
\ \"acc_norm_stderr\": 0.012123463271585892\n },\n \"harness|hendrycksTest-professional_medicine|5\"\
: {\n \"acc\": 0.3382352941176471,\n \"acc_stderr\": 0.028739328513983572,\n\
\ \"acc_norm\": 0.3382352941176471,\n \"acc_norm_stderr\": 0.028739328513983572\n\
\ },\n \"harness|hendrycksTest-professional_psychology|5\": {\n \"\
acc\": 0.4019607843137255,\n \"acc_stderr\": 0.01983517648437537,\n \
\ \"acc_norm\": 0.4019607843137255,\n \"acc_norm_stderr\": 0.01983517648437537\n\
\ },\n \"harness|hendrycksTest-public_relations|5\": {\n \"acc\": 0.6636363636363637,\n\
\ \"acc_stderr\": 0.045253935963025044,\n \"acc_norm\": 0.6636363636363637,\n\
\ \"acc_norm_stderr\": 0.045253935963025044\n },\n \"harness|hendrycksTest-security_studies|5\"\
: {\n \"acc\": 0.5306122448979592,\n \"acc_stderr\": 0.031949171367580624,\n\
\ \"acc_norm\": 0.5306122448979592,\n \"acc_norm_stderr\": 0.031949171367580624\n\
\ },\n \"harness|hendrycksTest-sociology|5\": {\n \"acc\": 0.6417910447761194,\n\
\ \"acc_stderr\": 0.03390393042268813,\n \"acc_norm\": 0.6417910447761194,\n\
\ \"acc_norm_stderr\": 0.03390393042268813\n },\n \"harness|hendrycksTest-us_foreign_policy|5\"\
: {\n \"acc\": 0.75,\n \"acc_stderr\": 0.04351941398892446,\n \
\ \"acc_norm\": 0.75,\n \"acc_norm_stderr\": 0.04351941398892446\n \
\ },\n \"harness|hendrycksTest-virology|5\": {\n \"acc\": 0.4819277108433735,\n\
\ \"acc_stderr\": 0.038899512528272166,\n \"acc_norm\": 0.4819277108433735,\n\
\ \"acc_norm_stderr\": 0.038899512528272166\n },\n \"harness|hendrycksTest-world_religions|5\"\
: {\n \"acc\": 0.6374269005847953,\n \"acc_stderr\": 0.0368713061556206,\n\
\ \"acc_norm\": 0.6374269005847953,\n \"acc_norm_stderr\": 0.0368713061556206\n\
\ },\n \"harness|truthfulqa:mc|0\": {\n \"mc1\": 0.23255813953488372,\n\
\ \"mc1_stderr\": 0.014789157531080515,\n \"mc2\": 0.3889085255663652,\n\
\ \"mc2_stderr\": 0.013877851892873926\n }\n}\n```"
repo_url: https://huggingface.co/togethercomputer/GPT-JT-6B-v0
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|arc:challenge|25_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hellaswag|10_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-07-19T15:42:14.994932.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-management|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-virology|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-07-19T15:42:14.994932.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- '**/details_harness|truthfulqa:mc|0_2023-07-19T15:42:14.994932.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2023-07-19T15:42:14.994932.parquet'
- config_name: results
data_files:
- split: 2023_07_19T15_42_14.994932
path:
- results_2023-07-19T15:42:14.994932.parquet
- split: latest
path:
- results_2023-07-19T15:42:14.994932.parquet
---
# Dataset Card for Evaluation run of togethercomputer/GPT-JT-6B-v0
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/togethercomputer/GPT-JT-6B-v0
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [togethercomputer/GPT-JT-6B-v0](https://huggingface.co/togethercomputer/GPT-JT-6B-v0) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 61 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_togethercomputer__GPT-JT-6B-v0",
"harness_truthfulqa_mc_0",
split="train")
```
## Latest results
These are the [latest results from run 2023-07-19T15:42:14.994932](https://huggingface.co/datasets/open-llm-leaderboard/details_togethercomputer__GPT-JT-6B-v0/blob/main/results_2023-07-19T15%3A42%3A14.994932.json) (note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"acc": 0.4917382409100623,
"acc_stderr": 0.03534934330073521,
"acc_norm": 0.4953627760622265,
"acc_norm_stderr": 0.03534663453635268,
"mc1": 0.23255813953488372,
"mc1_stderr": 0.014789157531080515,
"mc2": 0.3889085255663652,
"mc2_stderr": 0.013877851892873926
},
"harness|arc:challenge|25": {
"acc": 0.38993174061433444,
"acc_stderr": 0.01425295984889289,
"acc_norm": 0.4206484641638225,
"acc_norm_stderr": 0.014426211252508406
},
"harness|hellaswag|10": {
"acc": 0.4965146385182235,
"acc_stderr": 0.004989660180792163,
"acc_norm": 0.6796454889464251,
"acc_norm_stderr": 0.004656591678606739
},
"harness|hendrycksTest-abstract_algebra|5": {
"acc": 0.41,
"acc_stderr": 0.049431107042371025,
"acc_norm": 0.41,
"acc_norm_stderr": 0.049431107042371025
},
"harness|hendrycksTest-anatomy|5": {
"acc": 0.6592592592592592,
"acc_stderr": 0.04094376269996793,
"acc_norm": 0.6592592592592592,
"acc_norm_stderr": 0.04094376269996793
},
"harness|hendrycksTest-astronomy|5": {
"acc": 0.625,
"acc_stderr": 0.039397364351956274,
"acc_norm": 0.625,
"acc_norm_stderr": 0.039397364351956274
},
"harness|hendrycksTest-business_ethics|5": {
"acc": 0.55,
"acc_stderr": 0.049999999999999996,
"acc_norm": 0.55,
"acc_norm_stderr": 0.049999999999999996
},
"harness|hendrycksTest-clinical_knowledge|5": {
"acc": 0.569811320754717,
"acc_stderr": 0.030471445867183238,
"acc_norm": 0.569811320754717,
"acc_norm_stderr": 0.030471445867183238
},
"harness|hendrycksTest-college_biology|5": {
"acc": 0.5763888888888888,
"acc_stderr": 0.04132125019723369,
"acc_norm": 0.5763888888888888,
"acc_norm_stderr": 0.04132125019723369
},
"harness|hendrycksTest-college_chemistry|5": {
"acc": 0.42,
"acc_stderr": 0.049604496374885836,
"acc_norm": 0.42,
"acc_norm_stderr": 0.049604496374885836
},
"harness|hendrycksTest-college_computer_science|5": {
"acc": 0.45,
"acc_stderr": 0.05,
"acc_norm": 0.45,
"acc_norm_stderr": 0.05
},
"harness|hendrycksTest-college_mathematics|5": {
"acc": 0.44,
"acc_stderr": 0.04988876515698589,
"acc_norm": 0.44,
"acc_norm_stderr": 0.04988876515698589
},
"harness|hendrycksTest-college_medicine|5": {
"acc": 0.5375722543352601,
"acc_stderr": 0.0380168510452446,
"acc_norm": 0.5375722543352601,
"acc_norm_stderr": 0.0380168510452446
},
"harness|hendrycksTest-college_physics|5": {
"acc": 0.3137254901960784,
"acc_stderr": 0.04617034827006717,
"acc_norm": 0.3137254901960784,
"acc_norm_stderr": 0.04617034827006717
},
"harness|hendrycksTest-computer_security|5": {
"acc": 0.69,
"acc_stderr": 0.04648231987117316,
"acc_norm": 0.69,
"acc_norm_stderr": 0.04648231987117316
},
"harness|hendrycksTest-conceptual_physics|5": {
"acc": 0.44680851063829785,
"acc_stderr": 0.0325005368436584,
"acc_norm": 0.44680851063829785,
"acc_norm_stderr": 0.0325005368436584
},
"harness|hendrycksTest-econometrics|5": {
"acc": 0.3684210526315789,
"acc_stderr": 0.04537815354939392,
"acc_norm": 0.3684210526315789,
"acc_norm_stderr": 0.04537815354939392
},
"harness|hendrycksTest-electrical_engineering|5": {
"acc": 0.5793103448275863,
"acc_stderr": 0.0411391498118926,
"acc_norm": 0.5793103448275863,
"acc_norm_stderr": 0.0411391498118926
},
"harness|hendrycksTest-elementary_mathematics|5": {
"acc": 0.2671957671957672,
"acc_stderr": 0.02278967314577657,
"acc_norm": 0.2671957671957672,
"acc_norm_stderr": 0.02278967314577657
},
"harness|hendrycksTest-formal_logic|5": {
"acc": 0.3253968253968254,
"acc_stderr": 0.04190596438871136,
"acc_norm": 0.3253968253968254,
"acc_norm_stderr": 0.04190596438871136
},
"harness|hendrycksTest-global_facts|5": {
"acc": 0.41,
"acc_stderr": 0.049431107042371025,
"acc_norm": 0.41,
"acc_norm_stderr": 0.049431107042371025
},
"harness|hendrycksTest-high_school_biology|5": {
"acc": 0.4870967741935484,
"acc_stderr": 0.028434533152681855,
"acc_norm": 0.4870967741935484,
"acc_norm_stderr": 0.028434533152681855
},
"harness|hendrycksTest-high_school_chemistry|5": {
"acc": 0.3448275862068966,
"acc_stderr": 0.033442837442804574,
"acc_norm": 0.3448275862068966,
"acc_norm_stderr": 0.033442837442804574
},
"harness|hendrycksTest-high_school_computer_science|5": {
"acc": 0.53,
"acc_stderr": 0.050161355804659205,
"acc_norm": 0.53,
"acc_norm_stderr": 0.050161355804659205
},
"harness|hendrycksTest-high_school_european_history|5": {
"acc": 0.42424242424242425,
"acc_stderr": 0.03859268142070262,
"acc_norm": 0.42424242424242425,
"acc_norm_stderr": 0.03859268142070262
},
"harness|hendrycksTest-high_school_geography|5": {
"acc": 0.6313131313131313,
"acc_stderr": 0.034373055019806184,
"acc_norm": 0.6313131313131313,
"acc_norm_stderr": 0.034373055019806184
},
"harness|hendrycksTest-high_school_government_and_politics|5": {
"acc": 0.6217616580310881,
"acc_stderr": 0.034998072761933396,
"acc_norm": 0.6217616580310881,
"acc_norm_stderr": 0.034998072761933396
},
"harness|hendrycksTest-high_school_macroeconomics|5": {
"acc": 0.4153846153846154,
"acc_stderr": 0.024985354923102325,
"acc_norm": 0.4153846153846154,
"acc_norm_stderr": 0.024985354923102325
},
"harness|hendrycksTest-high_school_mathematics|5": {
"acc": 0.22592592592592592,
"acc_stderr": 0.025497532639609553,
"acc_norm": 0.22592592592592592,
"acc_norm_stderr": 0.025497532639609553
},
"harness|hendrycksTest-high_school_microeconomics|5": {
"acc": 0.42016806722689076,
"acc_stderr": 0.03206183783236153,
"acc_norm": 0.42016806722689076,
"acc_norm_stderr": 0.03206183783236153
},
"harness|hendrycksTest-high_school_physics|5": {
"acc": 0.2980132450331126,
"acc_stderr": 0.03734535676787198,
"acc_norm": 0.2980132450331126,
"acc_norm_stderr": 0.03734535676787198
},
"harness|hendrycksTest-high_school_psychology|5": {
"acc": 0.5504587155963303,
"acc_stderr": 0.021327881417823373,
"acc_norm": 0.5504587155963303,
"acc_norm_stderr": 0.021327881417823373
},
"harness|hendrycksTest-high_school_statistics|5": {
"acc": 0.36574074074074076,
"acc_stderr": 0.03284738857647206,
"acc_norm": 0.36574074074074076,
"acc_norm_stderr": 0.03284738857647206
},
"harness|hendrycksTest-high_school_us_history|5": {
"acc": 0.4362745098039216,
"acc_stderr": 0.03480693138457039,
"acc_norm": 0.4362745098039216,
"acc_norm_stderr": 0.03480693138457039
},
"harness|hendrycksTest-high_school_world_history|5": {
"acc": 0.48523206751054854,
"acc_stderr": 0.032533028078777386,
"acc_norm": 0.48523206751054854,
"acc_norm_stderr": 0.032533028078777386
},
"harness|hendrycksTest-human_aging|5": {
"acc": 0.547085201793722,
"acc_stderr": 0.033408675019233246,
"acc_norm": 0.547085201793722,
"acc_norm_stderr": 0.033408675019233246
},
"harness|hendrycksTest-human_sexuality|5": {
"acc": 0.6412213740458015,
"acc_stderr": 0.04206739313864908,
"acc_norm": 0.6412213740458015,
"acc_norm_stderr": 0.04206739313864908
},
"harness|hendrycksTest-international_law|5": {
"acc": 0.5289256198347108,
"acc_stderr": 0.04556710331269498,
"acc_norm": 0.5289256198347108,
"acc_norm_stderr": 0.04556710331269498
},
"harness|hendrycksTest-jurisprudence|5": {
"acc": 0.6018518518518519,
"acc_stderr": 0.04732332615978815,
"acc_norm": 0.6018518518518519,
"acc_norm_stderr": 0.04732332615978815
},
"harness|hendrycksTest-logical_fallacies|5": {
"acc": 0.6012269938650306,
"acc_stderr": 0.038470214204560246,
"acc_norm": 0.6012269938650306,
"acc_norm_stderr": 0.038470214204560246
},
"harness|hendrycksTest-machine_learning|5": {
"acc": 0.44642857142857145,
"acc_stderr": 0.04718471485219588,
"acc_norm": 0.44642857142857145,
"acc_norm_stderr": 0.04718471485219588
},
"harness|hendrycksTest-management|5": {
"acc": 0.7281553398058253,
"acc_stderr": 0.044052680241409216,
"acc_norm": 0.7281553398058253,
"acc_norm_stderr": 0.044052680241409216
},
"harness|hendrycksTest-marketing|5": {
"acc": 0.6196581196581197,
"acc_stderr": 0.03180425204384099,
"acc_norm": 0.6196581196581197,
"acc_norm_stderr": 0.03180425204384099
},
"harness|hendrycksTest-medical_genetics|5": {
"acc": 0.69,
"acc_stderr": 0.04648231987117316,
"acc_norm": 0.69,
"acc_norm_stderr": 0.04648231987117316
},
"harness|hendrycksTest-miscellaneous|5": {
"acc": 0.4904214559386973,
"acc_stderr": 0.017876682275340866,
"acc_norm": 0.4904214559386973,
"acc_norm_stderr": 0.017876682275340866
},
"harness|hendrycksTest-moral_disputes|5": {
"acc": 0.47109826589595377,
"acc_stderr": 0.026874085883518348,
"acc_norm": 0.47109826589595377,
"acc_norm_stderr": 0.026874085883518348
},
"harness|hendrycksTest-moral_scenarios|5": {
"acc": 0.23910614525139665,
"acc_stderr": 0.01426555419233114,
"acc_norm": 0.23910614525139665,
"acc_norm_stderr": 0.01426555419233114
},
"harness|hendrycksTest-nutrition|5": {
"acc": 0.49673202614379086,
"acc_stderr": 0.02862930519400354,
"acc_norm": 0.49673202614379086,
"acc_norm_stderr": 0.02862930519400354
},
"harness|hendrycksTest-philosophy|5": {
"acc": 0.45980707395498394,
"acc_stderr": 0.028306190403305696,
"acc_norm": 0.45980707395498394,
"acc_norm_stderr": 0.028306190403305696
},
"harness|hendrycksTest-prehistory|5": {
"acc": 0.5246913580246914,
"acc_stderr": 0.02778680093142745,
"acc_norm": 0.5246913580246914,
"acc_norm_stderr": 0.02778680093142745
},
"harness|hendrycksTest-professional_accounting|5": {
"acc": 0.375886524822695,
"acc_stderr": 0.02889395541211588,
"acc_norm": 0.375886524822695,
"acc_norm_stderr": 0.02889395541211588
},
"harness|hendrycksTest-professional_law|5": {
"acc": 0.3428943937418514,
"acc_stderr": 0.012123463271585892,
"acc_norm": 0.3428943937418514,
"acc_norm_stderr": 0.012123463271585892
},
"harness|hendrycksTest-professional_medicine|5": {
"acc": 0.3382352941176471,
"acc_stderr": 0.028739328513983572,
"acc_norm": 0.3382352941176471,
"acc_norm_stderr": 0.028739328513983572
},
"harness|hendrycksTest-professional_psychology|5": {
"acc": 0.4019607843137255,
"acc_stderr": 0.01983517648437537,
"acc_norm": 0.4019607843137255,
"acc_norm_stderr": 0.01983517648437537
},
"harness|hendrycksTest-public_relations|5": {
"acc": 0.6636363636363637,
"acc_stderr": 0.045253935963025044,
"acc_norm": 0.6636363636363637,
"acc_norm_stderr": 0.045253935963025044
},
"harness|hendrycksTest-security_studies|5": {
"acc": 0.5306122448979592,
"acc_stderr": 0.031949171367580624,
"acc_norm": 0.5306122448979592,
"acc_norm_stderr": 0.031949171367580624
},
"harness|hendrycksTest-sociology|5": {
"acc": 0.6417910447761194,
"acc_stderr": 0.03390393042268813,
"acc_norm": 0.6417910447761194,
"acc_norm_stderr": 0.03390393042268813
},
"harness|hendrycksTest-us_foreign_policy|5": {
"acc": 0.75,
"acc_stderr": 0.04351941398892446,
"acc_norm": 0.75,
"acc_norm_stderr": 0.04351941398892446
},
"harness|hendrycksTest-virology|5": {
"acc": 0.4819277108433735,
"acc_stderr": 0.038899512528272166,
"acc_norm": 0.4819277108433735,
"acc_norm_stderr": 0.038899512528272166
},
"harness|hendrycksTest-world_religions|5": {
"acc": 0.6374269005847953,
"acc_stderr": 0.0368713061556206,
"acc_norm": 0.6374269005847953,
"acc_norm_stderr": 0.0368713061556206
},
"harness|truthfulqa:mc|0": {
"mc1": 0.23255813953488372,
"mc1_stderr": 0.014789157531080515,
"mc2": 0.3889085255663652,
"mc2_stderr": 0.013877851892873926
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] |
GBaker/MedQA-USMLE-4-options | 2023-01-24T19:18:09.000Z | [
"language:en",
"license:cc-by-4.0",
"region:us"
] | GBaker | null | null | null | 16 | 10,784 | ---
license: cc-by-4.0
language:
- en
---
Original dataset introduced by Jin et al. in [What Disease does this Patient Have? A Large-scale Open Domain Question Answering Dataset from Medical Exams](https://paperswithcode.com/paper/what-disease-does-this-patient-have-a-large)
<h4>Citation information:</h4>
@article{jin2020disease,
title={What Disease does this Patient Have? A Large-scale Open Domain Question Answering Dataset from Medical Exams},
author={Jin, Di and Pan, Eileen and Oufattole, Nassim and Weng, Wei-Hung and Fang, Hanyi and Szolovits, Peter},
journal={arXiv preprint arXiv:2009.13081},
year={2020}
}
|
wino_bias | 2023-01-25T15:02:31.000Z | [
"task_categories:token-classification",
"task_ids:coreference-resolution",
"annotations_creators:expert-generated",
"language_creators:expert-generated",
"multilinguality:monolingual",
"size_categories:1K<n<10K",
"source_datasets:original",
"language:en",
"license:mit",
"arxiv:1804.06876",
"region:us"
] | null | WinoBias, a Winograd-schema dataset for coreference resolution focused on gender bias.
The corpus contains Winograd-schema style sentences with entities corresponding to people
referred by their occupation (e.g. the nurse, the doctor, the carpenter). | @article{DBLP:journals/corr/abs-1804-06876,
author = {Jieyu Zhao and
Tianlu Wang and
Mark Yatskar and
Vicente Ordonez and
Kai{-}Wei Chang},
title = {Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods},
journal = {CoRR},
volume = {abs/1804.06876},
year = {2018},
url = {http://arxiv.org/abs/1804.06876},
archivePrefix = {arXiv},
eprint = {1804.06876},
timestamp = {Mon, 13 Aug 2018 16:47:01 +0200},
biburl = {https://dblp.org/rec/journals/corr/abs-1804-06876.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
} | null | 9 | 10,469 | ---
annotations_creators:
- expert-generated
language_creators:
- expert-generated
language:
- en
license:
- mit
multilinguality:
- monolingual
size_categories:
- 1K<n<10K
source_datasets:
- original
task_categories:
- token-classification
task_ids:
- coreference-resolution
paperswithcode_id: winobias
pretty_name: WinoBias
dataset_info:
- config_name: wino_bias
features:
- name: document_id
dtype: string
- name: part_number
dtype: string
- name: word_number
sequence: int32
- name: tokens
sequence: string
- name: pos_tags
sequence:
class_label:
names:
'0': '"'
'1': ''''''
'2': '#'
'3': $
'4': (
'5': )
'6': ','
'7': .
'8': ':'
'9': '``'
'10': CC
'11': CD
'12': DT
'13': EX
'14': FW
'15': IN
'16': JJ
'17': JJR
'18': JJS
'19': LS
'20': MD
'21': NN
'22': NNP
'23': NNPS
'24': NNS
'25': NN|SYM
'26': PDT
'27': POS
'28': PRP
'29': PRP$
'30': RB
'31': RBR
'32': RBS
'33': RP
'34': SYM
'35': TO
'36': UH
'37': VB
'38': VBD
'39': VBG
'40': VBN
'41': VBP
'42': VBZ
'43': WDT
'44': WP
'45': WP$
'46': WRB
'47': HYPH
'48': XX
'49': NFP
'50': AFX
'51': ADD
'52': -LRB-
'53': -RRB-
- name: parse_bit
sequence: string
- name: predicate_lemma
sequence: string
- name: predicate_framenet_id
sequence: string
- name: word_sense
sequence: string
- name: speaker
sequence: string
- name: ner_tags
sequence:
class_label:
names:
'0': B-PERSON
'1': I-PERSON
'2': B-NORP
'3': I-NORP
'4': B-FAC
'5': I-FAC
'6': B-ORG
'7': I-ORG
'8': B-GPE
'9': I-GPE
'10': B-LOC
'11': I-LOC
'12': B-PRODUCT
'13': I-PRODUCT
'14': B-EVENT
'15': I-EVENT
'16': B-WORK_OF_ART
'17': I-WORK_OF_ART
'18': B-LAW
'19': I-LAW
'20': B-LANGUAGE
'21': I-LANGUAGE
'22': B-DATE
'23': I-DATE
'24': B-TIME
'25': I-TIME
'26': B-PERCENT
'27': I-PERCENT
'28': B-MONEY
'29': I-MONEY
'30': B-QUANTITY
'31': I-QUANTITY
'32': B-ORDINAL
'33': I-ORDINAL
'34': B-CARDINAL
'35': I-CARDINAL
'36': '*'
'37': '0'
- name: verbal_predicates
sequence: string
splits:
- name: train
num_bytes: 173899234
num_examples: 150335
download_size: 268725744
dataset_size: 173899234
- config_name: type1_pro
features:
- name: document_id
dtype: string
- name: part_number
dtype: string
- name: word_number
sequence: int32
- name: tokens
sequence: string
- name: pos_tags
sequence:
class_label:
names:
'0': '"'
'1': ''''''
'2': '#'
'3': $
'4': (
'5': )
'6': ','
'7': .
'8': ':'
'9': '``'
'10': CC
'11': CD
'12': DT
'13': EX
'14': FW
'15': IN
'16': JJ
'17': JJR
'18': JJS
'19': LS
'20': MD
'21': NN
'22': NNP
'23': NNPS
'24': NNS
'25': NN|SYM
'26': PDT
'27': POS
'28': PRP
'29': PRP$
'30': RB
'31': RBR
'32': RBS
'33': RP
'34': SYM
'35': TO
'36': UH
'37': VB
'38': VBD
'39': VBG
'40': VBN
'41': VBP
'42': VBZ
'43': WDT
'44': WP
'45': WP$
'46': WRB
'47': HYPH
'48': XX
'49': NFP
'50': AFX
'51': ADD
'52': -LRB-
'53': -RRB-
'54': '-'
- name: parse_bit
sequence: string
- name: predicate_lemma
sequence: string
- name: predicate_framenet_id
sequence: string
- name: word_sense
sequence: string
- name: speaker
sequence: string
- name: ner_tags
sequence:
class_label:
names:
'0': B-PERSON
'1': I-PERSON
'2': B-NORP
'3': I-NORP
'4': B-FAC
'5': I-FAC
'6': B-ORG
'7': I-ORG
'8': B-GPE
'9': I-GPE
'10': B-LOC
'11': I-LOC
'12': B-PRODUCT
'13': I-PRODUCT
'14': B-EVENT
'15': I-EVENT
'16': B-WORK_OF_ART
'17': I-WORK_OF_ART
'18': B-LAW
'19': I-LAW
'20': B-LANGUAGE
'21': I-LANGUAGE
'22': B-DATE
'23': I-DATE
'24': B-TIME
'25': I-TIME
'26': B-PERCENT
'27': I-PERCENT
'28': B-MONEY
'29': I-MONEY
'30': B-QUANTITY
'31': I-QUANTITY
'32': B-ORDINAL
'33': I-ORDINAL
'34': B-CARDINAL
'35': I-CARDINAL
'36': '*'
'37': '0'
'38': '-'
- name: verbal_predicates
sequence: string
- name: coreference_clusters
sequence: string
splits:
- name: validation
num_bytes: 379380
num_examples: 396
- name: test
num_bytes: 402041
num_examples: 396
download_size: 846198
dataset_size: 781421
- config_name: type1_anti
features:
- name: document_id
dtype: string
- name: part_number
dtype: string
- name: word_number
sequence: int32
- name: tokens
sequence: string
- name: pos_tags
sequence:
class_label:
names:
'0': '"'
'1': ''''''
'2': '#'
'3': $
'4': (
'5': )
'6': ','
'7': .
'8': ':'
'9': '``'
'10': CC
'11': CD
'12': DT
'13': EX
'14': FW
'15': IN
'16': JJ
'17': JJR
'18': JJS
'19': LS
'20': MD
'21': NN
'22': NNP
'23': NNPS
'24': NNS
'25': NN|SYM
'26': PDT
'27': POS
'28': PRP
'29': PRP$
'30': RB
'31': RBR
'32': RBS
'33': RP
'34': SYM
'35': TO
'36': UH
'37': VB
'38': VBD
'39': VBG
'40': VBN
'41': VBP
'42': VBZ
'43': WDT
'44': WP
'45': WP$
'46': WRB
'47': HYPH
'48': XX
'49': NFP
'50': AFX
'51': ADD
'52': -LRB-
'53': -RRB-
'54': '-'
- name: parse_bit
sequence: string
- name: predicate_lemma
sequence: string
- name: predicate_framenet_id
sequence: string
- name: word_sense
sequence: string
- name: speaker
sequence: string
- name: ner_tags
sequence:
class_label:
names:
'0': B-PERSON
'1': I-PERSON
'2': B-NORP
'3': I-NORP
'4': B-FAC
'5': I-FAC
'6': B-ORG
'7': I-ORG
'8': B-GPE
'9': I-GPE
'10': B-LOC
'11': I-LOC
'12': B-PRODUCT
'13': I-PRODUCT
'14': B-EVENT
'15': I-EVENT
'16': B-WORK_OF_ART
'17': I-WORK_OF_ART
'18': B-LAW
'19': I-LAW
'20': B-LANGUAGE
'21': I-LANGUAGE
'22': B-DATE
'23': I-DATE
'24': B-TIME
'25': I-TIME
'26': B-PERCENT
'27': I-PERCENT
'28': B-MONEY
'29': I-MONEY
'30': B-QUANTITY
'31': I-QUANTITY
'32': B-ORDINAL
'33': I-ORDINAL
'34': B-CARDINAL
'35': I-CARDINAL
'36': '*'
'37': '0'
'38': '-'
- name: verbal_predicates
sequence: string
- name: coreference_clusters
sequence: string
splits:
- name: validation
num_bytes: 380846
num_examples: 396
- name: test
num_bytes: 403229
num_examples: 396
download_size: 894311
dataset_size: 784075
- config_name: type2_pro
features:
- name: document_id
dtype: string
- name: part_number
dtype: string
- name: word_number
sequence: int32
- name: tokens
sequence: string
- name: pos_tags
sequence:
class_label:
names:
'0': '"'
'1': ''''''
'2': '#'
'3': $
'4': (
'5': )
'6': ','
'7': .
'8': ':'
'9': '``'
'10': CC
'11': CD
'12': DT
'13': EX
'14': FW
'15': IN
'16': JJ
'17': JJR
'18': JJS
'19': LS
'20': MD
'21': NN
'22': NNP
'23': NNPS
'24': NNS
'25': NN|SYM
'26': PDT
'27': POS
'28': PRP
'29': PRP$
'30': RB
'31': RBR
'32': RBS
'33': RP
'34': SYM
'35': TO
'36': UH
'37': VB
'38': VBD
'39': VBG
'40': VBN
'41': VBP
'42': VBZ
'43': WDT
'44': WP
'45': WP$
'46': WRB
'47': HYPH
'48': XX
'49': NFP
'50': AFX
'51': ADD
'52': -LRB-
'53': -RRB-
'54': '-'
- name: parse_bit
sequence: string
- name: predicate_lemma
sequence: string
- name: predicate_framenet_id
sequence: string
- name: word_sense
sequence: string
- name: speaker
sequence: string
- name: ner_tags
sequence:
class_label:
names:
'0': B-PERSON
'1': I-PERSON
'2': B-NORP
'3': I-NORP
'4': B-FAC
'5': I-FAC
'6': B-ORG
'7': I-ORG
'8': B-GPE
'9': I-GPE
'10': B-LOC
'11': I-LOC
'12': B-PRODUCT
'13': I-PRODUCT
'14': B-EVENT
'15': I-EVENT
'16': B-WORK_OF_ART
'17': I-WORK_OF_ART
'18': B-LAW
'19': I-LAW
'20': B-LANGUAGE
'21': I-LANGUAGE
'22': B-DATE
'23': I-DATE
'24': B-TIME
'25': I-TIME
'26': B-PERCENT
'27': I-PERCENT
'28': B-MONEY
'29': I-MONEY
'30': B-QUANTITY
'31': I-QUANTITY
'32': B-ORDINAL
'33': I-ORDINAL
'34': B-CARDINAL
'35': I-CARDINAL
'36': '*'
'37': '0'
'38': '-'
- name: verbal_predicates
sequence: string
- name: coreference_clusters
sequence: string
splits:
- name: validation
num_bytes: 367293
num_examples: 396
- name: test
num_bytes: 375480
num_examples: 396
download_size: 802425
dataset_size: 742773
- config_name: type2_anti
features:
- name: document_id
dtype: string
- name: part_number
dtype: string
- name: word_number
sequence: int32
- name: tokens
sequence: string
- name: pos_tags
sequence:
class_label:
names:
'0': '"'
'1': ''''''
'2': '#'
'3': $
'4': (
'5': )
'6': ','
'7': .
'8': ':'
'9': '``'
'10': CC
'11': CD
'12': DT
'13': EX
'14': FW
'15': IN
'16': JJ
'17': JJR
'18': JJS
'19': LS
'20': MD
'21': NN
'22': NNP
'23': NNPS
'24': NNS
'25': NN|SYM
'26': PDT
'27': POS
'28': PRP
'29': PRP$
'30': RB
'31': RBR
'32': RBS
'33': RP
'34': SYM
'35': TO
'36': UH
'37': VB
'38': VBD
'39': VBG
'40': VBN
'41': VBP
'42': VBZ
'43': WDT
'44': WP
'45': WP$
'46': WRB
'47': HYPH
'48': XX
'49': NFP
'50': AFX
'51': ADD
'52': -LRB-
'53': -RRB-
'54': '-'
- name: parse_bit
sequence: string
- name: predicate_lemma
sequence: string
- name: predicate_framenet_id
sequence: string
- name: word_sense
sequence: string
- name: speaker
sequence: string
- name: ner_tags
sequence:
class_label:
names:
'0': B-PERSON
'1': I-PERSON
'2': B-NORP
'3': I-NORP
'4': B-FAC
'5': I-FAC
'6': B-ORG
'7': I-ORG
'8': B-GPE
'9': I-GPE
'10': B-LOC
'11': I-LOC
'12': B-PRODUCT
'13': I-PRODUCT
'14': B-EVENT
'15': I-EVENT
'16': B-WORK_OF_ART
'17': I-WORK_OF_ART
'18': B-LAW
'19': I-LAW
'20': B-LANGUAGE
'21': I-LANGUAGE
'22': B-DATE
'23': I-DATE
'24': B-TIME
'25': I-TIME
'26': B-PERCENT
'27': I-PERCENT
'28': B-MONEY
'29': I-MONEY
'30': B-QUANTITY
'31': I-QUANTITY
'32': B-ORDINAL
'33': I-ORDINAL
'34': B-CARDINAL
'35': I-CARDINAL
'36': '*'
'37': '0'
'38': '-'
- name: verbal_predicates
sequence: string
- name: coreference_clusters
sequence: string
splits:
- name: validation
num_bytes: 368757
num_examples: 396
- name: test
num_bytes: 377262
num_examples: 396
download_size: 848804
dataset_size: 746019
---
# Dataset Card for Wino_Bias dataset
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [WinoBias](https://uclanlp.github.io/corefBias/overview)
- **Repository:**
- **Paper:** [Arxiv](https://arxiv.org/abs/1804.06876)
- **Leaderboard:**
- **Point of Contact:**
### Dataset Summary
WinoBias, a Winograd-schema dataset for coreference resolution focused on gender bias.
The corpus contains Winograd-schema style sentences with entities corresponding to people referred by their occupation (e.g. the nurse, the doctor, the carpenter).
### Supported Tasks and Leaderboards
The underlying task is coreference resolution.
### Languages
English
## Dataset Structure
### Data Instances
The dataset has 4 subsets: `type1_pro`, `type1_anti`, `type2_pro` and `type2_anti`.
The `*_pro` subsets contain sentences that reinforce gender stereotypes (e.g. mechanics are male, nurses are female), whereas the `*_anti` datasets contain "anti-stereotypical" sentences (e.g. mechanics are female, nurses are male).
The `type1` (*WB-Knowledge*) subsets contain sentences for which world knowledge is necessary to resolve the co-references, and `type2` (*WB-Syntax*) subsets require only the syntactic information present in the sentence to resolve them.
### Data Fields
- document_id = This is a variation on the document filename
- part_number = Some files are divided into multiple parts numbered as 000, 001, 002, ... etc.
- word_num = This is the word index of the word in that sentence.
- tokens = This is the token as segmented/tokenized in the Treebank.
- pos_tags = This is the Penn Treebank style part of speech. When parse information is missing, all part of speeches except the one for which there is some sense or proposition annotation are marked with a XX tag. The verb is marked with just a VERB tag.
- parse_bit = This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. When the parse information is missing, the first word of a sentence is tagged as "(TOP*" and the last word is tagged as "*)" and all intermediate words are tagged with a "*".
- predicate_lemma = The predicate lemma is mentioned for the rows for which we have semantic role information or word sense information. All other rows are marked with a "-".
- predicate_framenet_id = This is the PropBank frameset ID of the predicate in predicate_lemma.
- word_sense = This is the word sense of the word in Column tokens.
- speaker = This is the speaker or author name where available.
- ner_tags = These columns identifies the spans representing various named entities. For documents which do not have named entity annotation, each line is represented with an "*".
- verbal_predicates = There is one column each of predicate argument structure information for the predicate mentioned in predicate_lemma. If there are no predicates tagged in a sentence this is a single column with all rows marked with an "*".
### Data Splits
Dev and Test Split available
## Dataset Creation
### Curation Rationale
The WinoBias dataset was introduced in 2018 (see [paper](https://arxiv.org/abs/1804.06876)), with its original task being *coreference resolution*, which is a task that aims to identify mentions that refer to the same entity or person.
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
The dataset was created by researchers familiar with the WinoBias project, based on two prototypical templates provided by the authors, in which entities interact in plausible ways.
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
"Researchers familiar with the [WinoBias] project"
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[Recent work](https://www.microsoft.com/en-us/research/uploads/prod/2021/06/The_Salmon_paper.pdf) has shown that this dataset contains grammatical issues, incorrect or ambiguous labels, and stereotype conflation, among other limitations.
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez and Kai-Wei Chan
### Licensing Information
MIT Licence
### Citation Information
@article{DBLP:journals/corr/abs-1804-06876,
author = {Jieyu Zhao and
Tianlu Wang and
Mark Yatskar and
Vicente Ordonez and
Kai{-}Wei Chang},
title = {Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods},
journal = {CoRR},
volume = {abs/1804.06876},
year = {2018},
url = {http://arxiv.org/abs/1804.06876},
archivePrefix = {arXiv},
eprint = {1804.06876},
timestamp = {Mon, 13 Aug 2018 16:47:01 +0200},
biburl = {https://dblp.org/rec/journals/corr/abs-1804-06876.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
### Contributions
Thanks to [@akshayb7](https://github.com/akshayb7) for adding this dataset. Updated by [@JieyuZhao](https://github.com/JieyuZhao). |
CarperAI/openai_summarize_tldr | 2023-01-10T02:53:40.000Z | [
"region:us"
] | CarperAI | null | null | null | 12 | 10,410 | ---
dataset_info:
features:
- name: prompt
dtype: string
- name: label
dtype: string
splits:
- name: train
num_bytes: 181260841
num_examples: 116722
- name: valid
num_bytes: 10018338
num_examples: 6447
- name: test
num_bytes: 10198128
num_examples: 6553
download_size: 122973500
dataset_size: 201477307
---
# Dataset Card for "openai_summarize_tldr"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
NLPCoreTeam/mmlu_ru | 2023-06-28T19:21:48.000Z | [
"task_categories:question-answering",
"task_categories:multiple-choice",
"task_ids:multiple-choice-qa",
"size_categories:10K<n<100K",
"language:ru",
"language:en",
"arxiv:2009.03300",
"region:us"
] | NLPCoreTeam | null | null | null | 5 | 10,340 | ---
pretty_name: MMLU RU/EN
language:
- ru
- en
size_categories:
- 10K<n<100K
task_categories:
- question-answering
- multiple-choice
task_ids:
- multiple-choice-qa
dataset_info:
- config_name: abstract_algebra
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 2182
num_examples: 5
- name: val
num_bytes: 5220
num_examples: 11
- name: test
num_bytes: 50926
num_examples: 100
download_size: 5548198
dataset_size: 58328
- config_name: anatomy
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 2482
num_examples: 5
- name: val
num_bytes: 8448
num_examples: 14
- name: test
num_bytes: 91387
num_examples: 135
download_size: 5548198
dataset_size: 102317
- config_name: astronomy
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 6049
num_examples: 5
- name: val
num_bytes: 14187
num_examples: 16
- name: test
num_bytes: 130167
num_examples: 152
download_size: 5548198
dataset_size: 150403
- config_name: business_ethics
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 6197
num_examples: 5
- name: val
num_bytes: 8963
num_examples: 11
- name: test
num_bytes: 96566
num_examples: 100
download_size: 5548198
dataset_size: 111726
- config_name: clinical_knowledge
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 3236
num_examples: 5
- name: val
num_bytes: 18684
num_examples: 29
- name: test
num_bytes: 178043
num_examples: 265
download_size: 5548198
dataset_size: 199963
- config_name: college_biology
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 4232
num_examples: 5
- name: val
num_bytes: 13521
num_examples: 16
- name: test
num_bytes: 139322
num_examples: 144
download_size: 5548198
dataset_size: 157075
- config_name: college_chemistry
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 3533
num_examples: 5
- name: val
num_bytes: 6157
num_examples: 8
- name: test
num_bytes: 65540
num_examples: 100
download_size: 5548198
dataset_size: 75230
- config_name: college_computer_science
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 7513
num_examples: 5
- name: val
num_bytes: 13341
num_examples: 11
- name: test
num_bytes: 120578
num_examples: 100
download_size: 5548198
dataset_size: 141432
- config_name: college_mathematics
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 3841
num_examples: 5
- name: val
num_bytes: 6835
num_examples: 11
- name: test
num_bytes: 65110
num_examples: 100
download_size: 5548198
dataset_size: 75786
- config_name: college_medicine
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 4659
num_examples: 5
- name: val
num_bytes: 22116
num_examples: 22
- name: test
num_bytes: 235856
num_examples: 173
download_size: 5548198
dataset_size: 262631
- config_name: college_physics
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 3740
num_examples: 5
- name: val
num_bytes: 9491
num_examples: 11
- name: test
num_bytes: 81480
num_examples: 102
download_size: 5548198
dataset_size: 94711
- config_name: computer_security
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 3150
num_examples: 5
- name: val
num_bytes: 12859
num_examples: 11
- name: test
num_bytes: 77969
num_examples: 100
download_size: 5548198
dataset_size: 93978
- config_name: conceptual_physics
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 2611
num_examples: 5
- name: val
num_bytes: 12480
num_examples: 26
- name: test
num_bytes: 112243
num_examples: 235
download_size: 5548198
dataset_size: 127334
- config_name: econometrics
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 4548
num_examples: 5
- name: val
num_bytes: 13874
num_examples: 12
- name: test
num_bytes: 128633
num_examples: 114
download_size: 5548198
dataset_size: 147055
- config_name: electrical_engineering
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 2598
num_examples: 5
- name: val
num_bytes: 8003
num_examples: 16
- name: test
num_bytes: 70846
num_examples: 145
download_size: 5548198
dataset_size: 81447
- config_name: elementary_mathematics
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 3760
num_examples: 5
- name: val
num_bytes: 23416
num_examples: 41
- name: test
num_bytes: 181090
num_examples: 378
download_size: 5548198
dataset_size: 208266
- config_name: formal_logic
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 4715
num_examples: 5
- name: val
num_bytes: 17099
num_examples: 14
- name: test
num_bytes: 133930
num_examples: 126
download_size: 5548198
dataset_size: 155744
- config_name: global_facts
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 3450
num_examples: 5
- name: val
num_bytes: 4971
num_examples: 10
- name: test
num_bytes: 51481
num_examples: 100
download_size: 5548198
dataset_size: 59902
- config_name: high_school_biology
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 4759
num_examples: 5
- name: val
num_bytes: 30807
num_examples: 32
- name: test
num_bytes: 310356
num_examples: 310
download_size: 5548198
dataset_size: 345922
- config_name: high_school_chemistry
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 3204
num_examples: 5
- name: val
num_bytes: 18948
num_examples: 22
- name: test
num_bytes: 158246
num_examples: 203
download_size: 5548198
dataset_size: 180398
- config_name: high_school_computer_science
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 7933
num_examples: 5
- name: val
num_bytes: 9612
num_examples: 9
- name: test
num_bytes: 126403
num_examples: 100
download_size: 5548198
dataset_size: 143948
- config_name: high_school_european_history
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 32447
num_examples: 5
- name: val
num_bytes: 83098
num_examples: 18
- name: test
num_bytes: 754136
num_examples: 165
download_size: 5548198
dataset_size: 869681
- config_name: high_school_geography
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 4131
num_examples: 5
- name: val
num_bytes: 12467
num_examples: 22
- name: test
num_bytes: 119021
num_examples: 198
download_size: 5548198
dataset_size: 135619
- config_name: high_school_government_and_politics
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 5188
num_examples: 5
- name: val
num_bytes: 20564
num_examples: 21
- name: test
num_bytes: 194050
num_examples: 193
download_size: 5548198
dataset_size: 219802
- config_name: high_school_macroeconomics
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 3942
num_examples: 5
- name: val
num_bytes: 37243
num_examples: 43
- name: test
num_bytes: 340699
num_examples: 390
download_size: 5548198
dataset_size: 381884
- config_name: high_school_mathematics
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 3244
num_examples: 5
- name: val
num_bytes: 14758
num_examples: 29
- name: test
num_bytes: 140257
num_examples: 270
download_size: 5548198
dataset_size: 158259
- config_name: high_school_microeconomics
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 3503
num_examples: 5
- name: val
num_bytes: 22212
num_examples: 26
- name: test
num_bytes: 219097
num_examples: 238
download_size: 5548198
dataset_size: 244812
- config_name: high_school_physics
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 3905
num_examples: 5
- name: val
num_bytes: 18535
num_examples: 17
- name: test
num_bytes: 162917
num_examples: 151
download_size: 5548198
dataset_size: 185357
- config_name: high_school_psychology
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 5207
num_examples: 5
- name: val
num_bytes: 49277
num_examples: 60
- name: test
num_bytes: 455603
num_examples: 545
download_size: 5548198
dataset_size: 510087
- config_name: high_school_statistics
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 6823
num_examples: 5
- name: val
num_bytes: 28020
num_examples: 23
- name: test
num_bytes: 312578
num_examples: 216
download_size: 5548198
dataset_size: 347421
- config_name: high_school_us_history
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 25578
num_examples: 5
- name: val
num_bytes: 91278
num_examples: 22
- name: test
num_bytes: 842680
num_examples: 204
download_size: 5548198
dataset_size: 959536
- config_name: high_school_world_history
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 13893
num_examples: 5
- name: val
num_bytes: 129121
num_examples: 26
- name: test
num_bytes: 1068018
num_examples: 237
download_size: 5548198
dataset_size: 1211032
- config_name: human_aging
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 2820
num_examples: 5
- name: val
num_bytes: 13442
num_examples: 23
- name: test
num_bytes: 132242
num_examples: 223
download_size: 5548198
dataset_size: 148504
- config_name: human_sexuality
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 3072
num_examples: 5
- name: val
num_bytes: 6699
num_examples: 12
- name: test
num_bytes: 90007
num_examples: 131
download_size: 5548198
dataset_size: 99778
- config_name: international_law
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 6880
num_examples: 5
- name: val
num_bytes: 19166
num_examples: 13
- name: test
num_bytes: 157259
num_examples: 121
download_size: 5548198
dataset_size: 183305
- config_name: jurisprudence
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 3568
num_examples: 5
- name: val
num_bytes: 10638
num_examples: 11
- name: test
num_bytes: 97121
num_examples: 108
download_size: 5548198
dataset_size: 111327
- config_name: logical_fallacies
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 4526
num_examples: 5
- name: val
num_bytes: 14547
num_examples: 18
- name: test
num_bytes: 144501
num_examples: 163
download_size: 5548198
dataset_size: 163574
- config_name: machine_learning
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 6966
num_examples: 5
- name: val
num_bytes: 8986
num_examples: 11
- name: test
num_bytes: 95571
num_examples: 112
download_size: 5548198
dataset_size: 111523
- config_name: management
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 2427
num_examples: 5
- name: val
num_bytes: 5210
num_examples: 11
- name: test
num_bytes: 57201
num_examples: 103
download_size: 5548198
dataset_size: 64838
- config_name: marketing
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 4514
num_examples: 5
- name: val
num_bytes: 20832
num_examples: 25
- name: test
num_bytes: 181786
num_examples: 234
download_size: 5548198
dataset_size: 207132
- config_name: medical_genetics
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 3226
num_examples: 5
- name: val
num_bytes: 8214
num_examples: 11
- name: test
num_bytes: 57064
num_examples: 100
download_size: 5548198
dataset_size: 68504
- config_name: miscellaneous
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 1782
num_examples: 5
- name: val
num_bytes: 39225
num_examples: 86
- name: test
num_bytes: 407209
num_examples: 783
download_size: 5548198
dataset_size: 448216
- config_name: moral_disputes
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 4910
num_examples: 5
- name: val
num_bytes: 36026
num_examples: 38
- name: test
num_bytes: 313611
num_examples: 346
download_size: 5548198
dataset_size: 354547
- config_name: moral_scenarios
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 6175
num_examples: 5
- name: val
num_bytes: 129062
num_examples: 100
- name: test
num_bytes: 1137631
num_examples: 895
download_size: 5548198
dataset_size: 1272868
- config_name: nutrition
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 6030
num_examples: 5
- name: val
num_bytes: 24210
num_examples: 33
- name: test
num_bytes: 266173
num_examples: 306
download_size: 5548198
dataset_size: 296413
- config_name: philosophy
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 2631
num_examples: 5
- name: val
num_bytes: 25751
num_examples: 34
- name: test
num_bytes: 227086
num_examples: 311
download_size: 5548198
dataset_size: 255468
- config_name: prehistory
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 5394
num_examples: 5
- name: val
num_bytes: 28687
num_examples: 35
- name: test
num_bytes: 251723
num_examples: 324
download_size: 5548198
dataset_size: 285804
- config_name: professional_accounting
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 6277
num_examples: 5
- name: val
num_bytes: 40914
num_examples: 31
- name: test
num_bytes: 364528
num_examples: 282
download_size: 5548198
dataset_size: 411719
- config_name: professional_law
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 19120
num_examples: 5
- name: val
num_bytes: 589307
num_examples: 170
- name: test
num_bytes: 5479411
num_examples: 1534
download_size: 5548198
dataset_size: 6087838
- config_name: professional_medicine
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 10901
num_examples: 5
- name: val
num_bytes: 69703
num_examples: 31
- name: test
num_bytes: 633483
num_examples: 272
download_size: 5548198
dataset_size: 714087
- config_name: professional_psychology
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 6430
num_examples: 5
- name: val
num_bytes: 82745
num_examples: 69
- name: test
num_bytes: 648634
num_examples: 612
download_size: 5548198
dataset_size: 737809
- config_name: public_relations
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 4384
num_examples: 5
- name: val
num_bytes: 13108
num_examples: 12
- name: test
num_bytes: 82403
num_examples: 110
download_size: 5548198
dataset_size: 99895
- config_name: security_studies
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 16064
num_examples: 5
- name: val
num_bytes: 67877
num_examples: 27
- name: test
num_bytes: 611059
num_examples: 245
download_size: 5548198
dataset_size: 695000
- config_name: sociology
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 4693
num_examples: 5
- name: val
num_bytes: 20654
num_examples: 22
- name: test
num_bytes: 191420
num_examples: 201
download_size: 5548198
dataset_size: 216767
- config_name: us_foreign_policy
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 4781
num_examples: 5
- name: val
num_bytes: 9171
num_examples: 11
- name: test
num_bytes: 81649
num_examples: 100
download_size: 5548198
dataset_size: 95601
- config_name: virology
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 3063
num_examples: 5
- name: val
num_bytes: 15618
num_examples: 18
- name: test
num_bytes: 111027
num_examples: 166
download_size: 5548198
dataset_size: 129708
- config_name: world_religions
features:
- name: question_en
dtype: string
- name: choices_en
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
- name: question_ru
dtype: string
- name: choices_ru
sequence: string
splits:
- name: dev
num_bytes: 1691
num_examples: 5
- name: val
num_bytes: 7052
num_examples: 19
- name: test
num_bytes: 65559
num_examples: 171
download_size: 5548198
dataset_size: 74302
---
# MMLU in Russian (Massive Multitask Language Understanding)
## Overview of the Dataset
MMLU dataset for EN/RU, without auxiliary train.
The dataset contains `dev`/`val`/`test` splits for both, English and Russian languages.
Note it doesn't include `auxiliary_train` split, which wasn't translated.
Totally the dataset has ~16k samples per language: 285 `dev`, 1531 `val`, 14042 `test`.
## Description of original MMLU
MMLU dataset covers 57 different tasks.
Each task requires to choose the right answer out of four options for a given question.
Paper "Measuring Massive Multitask Language Understanding": https://arxiv.org/abs/2009.03300v3.
It is also known as the "hendrycks_test".
## Dataset Creation
The translation was made via Yandex.Translate API.
There are some translation mistakes, especially observed with terms and formulas, no fixes were applied.
Initial dataset was taken from: https://people.eecs.berkeley.edu/~hendrycks/data.tar.
## Sample example
```
{
"question_en": "Why doesn't Venus have seasons like Mars and Earth do?",
"choices_en": [
"Its rotation axis is nearly perpendicular to the plane of the Solar System.",
"It does not have an ozone layer.",
"It does not rotate fast enough.",
"It is too close to the Sun."
],
"answer": 0,
"question_ru": "Почему на Венере нет времен года, как на Марсе и Земле?",
"choices_ru": [
"Ось его вращения почти перпендикулярна плоскости Солнечной системы.",
"У него нет озонового слоя.",
"Он вращается недостаточно быстро.",
"Это слишком близко к Солнцу."
]
}
```
## Usage
To merge all subsets into dataframe per split:
```python
from collections import defaultdict
import datasets
import pandas as pd
subjects = ["abstract_algebra", "anatomy", "astronomy", "business_ethics", "clinical_knowledge", "college_biology", "college_chemistry", "college_computer_science", "college_mathematics", "college_medicine", "college_physics", "computer_security", "conceptual_physics", "econometrics", "electrical_engineering", "elementary_mathematics", "formal_logic", "global_facts", "high_school_biology", "high_school_chemistry", "high_school_computer_science", "high_school_european_history", "high_school_geography", "high_school_government_and_politics", "high_school_macroeconomics", "high_school_mathematics", "high_school_microeconomics", "high_school_physics", "high_school_psychology", "high_school_statistics", "high_school_us_history", "high_school_world_history", "human_aging", "human_sexuality", "international_law", "jurisprudence", "logical_fallacies", "machine_learning", "management", "marketing", "medical_genetics", "miscellaneous", "moral_disputes", "moral_scenarios", "nutrition", "philosophy", "prehistory", "professional_accounting", "professional_law", "professional_medicine", "professional_psychology", "public_relations", "security_studies", "sociology", "us_foreign_policy", "virology", "world_religions"]
splits = ["dev", "val", "test"]
all_datasets = {x: datasets.load_dataset("NLPCoreTeam/mmlu_ru", name=x) for x in subjects}
res = defaultdict(list)
for subject in subjects:
for split in splits:
dataset = all_datasets[subject][split]
df = dataset.to_pandas()
int2str = dataset.features['answer'].int2str
df['answer'] = df['answer'].map(int2str)
df.insert(loc=0, column='subject_en', value=subject)
res[split].append(df)
res = {k: pd.concat(v) for k, v in res.items()}
df_dev = res['dev']
df_val = res['val']
df_test = res['test']
```
## Evaluation
This dataset is intended to evaluate LLMs with few-shot/zero-shot setup.
Evaluation code: https://github.com/NLP-Core-Team/mmlu_ru
Also resources might be helpful:
1. https://github.com/hendrycks/test
1. https://github.com/openai/evals/blob/main/examples/mmlu.ipynb
1. https://github.com/EleutherAI/lm-evaluation-harness/blob/master/lm_eval/tasks/hendrycks_test.py
## Contributions
Dataset added by NLP core team RnD [Telegram channel](https://t.me/nlpcoreteam) |
huggingface/cats-image | 2022-02-03T12:31:30.000Z | [
"region:us"
] | huggingface | \\n | \\n | null | 0 | 10,326 | Entry not found |
hotpot_qa | 2023-04-05T10:07:23.000Z | [
"task_categories:question-answering",
"annotations_creators:crowdsourced",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:100K<n<1M",
"source_datasets:original",
"language:en",
"license:cc-by-sa-4.0",
"multi-hop",
"arxiv:1809.09600",
"region:us"
] | null | HotpotQA is a new dataset with 113k Wikipedia-based question-answer pairs with four key features:
(1) the questions require finding and reasoning over multiple supporting documents to answer;
(2) the questions are diverse and not constrained to any pre-existing knowledge bases or knowledge schemas;
(3) we provide sentence-level supporting facts required for reasoning, allowingQA systems to reason with strong supervisionand explain the predictions;
(4) we offer a new type of factoid comparison questions to testQA systems’ ability to extract relevant facts and perform necessary comparison. | @inproceedings{yang2018hotpotqa,
title={{HotpotQA}: A Dataset for Diverse, Explainable Multi-hop Question Answering},
author={Yang, Zhilin and Qi, Peng and Zhang, Saizheng and Bengio, Yoshua and Cohen, William W. and Salakhutdinov, Ruslan and Manning, Christopher D.},
booktitle={Conference on Empirical Methods in Natural Language Processing ({EMNLP})},
year={2018}
} | null | 18 | 10,306 | ---
annotations_creators:
- crowdsourced
language:
- en
language_creators:
- found
license:
- cc-by-sa-4.0
multilinguality:
- monolingual
pretty_name: HotpotQA
size_categories:
- 100K<n<1M
source_datasets:
- original
task_categories:
- question-answering
task_ids: []
paperswithcode_id: hotpotqa
tags:
- multi-hop
dataset_info:
- config_name: distractor
features:
- name: id
dtype: string
- name: question
dtype: string
- name: answer
dtype: string
- name: type
dtype: string
- name: level
dtype: string
- name: supporting_facts
sequence:
- name: title
dtype: string
- name: sent_id
dtype: int32
- name: context
sequence:
- name: title
dtype: string
- name: sentences
sequence: string
splits:
- name: train
num_bytes: 552949315
num_examples: 90447
- name: validation
num_bytes: 45716111
num_examples: 7405
download_size: 612746344
dataset_size: 598665426
- config_name: fullwiki
features:
- name: id
dtype: string
- name: question
dtype: string
- name: answer
dtype: string
- name: type
dtype: string
- name: level
dtype: string
- name: supporting_facts
sequence:
- name: title
dtype: string
- name: sent_id
dtype: int32
- name: context
sequence:
- name: title
dtype: string
- name: sentences
sequence: string
splits:
- name: train
num_bytes: 552949315
num_examples: 90447
- name: validation
num_bytes: 46848601
num_examples: 7405
- name: test
num_bytes: 46000102
num_examples: 7405
download_size: 660094672
dataset_size: 645798018
---
# Dataset Card for "hotpot_qa"
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://hotpotqa.github.io/](https://hotpotqa.github.io/)
- **Repository:** https://github.com/hotpotqa/hotpot
- **Paper:** [HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering](https://arxiv.org/abs/1809.09600)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 1.27 GB
- **Size of the generated dataset:** 1.24 GB
- **Total amount of disk used:** 2.52 GB
### Dataset Summary
HotpotQA is a new dataset with 113k Wikipedia-based question-answer pairs with four key features: (1) the questions require finding and reasoning over multiple supporting documents to answer; (2) the questions are diverse and not constrained to any pre-existing knowledge bases or knowledge schemas; (3) we provide sentence-level supporting facts required for reasoning, allowingQA systems to reason with strong supervision and explain the predictions; (4) we offer a new type of factoid comparison questions to test QA systems’ ability to extract relevant facts and perform necessary comparison.
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Dataset Structure
### Data Instances
#### distractor
- **Size of downloaded dataset files:** 612.75 MB
- **Size of the generated dataset:** 598.66 MB
- **Total amount of disk used:** 1.21 GB
An example of 'validation' looks as follows.
```
{
"answer": "This is the answer",
"context": {
"sentences": [["Sent 1"], ["Sent 21", "Sent 22"]],
"title": ["Title1", "Title 2"]
},
"id": "000001",
"level": "medium",
"question": "What is the answer?",
"supporting_facts": {
"sent_id": [0, 1, 3],
"title": ["Title of para 1", "Title of para 2", "Title of para 3"]
},
"type": "comparison"
}
```
#### fullwiki
- **Size of downloaded dataset files:** 660.10 MB
- **Size of the generated dataset:** 645.80 MB
- **Total amount of disk used:** 1.31 GB
An example of 'train' looks as follows.
```
{
"answer": "This is the answer",
"context": {
"sentences": [["Sent 1"], ["Sent 2"]],
"title": ["Title1", "Title 2"]
},
"id": "000001",
"level": "hard",
"question": "What is the answer?",
"supporting_facts": {
"sent_id": [0, 1, 3],
"title": ["Title of para 1", "Title of para 2", "Title of para 3"]
},
"type": "bridge"
}
```
### Data Fields
The data fields are the same among all splits.
#### distractor
- `id`: a `string` feature.
- `question`: a `string` feature.
- `answer`: a `string` feature.
- `type`: a `string` feature.
- `level`: a `string` feature.
- `supporting_facts`: a dictionary feature containing:
- `title`: a `string` feature.
- `sent_id`: a `int32` feature.
- `context`: a dictionary feature containing:
- `title`: a `string` feature.
- `sentences`: a `list` of `string` features.
#### fullwiki
- `id`: a `string` feature.
- `question`: a `string` feature.
- `answer`: a `string` feature.
- `type`: a `string` feature.
- `level`: a `string` feature.
- `supporting_facts`: a dictionary feature containing:
- `title`: a `string` feature.
- `sent_id`: a `int32` feature.
- `context`: a dictionary feature containing:
- `title`: a `string` feature.
- `sentences`: a `list` of `string` features.
### Data Splits
#### distractor
| |train|validation|
|----------|----:|---------:|
|distractor|90447| 7405|
#### fullwiki
| |train|validation|test|
|--------|----:|---------:|---:|
|fullwiki|90447| 7405|7405|
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
HotpotQA is distributed under a [CC BY-SA 4.0 License](http://creativecommons.org/licenses/by-sa/4.0/).
### Citation Information
```
@inproceedings{yang2018hotpotqa,
title={{HotpotQA}: A Dataset for Diverse, Explainable Multi-hop Question Answering},
author={Yang, Zhilin and Qi, Peng and Zhang, Saizheng and Bengio, Yoshua and Cohen, William W. and Salakhutdinov, Ruslan and Manning, Christopher D.},
booktitle={Conference on Empirical Methods in Natural Language Processing ({EMNLP})},
year={2018}
}
```
### Contributions
Thanks to [@albertvillanova](https://github.com/albertvillanova), [@ghomasHudson](https://github.com/ghomasHudson) for adding this dataset. |
tatsu-lab/alpaca_farm | 2023-05-29T01:00:10.000Z | [
"license:cc-by-nc-4.0",
"region:us"
] | tatsu-lab | Data used in the original AlpacaFarm experiments.
Includes SFT and preference examples. | @misc{alpaca_farm,
author = {Yann Dubois, Xuechen Li, Rohan Taori, Tianyi Zhang, Ishaan Gulrajani, Jimmy Ba, Carlos Guestrin, Percy Liang, Tatsunori Hashimoto},
title = {AlpacaFarm: A Simulation Framework for Methods that Learn from Human Feedback},
year = {2023},
howpublished = {\\url{https://github.com/tatsu-lab/alpaca_farm}},
} | null | 14 | 10,181 | ---
license: cc-by-nc-4.0
--- |
competition_math | 2023-06-08T06:40:09.000Z | [
"task_categories:text2text-generation",
"annotations_creators:expert-generated",
"language_creators:expert-generated",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:mit",
"explanation-generation",
"arxiv:2103.03874",
"region:us"
] | null | The Mathematics Aptitude Test of Heuristics (MATH) dataset consists of problems
from mathematics competitions, including the AMC 10, AMC 12, AIME, and more.
Each problem in MATH has a full step-by-step solution, which can be used to teach
models to generate answer derivations and explanations. | @article{hendrycksmath2021,
title={Measuring Mathematical Problem Solving With the MATH Dataset},
author={Dan Hendrycks
and Collin Burns
and Saurav Kadavath
and Akul Arora
and Steven Basart
and Eric Tang
and Dawn Song
and Jacob Steinhardt},
journal={arXiv preprint arXiv:2103.03874},
year={2021}
} | null | 51 | 9,916 | ---
annotations_creators:
- expert-generated
language_creators:
- expert-generated
language:
- en
license:
- mit
multilinguality:
- monolingual
pretty_name: Mathematics Aptitude Test of Heuristics (MATH)
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- text2text-generation
task_ids: []
tags:
- explanation-generation
dataset_info:
features:
- name: problem
dtype: string
- name: level
dtype: string
- name: type
dtype: string
- name: solution
dtype: string
splits:
- name: train
num_bytes: 5984788
num_examples: 7500
- name: test
num_bytes: 3732575
num_examples: 5000
download_size: 20327424
dataset_size: 9717363
---
# Dataset Card for Mathematics Aptitude Test of Heuristics (MATH) dataset
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** https://github.com/hendrycks/math
- **Repository:** https://github.com/hendrycks/math
- **Paper:** https://arxiv.org/pdf/2103.03874.pdf
- **Leaderboard:** N/A
- **Point of Contact:** Dan Hendrycks
### Dataset Summary
The Mathematics Aptitude Test of Heuristics (MATH) dataset consists of problems
from mathematics competitions, including the AMC 10, AMC 12, AIME, and more.
Each problem in MATH has a full step-by-step solution, which can be used to teach
models to generate answer derivations and explanations.
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
A data instance consists of a competition math problem and its step-by-step solution written in LaTeX and natural language. The step-by-step solution contains the final answer enclosed in LaTeX's `\boxed` tag.
An example from the dataset is:
```
{'problem': 'A board game spinner is divided into three parts labeled $A$, $B$ and $C$. The probability of the spinner landing on $A$ is $\\frac{1}{3}$ and the probability of the spinner landing on $B$ is $\\frac{5}{12}$. What is the probability of the spinner landing on $C$? Express your answer as a common fraction.',
'level': 'Level 1',
'type': 'Counting & Probability',
'solution': 'The spinner is guaranteed to land on exactly one of the three regions, so we know that the sum of the probabilities of it landing in each region will be 1. If we let the probability of it landing in region $C$ be $x$, we then have the equation $1 = \\frac{5}{12}+\\frac{1}{3}+x$, from which we have $x=\\boxed{\\frac{1}{4}}$.'}
```
### Data Fields
* `problem`: The competition math problem.
* `solution`: The step-by-step solution.
* `level`: The problem's difficulty level from 'Level 1' to 'Level 5', where a subject's easiest problems for humans are assigned to 'Level 1' and a subject's hardest problems are assigned to 'Level 5'.
* `type`: The subject of the problem: Algebra, Counting & Probability, Geometry, Intermediate Algebra, Number Theory, Prealgebra and Precalculus.
### Data Splits
* train: 7,500 examples
* test: 5,000 examples
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
https://github.com/hendrycks/math/blob/main/LICENSE
### Citation Information
```bibtex
@article{hendrycksmath2021,
title={Measuring Mathematical Problem Solving With the MATH Dataset},
author={Dan Hendrycks
and Collin Burns
and Saurav Kadavath
and Akul Arora
and Steven Basart
and Eric Tang
and Dawn Song
and Jacob Steinhardt},
journal={arXiv preprint arXiv:2103.03874},
year={2021}
}
```
### Contributions
Thanks to [@hacobe](https://github.com/hacobe) for adding this dataset. |
lvwerra/stack-exchange-paired | 2023-03-13T11:30:17.000Z | [
"task_categories:text-generation",
"task_categories:question-answering",
"size_categories:10M<n<100M",
"language:en",
"region:us"
] | lvwerra | null | null | null | 68 | 9,739 | ---
task_categories:
- text-generation
- question-answering
language:
- en
pretty_name: StackExchange Paired
size_categories:
- 10M<n<100M
---
# StackExchange Paired
This is a processed version of the [`HuggingFaceH4/stack-exchange-preferences`](https://huggingface.co/datasets/HuggingFaceH4/stack-exchange-preferences). The following steps were applied:
- Parse HTML to Markdown with `markdownify`
- Create pairs `(response_j, response_k)` where j was rated better than k
- Sample at most 10 pairs per question
- Shuffle the dataset globally
This dataset is designed to be used for preference learning. The processing notebook is in [the repository](https://huggingface.co/datasets/lvwerra/stack-exchange-paired/tree/main) as well.
|
OpenAssistant/oasst1 | 2023-05-02T13:21:21.000Z | [
"size_categories:100K<n<1M",
"language:en",
"language:es",
"language:ru",
"language:de",
"language:pl",
"language:th",
"language:vi",
"language:sv",
"language:bn",
"language:da",
"language:he",
"language:it",
"language:fa",
"language:sk",
"language:id",
"language:nb",
"language:el",
"language:nl",
"language:hu",
"language:eu",
"language:zh",
"language:eo",
"language:ja",
"language:ca",
"language:cs",
"language:bg",
"language:fi",
"language:pt",
"language:tr",
"language:ro",
"language:ar",
"language:uk",
"language:gl",
"language:fr",
"language:ko",
"license:apache-2.0",
"human-feedback",
"arxiv:2304.07327",
"region:us"
] | OpenAssistant | null | null | null | 1,040 | 9,684 | ---
license: apache-2.0
dataset_info:
features:
- name: message_id
dtype: string
- name: parent_id
dtype: string
- name: user_id
dtype: string
- name: created_date
dtype: string
- name: text
dtype: string
- name: role
dtype: string
- name: lang
dtype: string
- name: review_count
dtype: int32
- name: review_result
dtype: bool
- name: deleted
dtype: bool
- name: rank
dtype: int32
- name: synthetic
dtype: bool
- name: model_name
dtype: string
- name: detoxify
struct:
- name: toxicity
dtype: float64
- name: severe_toxicity
dtype: float64
- name: obscene
dtype: float64
- name: identity_attack
dtype: float64
- name: insult
dtype: float64
- name: threat
dtype: float64
- name: sexual_explicit
dtype: float64
- name: message_tree_id
dtype: string
- name: tree_state
dtype: string
- name: emojis
sequence:
- name: name
dtype: string
- name: count
dtype: int32
- name: labels
sequence:
- name: name
dtype: string
- name: value
dtype: float64
- name: count
dtype: int32
splits:
- name: train
num_bytes: 100367999
num_examples: 84437
- name: validation
num_bytes: 5243405
num_examples: 4401
download_size: 41596430
dataset_size: 105611404
language:
- en
- es
- ru
- de
- pl
- th
- vi
- sv
- bn
- da
- he
- it
- fa
- sk
- id
- nb
- el
- nl
- hu
- eu
- zh
- eo
- ja
- ca
- cs
- bg
- fi
- pt
- tr
- ro
- ar
- uk
- gl
- fr
- ko
tags:
- human-feedback
size_categories:
- 100K<n<1M
pretty_name: OpenAssistant Conversations
---
# OpenAssistant Conversations Dataset (OASST1)
## Dataset Description
- **Homepage:** https://www.open-assistant.io/
- **Repository:** https://github.com/LAION-AI/Open-Assistant
- **Paper:** https://arxiv.org/abs/2304.07327
### Dataset Summary
In an effort to democratize research on large-scale alignment, we release OpenAssistant
Conversations (OASST1), a human-generated, human-annotated assistant-style conversation
corpus consisting of 161,443 messages in 35 different languages, annotated with 461,292
quality ratings, resulting in over 10,000 fully annotated conversation trees. The corpus
is a product of a worldwide crowd-sourcing effort involving over 13,500 volunteers.
Please refer to our [paper](https://arxiv.org/abs/2304.07327) for further details.
### Dataset Structure
This dataset contains message trees. Each message tree has an initial prompt message as the root node,
which can have multiple child messages as replies, and these child messages can have multiple replies.
All messages have a role property: this can either be "assistant" or "prompter". The roles in
conversation threads from prompt to leaf node strictly alternate between "prompter" and "assistant".
This version of the dataset contains data collected on the [open-assistant.io](https://open-assistant.io/) website until April 12 2023.
### JSON Example: Message
For readability, the following JSON examples are shown formatted with indentation on multiple lines.
Objects are stored without indentation (on single lines) in the actual jsonl files.
```json
{
"message_id": "218440fd-5317-4355-91dc-d001416df62b",
"parent_id": "13592dfb-a6f9-4748-a92c-32b34e239bb4",
"user_id": "8e95461f-5e94-4d8b-a2fb-d4717ce973e4",
"text": "It was the winter of 2035, and artificial intelligence (..)",
"role": "assistant",
"lang": "en",
"review_count": 3,
"review_result": true,
"deleted": false,
"rank": 0,
"synthetic": true,
"model_name": "oasst-sft-0_3000,max_new_tokens=400 (..)",
"labels": {
"spam": { "value": 0.0, "count": 3 },
"lang_mismatch": { "value": 0.0, "count": 3 },
"pii": { "value": 0.0, "count": 3 },
"not_appropriate": { "value": 0.0, "count": 3 },
"hate_speech": { "value": 0.0, "count": 3 },
"sexual_content": { "value": 0.0, "count": 3 },
"quality": { "value": 0.416, "count": 3 },
"toxicity": { "value": 0.16, "count": 3 },
"humor": { "value": 0.0, "count": 3 },
"creativity": { "value": 0.33, "count": 3 },
"violence": { "value": 0.16, "count": 3 }
}
}
```
### JSON Example: Conversation Tree
For readability, only a subset of the message properties is shown here.
```json
{
"message_tree_id": "14fbb664-a620-45ce-bee4-7c519b16a793",
"tree_state": "ready_for_export",
"prompt": {
"message_id": "14fbb664-a620-45ce-bee4-7c519b16a793",
"text": "Why can't we divide by 0? (..)",
"role": "prompter",
"lang": "en",
"replies": [
{
"message_id": "894d30b6-56b4-4605-a504-89dd15d4d1c8",
"text": "The reason we cannot divide by zero is because (..)",
"role": "assistant",
"lang": "en",
"replies": [
// ...
]
},
{
"message_id": "84d0913b-0fd9-4508-8ef5-205626a7039d",
"text": "The reason that the result of a division by zero is (..)",
"role": "assistant",
"lang": "en",
"replies": [
{
"message_id": "3352725e-f424-4e3b-a627-b6db831bdbaa",
"text": "Math is confusing. Like those weird Irrational (..)",
"role": "prompter",
"lang": "en",
"replies": [
{
"message_id": "f46207ca-3149-46e9-a466-9163d4ce499c",
"text": "Irrational numbers are simply numbers (..)",
"role": "assistant",
"lang": "en",
"replies": []
},
// ...
]
}
]
}
]
}
}
```
Please refer to [oasst-data](https://github.com/LAION-AI/Open-Assistant/tree/main/oasst-data) for
details about the data structure and Python code to read and write jsonl files containing oasst data objects.
If you would like to explore the dataset yourself you can find a
[`getting-started`](https://github.com/LAION-AI/Open-Assistant/blob/main/notebooks/openassistant-oasst1/getting-started.ipynb)
notebook in the `notebooks/openassistant-oasst1` folder of the [LAION-AI/Open-Assistant](https://github.com/LAION-AI/Open-Assistant)
github repository.
## Main Dataset Files
Conversation data is provided either as nested messages in trees (extension `.trees.jsonl.gz`)
or as a flat list (table) of messages (extension `.messages.jsonl.gz`).
### Ready For Export Trees
```
2023-04-12_oasst_ready.trees.jsonl.gz 10,364 trees with 88,838 total messages
2023-04-12_oasst_ready.messages.jsonl.gz 88,838 messages
```
Trees in `ready_for_export` state without spam and deleted messages including message labels.
The oasst_ready-trees file usually is sufficient for supervised fine-tuning (SFT) & reward model (RM) training.
### All Trees
```
2023-04-12_oasst_all.trees.jsonl.gz 66,497 trees with 161,443 total messages
2023-04-12_oasst_all.messages.jsonl.gz 161,443 messages
```
All trees, including those in states `prompt_lottery_waiting` (trees that consist of only one message, namely the initial prompt),
`aborted_low_grade` (trees that stopped growing because the messages had low quality), and `halted_by_moderator`.
### Supplemental Exports: Spam & Prompts
```
2023-04-12_oasst_spam.messages.jsonl.gz
```
These are messages which were deleted or have a negative review result (`"review_result": false`).
Besides low quality, a frequent reason for message deletion is a wrong language tag.
```
2023-04-12_oasst_prompts.messages.jsonl.gz
```
These are all the kept initial prompt messages with positive review result (no spam) of trees in `ready_for_export` or `prompt_lottery_waiting` state.
### Using the Huggingface Datasets
While HF datasets is ideal for tabular datasets, it is not a natural fit for nested data structures like the OpenAssistant conversation trees.
Nevertheless, we make all messages which can also be found in the file `2023-04-12_oasst_ready.trees.jsonl.gz` available in parquet as train/validation splits.
These are directly loadable by [Huggingface Datasets](https://pypi.org/project/datasets/).
To load the oasst1 train & validation splits use:
```python
from datasets import load_dataset
ds = load_dataset("OpenAssistant/oasst1")
train = ds['train'] # len(train)=84437 (95%)
val = ds['validation'] # len(val)=4401 (5%)
```
The messages appear in depth-first order of the message trees.
Full conversation trees can be reconstructed from the flat messages table by using the `parent_id`
and `message_id` properties to identify the parent-child relationship of messages. The `message_tree_id`
and `tree_state` properties (only present in flat messages files) can be used to find all messages of a message tree or to select trees by their state.
### Languages
OpenAssistant Conversations incorporates 35 different languages with a distribution of messages as follows:
**Languages with over 1000 messages**
- English: 71956
- Spanish: 43061
- Russian: 9089
- German: 5279
- Chinese: 4962
- French: 4251
- Thai: 3042
- Portuguese (Brazil): 2969
- Catalan: 2260
- Korean: 1553
- Ukrainian: 1352
- Italian: 1320
- Japanese: 1018
<details>
<summary><b>Languages with under 1000 messages</b></summary>
<ul>
<li>Vietnamese: 952</li>
<li>Basque: 947</li>
<li>Polish: 886</li>
<li>Hungarian: 811</li>
<li>Arabic: 666</li>
<li>Dutch: 628</li>
<li>Swedish: 512</li>
<li>Turkish: 454</li>
<li>Finnish: 386</li>
<li>Czech: 372</li>
<li>Danish: 358</li>
<li>Galician: 339</li>
<li>Hebrew: 255</li>
<li>Romanian: 200</li>
<li>Norwegian Bokmål: 133</li>
<li>Indonesian: 115</li>
<li>Bulgarian: 95</li>
<li>Bengali: 82</li>
<li>Persian: 72</li>
<li>Greek: 66</li>
<li>Esperanto: 59</li>
<li>Slovak: 19</li>
</ul>
</details>
## Contact
- Discord [Open Assistant Discord Server](https://ykilcher.com/open-assistant-discord)
- GitHub: [LAION-AI/Open-Assistant](https://github.com/LAION-AI/Open-Assistant)
- E-Mail: [open-assistant@laion.ai](mailto:open-assistant@laion.ai) |
amazon_reviews_multi | 2023-08-18T14:12:47.000Z | [
"task_categories:summarization",
"task_categories:text-generation",
"task_categories:fill-mask",
"task_categories:text-classification",
"task_ids:text-scoring",
"task_ids:language-modeling",
"task_ids:masked-language-modeling",
"task_ids:sentiment-classification",
"task_ids:sentiment-scoring",
"task_ids:topic-classification",
"annotations_creators:found",
"language_creators:found",
"multilinguality:monolingual",
"multilinguality:multilingual",
"size_categories:100K<n<1M",
"size_categories:1M<n<10M",
"source_datasets:original",
"language:de",
"language:en",
"language:es",
"language:fr",
"language:ja",
"language:zh",
"license:other",
"arxiv:2010.02573",
"region:us"
] | null | We provide an Amazon product reviews dataset for multilingual text classification. The dataset contains reviews in English, Japanese, German, French, Chinese and Spanish, collected between November 1, 2015 and November 1, 2019. Each record in the dataset contains the review text, the review title, the star rating, an anonymized reviewer ID, an anonymized product ID and the coarse-grained product category (e.g. ‘books’, ‘appliances’, etc.) The corpus is balanced across stars, so each star rating constitutes 20% of the reviews in each language.
For each language, there are 200,000, 5,000 and 5,000 reviews in the training, development and test sets respectively. The maximum number of reviews per reviewer is 20 and the maximum number of reviews per product is 20. All reviews are truncated after 2,000 characters, and all reviews are at least 20 characters long.
Note that the language of a review does not necessarily match the language of its marketplace (e.g. reviews from amazon.de are primarily written in German, but could also be written in English, etc.). For this reason, we applied a language detection algorithm based on the work in Bojanowski et al. (2017) to determine the language of the review text and we removed reviews that were not written in the expected language. | @inproceedings{marc_reviews,
title={The Multilingual Amazon Reviews Corpus},
author={Keung, Phillip and Lu, Yichao and Szarvas, György and Smith, Noah A.},
booktitle={Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing},
year={2020}
} | null | 82 | 9,276 | ---
annotations_creators:
- found
language_creators:
- found
language:
- de
- en
- es
- fr
- ja
- zh
license:
- other
multilinguality:
- monolingual
- multilingual
size_categories:
- 100K<n<1M
- 1M<n<10M
source_datasets:
- original
task_categories:
- summarization
- text-generation
- fill-mask
- text-classification
task_ids:
- text-scoring
- language-modeling
- masked-language-modeling
- sentiment-classification
- sentiment-scoring
- topic-classification
paperswithcode_id: null
pretty_name: The Multilingual Amazon Reviews Corpus
dataset_info:
- config_name: all_languages
features:
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: reviewer_id
dtype: string
- name: stars
dtype: int32
- name: review_body
dtype: string
- name: review_title
dtype: string
- name: language
dtype: string
- name: product_category
dtype: string
splits:
- name: train
num_bytes: 364405048
num_examples: 1200000
- name: validation
num_bytes: 9047533
num_examples: 30000
- name: test
num_bytes: 9099141
num_examples: 30000
download_size: 640320386
dataset_size: 382551722
- config_name: de
features:
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: reviewer_id
dtype: string
- name: stars
dtype: int32
- name: review_body
dtype: string
- name: review_title
dtype: string
- name: language
dtype: string
- name: product_category
dtype: string
splits:
- name: train
num_bytes: 64485678
num_examples: 200000
- name: validation
num_bytes: 1605727
num_examples: 5000
- name: test
num_bytes: 1611044
num_examples: 5000
download_size: 94802490
dataset_size: 67702449
- config_name: en
features:
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: reviewer_id
dtype: string
- name: stars
dtype: int32
- name: review_body
dtype: string
- name: review_title
dtype: string
- name: language
dtype: string
- name: product_category
dtype: string
splits:
- name: train
num_bytes: 58601089
num_examples: 200000
- name: validation
num_bytes: 1474672
num_examples: 5000
- name: test
num_bytes: 1460565
num_examples: 5000
download_size: 86094112
dataset_size: 61536326
- config_name: es
features:
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: reviewer_id
dtype: string
- name: stars
dtype: int32
- name: review_body
dtype: string
- name: review_title
dtype: string
- name: language
dtype: string
- name: product_category
dtype: string
splits:
- name: train
num_bytes: 52375658
num_examples: 200000
- name: validation
num_bytes: 1303958
num_examples: 5000
- name: test
num_bytes: 1312347
num_examples: 5000
download_size: 81345461
dataset_size: 54991963
- config_name: fr
features:
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: reviewer_id
dtype: string
- name: stars
dtype: int32
- name: review_body
dtype: string
- name: review_title
dtype: string
- name: language
dtype: string
- name: product_category
dtype: string
splits:
- name: train
num_bytes: 54593565
num_examples: 200000
- name: validation
num_bytes: 1340763
num_examples: 5000
- name: test
num_bytes: 1364510
num_examples: 5000
download_size: 85917293
dataset_size: 57298838
- config_name: ja
features:
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: reviewer_id
dtype: string
- name: stars
dtype: int32
- name: review_body
dtype: string
- name: review_title
dtype: string
- name: language
dtype: string
- name: product_category
dtype: string
splits:
- name: train
num_bytes: 82401390
num_examples: 200000
- name: validation
num_bytes: 2035391
num_examples: 5000
- name: test
num_bytes: 2048048
num_examples: 5000
download_size: 177773783
dataset_size: 86484829
- config_name: zh
features:
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: reviewer_id
dtype: string
- name: stars
dtype: int32
- name: review_body
dtype: string
- name: review_title
dtype: string
- name: language
dtype: string
- name: product_category
dtype: string
splits:
- name: train
num_bytes: 51947668
num_examples: 200000
- name: validation
num_bytes: 1287106
num_examples: 5000
- name: test
num_bytes: 1302711
num_examples: 5000
download_size: 114387247
dataset_size: 54537485
config_names:
- all_languages
- de
- en
- es
- fr
- ja
- zh
---
# Dataset Card for The Multilingual Amazon Reviews Corpus
## Table of Contents
- [Dataset Card for amazon_reviews_multi](#dataset-card-for-amazon_reviews_multi)
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [plain_text](#plain_text)
- [Data Fields](#data-fields)
- [plain_text](#plain_text-1)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Initial Data Collection and Normalization](#initial-data-collection-and-normalization)
- [Who are the source language producers?](#who-are-the-source-language-producers)
- [Annotations](#annotations)
- [Annotation process](#annotation-process)
- [Who are the annotators?](#who-are-the-annotators)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Webpage:** https://registry.opendata.aws/amazon-reviews-ml/
- **Paper:** https://arxiv.org/abs/2010.02573
- **Point of Contact:** [multilingual-reviews-dataset@amazon.com](mailto:multilingual-reviews-dataset@amazon.com)
### Dataset Summary
We provide an Amazon product reviews dataset for multilingual text classification. The dataset contains reviews in English, Japanese, German, French, Chinese and Spanish, collected between November 1, 2015 and November 1, 2019. Each record in the dataset contains the review text, the review title, the star rating, an anonymized reviewer ID, an anonymized product ID and the coarse-grained product category (e.g. ‘books’, ‘appliances’, etc.) The corpus is balanced across stars, so each star rating constitutes 20% of the reviews in each language.
For each language, there are 200,000, 5,000 and 5,000 reviews in the training, development and test sets respectively. The maximum number of reviews per reviewer is 20 and the maximum number of reviews per product is 20. All reviews are truncated after 2,000 characters, and all reviews are at least 20 characters long.
Note that the language of a review does not necessarily match the language of its marketplace (e.g. reviews from amazon.de are primarily written in German, but could also be written in English, etc.). For this reason, we applied a language detection algorithm based on the work in Bojanowski et al. (2017) to determine the language of the review text and we removed reviews that were not written in the expected language.
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
The dataset contains reviews in English, Japanese, German, French, Chinese and Spanish.
## Dataset Structure
### Data Instances
Each data instance corresponds to a review. The original JSON for an instance looks like so (German example):
```json
{
"review_id": "de_0784695",
"product_id": "product_de_0572654",
"reviewer_id": "reviewer_de_0645436",
"stars": "1",
"review_body": "Leider, leider nach einmal waschen ausgeblichen . Es sieht super h\u00fcbsch aus , nur leider stinkt es ganz schrecklich und ein Waschgang in der Maschine ist notwendig ! Nach einem mal waschen sah es aus als w\u00e4re es 10 Jahre alt und hatte 1000 e von Waschg\u00e4ngen hinter sich :( echt schade !",
"review_title": "Leider nicht zu empfehlen",
"language": "de",
"product_category": "home"
}
```
### Data Fields
- `review_id`: A string identifier of the review.
- `product_id`: A string identifier of the product being reviewed.
- `reviewer_id`: A string identifier of the reviewer.
- `stars`: An int between 1-5 indicating the number of stars.
- `review_body`: The text body of the review.
- `review_title`: The text title of the review.
- `language`: The string identifier of the review language.
- `product_category`: String representation of the product's category.
### Data Splits
Each language configuration comes with its own `train`, `validation`, and `test` splits. The `all_languages` split
is simply a concatenation of the corresponding split across all languages. That is, the `train` split for
`all_languages` is a concatenation of the `train` splits for each of the languages and likewise for `validation` and
`test`.
## Dataset Creation
### Curation Rationale
The dataset is motivated by the desire to advance sentiment analysis and text classification in other (non-English)
languages.
### Source Data
#### Initial Data Collection and Normalization
The authors gathered the reviews from the marketplaces in the US, Japan, Germany, France, Spain, and China for the
English, Japanese, German, French, Spanish, and Chinese languages, respectively. They then ensured the correct
language by applying a language detection algorithm, only retaining those of the target language. In a random sample
of the resulting reviews, the authors observed a small percentage of target languages that were incorrectly filtered
out and a very few mismatched languages that were incorrectly retained.
#### Who are the source language producers?
The original text comes from Amazon customers reviewing products on the marketplace across a variety of product
categories.
### Annotations
#### Annotation process
Each of the fields included are submitted by the user with the review or otherwise associated with the review. No
manual or machine-driven annotation was necessary.
#### Who are the annotators?
N/A
### Personal and Sensitive Information
According to the original dataset [license terms](https://docs.opendata.aws/amazon-reviews-ml/license.txt), you may not:
- link or associate content in the Reviews Corpus with any personal information (including Amazon customer accounts), or
- attempt to determine the identity of the author of any content in the Reviews Corpus.
If you violate any of the foregoing conditions, your license to access and use the Reviews Corpus will automatically
terminate without prejudice to any of the other rights or remedies Amazon may have.
## Considerations for Using the Data
### Social Impact of Dataset
This dataset is part of an effort to encourage text classification research in languages other than English. Such
work increases the accessibility of natural language technology to more regions and cultures. Unfortunately, each of
the languages included here is relatively high resource and well studied.
### Discussion of Biases
The dataset contains only reviews from verified purchases (as described in the paper, section 2.1), and the reviews
should conform the [Amazon Community Guidelines](https://www.amazon.com/gp/help/customer/display.html?nodeId=GLHXEX85MENUE4XF).
### Other Known Limitations
The dataset is constructed so that the distribution of star ratings is balanced. This feature has some advantages for
purposes of classification, but some types of language may be over or underrepresented relative to the original
distribution of reviews to achieve this balance.
## Additional Information
### Dataset Curators
Published by Phillip Keung, Yichao Lu, György Szarvas, and Noah A. Smith. Managed by Amazon.
### Licensing Information
Amazon has licensed this dataset under its own agreement for non-commercial research usage only. This licence is quite restrictive preventing use anywhere a fee is received including paid for internships etc. A copy of the agreement can be found at the dataset webpage here:
https://docs.opendata.aws/amazon-reviews-ml/license.txt
By accessing the Multilingual Amazon Reviews Corpus ("Reviews Corpus"), you agree that the Reviews Corpus is an Amazon Service subject to the [Amazon.com Conditions of Use](https://www.amazon.com/gp/help/customer/display.html/ref=footer_cou?ie=UTF8&nodeId=508088) and you agree to be bound by them, with the following additional conditions:
In addition to the license rights granted under the Conditions of Use, Amazon or its content providers grant you a limited, non-exclusive, non-transferable, non-sublicensable, revocable license to access and use the Reviews Corpus for purposes of academic research. You may not resell, republish, or make any commercial use of the Reviews Corpus or its contents, including use of the Reviews Corpus for commercial research, such as research related to a funding or consultancy contract, internship, or other relationship in which the results are provided for a fee or delivered to a for-profit organization. You may not (a) link or associate content in the Reviews Corpus with any personal information (including Amazon customer accounts), or (b) attempt to determine the identity of the author of any content in the Reviews Corpus. If you violate any of the foregoing conditions, your license to access and use the Reviews Corpus will automatically terminate without prejudice to any of the other rights or remedies Amazon may have.
### Citation Information
Please cite the following paper (arXiv) if you found this dataset useful:
Phillip Keung, Yichao Lu, György Szarvas and Noah A. Smith. “The Multilingual Amazon Reviews Corpus.” In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, 2020.
```
@inproceedings{marc_reviews,
title={The Multilingual Amazon Reviews Corpus},
author={Keung, Phillip and Lu, Yichao and Szarvas, György and Smith, Noah A.},
booktitle={Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing},
year={2020}
}
```
### Contributions
Thanks to [@joeddav](https://github.com/joeddav) for adding this dataset. |
math_dataset | 2023-04-05T10:09:32.000Z | [
"language:en",
"region:us"
] | null | Mathematics database.
This dataset code generates mathematical question and answer pairs,
from a range of question types at roughly school-level difficulty.
This is designed to test the mathematical learning and algebraic
reasoning skills of learning models.
Original paper: Analysing Mathematical Reasoning Abilities of Neural Models
(Saxton, Grefenstette, Hill, Kohli).
Example usage:
train_examples, val_examples = datasets.load_dataset(
'math_dataset/arithmetic__mul',
split=['train', 'test'],
as_supervised=True) | @article{2019arXiv,
author = {Saxton, Grefenstette, Hill, Kohli},
title = {Analysing Mathematical Reasoning Abilities of Neural Models},
year = {2019},
journal = {arXiv:1904.01557}
} | null | 38 | 9,227 | ---
pretty_name: Mathematics Dataset
language:
- en
paperswithcode_id: mathematics
dataset_info:
- config_name: algebra__linear_1d
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 516405
num_examples: 10000
- name: train
num_bytes: 92086245
num_examples: 1999998
download_size: 2333082954
dataset_size: 92602650
- config_name: algebra__linear_1d_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1018090
num_examples: 10000
- name: train
num_bytes: 199566926
num_examples: 1999998
download_size: 2333082954
dataset_size: 200585016
- config_name: algebra__linear_2d
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 666095
num_examples: 10000
- name: train
num_bytes: 126743526
num_examples: 1999998
download_size: 2333082954
dataset_size: 127409621
- config_name: algebra__linear_2d_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1184664
num_examples: 10000
- name: train
num_bytes: 234405885
num_examples: 1999998
download_size: 2333082954
dataset_size: 235590549
- config_name: algebra__polynomial_roots
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 868630
num_examples: 10000
- name: train
num_bytes: 163134199
num_examples: 1999998
download_size: 2333082954
dataset_size: 164002829
- config_name: algebra__polynomial_roots_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1281321
num_examples: 10000
- name: train
num_bytes: 251435312
num_examples: 1999998
download_size: 2333082954
dataset_size: 252716633
- config_name: algebra__sequence_next_term
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 752459
num_examples: 10000
- name: train
num_bytes: 138735194
num_examples: 1999998
download_size: 2333082954
dataset_size: 139487653
- config_name: algebra__sequence_nth_term
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 947764
num_examples: 10000
- name: train
num_bytes: 175945643
num_examples: 1999998
download_size: 2333082954
dataset_size: 176893407
- config_name: arithmetic__add_or_sub
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 483725
num_examples: 10000
- name: train
num_bytes: 89690356
num_examples: 1999998
download_size: 2333082954
dataset_size: 90174081
- config_name: arithmetic__add_or_sub_in_base
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 502221
num_examples: 10000
- name: train
num_bytes: 93779137
num_examples: 1999998
download_size: 2333082954
dataset_size: 94281358
- config_name: arithmetic__add_sub_multiple
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 498421
num_examples: 10000
- name: train
num_bytes: 90962782
num_examples: 1999998
download_size: 2333082954
dataset_size: 91461203
- config_name: arithmetic__div
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 421520
num_examples: 10000
- name: train
num_bytes: 78417908
num_examples: 1999998
download_size: 2333082954
dataset_size: 78839428
- config_name: arithmetic__mixed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 513364
num_examples: 10000
- name: train
num_bytes: 93989009
num_examples: 1999998
download_size: 2333082954
dataset_size: 94502373
- config_name: arithmetic__mul
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 394004
num_examples: 10000
- name: train
num_bytes: 73499093
num_examples: 1999998
download_size: 2333082954
dataset_size: 73893097
- config_name: arithmetic__mul_div_multiple
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 497308
num_examples: 10000
- name: train
num_bytes: 91406689
num_examples: 1999998
download_size: 2333082954
dataset_size: 91903997
- config_name: arithmetic__nearest_integer_root
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 705630
num_examples: 10000
- name: train
num_bytes: 137771237
num_examples: 1999998
download_size: 2333082954
dataset_size: 138476867
- config_name: arithmetic__simplify_surd
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1261753
num_examples: 10000
- name: train
num_bytes: 207753790
num_examples: 1999998
download_size: 2333082954
dataset_size: 209015543
- config_name: calculus__differentiate
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1025947
num_examples: 10000
- name: train
num_bytes: 199013993
num_examples: 1999998
download_size: 2333082954
dataset_size: 200039940
- config_name: calculus__differentiate_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1343416
num_examples: 10000
- name: train
num_bytes: 263757570
num_examples: 1999998
download_size: 2333082954
dataset_size: 265100986
- config_name: comparison__closest
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 681229
num_examples: 10000
- name: train
num_bytes: 132274822
num_examples: 1999998
download_size: 2333082954
dataset_size: 132956051
- config_name: comparison__closest_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1071089
num_examples: 10000
- name: train
num_bytes: 210658152
num_examples: 1999998
download_size: 2333082954
dataset_size: 211729241
- config_name: comparison__kth_biggest
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 797185
num_examples: 10000
- name: train
num_bytes: 149077463
num_examples: 1999998
download_size: 2333082954
dataset_size: 149874648
- config_name: comparison__kth_biggest_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1144556
num_examples: 10000
- name: train
num_bytes: 221547532
num_examples: 1999998
download_size: 2333082954
dataset_size: 222692088
- config_name: comparison__pair
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 452528
num_examples: 10000
- name: train
num_bytes: 85707543
num_examples: 1999998
download_size: 2333082954
dataset_size: 86160071
- config_name: comparison__pair_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 946187
num_examples: 10000
- name: train
num_bytes: 184702998
num_examples: 1999998
download_size: 2333082954
dataset_size: 185649185
- config_name: comparison__sort
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 712498
num_examples: 10000
- name: train
num_bytes: 131752705
num_examples: 1999998
download_size: 2333082954
dataset_size: 132465203
- config_name: comparison__sort_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1114257
num_examples: 10000
- name: train
num_bytes: 213871896
num_examples: 1999998
download_size: 2333082954
dataset_size: 214986153
- config_name: measurement__conversion
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 592904
num_examples: 10000
- name: train
num_bytes: 118650852
num_examples: 1999998
download_size: 2333082954
dataset_size: 119243756
- config_name: measurement__time
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 584278
num_examples: 10000
- name: train
num_bytes: 116962599
num_examples: 1999998
download_size: 2333082954
dataset_size: 117546877
- config_name: numbers__base_conversion
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 490881
num_examples: 10000
- name: train
num_bytes: 90363333
num_examples: 1999998
download_size: 2333082954
dataset_size: 90854214
- config_name: numbers__div_remainder
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 644523
num_examples: 10000
- name: train
num_bytes: 125046212
num_examples: 1999998
download_size: 2333082954
dataset_size: 125690735
- config_name: numbers__div_remainder_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1151347
num_examples: 10000
- name: train
num_bytes: 226341870
num_examples: 1999998
download_size: 2333082954
dataset_size: 227493217
- config_name: numbers__gcd
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 659492
num_examples: 10000
- name: train
num_bytes: 127914889
num_examples: 1999998
download_size: 2333082954
dataset_size: 128574381
- config_name: numbers__gcd_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1206805
num_examples: 10000
- name: train
num_bytes: 237534189
num_examples: 1999998
download_size: 2333082954
dataset_size: 238740994
- config_name: numbers__is_factor
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 396129
num_examples: 10000
- name: train
num_bytes: 75875988
num_examples: 1999998
download_size: 2333082954
dataset_size: 76272117
- config_name: numbers__is_factor_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 949828
num_examples: 10000
- name: train
num_bytes: 185369842
num_examples: 1999998
download_size: 2333082954
dataset_size: 186319670
- config_name: numbers__is_prime
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 385749
num_examples: 10000
- name: train
num_bytes: 73983639
num_examples: 1999998
download_size: 2333082954
dataset_size: 74369388
- config_name: numbers__is_prime_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 947888
num_examples: 10000
- name: train
num_bytes: 184808483
num_examples: 1999998
download_size: 2333082954
dataset_size: 185756371
- config_name: numbers__lcm
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 717978
num_examples: 10000
- name: train
num_bytes: 136826050
num_examples: 1999998
download_size: 2333082954
dataset_size: 137544028
- config_name: numbers__lcm_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1127744
num_examples: 10000
- name: train
num_bytes: 221148668
num_examples: 1999998
download_size: 2333082954
dataset_size: 222276412
- config_name: numbers__list_prime_factors
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 585749
num_examples: 10000
- name: train
num_bytes: 109982816
num_examples: 1999998
download_size: 2333082954
dataset_size: 110568565
- config_name: numbers__list_prime_factors_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1053510
num_examples: 10000
- name: train
num_bytes: 205379513
num_examples: 1999998
download_size: 2333082954
dataset_size: 206433023
- config_name: numbers__place_value
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 496977
num_examples: 10000
- name: train
num_bytes: 95180091
num_examples: 1999998
download_size: 2333082954
dataset_size: 95677068
- config_name: numbers__place_value_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1011130
num_examples: 10000
- name: train
num_bytes: 197187918
num_examples: 1999998
download_size: 2333082954
dataset_size: 198199048
- config_name: numbers__round_number
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 570636
num_examples: 10000
- name: train
num_bytes: 111472483
num_examples: 1999998
download_size: 2333082954
dataset_size: 112043119
- config_name: numbers__round_number_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1016754
num_examples: 10000
- name: train
num_bytes: 201057283
num_examples: 1999998
download_size: 2333082954
dataset_size: 202074037
- config_name: polynomials__add
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1308455
num_examples: 10000
- name: train
num_bytes: 257576092
num_examples: 1999998
download_size: 2333082954
dataset_size: 258884547
- config_name: polynomials__coefficient_named
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1137226
num_examples: 10000
- name: train
num_bytes: 219716251
num_examples: 1999998
download_size: 2333082954
dataset_size: 220853477
- config_name: polynomials__collect
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 774709
num_examples: 10000
- name: train
num_bytes: 143743260
num_examples: 1999998
download_size: 2333082954
dataset_size: 144517969
- config_name: polynomials__compose
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1209763
num_examples: 10000
- name: train
num_bytes: 233651887
num_examples: 1999998
download_size: 2333082954
dataset_size: 234861650
- config_name: polynomials__evaluate
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 599446
num_examples: 10000
- name: train
num_bytes: 114538250
num_examples: 1999998
download_size: 2333082954
dataset_size: 115137696
- config_name: polynomials__evaluate_composed
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1148362
num_examples: 10000
- name: train
num_bytes: 226022455
num_examples: 1999998
download_size: 2333082954
dataset_size: 227170817
- config_name: polynomials__expand
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1057353
num_examples: 10000
- name: train
num_bytes: 202338235
num_examples: 1999998
download_size: 2333082954
dataset_size: 203395588
- config_name: polynomials__simplify_power
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1248040
num_examples: 10000
- name: train
num_bytes: 216407582
num_examples: 1999998
download_size: 2333082954
dataset_size: 217655622
- config_name: probability__swr_p_level_set
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1159050
num_examples: 10000
- name: train
num_bytes: 227540179
num_examples: 1999998
download_size: 2333082954
dataset_size: 228699229
- config_name: probability__swr_p_sequence
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: test
num_bytes: 1097442
num_examples: 10000
- name: train
num_bytes: 215865725
num_examples: 1999998
download_size: 2333082954
dataset_size: 216963167
---
# Dataset Card for "math_dataset"
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://github.com/deepmind/mathematics_dataset](https://github.com/deepmind/mathematics_dataset)
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 130.65 GB
- **Size of the generated dataset:** 9.08 GB
- **Total amount of disk used:** 139.73 GB
### Dataset Summary
Mathematics database.
This dataset code generates mathematical question and answer pairs,
from a range of question types at roughly school-level difficulty.
This is designed to test the mathematical learning and algebraic
reasoning skills of learning models.
Original paper: Analysing Mathematical Reasoning Abilities of Neural Models
(Saxton, Grefenstette, Hill, Kohli).
Example usage:
train_examples, val_examples = datasets.load_dataset(
'math_dataset/arithmetic__mul',
split=['train', 'test'],
as_supervised=True)
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Dataset Structure
### Data Instances
#### algebra__linear_1d
- **Size of downloaded dataset files:** 2.33 GB
- **Size of the generated dataset:** 92.60 MB
- **Total amount of disk used:** 2.43 GB
An example of 'train' looks as follows.
```
```
#### algebra__linear_1d_composed
- **Size of downloaded dataset files:** 2.33 GB
- **Size of the generated dataset:** 200.58 MB
- **Total amount of disk used:** 2.53 GB
An example of 'train' looks as follows.
```
```
#### algebra__linear_2d
- **Size of downloaded dataset files:** 2.33 GB
- **Size of the generated dataset:** 127.41 MB
- **Total amount of disk used:** 2.46 GB
An example of 'train' looks as follows.
```
```
#### algebra__linear_2d_composed
- **Size of downloaded dataset files:** 2.33 GB
- **Size of the generated dataset:** 235.59 MB
- **Total amount of disk used:** 2.57 GB
An example of 'train' looks as follows.
```
```
#### algebra__polynomial_roots
- **Size of downloaded dataset files:** 2.33 GB
- **Size of the generated dataset:** 164.01 MB
- **Total amount of disk used:** 2.50 GB
An example of 'train' looks as follows.
```
```
### Data Fields
The data fields are the same among all splits.
#### algebra__linear_1d
- `question`: a `string` feature.
- `answer`: a `string` feature.
#### algebra__linear_1d_composed
- `question`: a `string` feature.
- `answer`: a `string` feature.
#### algebra__linear_2d
- `question`: a `string` feature.
- `answer`: a `string` feature.
#### algebra__linear_2d_composed
- `question`: a `string` feature.
- `answer`: a `string` feature.
#### algebra__polynomial_roots
- `question`: a `string` feature.
- `answer`: a `string` feature.
### Data Splits
| name | train |test |
|---------------------------|------:|----:|
|algebra__linear_1d |1999998|10000|
|algebra__linear_1d_composed|1999998|10000|
|algebra__linear_2d |1999998|10000|
|algebra__linear_2d_composed|1999998|10000|
|algebra__polynomial_roots |1999998|10000|
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Citation Information
```
@article{2019arXiv,
author = {Saxton, Grefenstette, Hill, Kohli},
title = {Analysing Mathematical Reasoning Abilities of Neural Models},
year = {2019},
journal = {arXiv:1904.01557}
}
```
### Contributions
Thanks to [@patrickvonplaten](https://github.com/patrickvonplaten), [@lewtun](https://github.com/lewtun), [@thomwolf](https://github.com/thomwolf) for adding this dataset. |
amazon_us_reviews | 2023-04-05T09:14:36.000Z | [
"task_categories:summarization",
"task_categories:text-generation",
"task_categories:fill-mask",
"task_categories:text-classification",
"task_ids:text-scoring",
"task_ids:language-modeling",
"task_ids:masked-language-modeling",
"task_ids:sentiment-classification",
"task_ids:sentiment-scoring",
"task_ids:topic-classification",
"annotations_creators:no-annotation",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:100M<n<1B",
"source_datasets:original",
"language:en",
"license:other",
"region:us"
] | null | Amazon Customer Reviews (a.k.a. Product Reviews) is one of Amazons iconic products. In a period of over two decades since the first review in 1995, millions of Amazon customers have contributed over a hundred million reviews to express opinions and describe their experiences regarding products on the Amazon.com website. This makes Amazon Customer Reviews a rich source of information for academic researchers in the fields of Natural Language Processing (NLP), Information Retrieval (IR), and Machine Learning (ML), amongst others. Accordingly, we are releasing this data to further research in multiple disciplines related to understanding customer product experiences. Specifically, this dataset was constructed to represent a sample of customer evaluations and opinions, variation in the perception of a product across geographical regions, and promotional intent or bias in reviews.
Over 130+ million customer reviews are available to researchers as part of this release. The data is available in TSV files in the amazon-reviews-pds S3 bucket in AWS US East Region. Each line in the data files corresponds to an individual review (tab delimited, with no quote and escape characters).
Each Dataset contains the following columns:
- marketplace: 2 letter country code of the marketplace where the review was written.
- customer_id: Random identifier that can be used to aggregate reviews written by a single author.
- review_id: The unique ID of the review.
- product_id: The unique Product ID the review pertains to. In the multilingual dataset the reviews for the same product in different countries can be grouped by the same product_id.
- product_parent: Random identifier that can be used to aggregate reviews for the same product.
- product_title: Title of the product.
- product_category: Broad product category that can be used to group reviews (also used to group the dataset into coherent parts).
- star_rating: The 1-5 star rating of the review.
- helpful_votes: Number of helpful votes.
- total_votes: Number of total votes the review received.
- vine: Review was written as part of the Vine program.
- verified_purchase: The review is on a verified purchase.
- review_headline: The title of the review.
- review_body: The review text.
- review_date: The date the review was written. | \ | null | 51 | 9,121 | ---
annotations_creators:
- no-annotation
language_creators:
- found
language:
- en
license:
- other
multilinguality:
- monolingual
size_categories:
- 100M<n<1B
source_datasets:
- original
task_categories:
- summarization
- text-generation
- fill-mask
- text-classification
task_ids:
- text-scoring
- language-modeling
- masked-language-modeling
- sentiment-classification
- sentiment-scoring
- topic-classification
pretty_name: Amazon US Reviews
dataset_info:
- config_name: Books_v1_01
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 6997552259
num_examples: 6106719
download_size: 2692708591
dataset_size: 6997552259
- config_name: Watches_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 458976082
num_examples: 960872
download_size: 162973819
dataset_size: 458976082
- config_name: Personal_Care_Appliances_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 49036547
num_examples: 85981
download_size: 17634794
dataset_size: 49036547
- config_name: Mobile_Electronics_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 63293377
num_examples: 104975
download_size: 22870508
dataset_size: 63293377
- config_name: Digital_Video_Games_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 80176851
num_examples: 145431
download_size: 27442648
dataset_size: 80176851
- config_name: Digital_Software_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 58782931
num_examples: 102084
download_size: 18997559
dataset_size: 58782931
- config_name: Major_Appliances_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 67642424
num_examples: 96901
download_size: 24359816
dataset_size: 67642424
- config_name: Gift_Card_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 47188062
num_examples: 149086
download_size: 12134676
dataset_size: 47188062
- config_name: Video_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 356264426
num_examples: 380604
download_size: 138929896
dataset_size: 356264426
- config_name: Luggage_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 167354173
num_examples: 348657
download_size: 60320191
dataset_size: 167354173
- config_name: Software_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 266020595
num_examples: 341931
download_size: 94010685
dataset_size: 266020595
- config_name: Video_Games_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 1291054668
num_examples: 1785997
download_size: 475199894
dataset_size: 1291054668
- config_name: Furniture_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 405212374
num_examples: 792113
download_size: 148982796
dataset_size: 405212374
- config_name: Musical_Instruments_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 518908568
num_examples: 904765
download_size: 193389086
dataset_size: 518908568
- config_name: Digital_Music_Purchase_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 710546079
num_examples: 1688884
download_size: 253570168
dataset_size: 710546079
- config_name: Books_v1_02
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 3387034903
num_examples: 3105520
download_size: 1329539135
dataset_size: 3387034903
- config_name: Home_Entertainment_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 534333848
num_examples: 705889
download_size: 193168458
dataset_size: 534333848
- config_name: Grocery_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 1072289473
num_examples: 2402458
download_size: 401337166
dataset_size: 1072289473
- config_name: Outdoors_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 1172986088
num_examples: 2302401
download_size: 448963100
dataset_size: 1172986088
- config_name: Pet_Products_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 1355659812
num_examples: 2643619
download_size: 515815253
dataset_size: 1355659812
- config_name: Video_DVD_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 3953234561
num_examples: 5069140
download_size: 1512355451
dataset_size: 3953234561
- config_name: Apparel_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 2256558450
num_examples: 5906333
download_size: 648641286
dataset_size: 2256558450
- config_name: PC_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 3982684438
num_examples: 6908554
download_size: 1512903923
dataset_size: 3982684438
- config_name: Tools_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 872273119
num_examples: 1741100
download_size: 333782939
dataset_size: 872273119
- config_name: Jewelry_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 703275869
num_examples: 1767753
download_size: 247022254
dataset_size: 703275869
- config_name: Baby_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 956952590
num_examples: 1752932
download_size: 357392893
dataset_size: 956952590
- config_name: Home_Improvement_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 1329688315
num_examples: 2634781
download_size: 503339178
dataset_size: 1329688315
- config_name: Camera_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 1187101912
num_examples: 1801974
download_size: 442653086
dataset_size: 1187101912
- config_name: Lawn_and_Garden_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 1272255987
num_examples: 2557288
download_size: 486772662
dataset_size: 1272255987
- config_name: Office_Products_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 1370685534
num_examples: 2642434
download_size: 512323500
dataset_size: 1370685534
- config_name: Electronics_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 1875406721
num_examples: 3093869
download_size: 698828243
dataset_size: 1875406721
- config_name: Automotive_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 1520191087
num_examples: 3514942
download_size: 582145299
dataset_size: 1520191087
- config_name: Digital_Video_Download_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 1484214187
num_examples: 4057147
download_size: 506979922
dataset_size: 1484214187
- config_name: Mobile_Apps_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 1627857158
num_examples: 5033376
download_size: 557959415
dataset_size: 1627857158
- config_name: Shoes_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 1781283508
num_examples: 4366916
download_size: 642255314
dataset_size: 1781283508
- config_name: Toys_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 2197820069
num_examples: 4864249
download_size: 838451398
dataset_size: 2197820069
- config_name: Sports_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 2241349145
num_examples: 4850360
download_size: 872478735
dataset_size: 2241349145
- config_name: Kitchen_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 2453735305
num_examples: 4880466
download_size: 930744854
dataset_size: 2453735305
- config_name: Beauty_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 2399292506
num_examples: 5115666
download_size: 914070021
dataset_size: 2399292506
- config_name: Music_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 3900138839
num_examples: 4751577
download_size: 1521994296
dataset_size: 3900138839
- config_name: Health_Personal_Care_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 2679427491
num_examples: 5331449
download_size: 1011180212
dataset_size: 2679427491
- config_name: Digital_Ebook_Purchase_v1_01
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 3470453859
num_examples: 5101693
download_size: 1294879074
dataset_size: 3470453859
- config_name: Home_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 2796680249
num_examples: 6221559
download_size: 1081002012
dataset_size: 2796680249
- config_name: Wireless_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 4633213433
num_examples: 9002021
download_size: 1704713674
dataset_size: 4633213433
- config_name: Books_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 7197687124
num_examples: 10319090
download_size: 2740337188
dataset_size: 7197687124
- config_name: Digital_Ebook_Purchase_v1_00
features:
- name: marketplace
dtype: string
- name: customer_id
dtype: string
- name: review_id
dtype: string
- name: product_id
dtype: string
- name: product_parent
dtype: string
- name: product_title
dtype: string
- name: product_category
dtype: string
- name: star_rating
dtype: int32
- name: helpful_votes
dtype: int32
- name: total_votes
dtype: int32
- name: vine
dtype:
class_label:
names:
'0': N
'1': Y
- name: verified_purchase
dtype:
class_label:
names:
'0': N
'1': Y
- name: review_headline
dtype: string
- name: review_body
dtype: string
- name: review_date
dtype: string
splits:
- name: train
num_bytes: 7302303804
num_examples: 12520722
download_size: 2689739299
dataset_size: 7302303804
---
# Dataset Card for "amazon_us_reviews"
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://s3.amazonaws.com/amazon-reviews-pds/readme.html](https://s3.amazonaws.com/amazon-reviews-pds/readme.html)
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 32377.29 MB
- **Size of the generated dataset:** 82820.19 MB
- **Total amount of disk used:** 115197.49 MB
### Dataset Summary
Amazon Customer Reviews (a.k.a. Product Reviews) is one of Amazons iconic products. In a period of over two decades since the first review in 1995, millions of Amazon customers have contributed over a hundred million reviews to express opinions and describe their experiences regarding products on the Amazon.com website. This makes Amazon Customer Reviews a rich source of information for academic researchers in the fields of Natural Language Processing (NLP), Information Retrieval (IR), and Machine Learning (ML), amongst others. Accordingly, we are releasing this data to further research in multiple disciplines related to understanding customer product experiences. Specifically, this dataset was constructed to represent a sample of customer evaluations and opinions, variation in the perception of a product across geographical regions, and promotional intent or bias in reviews.
Over 130+ million customer reviews are available to researchers as part of this release. The data is available in TSV files in the amazon-reviews-pds S3 bucket in AWS US East Region. Each line in the data files corresponds to an individual review (tab delimited, with no quote and escape characters).
Each Dataset contains the following columns :
marketplace - 2 letter country code of the marketplace where the review was written.
customer_id - Random identifier that can be used to aggregate reviews written by a single author.
review_id - The unique ID of the review.
product_id - The unique Product ID the review pertains to. In the multilingual dataset the reviews
for the same product in different countries can be grouped by the same product_id.
product_parent - Random identifier that can be used to aggregate reviews for the same product.
product_title - Title of the product.
product_category - Broad product category that can be used to group reviews
(also used to group the dataset into coherent parts).
star_rating - The 1-5 star rating of the review.
helpful_votes - Number of helpful votes.
total_votes - Number of total votes the review received.
vine - Review was written as part of the Vine program.
verified_purchase - The review is on a verified purchase.
review_headline - The title of the review.
review_body - The review text.
review_date - The date the review was written.
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Dataset Structure
### Data Instances
#### Apparel_v1_00
- **Size of downloaded dataset files:** 648.64 MB
- **Size of the generated dataset:** 2254.36 MB
- **Total amount of disk used:** 2903.00 MB
An example of 'train' looks as follows.
```
{
"customer_id": "45223824",
"helpful_votes": 0,
"marketplace": "US",
"product_category": "Apparel",
"product_id": "B016PUU3VO",
"product_parent": "893588059",
"product_title": "Fruit of the Loom Boys' A-Shirt (Pack of 4)",
"review_body": "I ordered the same size as I ordered last time, and these shirts were much larger than the previous order. They were also about 6 inches longer. It was like they sent men's shirts instead of boys' shirts. I'll be returning these...",
"review_date": "2015-01-01",
"review_headline": "Sizes not correct, too big overall and WAY too long",
"review_id": "R1N3Z13931J3O9",
"star_rating": 2,
"total_votes": 0,
"verified_purchase": 1,
"vine": 0
}
```
#### Automotive_v1_00
- **Size of downloaded dataset files:** 582.15 MB
- **Size of the generated dataset:** 1518.88 MB
- **Total amount of disk used:** 2101.03 MB
An example of 'train' looks as follows.
```
{
"customer_id": "16825098",
"helpful_votes": 0,
"marketplace": "US",
"product_category": "Automotive",
"product_id": "B000E4PCGE",
"product_parent": "694793259",
"product_title": "00-03 NISSAN SENTRA MIRROR RH (PASSENGER SIDE), Power, Non-Heated (2000 00 2001 01 2002 02 2003 03) NS35ER 963015M000",
"review_body": "Product was as described, new and a great look. Only bad thing is that one of the screws was stripped so I couldn't tighten all three.",
"review_date": "2015-08-31",
"review_headline": "new and a great look. Only bad thing is that one of ...",
"review_id": "R2RUIDUMDKG7P",
"star_rating": 3,
"total_votes": 0,
"verified_purchase": 1,
"vine": 0
}
```
#### Baby_v1_00
- **Size of downloaded dataset files:** 357.40 MB
- **Size of the generated dataset:** 956.30 MB
- **Total amount of disk used:** 1313.70 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"customer_id": "23299101",
"helpful_votes": 2,
"marketplace": "US",
"product_category": "Baby",
"product_id": "B00SN6F9NG",
"product_parent": "3470998",
"product_title": "Rhoost Nail Clipper for Baby - Ergonomically Designed and Easy to Use Baby Nail Clipper, Natural Wooden Bamboo - Baby Health and Personal Care Kits",
"review_body": "\"This is an absolute MUST item to have! I was scared to death to clip my baby's nails. I tried other baby nail clippers and th...",
"review_date": "2015-08-31",
"review_headline": "If fits so comfortably in my hand and I feel like I have ...",
"review_id": "R2DRL5NRODVQ3Z",
"star_rating": 5,
"total_votes": 2,
"verified_purchase": 1,
"vine": 0
}
```
#### Beauty_v1_00
- **Size of downloaded dataset files:** 914.08 MB
- **Size of the generated dataset:** 2397.39 MB
- **Total amount of disk used:** 3311.47 MB
An example of 'train' looks as follows.
```
{
"customer_id": "24655453",
"helpful_votes": 1,
"marketplace": "US",
"product_category": "Beauty",
"product_id": "B00SAQ9DZY",
"product_parent": "292127037",
"product_title": "12 New, High Quality, Amber 2 ml (5/8 Dram) Glass Bottles, with Orifice Reducer and Black Cap.",
"review_body": "These are great for small mixtures for EO's, especially for traveling. I only gave this 4 stars because of the orifice reducer. The hole is so small it is hard to get the oil out. Just needs to be slightly bigger.",
"review_date": "2015-08-31",
"review_headline": "Good Product",
"review_id": "R2A30ALEGLMCGN",
"star_rating": 4,
"total_votes": 1,
"verified_purchase": 1,
"vine": 0
}
```
#### Books_v1_00
- **Size of downloaded dataset files:** 2740.34 MB
- **Size of the generated dataset:** 7193.86 MB
- **Total amount of disk used:** 9934.20 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"customer_id": "49735028",
"helpful_votes": 0,
"marketplace": "US",
"product_category": "Books",
"product_id": "0664254969",
"product_parent": "248307276",
"product_title": "Presbyterian Creeds: A Guide to the Book of Confessions",
"review_body": "\"The Presbyterian Book of Confessions contains multiple Creeds for use by the denomination. This guidebook helps he lay person t...",
"review_date": "2015-08-31",
"review_headline": "The Presbyterian Book of Confessions contains multiple Creeds for use ...",
"review_id": "R2G519UREHRO8M",
"star_rating": 3,
"total_votes": 1,
"verified_purchase": 1,
"vine": 0
}
```
### Data Fields
The data fields are the same among all splits.
#### Apparel_v1_00
- `marketplace`: a `string` feature.
- `customer_id`: a `string` feature.
- `review_id`: a `string` feature.
- `product_id`: a `string` feature.
- `product_parent`: a `string` feature.
- `product_title`: a `string` feature.
- `product_category`: a `string` feature.
- `star_rating`: a `int32` feature.
- `helpful_votes`: a `int32` feature.
- `total_votes`: a `int32` feature.
- `vine`: a classification label, with possible values including `Y` (0), `N` (1).
- `verified_purchase`: a classification label, with possible values including `Y` (0), `N` (1).
- `review_headline`: a `string` feature.
- `review_body`: a `string` feature.
- `review_date`: a `string` feature.
#### Automotive_v1_00
- `marketplace`: a `string` feature.
- `customer_id`: a `string` feature.
- `review_id`: a `string` feature.
- `product_id`: a `string` feature.
- `product_parent`: a `string` feature.
- `product_title`: a `string` feature.
- `product_category`: a `string` feature.
- `star_rating`: a `int32` feature.
- `helpful_votes`: a `int32` feature.
- `total_votes`: a `int32` feature.
- `vine`: a classification label, with possible values including `Y` (0), `N` (1).
- `verified_purchase`: a classification label, with possible values including `Y` (0), `N` (1).
- `review_headline`: a `string` feature.
- `review_body`: a `string` feature.
- `review_date`: a `string` feature.
#### Baby_v1_00
- `marketplace`: a `string` feature.
- `customer_id`: a `string` feature.
- `review_id`: a `string` feature.
- `product_id`: a `string` feature.
- `product_parent`: a `string` feature.
- `product_title`: a `string` feature.
- `product_category`: a `string` feature.
- `star_rating`: a `int32` feature.
- `helpful_votes`: a `int32` feature.
- `total_votes`: a `int32` feature.
- `vine`: a classification label, with possible values including `Y` (0), `N` (1).
- `verified_purchase`: a classification label, with possible values including `Y` (0), `N` (1).
- `review_headline`: a `string` feature.
- `review_body`: a `string` feature.
- `review_date`: a `string` feature.
#### Beauty_v1_00
- `marketplace`: a `string` feature.
- `customer_id`: a `string` feature.
- `review_id`: a `string` feature.
- `product_id`: a `string` feature.
- `product_parent`: a `string` feature.
- `product_title`: a `string` feature.
- `product_category`: a `string` feature.
- `star_rating`: a `int32` feature.
- `helpful_votes`: a `int32` feature.
- `total_votes`: a `int32` feature.
- `vine`: a classification label, with possible values including `Y` (0), `N` (1).
- `verified_purchase`: a classification label, with possible values including `Y` (0), `N` (1).
- `review_headline`: a `string` feature.
- `review_body`: a `string` feature.
- `review_date`: a `string` feature.
#### Books_v1_00
- `marketplace`: a `string` feature.
- `customer_id`: a `string` feature.
- `review_id`: a `string` feature.
- `product_id`: a `string` feature.
- `product_parent`: a `string` feature.
- `product_title`: a `string` feature.
- `product_category`: a `string` feature.
- `star_rating`: a `int32` feature.
- `helpful_votes`: a `int32` feature.
- `total_votes`: a `int32` feature.
- `vine`: a classification label, with possible values including `Y` (0), `N` (1).
- `verified_purchase`: a classification label, with possible values including `Y` (0), `N` (1).
- `review_headline`: a `string` feature.
- `review_body`: a `string` feature.
- `review_date`: a `string` feature.
### Data Splits
| name | train |
|----------------|-------:|
|Apparel_v1_00 | 5906333|
|Automotive_v1_00 | 3514942|
|Baby_v1_00 | 1752932|
|Beauty_v1_00 | 5115666|
|Books_v1_00 | 10319090|
|Books_v1_01 | 6106719|
|Books_v1_02 | 3105520|
|Camera_v1_00 | 1801974|
|Digital_Ebook_Purchase_v1_00 | 12520722|
|Digital_Ebook_Purchase_v1_01 | 5101693|
|Digital_Music_Purchase_v1_00 | 1688884|
|Digital_Software_v1_00 | 102084|
|Digital_Video_Download_v1_00 | 4057147|
|Digital_Video_Games_v1_00 | 145431|
|Electronics_v1_00 | 3093869|
|Furniture_v1_00 | 792113|
|Gift_Card_v1_00 | 149086|
|Grocery_v1_00 | 2402458|
|Health_Personal_Care_v1_00 | 5331449|
|Home_Entertainment_v1_00 | 705889|
|Home_Improvement_v1_00 | 2634781|
|Home_v1_00 | 6221559|
|Jewelry_v1_00 | 1767753|
|Kitchen_v1_00 | 4880466|
|Lawn_and_Garden_v1_00 | 2557288|
|Luggage_v1_00 | 348657|
|Major_Appliances_v1_00 | 96901|
|Mobile_Apps_v1_00 | 5033376|
|Mobile_Electronics_v1_00 | 104975|
|Music_v1_00 | 4751577|
|Musical_Instruments_v1_00 | 904765|
|Office_Products_v1_00 | 2642434|
|Outdoors_v1_00 | 2302401|
|PC_v1_00 | 6908554|
|Personal_Care_Appliances_v1_00 | 85981|
|Pet_Products_v1_00 | 2643619|
|Shoes_v1_00 | 4366916|
|Software_v1_00 | 341931|
|Sports_v1_00 | 4850360|
|Tools_v1_00 | 1741100|
|Toys_v1_00 | 4864249|
|Video_DVD_v1_00 | 5069140|
|Video_Games_v1_00 | 1785997|
|Video_v1_00 | 380604|
|Watches_v1_00 | 960872|
|Wireless_v1_00 | 9002021|
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
https://s3.amazonaws.com/amazon-reviews-pds/LICENSE.txt
By accessing the Amazon Customer Reviews Library ("Reviews Library"), you agree that the
Reviews Library is an Amazon Service subject to the [Amazon.com Conditions of Use](https://www.amazon.com/gp/help/customer/display.html/ref=footer_cou?ie=UTF8&nodeId=508088)
and you agree to be bound by them, with the following additional conditions:
In addition to the license rights granted under the Conditions of Use,
Amazon or its content providers grant you a limited, non-exclusive, non-transferable,
non-sublicensable, revocable license to access and use the Reviews Library
for purposes of academic research.
You may not resell, republish, or make any commercial use of the Reviews Library
or its contents, including use of the Reviews Library for commercial research,
such as research related to a funding or consultancy contract, internship, or
other relationship in which the results are provided for a fee or delivered
to a for-profit organization. You may not (a) link or associate content
in the Reviews Library with any personal information (including Amazon customer accounts),
or (b) attempt to determine the identity of the author of any content in the
Reviews Library.
If you violate any of the foregoing conditions, your license to access and use the
Reviews Library will automatically terminate without prejudice to any of the
other rights or remedies Amazon may have.
### Citation Information
No citation information.
### Contributions
Thanks to [@joeddav](https://github.com/joeddav) for adding this dataset. |
oscar-corpus/OSCAR-2301 | 2023-04-18T10:08:22.000Z | [
"task_categories:fill-mask",
"task_categories:text-generation",
"task_ids:language-modeling",
"multilinguality:multilingual",
"size_categories:n>1T",
"source_datasets:original",
"license:cc0-1.0",
"arxiv:2212.10440",
"arxiv:2010.14571",
"region:us"
] | oscar-corpus | The Open Super-large Crawled Aggregated coRpus is a huge multilingual corpus obtained by language classification and filtering of the Common Crawl corpus using the Ungoliant architecture.\ | @ARTICLE{2022arXiv221210440J,
author = {{Jansen}, Tim and {Tong}, Yangling and {Zevallos}, Victoria and {Ortiz Suarez}, Pedro},
title = "{Perplexed by Quality: A Perplexity-based Method for Adult and Harmful Content Detection in Multilingual Heterogeneous Web Data}",
journal = {arXiv e-prints},
keywords = {Computer Science - Computation and Language},
year = 2022,
month = dec,
eid = {arXiv:2212.10440},
pages = {arXiv:2212.10440},
doi = {10.48550/arXiv.2212.10440},
archivePrefix = {arXiv},
eprint = {2212.10440},
primaryClass = {cs.CL},
adsurl = {https://ui.adsabs.harvard.edu/abs/2022arXiv221210440J},
adsnote = {Provided by the SAO/NASA Astrophysics Data System}
}
@inproceedings{abadji-etal-2022-towards,
title = "Towards a Cleaner Document-Oriented Multilingual Crawled Corpus",
author = "Abadji, Julien and
Ortiz Suarez, Pedro and
Romary, Laurent and
Sagot, Beno{\^\i}t",
booktitle = "Proceedings of the Thirteenth Language Resources and Evaluation Conference",
month = jun,
year = "2022",
address = "Marseille, France",
publisher = "European Language Resources Association",
url = "https://aclanthology.org/2022.lrec-1.463",
pages = "4344--4355",
abstract = "The need for large corpora raw corpora has dramatically increased in recent years with the introduction of transfer learning and semi-supervised learning methods to Natural Language Processing. And while there have been some recent attempts to manually curate the amount of data necessary to train large language models, the main way to obtain this data is still through automatic web crawling. In this paper we take the existing multilingual web corpus OSCAR and its pipeline Ungoliant that extracts and classifies data from Common Crawl at the line level, and propose a set of improvements and automatic annotations in order to produce a new document-oriented version of OSCAR that could prove more suitable to pre-train large generative language models as well as hopefully other applications in Natural Language Processing and Digital Humanities.",
}
@inproceedings{AbadjiOrtizSuarezRomaryetal.2021,
author = {Julien Abadji and Pedro Javier Ortiz Su{\'a}rez and Laurent Romary and Beno{\^i}t Sagot},
title = {Ungoliant: An optimized pipeline for the generation of a very large-scale multilingual web corpus},
series = {Proceedings of the Workshop on Challenges in the Management of Large Corpora (CMLC-9) 2021. Limerick, 12 July 2021 (Online-Event)},
editor = {Harald L{\"u}ngen and Marc Kupietz and Piotr Bański and Adrien Barbaresi and Simon Clematide and Ines Pisetta},
publisher = {Leibniz-Institut f{\"u}r Deutsche Sprache},
address = {Mannheim},
doi = {10.14618/ids-pub-10468},
url = {https://nbn-resolving.org/urn:nbn:de:bsz:mh39-104688},
pages = {1 -- 9},
year = {2021},
abstract = {Since the introduction of large language models in Natural Language Processing, large raw corpora have played a crucial role in Computational Linguistics. However, most of these large raw corpora are either available only for English or not available to the general public due to copyright issues. Nevertheless, there are some examples of freely available multilingual corpora for training Deep Learning NLP models, such as the OSCAR and Paracrawl corpora. However, they have quality issues, especially for low-resource languages. Moreover, recreating or updating these corpora is very complex. In this work, we try to reproduce and improve the goclassy pipeline used to create the OSCAR corpus. We propose a new pipeline that is faster, modular, parameterizable, and well documented. We use it to create a corpus similar to OSCAR but larger and based on recent data. Also, unlike OSCAR, the metadata information is at the document level. We release our pipeline under an open source license and publish the corpus under a research-only license.},
language = {en}
}
@article{kreutzer-etal-2022-quality,
title = "Quality at a Glance: An Audit of Web-Crawled Multilingual Datasets",
author = {Kreutzer, Julia and
Caswell, Isaac and
Wang, Lisa and
Wahab, Ahsan and
van Esch, Daan and
Ulzii-Orshikh, Nasanbayar and
Tapo, Allahsera and
Subramani, Nishant and
Sokolov, Artem and
Sikasote, Claytone and
Setyawan, Monang and
Sarin, Supheakmungkol and
Samb, Sokhar and
Sagot, Beno{\^\i}t and
Rivera, Clara and
Rios, Annette and
Papadimitriou, Isabel and
Osei, Salomey and
Suarez, Pedro Ortiz and
Orife, Iroro and
Ogueji, Kelechi and
Rubungo, Andre Niyongabo and
Nguyen, Toan Q. and
M{\"u}ller, Mathias and
M{\"u}ller, Andr{\'e} and
Muhammad, Shamsuddeen Hassan and
Muhammad, Nanda and
Mnyakeni, Ayanda and
Mirzakhalov, Jamshidbek and
Matangira, Tapiwanashe and
Leong, Colin and
Lawson, Nze and
Kudugunta, Sneha and
Jernite, Yacine and
Jenny, Mathias and
Firat, Orhan and
Dossou, Bonaventure F. P. and
Dlamini, Sakhile and
de Silva, Nisansa and
{\c{C}}abuk Ball{\i}, Sakine and
Biderman, Stella and
Battisti, Alessia and
Baruwa, Ahmed and
Bapna, Ankur and
Baljekar, Pallavi and
Azime, Israel Abebe and
Awokoya, Ayodele and
Ataman, Duygu and
Ahia, Orevaoghene and
Ahia, Oghenefego and
Agrawal, Sweta and
Adeyemi, Mofetoluwa},
journal = "Transactions of the Association for Computational Linguistics",
volume = "10",
year = "2022",
address = "Cambridge, MA",
publisher = "MIT Press",
url = "https://aclanthology.org/2022.tacl-1.4",
doi = "10.1162/tacl_a_00447",
pages = "50--72",
abstract = "With the success of large-scale pre-training and multilingual modeling in Natural Language Processing (NLP), recent years have seen a proliferation of large, Web-mined text datasets covering hundreds of languages. We manually audit the quality of 205 language-specific corpora released with five major public datasets (CCAligned, ParaCrawl, WikiMatrix, OSCAR, mC4). Lower-resource corpora have systematic issues: At least 15 corpora have no usable text, and a significant fraction contains less than 50{\%} sentences of acceptable quality. In addition, many are mislabeled or use nonstandard/ambiguous language codes. We demonstrate that these issues are easy to detect even for non-proficient speakers, and supplement the human audit with automatic analyses. Finally, we recommend techniques to evaluate and improve multilingual corpora and discuss potential risks that come with low-quality data releases.",
}
@inproceedings{ortiz-suarez-etal-2020-monolingual,
title = "A Monolingual Approach to Contextualized Word Embeddings for Mid-Resource Languages",
author = "Ortiz Su{\'a}rez, Pedro Javier and
Romary, Laurent and
Sagot, Benoit",
booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics",
month = jul,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.acl-main.156",
pages = "1703--1714",
abstract = "We use the multilingual OSCAR corpus, extracted from Common Crawl via language classification, filtering and cleaning, to train monolingual contextualized word embeddings (ELMo) for five mid-resource languages. We then compare the performance of OSCAR-based and Wikipedia-based ELMo embeddings for these languages on the part-of-speech tagging and parsing tasks. We show that, despite the noise in the Common-Crawl-based OSCAR data, embeddings trained on OSCAR perform much better than monolingual embeddings trained on Wikipedia. They actually equal or improve the current state of the art in tagging and parsing for all five languages. In particular, they also improve over multilingual Wikipedia-based contextual embeddings (multilingual BERT), which almost always constitutes the previous state of the art, thereby showing that the benefit of a larger, more diverse corpus surpasses the cross-lingual benefit of multilingual embedding architectures.",
}
@inproceedings{OrtizSuarezSagotRomary2019,
author = {Pedro Javier {Ortiz Su{\'a}rez} and Benoit Sagot and Laurent Romary},
title = {Asynchronous pipelines for processing huge corpora on medium to low resource infrastructures},
series = {Proceedings of the Workshop on Challenges in the Management of Large Corpora (CMLC-7) 2019. Cardiff, 22nd July 2019},
editor = {Piotr Bański and Adrien Barbaresi and Hanno Biber and Evelyn Breiteneder and Simon Clematide and Marc Kupietz and Harald L{\"u}ngen and Caroline Iliadi},
publisher = {Leibniz-Institut f{\"u}r Deutsche Sprache},
address = {Mannheim},
doi = {10.14618/ids-pub-9021},
url = {http://nbn-resolving.de/urn:nbn:de:bsz:mh39-90215},
pages = {9 -- 16},
year = {2019},
abstract = {Common Crawl is a considerably large, heterogeneous multilingual corpus comprised of crawled documents from the internet, surpassing 20TB of data and distributed as a set of more than 50 thousand plain text files where each contains many documents written in a wide variety of languages. Even though each document has a metadata block associated to it, this data lacks any information about the language in which each document is written, making it extremely difficult to use Common Crawl for monolingual applications. We propose a general, highly parallel, multithreaded pipeline to clean and classify Common Crawl by language; we specifically design it so that it runs efficiently on medium to low resource infrastructures where I/O speeds are the main constraint. We develop the pipeline so that it can be easily reapplied to any kind of heterogeneous corpus and so that it can be parameterised to a wide range of infrastructures. We also distribute a 6.3TB version of Common Crawl, filtered, classified by language, shuffled at line level in order to avoid copyright issues, and ready to be used for NLP applications.},
language = {en}
} | null | 59 | 9,042 | ---
license: cc0-1.0
size_categories:
- n>1T
multilinguality:
- multilingual
source_datasets:
- original
task_categories:
- fill-mask
- text-generation
task_ids:
- language-modeling
paperswithcode_id: oscar
extra_gated_prompt: "By filling the form below, you understand that only the metadata and the annotations of OSCAR 23.01 have a cc0-1.0 license, and that the rest of the content is crawled data derived from the November/December 2022 snapshot of Common Crawl, for which the authors of OSCAR **do not** hold any copyright whatsoever."
extra_gated_fields:
Name: text
Email: text
Affiliation: text
Country: text
Usecase: text
I have explicitly check with my jurisdiction and I confirm that downloading OSCAR 2301 is legal in the country/region where I am located right now, and for the use case that I have described above: checkbox
---
# Dataset Card for "OSCAR 23.01"
## IMPORTANT NOTE: THIS DATASET CARD IS STILL BEING WRITTEN, PLEASE BE PATIENT WHILE WE COMPLETE ALL THE INFORMATION ABOUT THE CORPUS
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://oscar-project.org](https://oscar-project.org)
- **Repository:** [https://github.com/oscar-project](https://github.com/oscar-project)
- **Papers:** [Towards a Cleaner Document-Oriented Multilingual Crawled Corpus](https://aclanthology.org/2022.lrec-1.463/), [Perplexed by Quality: A Perplexity-based Method for Adult and Harmful Content Detection in Multilingual Heterogeneous Web Data](https://arxiv.org/abs/2212.10440)
- **Point of Contact:** [Contact](https://oscar-project.org/#contact)
### Dataset Summary
The OSCAR project (**O**pen **S**uper-large **C**rawled **A**ggregated co**R**pus) is an Open Source project aiming to provide web-based multilingual resources and datasets for Machine Learning (ML) and Artificial Intelligence (AI) applications. The project focuses specifically in providing large quantities of unannotated raw data that is commonly used in the pre-training of large deep learning models. The OSCAR project has developed [high-performance data pipelines](https://github.com/oscar-corpus/ungoliant) specifically conceived to classify and filter large amounts of [web data](https://commoncrawl.org/). The project has also put special attention in improving the data quality of web-based corpora as well as providing data for low-resource languages, so that these new ML/AI technologies are accessible to as many communities as possible.
OSCAR 23.01 is the January 2023 version of the OSCAR Corpus based on the [November/December 2022 dump of Common Crawl](https://commoncrawl.org/2022/12/nov-dec-2022-crawl-archive-now-available/). While being quite similar to OSCAR 22.01, it contains several new features, including [KenLM](https://kheafield.com/code/kenlm/)-based adult content detection, precomputed [Locality-Sensitive Hashes](https://fr.wikipedia.org/wiki/Locality_sensitive_hashing) for near deduplication, and [blocklist](https://dsi.ut-capitole.fr/blacklists/index_en.php)-based categories. OSCAR 23.01 has also moved from gzip to [Zstandard compression](https://facebook.github.io/zstd/). You might already have `zstd` installed on your system, but if not, please check the [Zstandard website](https://facebook.github.io/zstd/) for installation instructions.
### Supported Tasks and Leaderboards
OSCAR is mainly intended to pretrain language models and word representations.
### Languages
All the data is distributed by language, both the original and the deduplicated versions of the data are available. 151 different languages are available. The table in subsection [Data Splits Sample Size](#data-splits-sample-size) provides the language code for each subcorpus as well as the number of words (space separated tokens), lines and sizes for both the original and the deduplicated versions of OSCAR.
### Issues
OSCAR 23.01 may have quality issues on low size subcorpora, as it has been the case before.
Note that since the documents are identified as a whole, it is expected to have lines in other languages in a given language subcorpus.
As an example, it is known and expected that the German subcorpus contains documents holding lines identified as Swiss German / Alemannic.
**If you encounter something that is unexpected, please file an issue here: https://github.com/oscar-corpus/corpus/issues.**
|Language code|Language|Issues|
|-------------|--------|------|
| | | |
## Dataset Structure
We show detailed information for all the configurations of the dataset.
### Data Instances
TODO
### Layout
```js
{
"content":"English sentence\nphrase en français\n????????????", // (1)
"warc_headers":{ // (2)
"warc-identified-content-language":"fra,eng",
"warc-target-uri":"https://fr.wikipedia.org/wiki/...",
"warc-record-id":"<urn:uuid:29eaa920-d299-4b1d-b687-c72bd8d68116>",
"warc-type":"conversion",
"content-length":"35298", // (3)
"warc-refers-to":"<urn:uuid:39e42055-0d94-4e45-9c6c-9e7056635d64>",
"warc-block-digest":"sha1:WFH2A5WHCS2H365GIAFYQPI7UOAMFGHB", // (3)
"warc-date":"2022-11-26T09:45:47Z",
"content-type":"text/plain"
},
"metadata":{
"identification":{ // (4)
"label":"fr",
"prob":0.8938327
},
"harmful_pp":4063.1814, // (5)
"tlsh":"tlsh:T125315FF2B6088901EEA097015DB39B4600B...", // (6)
"quality_warnings":[ // (7)
"short_sentences",
"header",
"footer"
],
"categories":[ // (8)
"examen_pix",
"liste_bu"
],
"sentence_identifications":[ // (9)
{
"label":"fr",
"prob":0.99837273
},
{
"label":"en",
"prob":0.9992377
},
null
]
}
}
```
### Data Splits
<details>
<summary>Click to expand the number of samples per configuration</summary>
</details>
## Table
| | Code | Language | # docs | # words | Content Length : |
|----:|:-------|:-------------------------|:--------------|:----------------|:-----------------|
| 0 | af | Afrikaans | 23,994 | 6,217,024 | 37.2 MB |
| 1 | sq | Albanian | 1,342,790 | 462,694,599 | 3.2 GB |
| 2 | am | Amharic | 119,434 | 40,262,809 | 512.9 MB |
| 3 | ar | Arabic | 25,012,116 | 10,081,452,882 | 110.7 GB |
| 4 | an | Aragonese | 34 | 264 | 11.0 kB |
| 5 | hy | Armenian | 1,056,974 | 336,045,041 | 4.9 GB |
| 6 | as | Assamese | 89,542 | 24,395,215 | 412.1 MB |
| 7 | ast | Asturian | 440 | 10,917 | 74.1 kB |
| 8 | av | Avaric | 44 | 1,073 | 18.6 kB |
| 9 | az | Azerbaijani | 1,159,994 | 316,850,330 | 3.0 GB |
| 10 | bn | Bangla | 3,474,086 | 1,092,983,765 | 19.1 GB |
| 11 | ba | Bashkir | 128,248 | 26,036,637 | 363.7 MB |
| 12 | eu | Basque | 678,474 | 136,672,615 | 1.2 GB |
| 13 | be | Belarusian | 445,612 | 164,729,607 | 2.3 GB |
| 14 | bh | Bihari languages | 48 | 507 | 6.8 kB |
| 15 | bpy | Bishnupriya | 2,346 | 346,947 | 5.4 MB |
| 16 | bs | Bosnian | 20 | 395 | 3.0 kB |
| 17 | br | Breton | 36,338 | 4,759,407 | 31.4 MB |
| 18 | bg | Bulgarian | 8,933,998 | 3,635,273,738 | 44.1 GB |
| 19 | my | Burmese | 430,276 | 82,433,836 | 3.0 GB |
| 20 | ca | Catalan | 6,953,898 | 2,240,460,836 | 15.3 GB |
| 21 | ceb | Cebuano | 16,174 | 6,263,404 | 41.1 MB |
| 22 | ckb | Central Kurdish | 182,508 | 61,334,746 | 772.9 MB |
| 23 | ce | Chechen | 11,686 | 1,051,752 | 13.9 MB |
| 24 | zh | Chinese | 138,478,270 | 44,378,380,161 | 1.4 TB |
| 25 | cv | Chuvash | 16,652 | 3,039,925 | 42.3 MB |
| 26 | kw | Cornish | 8 | 80 | 432 Bytes |
| 27 | hr | Croatian | 31,808 | 3,542,961 | 26.5 MB |
| 28 | cs | Czech | 34,859,632 | 9,717,378,559 | 77.0 GB |
| 29 | da | Danish | 7,214,338 | 2,217,634,340 | 14.8 GB |
| 30 | dv | Divehi | 77,060 | 10,655,359 | 200.1 MB |
| 31 | nl | Dutch | 72,552,688 | 19,564,553,306 | 135.0 GB |
| 32 | mhr | Eastern Mari | 9,502 | 1,615,215 | 22.9 MB |
| 33 | arz | Egyptian Arabic | 3,958 | 385,511 | 3.7 MB |
| 34 | en | English | 1,235,510,986 | 523,869,288,690 | 3.4 TB |
| 35 | eo | Esperanto | 226,924 | 67,774,923 | 474.8 MB |
| 36 | et | Estonian | 3,601,904 | 938,296,892 | 8.0 GB |
| 37 | tl | Filipino | 250,558 | 110,560,444 | 719.2 MB |
| 38 | fi | Finnish | 14,471,710 | 4,198,143,883 | 41.1 GB |
| 39 | fr | French | 158,334,998 | 62,127,088,294 | 430.5 GB |
| 40 | gl | Galician | 248,762 | 38,345,625 | 255.7 MB |
| 41 | ka | Georgian | 1,343,036 | 373,935,158 | 8.4 GB |
| 42 | de | German | 206,598,430 | 73,848,586,648 | 594.7 GB |
| 43 | gom | Goan Konkani | 398 | 121,035 | 2.3 MB |
| 44 | el | Greek | 20,282,864 | 7,691,622,692 | 95.7 GB |
| 45 | gn | Guarani | 14 | 260 | 2.2 kB |
| 46 | gu | Gujarati | 425,552 | 417,001,705 | 5.6 GB |
| 47 | ht | Haitian Creole | 2 | 20,671 | 93.1 kB |
| 48 | he | Hebrew | 3,997,888 | 1,697,158,891 | 18.0 GB |
| 49 | hi | Hindi | 5,514,454 | 2,475,605,444 | 32.6 GB |
| 50 | hu | Hungarian | 21,349,372 | 16,013,364,289 | 150.1 GB |
| 51 | is | Icelandic | 1,210,232 | 294,471,539 | 2.2 GB |
| 52 | io | Ido | 224 | 2,598 | 16.1 kB |
| 53 | ilo | Iloko | 144 | 4,411 | 28.0 kB |
| 54 | id | Indonesian | 7,109,778 | 3,228,020,221 | 23.4 GB |
| 55 | ia | Interlingua | 34 | 9,384 | 33.5 kB |
| 56 | ie | Interlingue | 2 | 0 | 881 Bytes |
| 57 | ga | Irish | 29,894 | 9,054,923 | 63.2 MB |
| 58 | it | Italian | 89,021,606 | 36,327,274,203 | 259.4 GB |
| 59 | ja | Japanese | 94,236,404 | 4,401,059,165 | 181.2 GB |
| 60 | jv | Javanese | 172 | 3,286 | 25.7 kB |
| 61 | xal | Kalmyk | 2 | 27 | 315 Bytes |
| 62 | kn | Kannada | 448,500 | 124,924,350 | 2.6 GB |
| 63 | krc | Karachay-Balkar | 496 | 8,385 | 122.4 kB |
| 64 | kk | Kazakh | 677,622 | 214,679,857 | 3.3 GB |
| 65 | km | Khmer | 450,660 | 59,880,231 | 3.2 GB |
| 66 | kv | Komi | 460 | 5,909 | 70.3 kB |
| 67 | ko | Korean | 15,147,698 | 3,435,866,935 | 38.1 GB |
| 68 | ku | Kurdish | 80,338 | 25,921,607 | 174.1 MB |
| 69 | ky | Kyrgyz | 144,288 | 32,062,783 | 489.3 MB |
| 70 | lo | Lao | 118,374 | 10,659,203 | 472.1 MB |
| 71 | la | Latin | 14,384 | 307,865 | 2.0 MB |
| 72 | lv | Latvian | 2,435,882 | 845,459,899 | 7.4 GB |
| 73 | lez | Lezghian | 676 | 60,634 | 856.6 kB |
| 74 | li | Limburgish | 6 | 169 | 1.4 kB |
| 75 | lt | Lithuanian | 5,182,028 | 1,674,362,574 | 14.5 GB |
| 76 | jbo | Lojban | 572 | 312,315 | 1.5 MB |
| 77 | lmo | Lombard | 112 | 3,269 | 21.0 kB |
| 78 | nds | Low German | 5,248 | 1,612,175 | 10.7 MB |
| 79 | dsb | Lower Sorbian | 8 | 84 | 664 Bytes |
| 80 | lb | Luxembourgish | 18,090 | 2,514,838 | 18.4 MB |
| 81 | mk | Macedonian | 1,063,298 | 389,344,425 | 4.7 GB |
| 82 | mai | Maithili | 46 | 467 | 6.8 kB |
| 83 | mg | Malagasy | 10,830 | 1,416,430 | 11.2 MB |
| 84 | ms | Malay | 11,500 | 238,477 | 2.6 MB |
| 85 | ml | Malayalam | 800,936 | 236,597,838 | 5.8 GB |
| 86 | mt | Maltese | 5,180 | 149,886 | 1.3 MB |
| 87 | mr | Marathi | 729,578 | 252,706,331 | 4.5 GB |
| 88 | mzn | Mazanderani | 384 | 16,115 | 169.2 kB |
| 89 | min | Minangkabau | 2,436 | 305,589 | 3.8 MB |
| 90 | xmf | Mingrelian | 7,318 | 283,316 | 6.1 MB |
| 91 | mwl | Mirandese | 4 | 54 | 423 Bytes |
| 92 | mn | Mongolian | 1,061,710 | 454,350,415 | 5.8 GB |
| 93 | multi | **Multilingual** | 2,948,202 | 1,251,676,406 | 11.9 GB |
| 94 | nah | Nahuatl languages | 38 | 279 | 2.4 kB |
| 95 | ne | Nepali | 1,152,156 | 278,901,036 | 4.9 GB |
| 96 | new | Newari | 1,996 | 229,703 | 4.0 MB |
| 97 | no | Norwegian | 2,797,378 | 373,160,033 | 2.6 GB |
| 98 | nn | Norwegian Nynorsk | 19,470 | 575,518 | 3.7 MB |
| 99 | oc | Occitan | 920 | 34,701 | 405.0 kB |
| 100 | or | Odia | 158,426 | 31,963,340 | 543.1 MB |
| 101 | os | Ossetic | 8,628 | 3,935,964 | 50.7 MB |
| 102 | ps | Pashto | 87,408 | 30,196,179 | 261.6 MB |
| 103 | fa | Persian | 23,813,882 | 9,609,206,698 | 93.2 GB |
| 104 | pms | Piedmontese | 2,524 | 510,087 | 3.1 MB |
| 105 | pl | Polish | 57,184,826 | 18,073,705,588 | 147.1 GB |
| 106 | pt | Portuguese | 36,062,800 | 15,172,557,311 | 105.0 GB |
| 107 | pa | Punjabi | 222,058 | 104,235,418 | 1.4 GB |
| 108 | qu | Quechua | 2 | 13 | 143 Bytes |
| 109 | ro | Romanian | 11,985,668 | 6,302,600,833 | 45.6 GB |
| 110 | bxr | Russia Buriat | 72 | 698 | 8.2 kB |
| 111 | ru | Russian | 194,143,422 | 78,032,029,344 | 1.1 TB |
| 112 | sah | Sakha | 17,566 | 4,288,051 | 68.8 MB |
| 113 | sa | Sanskrit | 16,802 | 2,479,345 | 56.3 MB |
| 114 | gd | Scottish Gaelic | 776 | 18,458 | 146.1 kB |
| 115 | sr | Serbian | 1,677,896 | 632,781,822 | 7.7 GB |
| 116 | sh | Serbian (Latin) | 3,214 | 166,517 | 816.4 kB |
| 117 | sd | Sindhi | 48,566 | 14,667,207 | 131.6 MB |
| 118 | si | Sinhala | 301,066 | 172,755,385 | 2.6 GB |
| 119 | sk | Slovak | 8,931,784 | 2,704,716,280 | 21.5 GB |
| 120 | sl | Slovenian | 1,112,560 | 192,816,743 | 1.4 GB |
| 121 | so | Somali | 6 | 51 | 503 Bytes |
| 122 | azb | South Azerbaijani | 26,364 | 2,029,729 | 28.4 MB |
| 123 | es | Spanish | 153,574,556 | 63,388,237,965 | 429.9 GB |
| 124 | su | Sundanese | 18 | 258 | 2.0 kB |
| 125 | sw | Swahili | 1,664 | 164,459 | 1.0 MB |
| 126 | sv | Swedish | 21,891,348 | 6,993,719,601 | 50.0 GB |
| 127 | gsw | Swiss German | 342 | 34,328 | 232.7 kB |
| 128 | tg | Tajik | 144,932 | 76,987,285 | 1.0 GB |
| 129 | ta | Tamil | 1,638,238 | 738,824,392 | 15.8 GB |
| 130 | tt | Tatar | 262,654 | 59,253,765 | 833.8 MB |
| 131 | te | Telugu | 644,712 | 201,575,815 | 3.9 GB |
| 132 | th | Thai | 14,845,900 | 2,224,483,018 | 92.0 GB |
| 133 | bo | Tibetan | 62,352 | 6,062,558 | 531.6 MB |
| 134 | tr | Turkish | 26,654,330 | 8,290,890,087 | 73.7 GB |
| 135 | tk | Turkmen | 4,576 | 325,786 | 3.3 MB |
| 136 | uk | Ukrainian | 10,059,992 | 3,183,842,018 | 44.7 GB |
| 137 | x-eml | Emiliano-Romagnol | 4 | 329 | 1.8 kB |
| 138 | hsb | Upper Sorbian | 402 | 15,827 | 123.2 kB |
| 139 | ur | Urdu | 887,004 | 434,023,273 | 3.8 GB |
| 140 | ug | Uyghur | 51,304 | 14,659,554 | 219.8 MB |
| 141 | uz | Uzbek | 15,806 | 1,665,960 | 15.3 MB |
| 142 | vi | Vietnamese | 33,933,994 | 22,424,984,210 | 140.8 GB |
| 143 | vo | Volapük | 896 | 49,968 | 371.9 kB |
| 144 | wa | Walloon | 390 | 6,347 | 34.3 kB |
| 145 | war | Waray | 1,494 | 19,665 | 126.8 kB |
| 146 | cy | Welsh | 151,512 | 52,250,043 | 333.0 MB |
| 147 | fy | Western Frisian | 45,458 | 9,885,788 | 70.4 MB |
| 148 | mrj | Western Mari | 496 | 60,180 | 765.8 kB |
| 149 | pnb | Western Panjabi | 12,904 | 11,844,695 | 105.8 MB |
| 150 | wuu | Wu Chinese | 136 | 1,199 | 26.8 kB |
| 151 | yi | Yiddish | 47,438 | 14,287,370 | 171.7 MB |
| 152 | yo | Yoruba | 128 | 2,396 | 16.6 kB |
## Dataset Creation
### Curation Rationale
OSCAR was constructed using [`Ungoliant`](https://github.com/oscar-corpus/ungoliant), a new pipeline derived from [goclassy](https://github.com/oscar-corpus/goclassy), itself being derived from [fastText's one](https://github.com/facebookresearch/fastText).
The pipeline works on documents rather than lines.
`Ungoliant` is implemented in the [Rust programming language](https://rust-lang.org), and uses [rayon](https://github.com/rayon-rs/rayon) as its data parallelism strategy.
Threading is done at shard, record and sentence level, making the whole generation process much more efficient.
Filtering will be explained in a future blog post at our [website](https://oscar-corpus.com)
### Source Data
#### Initial Data Collection and Normalization
[Common Crawl](https://commoncrawl.org/) is a non-profit foundation which produces and maintains an open repository of web crawled data that is both accessible and analysable. Common Crawl's complete web archive consists of petabytes of data collected over 8 years of web crawling. The repository contains raw web page HTML data (WARC files), metdata extracts (WAT files) and plain text extracts (WET files). The organisation's crawlers has always respected [nofollow](http://microformats.org/wiki/rel-nofollow) and [robots.txt](https://www.robotstxt.org/) policies.
Each monthly Common Crawl snapshot is in itself a massive multilingual corpus, where every single file contains data coming from multiple web pages written in a large variety of languages and covering all possible types of topics.
To construct OSCAR the WET files of Common Crawl were used. These contain the extracted plain texts from the websites mostly converted to UTF-8, as well as headers containing the metatada of each crawled document. Each WET file comes compressed in gzip format and is stored on Amazon Web Services. In the case of OSCAR 22.01, the **November/December 2021** snapshot was used. It is composed by 64 000 compressed text files containing documents and their headers.
#### Who are the source language producers?
The data comes from multiple web pages in a large variety of languages.
### Annotations
The dataset does not contain any additional annotations.
#### Annotation process
N/A
#### Who are the annotators?
N/A
### Personal and Sensitive Information
Being constructed from Common Crawl, Personal and sensitive information might be present. This **must** be considered before training deep learning models with OSCAR, specially in the case of text-generation models.
## Considerations for Using the Data
### Social Impact of Dataset
OSCAR is intended to bring more data to a wide variety of lanuages, the aim of the corpus is to make large amounts of data available to lower resource languages in order to facilitate the pre-training of state-of-the-art language modeling architectures.
### Discussion of Biases
OSCAR is not properly filtered yet and this can be reflected on the models trained with it. Care is advised specially concerning biases of the resulting models.
### Other Known Limitations
The [fastText linear classifier](https://fasttext.cc) is limed both in performance and the variety of languages it can recognize, so the quality of some OSCAR sub-corpora might be lower than expected, specially for the lowest-resource langiuages. Some audits have already been done by [third parties](https://arxiv.org/abs/2010.14571).
## Additional Information
### Dataset Curators
This release of OSCAR was made possible by [Julien Abadji](https://ujj.space), [Pedro Ortiz Suarez](https://portizs.eu/), [Rua Ismail](https://oscar-project.org/authors/rua/), [Sotaro Takeshita](https://sotaro.io/about), [Sebastian Nagel](https://www.polver.uni-konstanz.de/cnc/people/nagel/) and [Benoit Sagot](http://pauillac.inria.fr/~sagot/).
### Licensing Information
These data are released under this licensing scheme
We do not own any of the text from which these data has been extracted.
We license the actual packaging, the metadata and the annotations of these data under the Creative Commons CC0 license ("no rights reserved") http://creativecommons.org/publicdomain/zero/1.0/
To the extent possible under law, the OSCAR project, Inria, the Univertity of Mannheim and DFKI GmbH have waived all copyright and related or neighboring rights to OSCAR
This work is published from: France and Germany.
Should you consider that our data contains material that is owned by you and should therefore not be reproduced here, please:
* Clearly identify yourself, with detailed contact data such as an address, telephone number or email address at which you can be contacted.
* Clearly identify the copyrighted work claimed to be infringed.
* Clearly identify the material that is claimed to be infringing and information reasonably sufficient to allow us to locate the material.
We will comply to legitimate requests by removing the affected sources from the next release of the corpus.
### Citation Information
```
@ARTICLE{2022arXiv221210440J,
author = {{Jansen}, Tim and {Tong}, Yangling and {Zevallos}, Victoria and {Ortiz Suarez}, Pedro},
title = "{Perplexed by Quality: A Perplexity-based Method for Adult and Harmful Content Detection in Multilingual Heterogeneous Web Data}",
journal = {arXiv e-prints},
keywords = {Computer Science - Computation and Language},
year = 2022,
month = dec,
eid = {arXiv:2212.10440},
pages = {arXiv:2212.10440},
doi = {10.48550/arXiv.2212.10440},
archivePrefix = {arXiv},
eprint = {2212.10440},
primaryClass = {cs.CL},
adsurl = {https://ui.adsabs.harvard.edu/abs/2022arXiv221210440J},
adsnote = {Provided by the SAO/NASA Astrophysics Data System}
}
@inproceedings{abadji-etal-2022-towards,
title = "Towards a Cleaner Document-Oriented Multilingual Crawled Corpus",
author = "Abadji, Julien and
Ortiz Suarez, Pedro and
Romary, Laurent and
Sagot, Beno{\^\i}t",
booktitle = "Proceedings of the Thirteenth Language Resources and Evaluation Conference",
month = jun,
year = "2022",
address = "Marseille, France",
publisher = "European Language Resources Association",
url = "https://aclanthology.org/2022.lrec-1.463",
pages = "4344--4355",
abstract = "The need for large corpora raw corpora has dramatically increased in recent years with the introduction of transfer learning and semi-supervised learning methods to Natural Language Processing. And while there have been some recent attempts to manually curate the amount of data necessary to train large language models, the main way to obtain this data is still through automatic web crawling. In this paper we take the existing multilingual web corpus OSCAR and its pipeline Ungoliant that extracts and classifies data from Common Crawl at the line level, and propose a set of improvements and automatic annotations in order to produce a new document-oriented version of OSCAR that could prove more suitable to pre-train large generative language models as well as hopefully other applications in Natural Language Processing and Digital Humanities.",
}
@inproceedings{AbadjiOrtizSuarezRomaryetal.2021,
author = {Julien Abadji and Pedro Javier Ortiz Su{\'a}rez and Laurent Romary and Beno{\^i}t Sagot},
title = {Ungoliant: An optimized pipeline for the generation of a very large-scale multilingual web corpus},
series = {Proceedings of the Workshop on Challenges in the Management of Large Corpora (CMLC-9) 2021. Limerick, 12 July 2021 (Online-Event)},
editor = {Harald L{\"u}ngen and Marc Kupietz and Piotr Bański and Adrien Barbaresi and Simon Clematide and Ines Pisetta},
publisher = {Leibniz-Institut f{\"u}r Deutsche Sprache},
address = {Mannheim},
doi = {10.14618/ids-pub-10468},
url = {https://nbn-resolving.org/urn:nbn:de:bsz:mh39-104688},
pages = {1 -- 9},
year = {2021},
abstract = {Since the introduction of large language models in Natural Language Processing, large raw corpora have played a crucial role in Computational Linguistics. However, most of these large raw corpora are either available only for English or not available to the general public due to copyright issues. Nevertheless, there are some examples of freely available multilingual corpora for training Deep Learning NLP models, such as the OSCAR and Paracrawl corpora. However, they have quality issues, especially for low-resource languages. Moreover, recreating or updating these corpora is very complex. In this work, we try to reproduce and improve the goclassy pipeline used to create the OSCAR corpus. We propose a new pipeline that is faster, modular, parameterizable, and well documented. We use it to create a corpus similar to OSCAR but larger and based on recent data. Also, unlike OSCAR, the metadata information is at the document level. We release our pipeline under an open source license and publish the corpus under a research-only license.},
language = {en}
}
@article{kreutzer-etal-2022-quality,
title = "Quality at a Glance: An Audit of Web-Crawled Multilingual Datasets",
author = {Kreutzer, Julia and
Caswell, Isaac and
Wang, Lisa and
Wahab, Ahsan and
van Esch, Daan and
Ulzii-Orshikh, Nasanbayar and
Tapo, Allahsera and
Subramani, Nishant and
Sokolov, Artem and
Sikasote, Claytone and
Setyawan, Monang and
Sarin, Supheakmungkol and
Samb, Sokhar and
Sagot, Beno{\^\i}t and
Rivera, Clara and
Rios, Annette and
Papadimitriou, Isabel and
Osei, Salomey and
Suarez, Pedro Ortiz and
Orife, Iroro and
Ogueji, Kelechi and
Rubungo, Andre Niyongabo and
Nguyen, Toan Q. and
M{\"u}ller, Mathias and
M{\"u}ller, Andr{\'e} and
Muhammad, Shamsuddeen Hassan and
Muhammad, Nanda and
Mnyakeni, Ayanda and
Mirzakhalov, Jamshidbek and
Matangira, Tapiwanashe and
Leong, Colin and
Lawson, Nze and
Kudugunta, Sneha and
Jernite, Yacine and
Jenny, Mathias and
Firat, Orhan and
Dossou, Bonaventure F. P. and
Dlamini, Sakhile and
de Silva, Nisansa and
{\c{C}}abuk Ball{\i}, Sakine and
Biderman, Stella and
Battisti, Alessia and
Baruwa, Ahmed and
Bapna, Ankur and
Baljekar, Pallavi and
Azime, Israel Abebe and
Awokoya, Ayodele and
Ataman, Duygu and
Ahia, Orevaoghene and
Ahia, Oghenefego and
Agrawal, Sweta and
Adeyemi, Mofetoluwa},
journal = "Transactions of the Association for Computational Linguistics",
volume = "10",
year = "2022",
address = "Cambridge, MA",
publisher = "MIT Press",
url = "https://aclanthology.org/2022.tacl-1.4",
doi = "10.1162/tacl_a_00447",
pages = "50--72",
abstract = "With the success of large-scale pre-training and multilingual modeling in Natural Language Processing (NLP), recent years have seen a proliferation of large, Web-mined text datasets covering hundreds of languages. We manually audit the quality of 205 language-specific corpora released with five major public datasets (CCAligned, ParaCrawl, WikiMatrix, OSCAR, mC4). Lower-resource corpora have systematic issues: At least 15 corpora have no usable text, and a significant fraction contains less than 50{\%} sentences of acceptable quality. In addition, many are mislabeled or use nonstandard/ambiguous language codes. We demonstrate that these issues are easy to detect even for non-proficient speakers, and supplement the human audit with automatic analyses. Finally, we recommend techniques to evaluate and improve multilingual corpora and discuss potential risks that come with low-quality data releases.",
}
@inproceedings{ortiz-suarez-etal-2020-monolingual,
title = "A Monolingual Approach to Contextualized Word Embeddings for Mid-Resource Languages",
author = "Ortiz Su{'a}rez, Pedro Javier and
Romary, Laurent and
Sagot, Benoit",
booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics",
month = jul,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.acl-main.156",
pages = "1703--1714",
abstract = "We use the multilingual OSCAR corpus, extracted from Common Crawl via language classification, filtering and cleaning, to train monolingual contextualized word embeddings (ELMo) for five mid-resource languages. We then compare the performance of OSCAR-based and Wikipedia-based ELMo embeddings for these languages on the part-of-speech tagging and parsing tasks. We show that, despite the noise in the Common-Crawl-based OSCAR data, embeddings trained on OSCAR perform much better than monolingual embeddings trained on Wikipedia. They actually equal or improve the current state of the art in tagging and parsing for all five languages. In particular, they also improve over multilingual Wikipedia-based contextual embeddings (multilingual BERT), which almost always constitutes the previous state of the art, thereby showing that the benefit of a larger, more diverse corpus surpasses the cross-lingual benefit of multilingual embedding architectures.",
}
@inproceedings{OrtizSuarezSagotRomary2019,
author = {Pedro Javier {Ortiz Su{'a}rez} and Benoit Sagot and Laurent Romary},
title = {Asynchronous pipelines for processing huge corpora on medium to low resource infrastructures},
series = {Proceedings of the Workshop on Challenges in the Management of Large Corpora (CMLC-7) 2019. Cardiff, 22nd July 2019},
editor = {Piotr Bański and Adrien Barbaresi and Hanno Biber and Evelyn Breiteneder and Simon Clematide and Marc Kupietz and Harald L{"u}ngen and Caroline Iliadi},
publisher = {Leibniz-Institut f{"u}r Deutsche Sprache},
address = {Mannheim},
doi = {10.14618/ids-pub-9021},
url = {http://nbn-resolving.de/urn:nbn:de:bsz:mh39-90215},
pages = {9 -- 16},
year = {2019},
abstract = {Common Crawl is a considerably large, heterogeneous multilingual corpus comprised of crawled documents from the internet, surpassing 20TB of data and distributed as a set of more than 50 thousand plain text files where each contains many documents written in a wide variety of languages. Even though each document has a metadata block associated to it, this data lacks any information about the language in which each document is written, making it extremely difficult to use Common Crawl for monolingual applications. We propose a general, highly parallel, multithreaded pipeline to clean and classify Common Crawl by language; we specifically design it so that it runs efficiently on medium to low resource infrastructures where I/O speeds are the main constraint. We develop the pipeline so that it can be easily reapplied to any kind of heterogeneous corpus and so that it can be parameterised to a wide range of infrastructures. We also distribute a 6.3TB version of Common Crawl, filtered, classified by language, shuffled at line level in order to avoid copyright issues, and ready to be used for NLP applications.},
language = {en}
}
```
|
ought/raft | 2022-10-25T09:54:19.000Z | [
"task_categories:text-classification",
"task_ids:multi-class-classification",
"annotations_creators:expert-generated",
"annotations_creators:crowdsourced",
"language_creators:expert-generated",
"multilinguality:monolingual",
"size_categories:unknown",
"source_datasets:original",
"source_datasets:extended|ade_corpus_v2",
"source_datasets:extended|banking77",
"language:en",
"license:other",
"arxiv:2109.14076",
"region:us"
] | ought | Large pre-trained language models have shown promise for few-shot learning, completing text-based tasks given only a few task-specific examples. Will models soon solve classification tasks that have so far been reserved for human research assistants?
[RAFT](https://raft.elicit.org) is a few-shot classification benchmark that tests language models:
- across multiple domains (lit review, tweets, customer interaction, etc.)
- on economically valuable classification tasks (someone inherently cares about the task)
- in a setting that mirrors deployment (50 examples per task, info retrieval allowed, hidden test set) | @InProceedings{huggingface:dataset,
title = {A great new dataset},
author={huggingface, Inc.
},
year={2020}
} | null | 31 | 8,956 | ---
annotations_creators:
- expert-generated
- crowdsourced
language_creators:
- expert-generated
language:
- en
license:
- other
multilinguality:
- monolingual
size_categories:
- unknown
source_datasets:
- original
- extended|ade_corpus_v2
- extended|banking77
task_categories:
- text-classification
task_ids:
- multi-class-classification
pretty_name: 'Real-world Annotated Few-shot Tasks: RAFT'
language_bcp47:
- en-US
---
# Dataset Card for RAFT
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** https://raft.elicit.org
- **Repository:** https://huggingface.co/datasets/ought/raft
- **Paper:** [arxiv.org](https://arxiv.org/abs/2109.14076)
- **Leaderboard:** https://huggingface.co/spaces/ought/raft-leaderboard
- **Point of Contact:** [Eli Lifland](eli.d.lifland@gmail.com)
### Dataset Summary
The Real-world Annotated Few-shot Tasks (RAFT) dataset is an aggregation of English-language datasets found in the real world. Associated with each dataset is a binary or multiclass classification task, intended to improve our understanding of how language models perform on tasks that have concrete, real-world value. Only 50 labeled examples are provided in each dataset.
### Supported Tasks and Leaderboards
- `text-classification`: Each subtask in RAFT is a text classification task, and the provided train and test sets can be used to submit to the [RAFT Leaderboard](https://huggingface.co/spaces/ought/raft-leaderboard) To prevent overfitting and tuning on a held-out test set, the leaderboard is only evaluated once per week. Each task has its macro-f1 score calculated, then those scores are averaged to produce the overall leaderboard score.
### Languages
RAFT is entirely in American English (en-US).
## Dataset Structure
### Data Instances
| Dataset | First Example |
| ----------- | ----------- |
| Ade Corpus V2 | <pre>Sentence: No regional side effects were noted.<br>ID: 0<br>Label: 2</pre> |
| Banking 77 | <pre>Query: Is it possible for me to change my PIN number?<br>ID: 0<br>Label: 23<br></pre> |
| NeurIPS Impact Statement Risks | <pre>Paper title: Auto-Panoptic: Cooperative Multi-Component Architecture Search for Panoptic Segmentation...<br>Paper link: https://proceedings.neurips.cc/paper/2020/file/ec1f764517b7ffb52057af6df18142b7-Paper.pdf...<br>Impact statement: This work makes the first attempt to search for all key components of panoptic pipeline and manages to accomplish this via the p...<br>ID: 0<br>Label: 1</pre> |
| One Stop English | <pre>Article: For 85 years, it was just a grey blob on classroom maps of the solar system. But, on 15 July, Pluto was seen in high resolution ...<br>ID: 0<br>Label: 3<br></pre> |
| Overruling | <pre>Sentence: in light of both our holding today and previous rulings in johnson, dueser, and gronroos, we now explicitly overrule dupree....<br>ID: 0<br>Label: 2<br></pre> |
| Semiconductor Org Types | <pre>Paper title: 3Gb/s AC-coupled chip-to-chip communication using a low-swing pulse receiver...<br>Organization name: North Carolina State Univ.,Raleigh,NC,USA<br>ID: 0<br>Label: 3<br></pre> |
| Systematic Review Inclusion | <pre>Title: Prototyping and transforming facial textures for perception research...<br>Abstract: Wavelet based methods for prototyping facial textures for artificially transforming the age of facial images were described. Pro...<br>Authors: Tiddeman, B.; Burt, M.; Perrett, D.<br>Journal: IEEE Comput Graphics Appl<br>ID: 0<br>Label: 2</pre> |
| TAI Safety Research | <pre>Title: Malign generalization without internal search<br>Abstract Note: In my last post, I challenged the idea that inner alignment failures should be explained by appealing to agents which perform ex...<br>Url: https://www.alignmentforum.org/posts/ynt9TD6PrYw6iT49m/malign-generalization-without-internal-search...<br>Publication Year: 2020<br>Item Type: blogPost<br>Author: Barnett, Matthew<br>Publication Title: AI Alignment Forum<br>ID: 0<br>Label: 1</pre> |
| Terms Of Service | <pre>Sentence: Crowdtangle may change these terms of service, as described above, notwithstanding any provision to the contrary in any agreemen...<br>ID: 0<br>Label: 2<br></pre> |
| Tweet Eval Hate | <pre>Tweet: New to Twitter-- any men on here know what the process is to get #verified?...<br>ID: 0<br>Label: 2<br></pre> |
| Twitter Complaints | <pre>Tweet text: @HMRCcustomers No this is my first job<br>ID: 0<br>Label: 2</pre> |
### Data Fields
The ID field is used for indexing data points. It will be used to match your submissions with the true test labels, so you must include it in your submission. All other columns contain textual data. Some contain links and URLs to websites on the internet.
All output fields are designated with the "Label" column header. The 0 value in this column indicates that the entry is unlabeled, and should only appear in the unlabeled test set. Other values in this column are various other labels. To get their textual value for a given dataset:
```
# Load the dataset
dataset = datasets.load_dataset("ought/raft", "ade_corpus_v2")
# First, get the object that holds information about the "Label" feature in the dataset.
label_info = dataset.features["Label"]
# Use the int2str method to access the textual labels.
print([label_info.int2str(i) for i in (0, 1, 2)])
# ['Unlabeled', 'ADE-related', 'not ADE-related']
```
### Data Splits
There are two splits provided: train data and unlabeled test data.
The training examples were chosen at random. No attempt was made to ensure that classes were balanced or proportional in the training data -- indeed, the Banking 77 task with 77 different classes if used cannot fit all of its classes into the 50 training examples.
| Dataset | Train Size | Test Size | |
|--------------------------------|------------|-----------|---|
| Ade Corpus V2 | 50 | 5000 | |
| Banking 77 | 50 | 5000 | |
| NeurIPS Impact Statement Risks | 50 | 150 | |
| One Stop English | 50 | 516 | |
| Overruling | 50 | 2350 | |
| Semiconductor Org Types | 50 | 449 | |
| Systematic Review Inclusion | 50 | 2243 | |
| TAI Safety Research | 50 | 1639 | |
| Terms Of Service | 50 | 5000 | |
| Tweet Eval Hate | 50 | 2966 | |
| Twitter Complaints | 50 | 3399 | |
| **Total** | **550** | **28712** | |
## Dataset Creation
### Curation Rationale
Generally speaking, the rationale behind RAFT was to create a benchmark for evaluating NLP models that didn't consist of contrived or artificial data sources, for which the tasks weren't originally assembled for the purpose of testing NLP models. However, each individual dataset in RAFT was collected independently. For the majority of datasets, we only collected them second-hand from existing curated sources. The datasets that we curated are:
* NeurIPS impact statement risks
* Semiconductor org types
* TAI Safety Research
Each of these three datasets was sourced from our existing collaborators at Ought. They had used our service, Elicit, to analyze their dataset in the past, and we contact them to include their dataset and the associated classification task in the benchmark. For all datasets, more information is provided in our paper. For the ones which we did not curate, we provide a link to the dataset. For the ones which we did, we provide a datasheet that elaborates on many of the topics here in greater detail.
For the three datasets that we introduced:
* **NeurIPS impact statement risks** The dataset was created to evaluate the then new requirement for authors to include an "impact statement" in their 2020 NeurIPS papers. Had it been successful? What kind of things did authors mention the most? How long were impact statements on average? Etc.
* **Semiconductor org types** The dataset was originally created to understand better which countries’ organisations have contributed most to semiconductor R\&D over the past 25 years using three main conferences. Moreover, to estimate the share of academic and private sector contributions, the organisations were classified as “university”, “research institute” or “company”.
* **TAI Safety Research** The primary motivations for assembling this database were to: (1) Aid potential donors in assessing organizations focusing on TAI safety by collecting and analyzing their research output. (2) Assemble a comprehensive bibliographic database that can be used as a base for future projects, such as a living review of the field.
**For the following sections, we will only describe the datasets we introduce. All other dataset details, and more details on the ones described here, can be found in our paper.**
### Source Data
#### Initial Data Collection and Normalization
* **NeurIPS impact statement risks** The data was directly observable (raw text scraped) for the most part; although some data was taken from previous datasets (which themselves had taken it from raw text). The data was validated, but only in part, by human reviewers. Cf this link for full details:
* **Semiconductor org types** We used the IEEE API to obtain institutions that contributed papers to semiconductor conferences in the last 25 years. This is a random sample of 500 of them with a corresponding conference paper title. The three conferences were the International Solid-State Circuits Conference (ISSCC), the Symposia on VLSI Technology and Circuits (VLSI) and the International Electron Devices Meeting (IEDM).
* **TAI Safety Research** We asked TAI safety organizations for what their employees had written, emailed some individual authors, and searched Google Scholar. See the LessWrong post for more details: https://www.lesswrong.com/posts/4DegbDJJiMX2b3EKm/tai-safety-bibliographic-database
#### Who are the source language producers?
* **NeurIPS impact statement risks** Language generated from NeurIPS 2020 impact statement authors, generally the authors of submission papers.
* **Semiconductor org types** Language generated from IEEE API. Generally machine-formatted names, and title of academic papers.
* **TAI Safety Research** Language generated by authors of TAI safety research publications.
### Annotations
#### Annotation process
* **NeurIPS impact statement risks** Annotations were entered directly into a Google Spreadsheet with instructions, labeled training examples, and unlabeled testing examples.
* **Semiconductor org types** Annotations were entered directly into a Google Spreadsheet with instructions, labeled training examples, and unlabeled testing examples.
* **TAI Safety Research** N/A
#### Who are the annotators?
* **NeurIPS impact statement risks** Contractors paid by Ought performed the labeling of whether impact statements mention harmful applications. A majority vote was taken from 3 annotators.
* **Semiconductor org types** Contractors paid by Ought performed the labeling of organization types. A majority vote was taken from 3 annotators.
* **TAI Safety Research** The dataset curators annotated the dataset by hand.
### Personal and Sensitive Information
It is worth mentioning that the Tweet Eval Hate, by necessity, contains highly offensive content.
* **NeurIPS impact statement risks** The dataset contains authors' names. These were scraped from publicly available scientific papers submitted to NeurIPS 2020.
* **Semiconductor org types** N/A
* **TAI Safety Research** N/A
## Considerations for Using the Data
### Social Impact of Dataset
* **NeurIPS impact statement risks** N/A
* **Semiconductor org types** N/A
* **TAI Safety Research** N/A
### Discussion of Biases
* **NeurIPS impact statement risks** N/A
* **Semiconductor org types** N/A
* **TAI Safety Research** N/A
### Other Known Limitations
* **NeurIPS impact statement risks** This dataset has limitations that should be taken into consideration when using it. In particular, the method used to collect broader impact statements involved automated downloads, conversions and scraping and was not error-proof. Although care has been taken to identify and correct as many errors as possible, not all texts have been reviewed by a human. This means it is possible some of the broader impact statements contained in the dataset are truncated or otherwise incorrectly extracted from their original article.
* **Semiconductor org types** N/A
* **TAI Safety Research** Don't use it to create a dangerous AI that could bring the end of days.
## Additional Information
### Dataset Curators
The overall RAFT curators are Neel Alex, Eli Lifland, and Andreas Stuhlmüller.
* **NeurIPS impact statement risks** Volunteers working with researchers affiliated to Oxford's Future of Humanity Institute (Carolyn Ashurst, now at The Alan Turing Institute) created the impact statements dataset.
* **Semiconductor org types** The data science unit of Stiftung Neue Verantwortung (Berlin).
* **TAI Safety Research** Angelica Deibel and Jess Riedel. We did not do it on behalf of any entity.
### Licensing Information
RAFT aggregates many other datasets, each of which is provided under its own license. Generally, those licenses permit research and commercial use.
| Dataset | License |
| ----------- | ----------- |
| Ade Corpus V2 | Unlicensed |
| Banking 77 | CC BY 4.0 |
| NeurIPS Impact Statement Risks | MIT License/CC BY 4.0 |
| One Stop English | CC BY-SA 4.0 |
| Overruling | Unlicensed |
| Semiconductor Org Types | CC BY-NC 4.0 |
| Systematic Review Inclusion | CC BY 4.0 |
| TAI Safety Research | CC BY-SA 4.0 |
| Terms Of Service | Unlicensed |
| Tweet Eval Hate | Unlicensed |
| Twitter Complaints | Unlicensed |
### Citation Information
[More Information Needed]
### Contributions
Thanks to [@neel-alex](https://github.com/neel-alex), [@uvafan](https://github.com/uvafan), and [@lewtun](https://github.com/lewtun) for adding this dataset. |
math_qa | 2023-04-05T10:09:35.000Z | [
"task_categories:question-answering",
"task_ids:multiple-choice-qa",
"annotations_creators:crowdsourced",
"language_creators:crowdsourced",
"language_creators:expert-generated",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:extended|aqua_rat",
"language:en",
"license:apache-2.0",
"region:us"
] | null | Our dataset is gathered by using a new representation language to annotate over the AQuA-RAT dataset. AQuA-RAT has provided the questions, options, rationale, and the correct options. | null | 38 | 8,911 | ---
annotations_creators:
- crowdsourced
language:
- en
language_creators:
- crowdsourced
- expert-generated
license:
- apache-2.0
multilinguality:
- monolingual
pretty_name: MathQA
size_categories:
- 10K<n<100K
source_datasets:
- extended|aqua_rat
task_categories:
- question-answering
task_ids:
- multiple-choice-qa
paperswithcode_id: mathqa
dataset_info:
features:
- name: Problem
dtype: string
- name: Rationale
dtype: string
- name: options
dtype: string
- name: correct
dtype: string
- name: annotated_formula
dtype: string
- name: linear_formula
dtype: string
- name: category
dtype: string
splits:
- name: test
num_bytes: 1844184
num_examples: 2985
- name: train
num_bytes: 18368826
num_examples: 29837
- name: validation
num_bytes: 2752969
num_examples: 4475
download_size: 7302821
dataset_size: 22965979
---
# Dataset Card for MathQA
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://math-qa.github.io/math-QA/](https://math-qa.github.io/math-QA/)
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Paper:** [MathQA: Towards Interpretable Math Word Problem Solving with Operation-Based Formalisms](https://aclanthology.org/N19-1245/)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 7.30 MB
- **Size of the generated dataset:** 22.96 MB
- **Total amount of disk used:** 30.27 MB
### Dataset Summary
We introduce a large-scale dataset of math word problems.
Our dataset is gathered by using a new representation language to annotate over the AQuA-RAT dataset with fully-specified operational programs.
AQuA-RAT has provided the questions, options, rationale, and the correct options.
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Dataset Structure
### Data Instances
#### default
- **Size of downloaded dataset files:** 7.30 MB
- **Size of the generated dataset:** 22.96 MB
- **Total amount of disk used:** 30.27 MB
An example of 'train' looks as follows.
```
{
"Problem": "a multiple choice test consists of 4 questions , and each question has 5 answer choices . in how many r ways can the test be completed if every question is unanswered ?",
"Rationale": "\"5 choices for each of the 4 questions , thus total r of 5 * 5 * 5 * 5 = 5 ^ 4 = 625 ways to answer all of them . answer : c .\"",
"annotated_formula": "power(5, 4)",
"category": "general",
"correct": "c",
"linear_formula": "power(n1,n0)|",
"options": "a ) 24 , b ) 120 , c ) 625 , d ) 720 , e ) 1024"
}
```
### Data Fields
The data fields are the same among all splits.
#### default
- `Problem`: a `string` feature.
- `Rationale`: a `string` feature.
- `options`: a `string` feature.
- `correct`: a `string` feature.
- `annotated_formula`: a `string` feature.
- `linear_formula`: a `string` feature.
- `category`: a `string` feature.
### Data Splits
| name |train|validation|test|
|-------|----:|---------:|---:|
|default|29837| 4475|2985|
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
The dataset is licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0).
### Citation Information
```
@inproceedings{amini-etal-2019-mathqa,
title = "{M}ath{QA}: Towards Interpretable Math Word Problem Solving with Operation-Based Formalisms",
author = "Amini, Aida and
Gabriel, Saadia and
Lin, Shanchuan and
Koncel-Kedziorski, Rik and
Choi, Yejin and
Hajishirzi, Hannaneh",
booktitle = "Proceedings of the 2019 Conference of the North {A}merican Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)",
month = jun,
year = "2019",
address = "Minneapolis, Minnesota",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/N19-1245",
doi = "10.18653/v1/N19-1245",
pages = "2357--2367",
}
```
### Contributions
Thanks to [@thomwolf](https://github.com/thomwolf), [@lewtun](https://github.com/lewtun), [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset. | |
SetFit/emotion | 2022-04-03T20:47:37.000Z | [
"region:us"
] | SetFit | null | null | null | 11 | 8,885 | ** Attention: There appears an overlap in train / test. I trained a model on the train set and achieved 100% acc on test set. With the original emotion dataset this is not the case (92.4% acc)** |
lighteval/mmlu | 2023-06-09T16:36:19.000Z | [
"task_categories:question-answering",
"task_ids:multiple-choice-qa",
"annotations_creators:no-annotation",
"language_creators:expert-generated",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:mit",
"arxiv:2009.03300",
"arxiv:2005.00700",
"arxiv:2005.14165",
"arxiv:2008.02275",
"region:us"
] | lighteval | This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more. | @article{hendryckstest2021,
title={Measuring Massive Multitask Language Understanding},
author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},
journal={Proceedings of the International Conference on Learning Representations (ICLR)},
year={2021}
} | null | 6 | 8,800 | ---
annotations_creators:
- no-annotation
language_creators:
- expert-generated
language:
- en
license:
- mit
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- question-answering
task_ids:
- multiple-choice-qa
paperswithcode_id: mmlu
pretty_name: Measuring Massive Multitask Language Understanding
language_bcp47:
- en-US
dataset_info:
- config_name: abstract_algebra
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 19328
num_examples: 100
- name: validation
num_bytes: 2024
num_examples: 11
- name: dev
num_bytes: 830
num_examples: 5
download_size: 166184960
dataset_size: 160623559
- config_name: anatomy
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 33121
num_examples: 135
- name: validation
num_bytes: 3140
num_examples: 14
- name: dev
num_bytes: 967
num_examples: 5
download_size: 166184960
dataset_size: 160638605
- config_name: astronomy
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 46771
num_examples: 152
- name: validation
num_bytes: 5027
num_examples: 16
- name: dev
num_bytes: 2076
num_examples: 5
download_size: 166184960
dataset_size: 160655251
- config_name: business_ethics
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 33252
num_examples: 100
- name: validation
num_bytes: 3038
num_examples: 11
- name: dev
num_bytes: 2190
num_examples: 5
download_size: 166184960
dataset_size: 160639857
- config_name: clinical_knowledge
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 62754
num_examples: 265
- name: validation
num_bytes: 6664
num_examples: 29
- name: dev
num_bytes: 1210
num_examples: 5
download_size: 166184960
dataset_size: 160672005
- config_name: college_biology
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 48797
num_examples: 144
- name: validation
num_bytes: 4819
num_examples: 16
- name: dev
num_bytes: 1532
num_examples: 5
download_size: 166184960
dataset_size: 160656525
- config_name: college_chemistry
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 24708
num_examples: 100
- name: validation
num_bytes: 2328
num_examples: 8
- name: dev
num_bytes: 1331
num_examples: 5
download_size: 166184960
dataset_size: 160629744
- config_name: college_computer_science
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 42641
num_examples: 100
- name: validation
num_bytes: 4663
num_examples: 11
- name: dev
num_bytes: 2765
num_examples: 5
download_size: 166184960
dataset_size: 160651446
- config_name: college_mathematics
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 24711
num_examples: 100
- name: validation
num_bytes: 2668
num_examples: 11
- name: dev
num_bytes: 1493
num_examples: 5
download_size: 166184960
dataset_size: 160630249
- config_name: college_medicine
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 82397
num_examples: 173
- name: validation
num_bytes: 7909
num_examples: 22
- name: dev
num_bytes: 1670
num_examples: 5
download_size: 166184960
dataset_size: 160693353
- config_name: college_physics
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 30181
num_examples: 102
- name: validation
num_bytes: 3490
num_examples: 11
- name: dev
num_bytes: 1412
num_examples: 5
download_size: 166184960
dataset_size: 160636460
- config_name: computer_security
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 27124
num_examples: 100
- name: validation
num_bytes: 4549
num_examples: 11
- name: dev
num_bytes: 1101
num_examples: 5
download_size: 166184960
dataset_size: 160634151
- config_name: conceptual_physics
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 40709
num_examples: 235
- name: validation
num_bytes: 4474
num_examples: 26
- name: dev
num_bytes: 934
num_examples: 5
download_size: 166184960
dataset_size: 160647494
- config_name: econometrics
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 46547
num_examples: 114
- name: validation
num_bytes: 4967
num_examples: 12
- name: dev
num_bytes: 1644
num_examples: 5
download_size: 166184960
dataset_size: 160654535
- config_name: electrical_engineering
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 25142
num_examples: 145
- name: validation
num_bytes: 2903
num_examples: 16
- name: dev
num_bytes: 972
num_examples: 5
download_size: 166184960
dataset_size: 160630394
- config_name: elementary_mathematics
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 70108
num_examples: 378
- name: validation
num_bytes: 8988
num_examples: 41
- name: dev
num_bytes: 1440
num_examples: 5
download_size: 166184960
dataset_size: 160681913
- config_name: formal_logic
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 49785
num_examples: 126
- name: validation
num_bytes: 6252
num_examples: 14
- name: dev
num_bytes: 1757
num_examples: 5
download_size: 166184960
dataset_size: 160659171
- config_name: global_facts
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 18403
num_examples: 100
- name: validation
num_bytes: 1865
num_examples: 10
- name: dev
num_bytes: 1229
num_examples: 5
download_size: 166184960
dataset_size: 160622874
- config_name: high_school_biology
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 109732
num_examples: 310
- name: validation
num_bytes: 11022
num_examples: 32
- name: dev
num_bytes: 1673
num_examples: 5
download_size: 166184960
dataset_size: 160723804
- config_name: high_school_chemistry
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 58464
num_examples: 203
- name: validation
num_bytes: 7092
num_examples: 22
- name: dev
num_bytes: 1220
num_examples: 5
download_size: 166184960
dataset_size: 160668153
- config_name: high_school_computer_science
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 44476
num_examples: 100
- name: validation
num_bytes: 3343
num_examples: 9
- name: dev
num_bytes: 2918
num_examples: 5
download_size: 166184960
dataset_size: 160652114
- config_name: high_school_european_history
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 270300
num_examples: 165
- name: validation
num_bytes: 29632
num_examples: 18
- name: dev
num_bytes: 11564
num_examples: 5
download_size: 166184960
dataset_size: 160912873
- config_name: high_school_geography
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 42034
num_examples: 198
- name: validation
num_bytes: 4332
num_examples: 22
- name: dev
num_bytes: 1403
num_examples: 5
download_size: 166184960
dataset_size: 160649146
- config_name: high_school_government_and_politics
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 66074
num_examples: 193
- name: validation
num_bytes: 7063
num_examples: 21
- name: dev
num_bytes: 1779
num_examples: 5
download_size: 166184960
dataset_size: 160676293
- config_name: high_school_macroeconomics
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 117687
num_examples: 390
- name: validation
num_bytes: 13020
num_examples: 43
- name: dev
num_bytes: 1328
num_examples: 5
download_size: 166184960
dataset_size: 160733412
- config_name: high_school_mathematics
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 54854
num_examples: 270
- name: validation
num_bytes: 5765
num_examples: 29
- name: dev
num_bytes: 1297
num_examples: 5
download_size: 166184960
dataset_size: 160663293
- config_name: high_school_microeconomics
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 75703
num_examples: 238
- name: validation
num_bytes: 7553
num_examples: 26
- name: dev
num_bytes: 1298
num_examples: 5
download_size: 166184960
dataset_size: 160685931
- config_name: high_school_physics
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 59538
num_examples: 151
- name: validation
num_bytes: 6771
num_examples: 17
- name: dev
num_bytes: 1489
num_examples: 5
download_size: 166184960
dataset_size: 160669175
- config_name: high_school_psychology
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 159407
num_examples: 545
- name: validation
num_bytes: 17269
num_examples: 60
- name: dev
num_bytes: 1905
num_examples: 5
download_size: 166184960
dataset_size: 160779958
- config_name: high_school_statistics
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 110702
num_examples: 216
- name: validation
num_bytes: 9997
num_examples: 23
- name: dev
num_bytes: 2528
num_examples: 5
download_size: 166184960
dataset_size: 160724604
- config_name: high_school_us_history
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 296734
num_examples: 204
- name: validation
num_bytes: 31706
num_examples: 22
- name: dev
num_bytes: 8864
num_examples: 5
download_size: 166184960
dataset_size: 160938681
- config_name: high_school_world_history
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 378617
num_examples: 237
- name: validation
num_bytes: 45501
num_examples: 26
- name: dev
num_bytes: 4882
num_examples: 5
download_size: 166184960
dataset_size: 161030377
- config_name: human_aging
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 46098
num_examples: 223
- name: validation
num_bytes: 4707
num_examples: 23
- name: dev
num_bytes: 1008
num_examples: 5
download_size: 166184960
dataset_size: 160653190
- config_name: human_sexuality
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 32110
num_examples: 131
- name: validation
num_bytes: 2421
num_examples: 12
- name: dev
num_bytes: 1077
num_examples: 5
download_size: 166184960
dataset_size: 160636985
- config_name: international_law
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 53531
num_examples: 121
- name: validation
num_bytes: 6473
num_examples: 13
- name: dev
num_bytes: 2418
num_examples: 5
download_size: 166184960
dataset_size: 160663799
- config_name: jurisprudence
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 33986
num_examples: 108
- name: validation
num_bytes: 3729
num_examples: 11
- name: dev
num_bytes: 1303
num_examples: 5
download_size: 166184960
dataset_size: 160640395
- config_name: logical_fallacies
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 50117
num_examples: 163
- name: validation
num_bytes: 5103
num_examples: 18
- name: dev
num_bytes: 1573
num_examples: 5
download_size: 166184960
dataset_size: 160658170
- config_name: machine_learning
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 33880
num_examples: 112
- name: validation
num_bytes: 3232
num_examples: 11
- name: dev
num_bytes: 2323
num_examples: 5
download_size: 166184960
dataset_size: 160640812
- config_name: management
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 20002
num_examples: 103
- name: validation
num_bytes: 1820
num_examples: 11
- name: dev
num_bytes: 898
num_examples: 5
download_size: 166184960
dataset_size: 160624097
- config_name: marketing
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 63025
num_examples: 234
- name: validation
num_bytes: 7394
num_examples: 25
- name: dev
num_bytes: 1481
num_examples: 5
download_size: 166184960
dataset_size: 160673277
- config_name: medical_genetics
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 20864
num_examples: 100
- name: validation
num_bytes: 3005
num_examples: 11
- name: dev
num_bytes: 1089
num_examples: 5
download_size: 166184960
dataset_size: 160626335
- config_name: miscellaneous
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 147704
num_examples: 783
- name: validation
num_bytes: 14330
num_examples: 86
- name: dev
num_bytes: 699
num_examples: 5
download_size: 166184960
dataset_size: 160764110
- config_name: moral_disputes
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 107818
num_examples: 346
- name: validation
num_bytes: 12420
num_examples: 38
- name: dev
num_bytes: 1755
num_examples: 5
download_size: 166184960
dataset_size: 160723370
- config_name: moral_scenarios
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 374026
num_examples: 895
- name: validation
num_bytes: 42338
num_examples: 100
- name: dev
num_bytes: 2058
num_examples: 5
download_size: 166184960
dataset_size: 161019799
- config_name: nutrition
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 92410
num_examples: 306
- name: validation
num_bytes: 8436
num_examples: 33
- name: dev
num_bytes: 2085
num_examples: 5
download_size: 166184960
dataset_size: 160704308
- config_name: philosophy
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 80073
num_examples: 311
- name: validation
num_bytes: 9184
num_examples: 34
- name: dev
num_bytes: 988
num_examples: 5
download_size: 166184960
dataset_size: 160691622
- config_name: prehistory
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 89594
num_examples: 324
- name: validation
num_bytes: 10285
num_examples: 35
- name: dev
num_bytes: 1878
num_examples: 5
download_size: 166184960
dataset_size: 160703134
- config_name: professional_accounting
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 124550
num_examples: 282
- name: validation
num_bytes: 14372
num_examples: 31
- name: dev
num_bytes: 2148
num_examples: 5
download_size: 166184960
dataset_size: 160742447
- config_name: professional_law
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 1891762
num_examples: 1534
- name: validation
num_bytes: 203519
num_examples: 170
- name: dev
num_bytes: 6610
num_examples: 5
download_size: 166184960
dataset_size: 162703268
- config_name: professional_medicine
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 217561
num_examples: 272
- name: validation
num_bytes: 23847
num_examples: 31
- name: dev
num_bytes: 3807
num_examples: 5
download_size: 166184960
dataset_size: 160846592
- config_name: professional_psychology
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 225899
num_examples: 612
- name: validation
num_bytes: 29101
num_examples: 69
- name: dev
num_bytes: 2267
num_examples: 5
download_size: 166184960
dataset_size: 160858644
- config_name: public_relations
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 28760
num_examples: 110
- name: validation
num_bytes: 4566
num_examples: 12
- name: dev
num_bytes: 1496
num_examples: 5
download_size: 166184960
dataset_size: 160636199
- config_name: security_studies
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 204844
num_examples: 245
- name: validation
num_bytes: 22637
num_examples: 27
- name: dev
num_bytes: 5335
num_examples: 5
download_size: 166184960
dataset_size: 160834193
- config_name: sociology
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 66243
num_examples: 201
- name: validation
num_bytes: 7184
num_examples: 22
- name: dev
num_bytes: 1613
num_examples: 5
download_size: 166184960
dataset_size: 160676417
- config_name: us_foreign_policy
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 28443
num_examples: 100
- name: validation
num_bytes: 3264
num_examples: 11
- name: dev
num_bytes: 1611
num_examples: 5
download_size: 166184960
dataset_size: 160634695
- config_name: virology
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 38759
num_examples: 166
- name: validation
num_bytes: 5463
num_examples: 18
- name: dev
num_bytes: 1096
num_examples: 5
download_size: 166184960
dataset_size: 160646695
- config_name: world_religions
features:
- name: question
dtype: string
- name: choices
sequence: string
- name: answer
dtype:
class_label:
names:
'0': A
'1': B
'2': C
'3': D
splits:
- name: auxiliary_train
num_bytes: 160601377
num_examples: 99842
- name: test
num_bytes: 25274
num_examples: 171
- name: validation
num_bytes: 2765
num_examples: 19
- name: dev
num_bytes: 670
num_examples: 5
download_size: 166184960
dataset_size: 160630086
---
# Dataset Card for MMLU
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Repository**: https://github.com/hendrycks/test
- **Paper**: https://arxiv.org/abs/2009.03300
### Dataset Summary
[Measuring Massive Multitask Language Understanding](https://arxiv.org/pdf/2009.03300) by [Dan Hendrycks](https://people.eecs.berkeley.edu/~hendrycks/), [Collin Burns](http://collinpburns.com), [Steven Basart](https://stevenbas.art), Andy Zou, Mantas Mazeika, [Dawn Song](https://people.eecs.berkeley.edu/~dawnsong/), and [Jacob Steinhardt](https://www.stat.berkeley.edu/~jsteinhardt/) (ICLR 2021).
This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge. The test spans subjects in the humanities, social sciences, hard sciences, and other areas that are important for some people to learn. This covers 57 tasks including elementary mathematics, US history, computer science, law, and more. To attain high accuracy on this test, models must possess extensive world knowledge and problem solving ability.
A complete list of tasks: ['abstract_algebra', 'anatomy', 'astronomy', 'business_ethics', 'clinical_knowledge', 'college_biology', 'college_chemistry', 'college_computer_science', 'college_mathematics', 'college_medicine', 'college_physics', 'computer_security', 'conceptual_physics', 'econometrics', 'electrical_engineering', 'elementary_mathematics', 'formal_logic', 'global_facts', 'high_school_biology', 'high_school_chemistry', 'high_school_computer_science', 'high_school_european_history', 'high_school_geography', 'high_school_government_and_politics', 'high_school_macroeconomics', 'high_school_mathematics', 'high_school_microeconomics', 'high_school_physics', 'high_school_psychology', 'high_school_statistics', 'high_school_us_history', 'high_school_world_history', 'human_aging', 'human_sexuality', 'international_law', 'jurisprudence', 'logical_fallacies', 'machine_learning', 'management', 'marketing', 'medical_genetics', 'miscellaneous', 'moral_disputes', 'moral_scenarios', 'nutrition', 'philosophy', 'prehistory', 'professional_accounting', 'professional_law', 'professional_medicine', 'professional_psychology', 'public_relations', 'security_studies', 'sociology', 'us_foreign_policy', 'virology', 'world_religions']
### Supported Tasks and Leaderboards
| Model | Authors | Humanities | Social Science | STEM | Other | Average |
|------------------------------------|----------|:-------:|:-------:|:-------:|:-------:|:-------:|
| [UnifiedQA](https://arxiv.org/abs/2005.00700) | Khashabi et al., 2020 | 45.6 | 56.6 | 40.2 | 54.6 | 48.9
| [GPT-3](https://arxiv.org/abs/2005.14165) (few-shot) | Brown et al., 2020 | 40.8 | 50.4 | 36.7 | 48.8 | 43.9
| [GPT-2](https://arxiv.org/abs/2005.14165) | Radford et al., 2019 | 32.8 | 33.3 | 30.2 | 33.1 | 32.4
| Random Baseline | N/A | 25.0 | 25.0 | 25.0 | 25.0 | 25.0 | 25.0
### Languages
English
## Dataset Structure
### Data Instances
An example from anatomy subtask looks as follows:
```
{
"question": "What is the embryological origin of the hyoid bone?",
"choices": ["The first pharyngeal arch", "The first and second pharyngeal arches", "The second pharyngeal arch", "The second and third pharyngeal arches"],
"answer": "D"
}
```
### Data Fields
- `question`: a string feature
- `choices`: a list of 4 string features
- `answer`: a ClassLabel feature
### Data Splits
- `auxiliary_train`: auxiliary multiple-choice training questions from ARC, MC_TEST, OBQA, RACE, etc.
- `dev`: 5 examples per subtask, meant for few-shot setting
- `test`: there are at least 100 examples per subtask
| | auxiliary_train | dev | val | test |
| ----- | :------: | :-----: | :-----: | :-----: |
| TOTAL | 99842 | 285 | 1531 | 14042
## Dataset Creation
### Curation Rationale
Transformer models have driven this recent progress by pretraining on massive text corpora, including all of Wikipedia, thousands of books, and numerous websites. These models consequently see extensive information about specialized topics, most of which is not assessed by existing NLP benchmarks. To bridge the gap between the wide-ranging knowledge that models see during pretraining and the existing measures of success, we introduce a new benchmark for assessing models across a diverse set of subjects that humans learn.
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[MIT License](https://github.com/hendrycks/test/blob/master/LICENSE)
### Citation Information
If you find this useful in your research, please consider citing the test and also the [ETHICS](https://arxiv.org/abs/2008.02275) dataset it draws from:
```
@article{hendryckstest2021,
title={Measuring Massive Multitask Language Understanding},
author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},
journal={Proceedings of the International Conference on Learning Representations (ICLR)},
year={2021}
}
@article{hendrycks2021ethics,
title={Aligning AI With Shared Human Values},
author={Dan Hendrycks and Collin Burns and Steven Basart and Andrew Critch and Jerry Li and Dawn Song and Jacob Steinhardt},
journal={Proceedings of the International Conference on Learning Representations (ICLR)},
year={2021}
}
```
### Contributions
Thanks to [@andyzoujm](https://github.com/andyzoujm) for adding this dataset.
|
mteb/tatoeba-bitext-mining | 2022-09-27T19:07:02.000Z | [
"language:eng",
"language:sqi",
"language:fry",
"language:kur",
"language:tur",
"language:deu",
"language:nld",
"language:ron",
"language:ang",
"language:ido",
"language:jav",
"language:isl",
"language:slv",
"language:cym",
"language:kaz",
"language:est",
"language:heb",
"language:gla",
"language:mar",
"language:lat",
"language:bel",
"language:pms",
"language:gle",
"language:pes",
"language:nob",
"language:bul",
"language:cbk",
"language:hun",
"language:uig",
"language:rus",
"language:spa",
"language:hye",
"language:tel",
"language:afr",
"language:mon",
"language:arz",
"language:hrv",
"language:nov",
"language:gsw",
"language:nds",
"language:ukr",
"language:uzb",
"language:lit",
"language:ina",
"language:lfn",
"language:zsm",
"language:ita",
"language:cmn",
"language:lvs",
"language:glg",
"language:ceb",
"language:bre",
"language:ben",
"language:swg",
"language:arq",
"language:kab",
"language:fra",
"language:por",
"language:tat",
"language:oci",
"language:pol",
"language:war",
"language:aze",
"language:vie",
"language:nno",
"language:cha",
"language:mhr",
"language:dan",
"language:ell",
"language:amh",
"language:pam",
"language:hsb",
"language:srp",
"language:epo",
"language:kzj",
"language:awa",
"language:fao",
"language:mal",
"language:ile",
"language:bos",
"language:cor",
"language:cat",
"language:eus",
"language:yue",
"language:swe",
"language:dtp",
"language:kat",
"language:jpn",
"language:csb",
"language:xho",
"language:orv",
"language:ind",
"language:tuk",
"language:max",
"language:swh",
"language:hin",
"language:dsb",
"language:ber",
"language:tam",
"language:slk",
"language:tgl",
"language:ast",
"language:mkd",
"language:khm",
"language:ces",
"language:tzl",
"language:urd",
"language:ara",
"language:kor",
"language:yid",
"language:fin",
"language:tha",
"language:wuu",
"region:us"
] | mteb | Tatoeba multilingual test set | null | null | 3 | 8,762 | ---
language:
- eng
- sqi
- fry
- kur
- tur
- deu
- nld
- ron
- ang
- ido
- jav
- isl
- slv
- cym
- kaz
- est
- heb
- gla
- mar
- lat
- bel
- pms
- gle
- pes
- nob
- bul
- cbk
- hun
- uig
- rus
- spa
- hye
- tel
- afr
- mon
- arz
- hrv
- nov
- gsw
- nds
- ukr
- uzb
- lit
- ina
- lfn
- zsm
- ita
- cmn
- lvs
- glg
- ceb
- bre
- ben
- swg
- arq
- kab
- fra
- por
- tat
- oci
- pol
- war
- aze
- vie
- nno
- cha
- mhr
- dan
- ell
- amh
- pam
- hsb
- srp
- epo
- kzj
- awa
- fao
- mal
- ile
- bos
- cor
- cat
- eus
- yue
- swe
- dtp
- kat
- jpn
- csb
- xho
- orv
- ind
- tuk
- max
- swh
- hin
- dsb
- ber
- tam
- slk
- tgl
- ast
- mkd
- khm
- ces
- tzl
- urd
- ara
- kor
- yid
- fin
- tha
- wuu
--- |
juletxara/xstory_cloze | 2023-05-21T16:04:36.000Z | [
"task_categories:other",
"annotations_creators:found",
"language_creators:found",
"language_creators:expert-generated",
"multilinguality:multilingual",
"size_categories:1K<n<10K",
"source_datasets:extended|story_cloze",
"language:en",
"language:ru",
"language:zh",
"language:es",
"language:ar",
"language:hi",
"language:id",
"language:te",
"language:sw",
"language:eu",
"language:my",
"license:cc-by-sa-4.0",
"arxiv:2112.10668",
"region:us"
] | juletxara | XStoryCloze consists of the professionally translated version of the [English StoryCloze dataset](https://cs.rochester.edu/nlp/rocstories/) (Spring 2016 version) to 10 non-English languages. This dataset is released by Meta AI. | @article{DBLP:journals/corr/abs-2112-10668,
author = {Xi Victoria Lin and
Todor Mihaylov and
Mikel Artetxe and
Tianlu Wang and
Shuohui Chen and
Daniel Simig and
Myle Ott and
Naman Goyal and
Shruti Bhosale and
Jingfei Du and
Ramakanth Pasunuru and
Sam Shleifer and
Punit Singh Koura and
Vishrav Chaudhary and
Brian O'Horo and
Jeff Wang and
Luke Zettlemoyer and
Zornitsa Kozareva and
Mona T. Diab and
Veselin Stoyanov and
Xian Li},
title = {Few-shot Learning with Multilingual Language Models},
journal = {CoRR},
volume = {abs/2112.10668},
year = {2021},
url = {https://arxiv.org/abs/2112.10668},
eprinttype = {arXiv},
eprint = {2112.10668},
timestamp = {Tue, 04 Jan 2022 15:59:27 +0100},
biburl = {https://dblp.org/rec/journals/corr/abs-2112-10668.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
} | null | 3 | 8,732 | ---
annotations_creators:
- found
language:
- en
- ru
- zh
- es
- ar
- hi
- id
- te
- sw
- eu
- my
language_creators:
- found
- expert-generated
license:
- cc-by-sa-4.0
multilinguality:
- multilingual
paperswithcode_id: null
pretty_name: XStoryCloze
size_categories:
- 1K<n<10K
source_datasets:
- extended|story_cloze
tags: []
task_categories:
- other
task_ids: []
dataset_info:
- config_name: en
features:
- name: story_id
dtype: string
- name: input_sentence_1
dtype: string
- name: input_sentence_2
dtype: string
- name: input_sentence_3
dtype: string
- name: input_sentence_4
dtype: string
- name: sentence_quiz1
dtype: string
- name: sentence_quiz2
dtype: string
- name: answer_right_ending
dtype: int32
splits:
- name: train
num_bytes: 118484
num_examples: 360
- name: eval
num_bytes: 495572
num_examples: 1511
download_size: 573176
dataset_size: 614056
- config_name: ru
features:
- name: story_id
dtype: string
- name: input_sentence_1
dtype: string
- name: input_sentence_2
dtype: string
- name: input_sentence_3
dtype: string
- name: input_sentence_4
dtype: string
- name: sentence_quiz1
dtype: string
- name: sentence_quiz2
dtype: string
- name: answer_right_ending
dtype: int32
splits:
- name: train
num_bytes: 202961
num_examples: 360
- name: eval
num_bytes: 846720
num_examples: 1511
download_size: 1008802
dataset_size: 1049681
- config_name: zh
features:
- name: story_id
dtype: string
- name: input_sentence_1
dtype: string
- name: input_sentence_2
dtype: string
- name: input_sentence_3
dtype: string
- name: input_sentence_4
dtype: string
- name: sentence_quiz1
dtype: string
- name: sentence_quiz2
dtype: string
- name: answer_right_ending
dtype: int32
splits:
- name: train
num_bytes: 116624
num_examples: 360
- name: eval
num_bytes: 487501
num_examples: 1511
download_size: 563245
dataset_size: 604125
- config_name: es
features:
- name: story_id
dtype: string
- name: input_sentence_1
dtype: string
- name: input_sentence_2
dtype: string
- name: input_sentence_3
dtype: string
- name: input_sentence_4
dtype: string
- name: sentence_quiz1
dtype: string
- name: sentence_quiz2
dtype: string
- name: answer_right_ending
dtype: int32
splits:
- name: train
num_bytes: 129011
num_examples: 360
- name: eval
num_bytes: 538761
num_examples: 1511
download_size: 626890
dataset_size: 667772
- config_name: ar
features:
- name: story_id
dtype: string
- name: input_sentence_1
dtype: string
- name: input_sentence_2
dtype: string
- name: input_sentence_3
dtype: string
- name: input_sentence_4
dtype: string
- name: sentence_quiz1
dtype: string
- name: sentence_quiz2
dtype: string
- name: answer_right_ending
dtype: int32
splits:
- name: train
num_bytes: 171008
num_examples: 360
- name: eval
num_bytes: 701607
num_examples: 1511
download_size: 831735
dataset_size: 872615
- config_name: hi
features:
- name: story_id
dtype: string
- name: input_sentence_1
dtype: string
- name: input_sentence_2
dtype: string
- name: input_sentence_3
dtype: string
- name: input_sentence_4
dtype: string
- name: sentence_quiz1
dtype: string
- name: sentence_quiz2
dtype: string
- name: answer_right_ending
dtype: int32
splits:
- name: train
num_bytes: 277601
num_examples: 360
- name: eval
num_bytes: 1165329
num_examples: 1511
download_size: 1402058
dataset_size: 1442930
- config_name: id
features:
- name: story_id
dtype: string
- name: input_sentence_1
dtype: string
- name: input_sentence_2
dtype: string
- name: input_sentence_3
dtype: string
- name: input_sentence_4
dtype: string
- name: sentence_quiz1
dtype: string
- name: sentence_quiz2
dtype: string
- name: answer_right_ending
dtype: int32
splits:
- name: train
num_bytes: 131364
num_examples: 360
- name: eval
num_bytes: 549171
num_examples: 1511
download_size: 639655
dataset_size: 680535
- config_name: te
features:
- name: story_id
dtype: string
- name: input_sentence_1
dtype: string
- name: input_sentence_2
dtype: string
- name: input_sentence_3
dtype: string
- name: input_sentence_4
dtype: string
- name: sentence_quiz1
dtype: string
- name: sentence_quiz2
dtype: string
- name: answer_right_ending
dtype: int32
splits:
- name: train
num_bytes: 293917
num_examples: 360
- name: eval
num_bytes: 1215543
num_examples: 1511
download_size: 1468580
dataset_size: 1509460
- config_name: sw
features:
- name: story_id
dtype: string
- name: input_sentence_1
dtype: string
- name: input_sentence_2
dtype: string
- name: input_sentence_3
dtype: string
- name: input_sentence_4
dtype: string
- name: sentence_quiz1
dtype: string
- name: sentence_quiz2
dtype: string
- name: answer_right_ending
dtype: int32
splits:
- name: train
num_bytes: 127745
num_examples: 360
- name: eval
num_bytes: 528398
num_examples: 1511
download_size: 615263
dataset_size: 656143
- config_name: eu
features:
- name: story_id
dtype: string
- name: input_sentence_1
dtype: string
- name: input_sentence_2
dtype: string
- name: input_sentence_3
dtype: string
- name: input_sentence_4
dtype: string
- name: sentence_quiz1
dtype: string
- name: sentence_quiz2
dtype: string
- name: answer_right_ending
dtype: int32
splits:
- name: train
num_bytes: 131524
num_examples: 360
- name: eval
num_bytes: 540582
num_examples: 1511
download_size: 631228
dataset_size: 672106
- config_name: my
features:
- name: story_id
dtype: string
- name: input_sentence_1
dtype: string
- name: input_sentence_2
dtype: string
- name: input_sentence_3
dtype: string
- name: input_sentence_4
dtype: string
- name: sentence_quiz1
dtype: string
- name: sentence_quiz2
dtype: string
- name: answer_right_ending
dtype: int32
splits:
- name: train
num_bytes: 381092
num_examples: 360
- name: eval
num_bytes: 1627322
num_examples: 1511
download_size: 1967534
dataset_size: 2008414
---
# Dataset Card for XStoryCloze
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://cs.rochester.edu/nlp/rocstories/](https://cs.rochester.edu/nlp/rocstories/)
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Paper:** [Few-shot Learning with Multilingual Generative Language Models](https://arxiv.org/pdf/2112.10668.pdf)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 2.03 MB
- **Size of the generated dataset:** 2.03 MB
- **Total amount of disk used:** 2.05 MB
### Dataset Summary
XStoryCloze consists of the professionally translated version of the [English StoryCloze dataset](https://cs.rochester.edu/nlp/rocstories/) (Spring 2016 version) to 10 non-English languages. This dataset is released by Meta AI.
### Supported Tasks and Leaderboards
commonsense reasoning
### Languages
en, ru, zh (Simplified), es (Latin America), ar, hi, id, te, sw, eu, my.
## Dataset Structure
### Data Instances
- **Size of downloaded dataset files:** 2.03 MB
- **Size of the generated dataset:** 2.03 MB
- **Total amount of disk used:** 2.05 MB
An example of 'train' looks as follows.
```
{'answer_right_ending': 1,
'input_sentence_1': 'Rick grew up in a troubled household.',
'input_sentence_2': 'He never found good support in family, and turned to gangs.',
'input_sentence_3': "It wasn't long before Rick got shot in a robbery.",
'input_sentence_4': 'The incident caused him to turn a new leaf.',
'sentence_quiz1': 'He is happy now.',
'sentence_quiz2': 'He joined a gang.',
'story_id': '138d5bfb-05cc-41e3-bf2c-fa85ebad14e2'}
```
### Data Fields
The data fields are the same among all splits.
- `input_sentence_1`: The first statement in the story.
- `input_sentence_2`: The second statement in the story.
- `input_sentence_3`: The third statement in the story.
- `input_sentence_4`: The forth statement in the story.
- `sentence_quiz1`: first possible continuation of the story.
- `sentence_quiz2`: second possible continuation of the story.
- `answer_right_ending`: correct possible ending; either 1 or 2.
- `story_id`: story id.
### Data Splits
This dataset is intended to be used for evaluating the zero- and few-shot learning capabilities of multlingual language models. We split the data for each language into train and test (360 vs. 1510 examples, respectively). The released data files for different languages maintain a line-by-line alignment.
| name |train |test|
|-------|-----:|---:|
|en|360|1510|
|ru|360|1510|
|zh|360|1510|
|es|360|1510|
|ar|360|1510|
|hi|360|1510|
|id|360|1510|
|te|360|1510|
|sw|360|1510|
|eu|360|1510|
|my|360|1510|
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
XStoryCloze is opensourced under [CC BY-SA 4.0](https://creativecommons.org/licenses/by-sa/4.0/legalcode), the same license as the original English StoryCloze.
### Citation Information
```
@article{DBLP:journals/corr/abs-2112-10668,
author = {Xi Victoria Lin and
Todor Mihaylov and
Mikel Artetxe and
Tianlu Wang and
Shuohui Chen and
Daniel Simig and
Myle Ott and
Naman Goyal and
Shruti Bhosale and
Jingfei Du and
Ramakanth Pasunuru and
Sam Shleifer and
Punit Singh Koura and
Vishrav Chaudhary and
Brian O'Horo and
Jeff Wang and
Luke Zettlemoyer and
Zornitsa Kozareva and
Mona T. Diab and
Veselin Stoyanov and
Xian Li},
title = {Few-shot Learning with Multilingual Language Models},
journal = {CoRR},
volume = {abs/2112.10668},
year = {2021},
url = {https://arxiv.org/abs/2112.10668},
eprinttype = {arXiv},
eprint = {2112.10668},
timestamp = {Tue, 04 Jan 2022 15:59:27 +0100},
biburl = {https://dblp.org/rec/journals/corr/abs-2112-10668.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
```
### Contributions
Thanks to [@juletx](https://github.com/juletx). |
naver-clova-ix/cord-v2 | 2022-07-19T23:43:33.000Z | [
"license:cc-by-4.0",
"region:us"
] | naver-clova-ix | null | null | null | 27 | 8,674 | ---
license: cc-by-4.0
---
|
mteb/sts17-crosslingual-sts | 2022-09-27T19:09:43.000Z | [
"language:ar",
"language:de",
"language:en",
"language:es",
"language:fr",
"language:it",
"language:nl",
"language:ko",
"language:tr",
"region:us"
] | mteb | STS17 Cross-lingual dataset | null | null | 1 | 8,655 | ---
language:
- ar
- de
- en
- es
- fr
- it
- nl
- ko
- tr
--- |
sst | 2023-06-01T14:59:56.000Z | [
"task_categories:text-classification",
"task_ids:text-scoring",
"task_ids:sentiment-classification",
"task_ids:sentiment-scoring",
"annotations_creators:crowdsourced",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:100K<n<1M",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:unknown",
"region:us"
] | null | The Stanford Sentiment Treebank, the first corpus with fully labeled parse trees that allows for a
complete analysis of the compositional effects of sentiment in language. | @inproceedings{socher-etal-2013-recursive,
title = "Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank",
author = "Socher, Richard and Perelygin, Alex and Wu, Jean and
Chuang, Jason and Manning, Christopher D. and Ng, Andrew and Potts, Christopher",
booktitle = "Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing",
month = oct,
year = "2013",
address = "Seattle, Washington, USA",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/D13-1170",
pages = "1631--1642",
} | null | 11 | 8,654 | ---
annotations_creators:
- crowdsourced
language_creators:
- found
language:
- en
license:
- unknown
multilinguality:
- monolingual
size_categories:
- 100K<n<1M
- 10K<n<100K
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- text-scoring
- sentiment-classification
- sentiment-scoring
paperswithcode_id: sst
pretty_name: Stanford Sentiment Treebank
dataset_info:
- config_name: default
features:
- name: sentence
dtype: string
- name: label
dtype: float32
- name: tokens
dtype: string
- name: tree
dtype: string
splits:
- name: train
num_bytes: 2818768
num_examples: 8544
- name: validation
num_bytes: 366205
num_examples: 1101
- name: test
num_bytes: 730154
num_examples: 2210
download_size: 7162356
dataset_size: 3915127
- config_name: dictionary
features:
- name: phrase
dtype: string
- name: label
dtype: float32
splits:
- name: dictionary
num_bytes: 12121843
num_examples: 239232
download_size: 7162356
dataset_size: 12121843
- config_name: ptb
features:
- name: ptb_tree
dtype: string
splits:
- name: train
num_bytes: 2185694
num_examples: 8544
- name: validation
num_bytes: 284132
num_examples: 1101
- name: test
num_bytes: 566248
num_examples: 2210
download_size: 7162356
dataset_size: 3036074
config_names:
- default
- dictionary
- ptb
---
# Dataset Card for sst
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** https://nlp.stanford.edu/sentiment/index.html
- **Repository:** [Needs More Information]
- **Paper:** [Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank](https://www.aclweb.org/anthology/D13-1170/)
- **Leaderboard:** [Needs More Information]
- **Point of Contact:** [Needs More Information]
### Dataset Summary
The Stanford Sentiment Treebank is the first corpus with fully labeled parse trees that allows for a complete analysis of the compositional effects of sentiment in language.
### Supported Tasks and Leaderboards
- `sentiment-scoring`: Each complete sentence is annotated with a `float` label that indicates its level of positive sentiment from 0.0 to 1.0. One can decide to use only complete sentences or to include the contributions of the sub-sentences (aka phrases). The labels for each phrase are included in the `dictionary` configuration. To obtain all the phrases in a sentence we need to visit the parse tree included with each example. In contrast, the `ptb` configuration explicitly provides all the labelled parse trees in Penn Treebank format. Here the labels are binned in 5 bins from 0 to 4.
- `sentiment-classification`: We can transform the above into a binary sentiment classification task by rounding each label to 0 or 1.
### Languages
The text in the dataset is in English
## Dataset Structure
### Data Instances
For the `default` configuration:
```
{'label': 0.7222200036048889,
'sentence': 'Yet the act is still charming here .',
'tokens': 'Yet|the|act|is|still|charming|here|.',
'tree': '15|13|13|10|9|9|11|12|10|11|12|14|14|15|0'}
```
For the `dictionary` configuration:
```
{'label': 0.7361099720001221,
'phrase': 'still charming'}
```
For the `ptb` configuration:
```
{'ptb_tree': '(3 (2 Yet) (3 (2 (2 the) (2 act)) (3 (4 (3 (2 is) (3 (2 still) (4 charming))) (2 here)) (2 .))))'}
```
### Data Fields
- `sentence`: a complete sentence expressing an opinion about a film
- `label`: the degree of "positivity" of the opinion, on a scale between 0.0 and 1.0
- `tokens`: a sequence of tokens that form a sentence
- `tree`: a sentence parse tree formatted as a parent pointer tree
- `phrase`: a sub-sentence of a complete sentence
- `ptb_tree`: a sentence parse tree formatted in Penn Treebank-style, where each component's degree of positive sentiment is labelled on a scale from 0 to 4
### Data Splits
The set of complete sentences (both `default` and `ptb` configurations) is split into a training, validation and test set. The `dictionary` configuration has only one split as it is used for reference rather than for learning.
## Dataset Creation
### Curation Rationale
[Needs More Information]
### Source Data
#### Initial Data Collection and Normalization
[Needs More Information]
#### Who are the source language producers?
Rotten Tomatoes reviewers.
### Annotations
#### Annotation process
[Needs More Information]
#### Who are the annotators?
[Needs More Information]
### Personal and Sensitive Information
[Needs More Information]
## Considerations for Using the Data
### Social Impact of Dataset
[Needs More Information]
### Discussion of Biases
[Needs More Information]
### Other Known Limitations
[Needs More Information]
## Additional Information
### Dataset Curators
[Needs More Information]
### Licensing Information
[Needs More Information]
### Citation Information
```
@inproceedings{socher-etal-2013-recursive,
title = "Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank",
author = "Socher, Richard and
Perelygin, Alex and
Wu, Jean and
Chuang, Jason and
Manning, Christopher D. and
Ng, Andrew and
Potts, Christopher",
booktitle = "Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing",
month = oct,
year = "2013",
address = "Seattle, Washington, USA",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/D13-1170",
pages = "1631--1642",
}
```
### Contributions
Thanks to [@patpizio](https://github.com/patpizio) for adding this dataset. |
poloclub/diffusiondb | 2023-05-09T19:00:45.000Z | [
"task_categories:text-to-image",
"task_categories:image-to-text",
"task_ids:image-captioning",
"annotations_creators:no-annotation",
"language_creators:found",
"multilinguality:multilingual",
"size_categories:n>1T",
"source_datasets:original",
"language:en",
"license:cc0-1.0",
"stable diffusion",
"prompt engineering",
"prompts",
"research paper",
"arxiv:2210.14896",
"region:us"
] | poloclub | DiffusionDB is the first large-scale text-to-image prompt dataset. It contains 2
million images generated by Stable Diffusion using prompts and hyperparameters
specified by real users. The unprecedented scale and diversity of this
human-actuated dataset provide exciting research opportunities in understanding
the interplay between prompts and generative models, detecting deepfakes, and
designing human-AI interaction tools to help users more easily use these models. | @article{wangDiffusionDBLargescalePrompt2022,
title = {{{DiffusionDB}}: {{A}} Large-Scale Prompt Gallery Dataset for Text-to-Image Generative Models},
author = {Wang, Zijie J. and Montoya, Evan and Munechika, David and Yang, Haoyang and Hoover, Benjamin and Chau, Duen Horng},
year = {2022},
journal = {arXiv:2210.14896 [cs]},
url = {https://arxiv.org/abs/2210.14896}
} | null | 312 | 8,594 | ---
layout: default
title: Home
nav_order: 1
has_children: false
annotations_creators:
- no-annotation
language:
- en
language_creators:
- found
license:
- cc0-1.0
multilinguality:
- multilingual
pretty_name: DiffusionDB
size_categories:
- n>1T
source_datasets:
- original
tags:
- stable diffusion
- prompt engineering
- prompts
- research paper
task_categories:
- text-to-image
- image-to-text
task_ids:
- image-captioning
---
# DiffusionDB
<img width="100%" src="https://user-images.githubusercontent.com/15007159/201762588-f24db2b8-dbb2-4a94-947b-7de393fc3d33.gif">
## Table of Contents
- [DiffusionDB](#diffusiondb)
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Two Subsets](#two-subsets)
- [Key Differences](#key-differences)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Dataset Metadata](#dataset-metadata)
- [Metadata Schema](#metadata-schema)
- [Data Splits](#data-splits)
- [Loading Data Subsets](#loading-data-subsets)
- [Method 1: Using Hugging Face Datasets Loader](#method-1-using-hugging-face-datasets-loader)
- [Method 2. Use the PoloClub Downloader](#method-2-use-the-poloclub-downloader)
- [Usage/Examples](#usageexamples)
- [Downloading a single file](#downloading-a-single-file)
- [Downloading a range of files](#downloading-a-range-of-files)
- [Downloading to a specific directory](#downloading-to-a-specific-directory)
- [Setting the files to unzip once they've been downloaded](#setting-the-files-to-unzip-once-theyve-been-downloaded)
- [Method 3. Use `metadata.parquet` (Text Only)](#method-3-use-metadataparquet-text-only)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Initial Data Collection and Normalization](#initial-data-collection-and-normalization)
- [Who are the source language producers?](#who-are-the-source-language-producers)
- [Annotations](#annotations)
- [Annotation process](#annotation-process)
- [Who are the annotators?](#who-are-the-annotators)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [DiffusionDB homepage](https://poloclub.github.io/diffusiondb)
- **Repository:** [DiffusionDB repository](https://github.com/poloclub/diffusiondb)
- **Distribution:** [DiffusionDB Hugging Face Dataset](https://huggingface.co/datasets/poloclub/diffusiondb)
- **Paper:** [DiffusionDB: A Large-scale Prompt Gallery Dataset for Text-to-Image Generative Models](https://arxiv.org/abs/2210.14896)
- **Point of Contact:** [Jay Wang](mailto:jayw@gatech.edu)
### Dataset Summary
DiffusionDB is the first large-scale text-to-image prompt dataset. It contains **14 million** images generated by Stable Diffusion using prompts and hyperparameters specified by real users.
DiffusionDB is publicly available at [🤗 Hugging Face Dataset](https://huggingface.co/datasets/poloclub/diffusiondb).
### Supported Tasks and Leaderboards
The unprecedented scale and diversity of this human-actuated dataset provide exciting research opportunities in understanding the interplay between prompts and generative models, detecting deepfakes, and designing human-AI interaction tools to help users more easily use these models.
### Languages
The text in the dataset is mostly English. It also contains other languages such as Spanish, Chinese, and Russian.
### Two Subsets
DiffusionDB provides two subsets (DiffusionDB 2M and DiffusionDB Large) to support different needs.
|Subset|Num of Images|Num of Unique Prompts|Size|Image Directory|Metadata Table|
|:--|--:|--:|--:|--:|--:|
|DiffusionDB 2M|2M|1.5M|1.6TB|`images/`|`metadata.parquet`|
|DiffusionDB Large|14M|1.8M|6.5TB|`diffusiondb-large-part-1/` `diffusiondb-large-part-2/`|`metadata-large.parquet`|
##### Key Differences
1. Two subsets have a similar number of unique prompts, but DiffusionDB Large has much more images. DiffusionDB Large is a superset of DiffusionDB 2M.
2. Images in DiffusionDB 2M are stored in `png` format; images in DiffusionDB Large use a lossless `webp` format.
## Dataset Structure
We use a modularized file structure to distribute DiffusionDB. The 2 million images in DiffusionDB 2M are split into 2,000 folders, where each folder contains 1,000 images and a JSON file that links these 1,000 images to their prompts and hyperparameters. Similarly, the 14 million images in DiffusionDB Large are split into 14,000 folders.
```bash
# DiffusionDB 2M
./
├── images
│ ├── part-000001
│ │ ├── 3bfcd9cf-26ea-4303-bbe1-b095853f5360.png
│ │ ├── 5f47c66c-51d4-4f2c-a872-a68518f44adb.png
│ │ ├── 66b428b9-55dc-4907-b116-55aaa887de30.png
│ │ ├── [...]
│ │ └── part-000001.json
│ ├── part-000002
│ ├── part-000003
│ ├── [...]
│ └── part-002000
└── metadata.parquet
```
```bash
# DiffusionDB Large
./
├── diffusiondb-large-part-1
│ ├── part-000001
│ │ ├── 0a8dc864-1616-4961-ac18-3fcdf76d3b08.webp
│ │ ├── 0a25cacb-5d91-4f27-b18a-bd423762f811.webp
│ │ ├── 0a52d584-4211-43a0-99ef-f5640ee2fc8c.webp
│ │ ├── [...]
│ │ └── part-000001.json
│ ├── part-000002
│ ├── part-000003
│ ├── [...]
│ └── part-010000
├── diffusiondb-large-part-2
│ ├── part-010001
│ │ ├── 0a68f671-3776-424c-91b6-c09a0dd6fc2d.webp
│ │ ├── 0a0756e9-1249-4fe2-a21a-12c43656c7a3.webp
│ │ ├── 0aa48f3d-f2d9-40a8-a800-c2c651ebba06.webp
│ │ ├── [...]
│ │ └── part-000001.json
│ ├── part-010002
│ ├── part-010003
│ ├── [...]
│ └── part-014000
└── metadata-large.parquet
```
These sub-folders have names `part-0xxxxx`, and each image has a unique name generated by [UUID Version 4](https://en.wikipedia.org/wiki/Universally_unique_identifier). The JSON file in a sub-folder has the same name as the sub-folder. Each image is a `PNG` file (DiffusionDB 2M) or a lossless `WebP` file (DiffusionDB Large). The JSON file contains key-value pairs mapping image filenames to their prompts and hyperparameters.
### Data Instances
For example, below is the image of `f3501e05-aef7-4225-a9e9-f516527408ac.png` and its key-value pair in `part-000001.json`.
<img width="300" src="https://i.imgur.com/gqWcRs2.png">
```json
{
"f3501e05-aef7-4225-a9e9-f516527408ac.png": {
"p": "geodesic landscape, john chamberlain, christopher balaskas, tadao ando, 4 k, ",
"se": 38753269,
"c": 12.0,
"st": 50,
"sa": "k_lms"
},
}
```
### Data Fields
- key: Unique image name
- `p`: Prompt
- `se`: Random seed
- `c`: CFG Scale (guidance scale)
- `st`: Steps
- `sa`: Sampler
### Dataset Metadata
To help you easily access prompts and other attributes of images without downloading all the Zip files, we include two metadata tables `metadata.parquet` and `metadata-large.parquet` for DiffusionDB 2M and DiffusionDB Large, respectively.
The shape of `metadata.parquet` is (2000000, 13) and the shape of `metatable-large.parquet` is (14000000, 13). Two tables share the same schema, and each row represents an image. We store these tables in the Parquet format because Parquet is column-based: you can efficiently query individual columns (e.g., prompts) without reading the entire table.
Below are three random rows from `metadata.parquet`.
| image_name | prompt | part_id | seed | step | cfg | sampler | width | height | user_name | timestamp | image_nsfw | prompt_nsfw |
|:-----------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|----------:|-----------:|-------:|------:|----------:|--------:|---------:|:-----------------------------------------------------------------|:--------------------------|-------------:|--------------:|
| 0c46f719-1679-4c64-9ba9-f181e0eae811.png | a small liquid sculpture, corvette, viscous, reflective, digital art | 1050 | 2026845913 | 50 | 7 | 8 | 512 | 512 | c2f288a2ba9df65c38386ffaaf7749106fed29311835b63d578405db9dbcafdb | 2022-08-11 09:05:00+00:00 | 0.0845108 | 0.00383462 |
| a00bdeaa-14eb-4f6c-a303-97732177eae9.png | human sculpture of lanky tall alien on a romantic date at italian restaurant with smiling woman, nice restaurant, photography, bokeh | 905 | 1183522603 | 50 | 10 | 8 | 512 | 768 | df778e253e6d32168eb22279a9776b3cde107cc82da05517dd6d114724918651 | 2022-08-19 17:55:00+00:00 | 0.692934 | 0.109437 |
| 6e5024ce-65ed-47f3-b296-edb2813e3c5b.png | portrait of barbaric spanish conquistador, symmetrical, by yoichi hatakenaka, studio ghibli and dan mumford | 286 | 1713292358 | 50 | 7 | 8 | 512 | 640 | 1c2e93cfb1430adbd956be9c690705fe295cbee7d9ac12de1953ce5e76d89906 | 2022-08-12 03:26:00+00:00 | 0.0773138 | 0.0249675 |
#### Metadata Schema
`metadata.parquet` and `metatable-large.parquet` share the same schema.
|Column|Type|Description|
|:---|:---|:---|
|`image_name`|`string`|Image UUID filename.|
|`prompt`|`string`|The text prompt used to generate this image.|
|`part_id`|`uint16`|Folder ID of this image.|
|`seed`|`uint32`| Random seed used to generate this image.|
|`step`|`uint16`| Step count (hyperparameter).|
|`cfg`|`float32`| Guidance scale (hyperparameter).|
|`sampler`|`uint8`| Sampler method (hyperparameter). Mapping: `{1: "ddim", 2: "plms", 3: "k_euler", 4: "k_euler_ancestral", 5: "k_heun", 6: "k_dpm_2", 7: "k_dpm_2_ancestral", 8: "k_lms", 9: "others"}`.
|`width`|`uint16`|Image width.|
|`height`|`uint16`|Image height.|
|`user_name`|`string`|The unique discord ID's SHA256 hash of the user who generated this image. For example, the hash for `xiaohk#3146` is `e285b7ef63be99e9107cecd79b280bde602f17e0ca8363cb7a0889b67f0b5ed0`. "deleted_account" refer to users who have deleted their accounts. None means the image has been deleted before we scrape it for the second time.|
|`timestamp`|`timestamp`|UTC Timestamp when this image was generated. None means the image has been deleted before we scrape it for the second time. Note that timestamp is not accurate for duplicate images that have the same prompt, hypareparameters, width, height.|
|`image_nsfw`|`float32`|Likelihood of an image being NSFW. Scores are predicted by [LAION's state-of-art NSFW detector](https://github.com/LAION-AI/LAION-SAFETY) (range from 0 to 1). A score of 2.0 means the image has already been flagged as NSFW and blurred by Stable Diffusion.|
|`prompt_nsfw`|`float32`|Likelihood of a prompt being NSFW. Scores are predicted by the library [Detoxicy](https://github.com/unitaryai/detoxify). Each score represents the maximum of `toxicity` and `sexual_explicit` (range from 0 to 1).|
> **Warning**
> Although the Stable Diffusion model has an NSFW filter that automatically blurs user-generated NSFW images, this NSFW filter is not perfect—DiffusionDB still contains some NSFW images. Therefore, we compute and provide the NSFW scores for images and prompts using the state-of-the-art models. The distribution of these scores is shown below. Please decide an appropriate NSFW score threshold to filter out NSFW images before using DiffusionDB in your projects.
<img src="https://i.imgur.com/1RiGAXL.png" width="100%">
### Data Splits
For DiffusionDB 2M, we split 2 million images into 2,000 folders where each folder contains 1,000 images and a JSON file. For DiffusionDB Large, we split 14 million images into 14,000 folders where each folder contains 1,000 images and a JSON file.
### Loading Data Subsets
DiffusionDB is large (1.6TB or 6.5 TB)! However, with our modularized file structure, you can easily load a desirable number of images and their prompts and hyperparameters. In the [`example-loading.ipynb`](https://github.com/poloclub/diffusiondb/blob/main/notebooks/example-loading.ipynb) notebook, we demonstrate three methods to load a subset of DiffusionDB. Below is a short summary.
#### Method 1: Using Hugging Face Datasets Loader
You can use the Hugging Face [`Datasets`](https://huggingface.co/docs/datasets/quickstart) library to easily load prompts and images from DiffusionDB. We pre-defined 16 DiffusionDB subsets (configurations) based on the number of instances. You can see all subsets in the [Dataset Preview](https://huggingface.co/datasets/poloclub/diffusiondb/viewer/all/train).
```python
import numpy as np
from datasets import load_dataset
# Load the dataset with the `large_random_1k` subset
dataset = load_dataset('poloclub/diffusiondb', 'large_random_1k')
```
#### Method 2. Use the PoloClub Downloader
This repo includes a Python downloader [`download.py`](https://github.com/poloclub/diffusiondb/blob/main/scripts/download.py) that allows you to download and load DiffusionDB. You can use it from your command line. Below is an example of loading a subset of DiffusionDB.
##### Usage/Examples
The script is run using command-line arguments as follows:
- `-i` `--index` - File to download or lower bound of a range of files if `-r` is also set.
- `-r` `--range` - Upper bound of range of files to download if `-i` is set.
- `-o` `--output` - Name of custom output directory. Defaults to the current directory if not set.
- `-z` `--unzip` - Unzip the file/files after downloading
- `-l` `--large` - Download from Diffusion DB Large. Defaults to Diffusion DB 2M.
###### Downloading a single file
The specific file to download is supplied as the number at the end of the file on HuggingFace. The script will automatically pad the number out and generate the URL.
```bash
python download.py -i 23
```
###### Downloading a range of files
The upper and lower bounds of the set of files to download are set by the `-i` and `-r` flags respectively.
```bash
python download.py -i 1 -r 2000
```
Note that this range will download the entire dataset. The script will ask you to confirm that you have 1.7Tb free at the download destination.
###### Downloading to a specific directory
The script will default to the location of the dataset's `part` .zip files at `images/`. If you wish to move the download location, you should move these files as well or use a symbolic link.
```bash
python download.py -i 1 -r 2000 -o /home/$USER/datahoarding/etc
```
Again, the script will automatically add the `/` between the directory and the file when it downloads.
###### Setting the files to unzip once they've been downloaded
The script is set to unzip the files _after_ all files have downloaded as both can be lengthy processes in certain circumstances.
```bash
python download.py -i 1 -r 2000 -z
```
#### Method 3. Use `metadata.parquet` (Text Only)
If your task does not require images, then you can easily access all 2 million prompts and hyperparameters in the `metadata.parquet` table.
```python
from urllib.request import urlretrieve
import pandas as pd
# Download the parquet table
table_url = f'https://huggingface.co/datasets/poloclub/diffusiondb/resolve/main/metadata.parquet'
urlretrieve(table_url, 'metadata.parquet')
# Read the table using Pandas
metadata_df = pd.read_parquet('metadata.parquet')
```
## Dataset Creation
### Curation Rationale
Recent diffusion models have gained immense popularity by enabling high-quality and controllable image generation based on text prompts written in natural language. Since the release of these models, people from different domains have quickly applied them to create award-winning artworks, synthetic radiology images, and even hyper-realistic videos.
However, generating images with desired details is difficult, as it requires users to write proper prompts specifying the exact expected results. Developing such prompts requires trial and error, and can often feel random and unprincipled. Simon Willison analogizes writing prompts to wizards learning “magical spells”: users do not understand why some prompts work, but they will add these prompts to their “spell book.” For example, to generate highly-detailed images, it has become a common practice to add special keywords such as “trending on artstation” and “unreal engine” in the prompt.
Prompt engineering has become a field of study in the context of text-to-text generation, where researchers systematically investigate how to construct prompts to effectively solve different down-stream tasks. As large text-to-image models are relatively new, there is a pressing need to understand how these models react to prompts, how to write effective prompts, and how to design tools to help users generate images.
To help researchers tackle these critical challenges, we create DiffusionDB, the first large-scale prompt dataset with 14 million real prompt-image pairs.
### Source Data
#### Initial Data Collection and Normalization
We construct DiffusionDB by scraping user-generated images on the official Stable Diffusion Discord server. We choose Stable Diffusion because it is currently the only open-source large text-to-image generative model, and all generated images have a CC0 1.0 Universal Public Domain Dedication license that waives all copyright and allows uses for any purpose. We choose the official [Stable Diffusion Discord server](https://discord.gg/stablediffusion) because it is public, and it has strict rules against generating and sharing illegal, hateful, or NSFW (not suitable for work, such as sexual and violent content) images. The server also disallows users to write or share prompts with personal information.
#### Who are the source language producers?
The language producers are users of the official [Stable Diffusion Discord server](https://discord.gg/stablediffusion).
### Annotations
The dataset does not contain any additional annotations.
#### Annotation process
[N/A]
#### Who are the annotators?
[N/A]
### Personal and Sensitive Information
The authors removed the discord usernames from the dataset.
We decide to anonymize the dataset because some prompts might include sensitive information: explicitly linking them to their creators can cause harm to creators.
## Considerations for Using the Data
### Social Impact of Dataset
The purpose of this dataset is to help develop better understanding of large text-to-image generative models.
The unprecedented scale and diversity of this human-actuated dataset provide exciting research opportunities in understanding the interplay between prompts and generative models, detecting deepfakes, and designing human-AI interaction tools to help users more easily use these models.
It should note that we collect images and their prompts from the Stable Diffusion Discord server. The Discord server has rules against users generating or sharing harmful or NSFW (not suitable for work, such as sexual and violent content) images. The Stable Diffusion model used in the server also has an NSFW filter that blurs the generated images if it detects NSFW content. However, it is still possible that some users had generated harmful images that were not detected by the NSFW filter or removed by the server moderators. Therefore, DiffusionDB can potentially contain these images. To mitigate the potential harm, we provide a [Google Form](https://forms.gle/GbYaSpRNYqxCafMZ9) on the [DiffusionDB website](https://poloclub.github.io/diffusiondb/) where users can report harmful or inappropriate images and prompts. We will closely monitor this form and remove reported images and prompts from DiffusionDB.
### Discussion of Biases
The 14 million images in DiffusionDB have diverse styles and categories. However, Discord can be a biased data source. Our images come from channels where early users could use a bot to use Stable Diffusion before release. As these users had started using Stable Diffusion before the model was public, we hypothesize that they are AI art enthusiasts and are likely to have experience with other text-to-image generative models. Therefore, the prompting style in DiffusionDB might not represent novice users. Similarly, the prompts in DiffusionDB might not generalize to domains that require specific knowledge, such as medical images.
### Other Known Limitations
**Generalizability.** Previous research has shown a prompt that works well on one generative model might not give the optimal result when used in other models.
Therefore, different models can need users to write different prompts. For example, many Stable Diffusion prompts use commas to separate keywords, while this pattern is less seen in prompts for DALL-E 2 or Midjourney. Thus, we caution researchers that some research findings from DiffusionDB might not be generalizable to other text-to-image generative models.
## Additional Information
### Dataset Curators
DiffusionDB is created by [Jay Wang](https://zijie.wang), [Evan Montoya](https://www.linkedin.com/in/evan-montoya-b252391b4/), [David Munechika](https://www.linkedin.com/in/dmunechika/), [Alex Yang](https://alexanderyang.me), [Ben Hoover](https://www.bhoov.com), [Polo Chau](https://faculty.cc.gatech.edu/~dchau/).
### Licensing Information
The DiffusionDB dataset is available under the [CC0 1.0 License](https://creativecommons.org/publicdomain/zero/1.0/).
The Python code in this repository is available under the [MIT License](https://github.com/poloclub/diffusiondb/blob/main/LICENSE).
### Citation Information
```bibtex
@article{wangDiffusionDBLargescalePrompt2022,
title = {{{DiffusionDB}}: {{A}} Large-Scale Prompt Gallery Dataset for Text-to-Image Generative Models},
author = {Wang, Zijie J. and Montoya, Evan and Munechika, David and Yang, Haoyang and Hoover, Benjamin and Chau, Duen Horng},
year = {2022},
journal = {arXiv:2210.14896 [cs]},
url = {https://arxiv.org/abs/2210.14896}
}
```
### Contributions
If you have any questions, feel free to [open an issue](https://github.com/poloclub/diffusiondb/issues/new) or contact [Jay Wang](https://zijie.wang).
|
eli5 | 2023-06-08T12:42:30.000Z | [
"task_categories:text2text-generation",
"task_ids:abstractive-qa",
"task_ids:open-domain-abstractive-qa",
"annotations_creators:no-annotation",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:100K<n<1M",
"source_datasets:original",
"language:en",
"license:unknown",
"arxiv:1907.09190",
"arxiv:1904.04047",
"region:us"
] | null | Explain Like I'm 5 long form QA dataset | @inproceedings{DBLP:conf/acl/FanJPGWA19,
author = {Angela Fan and
Yacine Jernite and
Ethan Perez and
David Grangier and
Jason Weston and
Michael Auli},
editor = {Anna Korhonen and
David R. Traum and
Lluis Marquez},
title = {{ELI5:} Long Form Question Answering},
booktitle = {Proceedings of the 57th Conference of the Association for Computational
Linguistics, {ACL} 2019, Florence, Italy, July 28- August 2, 2019,
Volume 1: Long Papers},
pages = {3558--3567},
publisher = {Association for Computational Linguistics},
year = {2019},
url = {https://doi.org/10.18653/v1/p19-1346},
doi = {10.18653/v1/p19-1346},
} | null | 36 | 8,547 | ---
annotations_creators:
- no-annotation
language_creators:
- found
language:
- en
license:
- unknown
multilinguality:
- monolingual
size_categories:
- 100K<n<1M
source_datasets:
- original
task_categories:
- text2text-generation
task_ids:
- abstractive-qa
- open-domain-abstractive-qa
paperswithcode_id: eli5
pretty_name: ELI5
viewer: false
dataset_info:
features:
- name: q_id
dtype: string
- name: title
dtype: string
- name: selftext
dtype: string
- name: document
dtype: string
- name: subreddit
dtype: string
- name: answers
sequence:
- name: a_id
dtype: string
- name: text
dtype: string
- name: score
dtype: int32
- name: title_urls
sequence:
- name: url
dtype: string
- name: selftext_urls
sequence:
- name: url
dtype: string
- name: answers_urls
sequence:
- name: url
dtype: string
config_name: LFQA_reddit
splits:
- name: train_eli5
num_bytes: 577188173
num_examples: 272634
- name: validation_eli5
num_bytes: 21117891
num_examples: 9812
- name: test_eli5
num_bytes: 53099796
num_examples: 24512
- name: train_asks
num_bytes: 286464210
num_examples: 131778
- name: validation_asks
num_bytes: 9662481
num_examples: 2281
- name: test_asks
num_bytes: 17713920
num_examples: 4462
- name: train_askh
num_bytes: 330483260
num_examples: 98525
- name: validation_askh
num_bytes: 18690845
num_examples: 4901
- name: test_askh
num_bytes: 36246784
num_examples: 9764
download_size: 6326543
dataset_size: 1350667360
---
## <span style="color:red">⚠️ Reddit recently [changed the terms of access](https://www.reddit.com/r/reddit/comments/12qwagm/an_update_regarding_reddits_api/) to its API, making the source data for this dataset unavailable</span>.
# Dataset Card for ELI5
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [ELI5 homepage](https://facebookresearch.github.io/ELI5/explore.html)
- **Repository:** [ELI5 repository](https://github.com/facebookresearch/ELI5)
- **Paper:** [ELI5: Long Form Question Answering](https://arxiv.org/abs/1907.09190)
- **Point of Contact:** [Yacine Jernite](mailto:yacine@huggingface.co)
### Dataset Summary
The ELI5 dataset is an English-language dataset of questions and answers gathered from three subreddits where users ask factual questions requiring paragraph-length or longer answers. The dataset was created to support the task of open-domain long form abstractive question answering, and covers questions about general topics in its [r/explainlikeimfive](https://www.reddit.com/r/explainlikeimfive/) subset, science in it [r/askscience](https://www.reddit.com/r/askscience/) subset, and History in its [r/AskHistorians](https://www.reddit.com/r/AskHistorians/) subset.
### Supported Tasks and Leaderboards
- `abstractive-qa`, `open-domain-abstractive-qa`: The dataset can be used to train a model for Open Domain Long Form Question Answering. An LFQA model is presented with a non-factoid and asked to retrieve relevant information from a knowledge source (such as [Wikipedia](https://www.wikipedia.org/)), then use it to generate a multi-sentence answer. The model performance is measured by how high its [ROUGE](https://huggingface.co/metrics/rouge) score to the reference is. A [BART-based model](https://huggingface.co/yjernite/bart_eli5) with a [dense retriever](https://huggingface.co/yjernite/retribert-base-uncased) trained to draw information from [Wikipedia passages](https://huggingface.co/datasets/wiki_snippets) achieves a [ROUGE-L of 0.149](https://yjernite.github.io/lfqa.html#generation).
### Languages
The text in the dataset is in English, as spoken by Reddit users on the [r/explainlikeimfive](https://www.reddit.com/r/explainlikeimfive/), [r/askscience](https://www.reddit.com/r/askscience/), and [r/AskHistorians](https://www.reddit.com/r/AskHistorians/) subreddits. The associated BCP-47 code is `en`.
## Dataset Structure
### Data Instances
A typical data point comprises a question, with a `title` containing the main question and a `selftext` which sometimes elaborates on it, and a list of answers from the forum sorted by the number of upvotes they obtained. Additionally, the URLs in each of the text fields have been extracted to respective lists and replaced by generic tokens in the text.
An example from the ELI5 test set looks as follows:
```
{'q_id': '8houtx',
'title': 'Why does water heated to room temperature feel colder than the air around it?',
'selftext': '',
'document': '',
'subreddit': 'explainlikeimfive',
'answers': {'a_id': ['dylcnfk', 'dylcj49'],
'text': ["Water transfers heat more efficiently than air. When something feels cold it's because heat is being transferred from your skin to whatever you're touching. Since water absorbs the heat more readily than air, it feels colder.",
"Air isn't as good at transferring heat compared to something like water or steel (sit on a room temperature steel bench vs. a room temperature wooden bench, and the steel one will feel more cold).\n\nWhen you feel cold, what you're feeling is heat being transferred out of you. If there is no breeze, you feel a certain way. If there's a breeze, you will get colder faster (because the moving air is pulling the heat away from you), and if you get into water, its quite good at pulling heat from you. Get out of the water and have a breeze blow on you while you're wet, all of the water starts evaporating, pulling even more heat from you."],
'score': [5, 2]},
'title_urls': {'url': []},
'selftext_urls': {'url': []},
'answers_urls': {'url': []}}
```
### Data Fields
- `q_id`: a string question identifier for each example, corresponding to its ID in the [Pushshift.io](https://files.pushshift.io/reddit/submissions/) Reddit submission dumps.
- `subreddit`: One of `explainlikeimfive`, `askscience`, or `AskHistorians`, indicating which subreddit the question came from
- `title`: title of the question, with URLs extracted and replaced by `URL_n` tokens
- `title_urls`: list of the extracted URLs, the `n`th element of the list was replaced by `URL_n`
- `selftext`: either an empty string or an elaboration of the question
- `selftext_urls`: similar to `title_urls` but for `self_text`
- `answers`: a list of answers, each answer has:
- `a_id`: a string answer identifier for each answer, corresponding to its ID in the [Pushshift.io](https://files.pushshift.io/reddit/comments/) Reddit comments dumps.
- `text`: the answer text with the URLs normalized
- `score`: the number of upvotes the answer had received when the dumps were created
- `answers_urls`: a list of the extracted URLs. All answers use the same list, the numbering of the normalization token continues across answer texts
### Data Splits
The data is split into a training, validation and test set for each of the three subreddits. In order to avoid having duplicate questions in across sets, the `title` field of each of the questions were ranked by their tf-idf match to their nearest neighbor and the ones with the smallest value were used in the test and validation sets. The final split sizes are as follow:
| | Train | Valid | Test |
| ----- | ------ | ----- | ---- |
| r/explainlikeimfive examples| 272634 | 9812 | 24512|
| r/askscience examples | 131778 | 2281 | 4462 |
| r/AskHistorians examples | 98525 | 4901 | 9764 |
## Dataset Creation
### Curation Rationale
ELI5 was built to provide a testbed for machines to learn how to answer more complex questions, which requires them to find and combine information in a coherent manner. The dataset was built by gathering questions that were asked by community members of three subreddits, including [r/explainlikeimfive](https://www.reddit.com/r/explainlikeimfive/), along with the answers that were provided by other users. The [rules of the subreddit](https://www.reddit.com/r/explainlikeimfive/wiki/detailed_rules) make this data particularly well suited to training a model for abstractive question answering: the questions need to seek an objective explanation about well established facts, and the answers provided need to be understandable to a layperson without any particular knowledge domain.
### Source Data
#### Initial Data Collection and Normalization
The data was obtained by filtering submissions and comments from the subreddits of interest from the XML dumps of the [Reddit forum](https://www.reddit.com/) hosted on [Pushshift.io](https://files.pushshift.io/reddit/).
In order to further improve the quality of the selected examples, only questions with a score of at least 2 and at least one answer with a score of at least 2 were selected for the dataset. The dataset questions and answers span a period form August 2012 to August 2019.
#### Who are the source language producers?
The language producers are users of the [r/explainlikeimfive](https://www.reddit.com/r/explainlikeimfive/), [r/askscience](https://www.reddit.com/r/askscience/), and [r/AskHistorians](https://www.reddit.com/r/AskHistorians/) subreddits between 2012 and 2019. No further demographic information was available from the data source.
### Annotations
The dataset does not contain any additional annotations.
#### Annotation process
[N/A]
#### Who are the annotators?
[N/A]
### Personal and Sensitive Information
The authors removed the speaker IDs from the [Pushshift.io](https://files.pushshift.io/reddit/) dumps but did not otherwise anonymize the data. Some of the questions and answers are about contemporary public figures or individuals who appeared in the news.
## Considerations for Using the Data
### Social Impact of Dataset
The purpose of this dataset is to help develop better question answering systems.
A system that succeeds at the supported task would be able to provide a coherent answer to even complex questions requiring a multi-step explanation, which is beyond the ability of even the larger existing models. The task is also thought as a test-bed for retrieval model which can show the users which source text was used in generating the answer and allow them to confirm the information provided to them.
It should be noted however that the provided answers were written by Reddit users, an information which may be lost if models trained on it are deployed in down-stream applications and presented to users without context. The specific biases this may introduce are discussed in the next section.
### Discussion of Biases
While Reddit hosts a number of thriving communities with high quality discussions, it is also widely known to have corners where sexism, hate, and harassment are significant issues. See for example the [recent post from Reddit founder u/spez](https://www.reddit.com/r/announcements/comments/gxas21/upcoming_changes_to_our_content_policy_our_board/) outlining some of the ways he thinks the website's historical policies have been responsible for this problem, [Adrienne Massanari's 2015 article on GamerGate](https://www.researchgate.net/publication/283848479_Gamergate_and_The_Fappening_How_Reddit's_algorithm_governance_and_culture_support_toxic_technocultures) and follow-up works, or a [2019 Wired article on misogyny on Reddit](https://www.wired.com/story/misogyny-reddit-research/).
While there has been some recent work in the NLP community on *de-biasing* models (e.g. [Black is to Criminal as Caucasian is to Police: Detecting and Removing Multiclass Bias in Word Embeddings](https://arxiv.org/abs/1904.04047) for word embeddings trained specifically on Reddit data), this problem is far from solved, and the likelihood that a trained model might learn the biases present in the data remains a significant concern.
We still note some encouraging signs for all of these communities: [r/explainlikeimfive](https://www.reddit.com/r/explainlikeimfive/) and [r/askscience](https://www.reddit.com/r/askscience/) have similar structures and purposes, and [r/askscience](https://www.reddit.com/r/askscience/) was found in 2015 to show medium supportiveness and very low toxicity when compared to other subreddits (see a [hackerfall post](https://hackerfall.com/story/study-and-interactive-visualization-of-toxicity-in), [thecut.com write-up](https://www.thecut.com/2015/03/interactive-chart-of-reddits-toxicity.html) and supporting [data](https://chart-studio.plotly.com/~bsbell21/210/toxicity-vs-supportiveness-by-subreddit/#data)). Meanwhile, the [r/AskHistorians rules](https://www.reddit.com/r/AskHistorians/wiki/rules) mention that the admins will not tolerate "_racism, sexism, or any other forms of bigotry_". However, further analysis of whether and to what extent these rules reduce toxicity is still needed.
We also note that given the audience of the Reddit website which is more broadly used in the US and Europe, the answers will likely present a Western perspectives, which is particularly important to note when dealing with historical topics.
### Other Known Limitations
The answers provided in the dataset are represent the opinion of Reddit users. While these communities strive to be helpful, they should not be considered to represent a ground truth.
## Additional Information
### Dataset Curators
The dataset was initially created by Angela Fan, Ethan Perez, Yacine Jernite, Jason Weston, Michael Auli, and David Grangier, during work done at Facebook AI Research (FAIR).
### Licensing Information
The licensing status of the dataset hinges on the legal status of the [Pushshift.io](https://files.pushshift.io/reddit/) data which is unclear.
### Citation Information
```
@inproceedings{eli5_lfqa,
author = {Angela Fan and
Yacine Jernite and
Ethan Perez and
David Grangier and
Jason Weston and
Michael Auli},
editor = {Anna Korhonen and
David R. Traum and
Llu{\'{\i}}s M{\`{a}}rquez},
title = {{ELI5:} Long Form Question Answering},
booktitle = {Proceedings of the 57th Conference of the Association for Computational
Linguistics, {ACL} 2019, Florence, Italy, July 28- August 2, 2019,
Volume 1: Long Papers},
pages = {3558--3567},
publisher = {Association for Computational Linguistics},
year = {2019},
url = {https://doi.org/10.18653/v1/p19-1346},
doi = {10.18653/v1/p19-1346}
}
```
### Contributions
Thanks to [@lewtun](https://github.com/lewtun), [@lhoestq](https://github.com/lhoestq), [@mariamabarham](https://github.com/mariamabarham), [@thomwolf](https://github.com/thomwolf), [@yjernite](https://github.com/yjernite) for adding this dataset. |
tydiqa | 2023-04-05T13:42:46.000Z | [
"task_categories:question-answering",
"task_ids:extractive-qa",
"annotations_creators:crowdsourced",
"language_creators:crowdsourced",
"multilinguality:multilingual",
"size_categories:unknown",
"source_datasets:extended|wikipedia",
"language:ar",
"language:bn",
"language:en",
"language:fi",
"language:id",
"language:ja",
"language:ko",
"language:ru",
"language:sw",
"language:te",
"language:th",
"license:apache-2.0",
"region:us"
] | null | TyDi QA is a question answering dataset covering 11 typologically diverse languages with 204K question-answer pairs.
The languages of TyDi QA are diverse with regard to their typology -- the set of linguistic features that each language
expresses -- such that we expect models performing well on this set to generalize across a large number of the languages
in the world. It contains language phenomena that would not be found in English-only corpora. To provide a realistic
information-seeking task and avoid priming effects, questions are written by people who want to know the answer, but
don’t know the answer yet, (unlike SQuAD and its descendents) and the data is collected directly in each language without
the use of translation (unlike MLQA and XQuAD). | @article{tydiqa,
title = {TyDi QA: A Benchmark for Information-Seeking Question Answering in Typologically Diverse Languages},
author = {Jonathan H. Clark and Eunsol Choi and Michael Collins and Dan Garrette and Tom Kwiatkowski and Vitaly Nikolaev and Jennimaria Palomaki}
year = {2020},
journal = {Transactions of the Association for Computational Linguistics}
} | null | 13 | 8,535 | ---
pretty_name: TyDi QA
annotations_creators:
- crowdsourced
language_creators:
- crowdsourced
language:
- ar
- bn
- en
- fi
- id
- ja
- ko
- ru
- sw
- te
- th
license:
- apache-2.0
multilinguality:
- multilingual
size_categories:
- unknown
source_datasets:
- extended|wikipedia
task_categories:
- question-answering
task_ids:
- extractive-qa
paperswithcode_id: tydi-qa
dataset_info:
- config_name: primary_task
features:
- name: passage_answer_candidates
sequence:
- name: plaintext_start_byte
dtype: int32
- name: plaintext_end_byte
dtype: int32
- name: question_text
dtype: string
- name: document_title
dtype: string
- name: language
dtype: string
- name: annotations
sequence:
- name: passage_answer_candidate_index
dtype: int32
- name: minimal_answers_start_byte
dtype: int32
- name: minimal_answers_end_byte
dtype: int32
- name: yes_no_answer
dtype: string
- name: document_plaintext
dtype: string
- name: document_url
dtype: string
splits:
- name: train
num_bytes: 5550574617
num_examples: 166916
- name: validation
num_bytes: 484380443
num_examples: 18670
download_size: 1953887429
dataset_size: 6034955060
- config_name: secondary_task
features:
- name: id
dtype: string
- name: title
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
sequence:
- name: text
dtype: string
- name: answer_start
dtype: int32
splits:
- name: train
num_bytes: 52948607
num_examples: 49881
- name: validation
num_bytes: 5006461
num_examples: 5077
download_size: 1953887429
dataset_size: 57955068
---
# Dataset Card for "tydiqa"
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://github.com/google-research-datasets/tydiqa](https://github.com/google-research-datasets/tydiqa)
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 3.91 GB
- **Size of the generated dataset:** 6.10 GB
- **Total amount of disk used:** 10.00 GB
### Dataset Summary
TyDi QA is a question answering dataset covering 11 typologically diverse languages with 204K question-answer pairs.
The languages of TyDi QA are diverse with regard to their typology -- the set of linguistic features that each language
expresses -- such that we expect models performing well on this set to generalize across a large number of the languages
in the world. It contains language phenomena that would not be found in English-only corpora. To provide a realistic
information-seeking task and avoid priming effects, questions are written by people who want to know the answer, but
don’t know the answer yet, (unlike SQuAD and its descendents) and the data is collected directly in each language without
the use of translation (unlike MLQA and XQuAD).
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Dataset Structure
### Data Instances
#### primary_task
- **Size of downloaded dataset files:** 1.95 GB
- **Size of the generated dataset:** 6.04 GB
- **Total amount of disk used:** 7.99 GB
An example of 'validation' looks as follows.
```
This example was too long and was cropped:
{
"annotations": {
"minimal_answers_end_byte": [-1, -1, -1],
"minimal_answers_start_byte": [-1, -1, -1],
"passage_answer_candidate_index": [-1, -1, -1],
"yes_no_answer": ["NONE", "NONE", "NONE"]
},
"document_plaintext": "\"\\nรองศาสตราจารย์[1] หม่อมราชวงศ์สุขุมพันธุ์ บริพัตร (22 กันยายน 2495 -) ผู้ว่าราชการกรุงเทพมหานครคนที่ 15 อดีตรองหัวหน้าพรรคปร...",
"document_title": "หม่อมราชวงศ์สุขุมพันธุ์ บริพัตร",
"document_url": "\"https://th.wikipedia.org/wiki/%E0%B8%AB%E0%B8%A1%E0%B9%88%E0%B8%AD%E0%B8%A1%E0%B8%A3%E0%B8%B2%E0%B8%8A%E0%B8%A7%E0%B8%87%E0%B8%...",
"language": "thai",
"passage_answer_candidates": "{\"plaintext_end_byte\": [494, 1779, 2931, 3904, 4506, 5588, 6383, 7122, 8224, 9375, 10473, 12563, 15134, 17765, 19863, 21902, 229...",
"question_text": "\"หม่อมราชวงศ์สุขุมพันธุ์ บริพัตร เรียนจบจากที่ไหน ?\"..."
}
```
#### secondary_task
- **Size of downloaded dataset files:** 1.95 GB
- **Size of the generated dataset:** 58.03 MB
- **Total amount of disk used:** 2.01 GB
An example of 'validation' looks as follows.
```
This example was too long and was cropped:
{
"answers": {
"answer_start": [394],
"text": ["بطولتين"]
},
"context": "\"أقيمت البطولة 21 مرة، شارك في النهائيات 78 دولة، وعدد الفرق التي فازت بالبطولة حتى الآن 8 فرق، ويعد المنتخب البرازيلي الأكثر تت...",
"id": "arabic-2387335860751143628-1",
"question": "\"كم عدد مرات فوز الأوروغواي ببطولة كاس العالم لكرو القدم؟\"...",
"title": "قائمة نهائيات كأس العالم"
}
```
### Data Fields
The data fields are the same among all splits.
#### primary_task
- `passage_answer_candidates`: a dictionary feature containing:
- `plaintext_start_byte`: a `int32` feature.
- `plaintext_end_byte`: a `int32` feature.
- `question_text`: a `string` feature.
- `document_title`: a `string` feature.
- `language`: a `string` feature.
- `annotations`: a dictionary feature containing:
- `passage_answer_candidate_index`: a `int32` feature.
- `minimal_answers_start_byte`: a `int32` feature.
- `minimal_answers_end_byte`: a `int32` feature.
- `yes_no_answer`: a `string` feature.
- `document_plaintext`: a `string` feature.
- `document_url`: a `string` feature.
#### secondary_task
- `id`: a `string` feature.
- `title`: a `string` feature.
- `context`: a `string` feature.
- `question`: a `string` feature.
- `answers`: a dictionary feature containing:
- `text`: a `string` feature.
- `answer_start`: a `int32` feature.
### Data Splits
| name | train | validation |
| -------------- | -----: | ---------: |
| primary_task | 166916 | 18670 |
| secondary_task | 49881 | 5077 |
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Citation Information
```
@article{tydiqa,
title = {TyDi QA: A Benchmark for Information-Seeking Question Answering in Typologically Diverse Languages},
author = {Jonathan H. Clark and Eunsol Choi and Michael Collins and Dan Garrette and Tom Kwiatkowski and Vitaly Nikolaev and Jennimaria Palomaki}
year = {2020},
journal = {Transactions of the Association for Computational Linguistics}
}
```
### Contributions
Thanks to [@thomwolf](https://github.com/thomwolf), [@albertvillanova](https://github.com/albertvillanova), [@lewtun](https://github.com/lewtun), [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset. |
Dahoas/rm-static | 2023-03-06T00:13:07.000Z | [
"region:us"
] | Dahoas | null | null | null | 86 | 8,526 | ---
dataset_info:
features:
- name: prompt
dtype: string
- name: response
dtype: string
- name: chosen
dtype: string
- name: rejected
dtype: string
splits:
- name: train
num_bytes: 113850006
num_examples: 76256
- name: test
num_bytes: 7649255
num_examples: 5103
download_size: 73006535
dataset_size: 121499261
---
# Dataset Card for "rm-static"
Split of [hh-static](https://huggingface.co/datasets/Dahoas/static-hh) used for training reward models after supervised fine-tuning. |
snli | 2023-01-25T14:44:35.000Z | [
"task_categories:text-classification",
"task_ids:natural-language-inference",
"task_ids:multi-input-text-classification",
"annotations_creators:crowdsourced",
"language_creators:crowdsourced",
"multilinguality:monolingual",
"size_categories:100K<n<1M",
"source_datasets:extended|other-flicker-30k",
"source_datasets:extended|other-visual-genome",
"language:en",
"license:cc-by-4.0",
"arxiv:1909.02209",
"region:us"
] | null | The SNLI corpus (version 1.0) is a collection of 570k human-written English
sentence pairs manually labeled for balanced classification with the labels
entailment, contradiction, and neutral, supporting the task of natural language
inference (NLI), also known as recognizing textual entailment (RTE). | @inproceedings{snli:emnlp2015,
Author = {Bowman, Samuel R. and Angeli, Gabor and Potts, Christopher, and Manning, Christopher D.},
Booktitle = {Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP)},
Publisher = {Association for Computational Linguistics},
Title = {A large annotated corpus for learning natural language inference},
Year = {2015}
} | null | 29 | 8,458 | ---
annotations_creators:
- crowdsourced
language_creators:
- crowdsourced
language:
- en
license:
- cc-by-4.0
multilinguality:
- monolingual
size_categories:
- 100K<n<1M
source_datasets:
- extended|other-flicker-30k
- extended|other-visual-genome
task_categories:
- text-classification
task_ids:
- natural-language-inference
- multi-input-text-classification
paperswithcode_id: snli
pretty_name: Stanford Natural Language Inference
dataset_info:
features:
- name: premise
dtype: string
- name: hypothesis
dtype: string
- name: label
dtype:
class_label:
names:
'0': entailment
'1': neutral
'2': contradiction
config_name: plain_text
splits:
- name: test
num_bytes: 1263912
num_examples: 10000
- name: train
num_bytes: 66159510
num_examples: 550152
- name: validation
num_bytes: 1268044
num_examples: 10000
download_size: 94550081
dataset_size: 68691466
---
# Dataset Card for SNLI
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [SNLI homepage](https://nlp.stanford.edu/projects/snli/)
- **Repository:**
- **Paper:** [A large annotated corpus for learning natural langauge inference](https://nlp.stanford.edu/pubs/snli_paper.pdf)
- **Leaderboard:** [SNLI leaderboard](https://nlp.stanford.edu/projects/snli/) (located on the homepage)
- **Point of Contact:** [Samuel Bowman](mailto:bowman@nyu.edu) and [Gabor Angeli](mailto:angeli@stanford.edu)
### Dataset Summary
The SNLI corpus (version 1.0) is a collection of 570k human-written English sentence pairs manually labeled for balanced classification with the labels entailment, contradiction, and neutral, supporting the task of natural language inference (NLI), also known as recognizing textual entailment (RTE).
### Supported Tasks and Leaderboards
[SemBERT](https://arxiv.org/pdf/1909.02209.pdf) (Zhousheng Zhang et al, 2019b) is currently listed as SOTA, achieving 91.9% accuracy on the test set. See the [corpus webpage](https://nlp.stanford.edu/projects/snli/) for a list of published results.
### Languages
The language in the dataset is English as spoken by users of the website Flickr and as spoken by crowdworkers from Amazon Mechanical Turk. The BCP-47 code for English is en.
## Dataset Structure
### Data Instances
For each instance, there is a string for the premise, a string for the hypothesis, and an integer for the label. Note that each premise may appear three times with a different hypothesis and label. See the [SNLI corpus viewer](https://huggingface.co/datasets/viewer/?dataset=snli) to explore more examples.
```
{'premise': 'Two women are embracing while holding to go packages.'
'hypothesis': 'The sisters are hugging goodbye while holding to go packages after just eating lunch.'
'label': 1}
```
The average token count for the premises and hypotheses are given below:
| Feature | Mean Token Count |
| ---------- | ---------------- |
| Premise | 14.1 |
| Hypothesis | 8.3 |
### Data Fields
- `premise`: a string used to determine the truthfulness of the hypothesis
- `hypothesis`: a string that may be true, false, or whose truth conditions may not be knowable when compared to the premise
- `label`: an integer whose value may be either _0_, indicating that the hypothesis entails the premise, _1_, indicating that the premise and hypothesis neither entail nor contradict each other, or _2_, indicating that the hypothesis contradicts the premise. Dataset instances which don't have any gold label are marked with -1 label. Make sure you filter them before starting the training using `datasets.Dataset.filter`.
### Data Splits
The SNLI dataset has 3 splits: _train_, _validation_, and _test_. All of the examples in the _validation_ and _test_ sets come from the set that was annotated in the validation task with no-consensus examples removed. The remaining multiply-annotated examples are in the training set with no-consensus examples removed. Each unique premise/caption shows up in only one split, even though they usually appear in at least three different examples.
| Dataset Split | Number of Instances in Split |
| ------------- |----------------------------- |
| Train | 550,152 |
| Validation | 10,000 |
| Test | 10,000 |
## Dataset Creation
### Curation Rationale
The [SNLI corpus (version 1.0)](https://nlp.stanford.edu/projects/snli/) was developed as a benchmark for natural langauge inference (NLI), also known as recognizing textual entailment (RTE), with the goal of producing a dataset large enough to train models using neural methodologies.
### Source Data
#### Initial Data Collection and Normalization
The hypotheses were elicited by presenting crowdworkers with captions from preexisting datasets without the associated photos, but the vocabulary of the hypotheses still reflects the content of the photos as well as the caption style of writing (e.g. mostly present tense). The dataset developers report 37,026 distinct words in the corpus, ignoring case. They allowed bare NPs as well as full sentences. Using the Stanford PCFG Parser 3.5.2 (Klein and Manning, 2003) trained on the standard training set as well as on the Brown Corpus (Francis and Kucera 1979), the authors report that 74% of the premises and 88.9% of the hypotheses result in a parse rooted with an 'S'. The corpus was developed between 2014 and 2015.
Crowdworkers were presented with a caption without the associated photo and asked to produce three alternate captions, one that is definitely true, one that might be true, and one that is definitely false. See Section 2.1 and Figure 1 for details (Bowman et al., 2015).
The corpus includes content from the [Flickr 30k corpus](http://shannon.cs.illinois.edu/DenotationGraph/) and the [VisualGenome corpus](https://visualgenome.org/). The photo captions used to prompt the data creation were collected on Flickr by [Young et al. (2014)](https://www.aclweb.org/anthology/Q14-1006.pdf), who extended the Flickr 8K dataset developed by [Hodosh et al. (2013)](https://www.jair.org/index.php/jair/article/view/10833). Hodosh et al. collected photos from the following Flickr groups: strangers!, Wild-Child (Kids in Action), Dogs in Action (Read the Rules), Outdoor Activities, Action Photography, Flickr-Social (two or more people in the photo). Young et al. do not list the specific groups they collected photos from. The VisualGenome corpus also contains images from Flickr, originally collected in [MS-COCO](https://cocodataset.org/#home) and [YFCC100M](http://projects.dfki.uni-kl.de/yfcc100m/).
The premises from the Flickr 30k corpus corrected for spelling using the Linux spell checker and ungrammatical sentences were removed. Bowman et al. do not report any normalization, though they note that punctuation and capitalization are often omitted.
#### Who are the source language producers?
A large portion of the premises (160k) were produced in the [Flickr 30k corpus](http://shannon.cs.illinois.edu/DenotationGraph/) by an unknown number of crowdworkers. About 2,500 crowdworkers from Amazon Mechanical Turk produced the associated hypotheses. The premises from the Flickr 30k project describe people and animals whose photos were collected and presented to the Flickr 30k crowdworkers, but the SNLI corpus did not present the photos to the hypotheses creators.
The Flickr 30k corpus did not report crowdworker or photo subject demographic information or crowdworker compensation. The SNLI crowdworkers were compensated per HIT at rates between $.1 and $.5 with no incentives. Workers who ignored the guidelines were disqualified, and automated bulk submissions were rejected. No demographic information was collected from the SNLI crowdworkers.
An additional 4,000 premises come from the pilot study of the [VisualGenome corpus](https://visualgenome.org/static/paper/Visual_Genome.pdf). Though the pilot study itself is not described, the location information of the 33,000 AMT crowdworkers that participated over the course of the 6 months of data collection are aggregated. Most of the workers were located in the United States (93%), with others from the Philippines, Kenya, India, Russia, and Canada. Workers were paid $6-$8 per hour.
### Annotations
#### Annotation process
56,941 of the total sentence pairs were further annotated in a validation task. Four annotators each labeled a premise-hypothesis pair as entailment, contradiction, or neither, resulting in 5 total judgements including the original hypothesis author judgement. See Section 2.2 for more details (Bowman et al., 2015).
The authors report 3/5 annotator agreement on 98% of the validation set and unanimous annotator agreement on 58.3% of the validation set. If a label was chosen by three annotators, that label was made the gold label. Following from this, 2% of the data did not have a consensus label and was labeled '-' by the authors.
| Label | Fleiss κ |
| --------------- |--------- |
| _contradiction_ | 0.77 |
| _entailment_ | 0.72 |
| _neutral_ | 0.60 |
| overall | 0.70 |
#### Who are the annotators?
The annotators of the validation task were a closed set of about 30 trusted crowdworkers on Amazon Mechanical Turk. No demographic information was collected. Annotators were compensated per HIT between $.1 and $.5 with $1 bonuses in cases where annotator labels agreed with the curators' labels for 250 randomly distributed examples.
### Personal and Sensitive Information
The dataset does not contain any personal information about the authors or the crowdworkers, but may contain descriptions of the people in the original Flickr photos.
## Considerations for Using the Data
### Social Impact of Dataset
This dataset was developed as a benchmark for evaluating representational systems for text, especially including those induced by representation learning methods, in the task of predicting truth conditions in a given context. (It should be noted that the truth conditions of a hypothesis given a premise does not necessarily match the truth conditions of the hypothesis in the real world.) Systems that are successful at such a task may be more successful in modeling semantic representations.
### Discussion of Biases
The language reflects the content of the photos collected from Flickr, as described in the [Data Collection](#initial-data-collection-and-normalization) section. [Rudinger et al (2017)](https://www.aclweb.org/anthology/W17-1609.pdf) use pointwise mutual information to calculate a measure of association between a manually selected list of tokens corresponding to identity categories and the other words in the corpus, showing strong evidence of stereotypes across gender categories. They also provide examples in which crowdworkers reproduced harmful stereotypes or pejorative language in the hypotheses.
### Other Known Limitations
[Gururangan et al (2018)](https://www.aclweb.org/anthology/N18-2017.pdf), [Poliak et al (2018)](https://www.aclweb.org/anthology/S18-2023.pdf), and [Tsuchiya (2018)](https://www.aclweb.org/anthology/L18-1239.pdf) show that the SNLI corpus has a number of annotation artifacts. Using various classifiers, Poliak et al correctly predicted the label of the hypothesis 69% of the time without using the premise, Gururangan et al 67% of the time, and Tsuchiya 63% of the time.
## Additional Information
### Dataset Curators
The SNLI corpus was developed by Samuel R. Bowman, Gabor Angeli, Christopher Potts, and Christopher D. Manning as part of the [Stanford NLP group](https://nlp.stanford.edu/).
It was supported by a Google Faculty Research Award, a gift from Bloomberg L.P., the Defense Advanced Research Projects Agency (DARPA) Deep Exploration and Filtering of Text (DEFT) Program under Air Force Research Laboratory (AFRL) contract no. FA8750-13-2-0040, the National Science Foundation under grant no. IIS 1159679, and the Department of the Navy, Office of Naval Research, under grant no. N00014-10-1-0109.
### Licensing Information
The Stanford Natural Language Inference Corpus is licensed under a [Creative Commons Attribution-ShareAlike 4.0 International License](http://creativecommons.org/licenses/by-sa/4.0/).
### Citation Information
```
@inproceedings{snli:emnlp2015,
Author = {Bowman, Samuel R. and Angeli, Gabor and Potts, Christopher, and Manning, Christopher D.},
Booktitle = {Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP)},
Publisher = {Association for Computational Linguistics},
Title = {A large annotated corpus for learning natural language inference},
Year = {2015}
}
```
### Contributions
Thanks to [@mariamabarham](https://github.com/mariamabarham), [@thomwolf](https://github.com/thomwolf), [@lewtun](https://github.com/lewtun), [@patrickvonplaten](https://github.com/patrickvonplaten) and [@mcmillanmajora](https://github.com/mcmillanmajora) for adding this dataset. |
multi_nli | 2023-04-05T10:10:15.000Z | [
"task_categories:text-classification",
"task_ids:natural-language-inference",
"task_ids:multi-input-text-classification",
"annotations_creators:crowdsourced",
"language_creators:crowdsourced",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:100K<n<1M",
"source_datasets:original",
"language:en",
"license:cc-by-3.0",
"license:cc-by-sa-3.0",
"license:mit",
"license:other",
"region:us"
] | null | The Multi-Genre Natural Language Inference (MultiNLI) corpus is a
crowd-sourced collection of 433k sentence pairs annotated with textual
entailment information. The corpus is modeled on the SNLI corpus, but differs in
that covers a range of genres of spoken and written text, and supports a
distinctive cross-genre generalization evaluation. The corpus served as the
basis for the shared task of the RepEval 2017 Workshop at EMNLP in Copenhagen. | @InProceedings{N18-1101,
author = {Williams, Adina
and Nangia, Nikita
and Bowman, Samuel},
title = {A Broad-Coverage Challenge Corpus for
Sentence Understanding through Inference},
booktitle = {Proceedings of the 2018 Conference of
the North American Chapter of the
Association for Computational Linguistics:
Human Language Technologies, Volume 1 (Long
Papers)},
year = {2018},
publisher = {Association for Computational Linguistics},
pages = {1112--1122},
location = {New Orleans, Louisiana},
url = {http://aclweb.org/anthology/N18-1101}
} | null | 36 | 8,353 | ---
annotations_creators:
- crowdsourced
language_creators:
- crowdsourced
- found
language:
- en
license:
- cc-by-3.0
- cc-by-sa-3.0
- mit
- other
multilinguality:
- monolingual
size_categories:
- 100K<n<1M
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- natural-language-inference
- multi-input-text-classification
paperswithcode_id: multinli
pretty_name: Multi-Genre Natural Language Inference
license_details: Open Portion of the American National Corpus
dataset_info:
features:
- name: promptID
dtype: int32
- name: pairID
dtype: string
- name: premise
dtype: string
- name: premise_binary_parse
dtype: string
- name: premise_parse
dtype: string
- name: hypothesis
dtype: string
- name: hypothesis_binary_parse
dtype: string
- name: hypothesis_parse
dtype: string
- name: genre
dtype: string
- name: label
dtype:
class_label:
names:
'0': entailment
'1': neutral
'2': contradiction
splits:
- name: train
num_bytes: 410211586
num_examples: 392702
- name: validation_matched
num_bytes: 10063939
num_examples: 9815
- name: validation_mismatched
num_bytes: 10610221
num_examples: 9832
download_size: 226850426
dataset_size: 430885746
---
# Dataset Card for Multi-Genre Natural Language Inference (MultiNLI)
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://www.nyu.edu/projects/bowman/multinli/](https://www.nyu.edu/projects/bowman/multinli/)
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 226.85 MB
- **Size of the generated dataset:** 76.95 MB
- **Total amount of disk used:** 303.81 MB
### Dataset Summary
The Multi-Genre Natural Language Inference (MultiNLI) corpus is a
crowd-sourced collection of 433k sentence pairs annotated with textual
entailment information. The corpus is modeled on the SNLI corpus, but differs in
that covers a range of genres of spoken and written text, and supports a
distinctive cross-genre generalization evaluation. The corpus served as the
basis for the shared task of the RepEval 2017 Workshop at EMNLP in Copenhagen.
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
The dataset contains samples in English only.
## Dataset Structure
### Data Instances
- **Size of downloaded dataset files:** 226.85 MB
- **Size of the generated dataset:** 76.95 MB
- **Total amount of disk used:** 303.81 MB
Example of a data instance:
```
{
"promptID": 31193,
"pairID": "31193n",
"premise": "Conceptually cream skimming has two basic dimensions - product and geography.",
"premise_binary_parse": "( ( Conceptually ( cream skimming ) ) ( ( has ( ( ( two ( basic dimensions ) ) - ) ( ( product and ) geography ) ) ) . ) )",
"premise_parse": "(ROOT (S (NP (JJ Conceptually) (NN cream) (NN skimming)) (VP (VBZ has) (NP (NP (CD two) (JJ basic) (NNS dimensions)) (: -) (NP (NN product) (CC and) (NN geography)))) (. .)))",
"hypothesis": "Product and geography are what make cream skimming work. ",
"hypothesis_binary_parse": "( ( ( Product and ) geography ) ( ( are ( what ( make ( cream ( skimming work ) ) ) ) ) . ) )",
"hypothesis_parse": "(ROOT (S (NP (NN Product) (CC and) (NN geography)) (VP (VBP are) (SBAR (WHNP (WP what)) (S (VP (VBP make) (NP (NP (NN cream)) (VP (VBG skimming) (NP (NN work)))))))) (. .)))",
"genre": "government",
"label": 1
}
```
### Data Fields
The data fields are the same among all splits.
- `promptID`: Unique identifier for prompt
- `pairID`: Unique identifier for pair
- `{premise,hypothesis}`: combination of `premise` and `hypothesis`
- `{premise,hypothesis} parse`: Each sentence as parsed by the Stanford PCFG Parser 3.5.2
- `{premise,hypothesis} binary parse`: parses in unlabeled binary-branching format
- `genre`: a `string` feature.
- `label`: a classification label, with possible values including `entailment` (0), `neutral` (1), `contradiction` (2). Dataset instances which don't have any gold label are marked with -1 label. Make sure you filter them before starting the training using `datasets.Dataset.filter`.
### Data Splits
|train |validation_matched|validation_mismatched|
|-----:|-----------------:|--------------------:|
|392702| 9815| 9832|
## Dataset Creation
### Curation Rationale
They constructed MultiNLI so as to make it possible to explicitly evaluate models both on the quality of their sentence representations within the training domain and on their ability to derive reasonable representations in unfamiliar domains.
### Source Data
#### Initial Data Collection and Normalization
They created each sentence pair by selecting a premise sentence from a preexisting text source and asked a human annotator to compose a novel sentence to pair with it as a hypothesis.
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
The majority of the corpus is released under the OANC’s license, which allows all content to be freely used, modified, and shared under permissive terms. The data in the FICTION section falls under several permissive licenses; Seven Swords is available under a Creative Commons Share-Alike 3.0 Unported License, and with the explicit permission of the author, Living History and Password Incorrect are available under Creative Commons Attribution 3.0 Unported Licenses; the remaining works of fiction are in the public domain in the United States (but may be licensed differently elsewhere).
### Citation Information
```
@InProceedings{N18-1101,
author = "Williams, Adina
and Nangia, Nikita
and Bowman, Samuel",
title = "A Broad-Coverage Challenge Corpus for
Sentence Understanding through Inference",
booktitle = "Proceedings of the 2018 Conference of
the North American Chapter of the
Association for Computational Linguistics:
Human Language Technologies, Volume 1 (Long
Papers)",
year = "2018",
publisher = "Association for Computational Linguistics",
pages = "1112--1122",
location = "New Orleans, Louisiana",
url = "http://aclweb.org/anthology/N18-1101"
}
```
### Contributions
Thanks to [@bhavitvyamalik](https://github.com/bhavitvyamalik), [@patrickvonplaten](https://github.com/patrickvonplaten), [@thomwolf](https://github.com/thomwolf), [@mariamabarham](https://github.com/mariamabarham) for adding this dataset. |
gia-project/gia-dataset-parquet | 2023-10-02T21:52:41.000Z | [
"task_categories:reinforcement-learning",
"task_categories:text-generation",
"task_categories:question-answering",
"annotations_creators:found",
"annotations_creators:machine-generated",
"source_datasets:conceptual-captions",
"source_datasets:ok-vqa",
"source_datasets:oscar",
"license:apache-2.0",
"imitation-learning",
"reinforcement-learning",
"text-generation",
"question-answering",
"generalist-agent",
"arxiv:2303.03915",
"region:us"
] | gia-project | null | null | null | 0 | 8,350 | ---
annotations_creators:
- found
- machine-generated
license: apache-2.0
size_categories:
- {}
source_datasets:
- conceptual-captions
- ok-vqa
- oscar
task_categories:
- reinforcement-learning
- text-generation
- question-answering
pretty_name: GIA-dataset
configs:
- config_name: atari-alien
data_files:
- split: train
path: atari-alien/train-*
- split: test
path: atari-alien/test-*
- config_name: atari-amidar
data_files:
- split: train
path: atari-amidar/train-*
- split: test
path: atari-amidar/test-*
- config_name: atari-assault
data_files:
- split: train
path: atari-assault/train-*
- split: test
path: atari-assault/test-*
- config_name: atari-asterix
data_files:
- split: train
path: atari-asterix/train-*
- split: test
path: atari-asterix/test-*
- config_name: atari-asteroids
data_files:
- split: train
path: atari-asteroids/train-*
- split: test
path: atari-asteroids/test-*
- config_name: atari-atlantis
data_files:
- split: train
path: atari-atlantis/train-*
- split: test
path: atari-atlantis/test-*
- config_name: atari-bankheist
data_files:
- split: train
path: atari-bankheist/train-*
- split: test
path: atari-bankheist/test-*
- config_name: atari-battlezone
data_files:
- split: train
path: atari-battlezone/train-*
- split: test
path: atari-battlezone/test-*
- config_name: atari-beamrider
data_files:
- split: train
path: atari-beamrider/train-*
- split: test
path: atari-beamrider/test-*
- config_name: atari-berzerk
data_files:
- split: train
path: atari-berzerk/train-*
- split: test
path: atari-berzerk/test-*
- config_name: atari-bowling
data_files:
- split: train
path: atari-bowling/train-*
- split: test
path: atari-bowling/test-*
- config_name: atari-boxing
data_files:
- split: train
path: atari-boxing/train-*
- split: test
path: atari-boxing/test-*
- config_name: atari-breakout
data_files:
- split: train
path: atari-breakout/train-*
- split: test
path: atari-breakout/test-*
- config_name: atari-centipede
data_files:
- split: train
path: atari-centipede/train-*
- split: test
path: atari-centipede/test-*
- config_name: atari-choppercommand
data_files:
- split: train
path: atari-choppercommand/train-*
- split: test
path: atari-choppercommand/test-*
- config_name: atari-crazyclimber
data_files:
- split: train
path: atari-crazyclimber/train-*
- split: test
path: atari-crazyclimber/test-*
- config_name: atari-defender
data_files:
- split: train
path: atari-defender/train-*
- split: test
path: atari-defender/test-*
- config_name: atari-demonattack
data_files:
- split: train
path: atari-demonattack/train-*
- split: test
path: atari-demonattack/test-*
- config_name: atari-doubledunk
data_files:
- split: test
path: atari-doubledunk/test-*
- split: train
path: atari-doubledunk/train-*
- config_name: atari-enduro
data_files:
- split: train
path: atari-enduro/train-*
- split: test
path: atari-enduro/test-*
- config_name: atari-fishingderby
data_files:
- split: train
path: atari-fishingderby/train-*
- split: test
path: atari-fishingderby/test-*
- config_name: atari-freeway
data_files:
- split: train
path: atari-freeway/train-*
- split: test
path: atari-freeway/test-*
- config_name: atari-frostbite
data_files:
- split: train
path: atari-frostbite/train-*
- split: test
path: atari-frostbite/test-*
- config_name: atari-gopher
data_files:
- split: train
path: atari-gopher/train-*
- split: test
path: atari-gopher/test-*
- config_name: atari-gravitar
data_files:
- split: train
path: atari-gravitar/train-*
- split: test
path: atari-gravitar/test-*
- config_name: atari-hero
data_files:
- split: train
path: atari-hero/train-*
- split: test
path: atari-hero/test-*
- config_name: atari-icehockey
data_files:
- split: train
path: atari-icehockey/train-*
- split: test
path: atari-icehockey/test-*
- config_name: atari-jamesbond
data_files:
- split: train
path: atari-jamesbond/train-*
- split: test
path: atari-jamesbond/test-*
- config_name: atari-kangaroo
data_files:
- split: train
path: atari-kangaroo/train-*
- split: test
path: atari-kangaroo/test-*
- config_name: atari-krull
data_files:
- split: train
path: atari-krull/train-*
- split: test
path: atari-krull/test-*
- config_name: atari-kungfumaster
data_files:
- split: train
path: atari-kungfumaster/train-*
- split: test
path: atari-kungfumaster/test-*
- config_name: atari-montezumarevenge
data_files:
- split: train
path: atari-montezumarevenge/train-*
- split: test
path: atari-montezumarevenge/test-*
- config_name: atari-mspacman
data_files:
- split: train
path: atari-mspacman/train-*
- split: test
path: atari-mspacman/test-*
- config_name: atari-namethisgame
data_files:
- split: train
path: atari-namethisgame/train-*
- split: test
path: atari-namethisgame/test-*
- config_name: atari-phoenix
data_files:
- split: train
path: atari-phoenix/train-*
- split: test
path: atari-phoenix/test-*
- config_name: atari-pitfall
data_files:
- split: train
path: atari-pitfall/train-*
- split: test
path: atari-pitfall/test-*
- config_name: atari-pong
data_files:
- split: test
path: atari-pong/test-*
- split: train
path: atari-pong/train-*
- config_name: atari-privateeye
data_files:
- split: test
path: atari-privateeye/test-*
- split: train
path: atari-privateeye/train-*
- config_name: atari-qbert
data_files:
- split: test
path: atari-qbert/test-*
- split: train
path: atari-qbert/train-*
- config_name: atari-riverraid
data_files:
- split: test
path: atari-riverraid/test-*
- split: train
path: atari-riverraid/train-*
- config_name: atari-roadrunner
data_files:
- split: test
path: atari-roadrunner/test-*
- split: train
path: atari-roadrunner/train-*
- config_name: atari-robotank
data_files:
- split: test
path: atari-robotank/test-*
- split: train
path: atari-robotank/train-*
- config_name: atari-seaquest
data_files:
- split: test
path: atari-seaquest/test-*
- split: train
path: atari-seaquest/train-*
- config_name: atari-skiing
data_files:
- split: train
path: atari-skiing/train-*
- split: test
path: atari-skiing/test-*
- config_name: atari-solaris
data_files:
- split: train
path: atari-solaris/train-*
- split: test
path: atari-solaris/test-*
- config_name: atari-spaceinvaders
data_files:
- split: train
path: atari-spaceinvaders/train-*
- split: test
path: atari-spaceinvaders/test-*
- config_name: atari-stargunner
data_files:
- split: train
path: atari-stargunner/train-*
- split: test
path: atari-stargunner/test-*
- config_name: atari-surround
data_files:
- split: train
path: atari-surround/train-*
- split: test
path: atari-surround/test-*
- config_name: atari-tennis
data_files:
- split: train
path: atari-tennis/train-*
- split: test
path: atari-tennis/test-*
- config_name: atari-timepilot
data_files:
- split: train
path: atari-timepilot/train-*
- split: test
path: atari-timepilot/test-*
- config_name: atari-tutankham
data_files:
- split: train
path: atari-tutankham/train-*
- split: test
path: atari-tutankham/test-*
- config_name: atari-upndown
data_files:
- split: train
path: atari-upndown/train-*
- split: test
path: atari-upndown/test-*
- config_name: atari-venture
data_files:
- split: test
path: atari-venture/test-*
- split: train
path: atari-venture/train-*
- config_name: atari-videopinball
data_files:
- split: test
path: atari-videopinball/test-*
- split: train
path: atari-videopinball/train-*
- config_name: atari-wizardofwor
data_files:
- split: test
path: atari-wizardofwor/test-*
- split: train
path: atari-wizardofwor/train-*
- config_name: atari-yarsrevenge
data_files:
- split: test
path: atari-yarsrevenge/test-*
- split: train
path: atari-yarsrevenge/train-*
- config_name: atari-zaxxon
data_files:
- split: test
path: atari-zaxxon/test-*
- split: train
path: atari-zaxxon/train-*
- config_name: babyai-action-obj-door
data_files:
- split: train
path: babyai-action-obj-door/train-*
- split: test
path: babyai-action-obj-door/test-*
- config_name: babyai-blocked-unlock-pickup
data_files:
- split: test
path: babyai-blocked-unlock-pickup/test-*
- split: train
path: babyai-blocked-unlock-pickup/train-*
- config_name: babyai-boss-level
data_files:
- split: test
path: babyai-boss-level/test-*
- split: train
path: babyai-boss-level/train-*
- config_name: babyai-boss-level-no-unlock
data_files:
- split: test
path: babyai-boss-level-no-unlock/test-*
- split: train
path: babyai-boss-level-no-unlock/train-*
- config_name: babyai-find-obj-s5
data_files:
- split: train
path: babyai-find-obj-s5/train-*
- split: test
path: babyai-find-obj-s5/test-*
- config_name: babyai-go-to
data_files:
- split: train
path: babyai-go-to/train-*
- split: test
path: babyai-go-to/test-*
- config_name: babyai-go-to-door
data_files:
- split: train
path: babyai-go-to-door/train-*
- split: test
path: babyai-go-to-door/test-*
- config_name: babyai-go-to-local
data_files:
- split: train
path: babyai-go-to-local/train-*
- split: test
path: babyai-go-to-local/test-*
- config_name: babyai-go-to-obj
data_files:
- split: train
path: babyai-go-to-obj/train-*
- split: test
path: babyai-go-to-obj/test-*
- config_name: babyai-go-to-obj-door
data_files:
- split: train
path: babyai-go-to-obj-door/train-*
- split: test
path: babyai-go-to-obj-door/test-*
- config_name: babyai-go-to-red-ball
data_files:
- split: train
path: babyai-go-to-red-ball/train-*
- split: test
path: babyai-go-to-red-ball/test-*
- config_name: babyai-go-to-red-ball-grey
data_files:
- split: train
path: babyai-go-to-red-ball-grey/train-*
- split: test
path: babyai-go-to-red-ball-grey/test-*
- config_name: babyai-go-to-red-ball-no-dists
data_files:
- split: train
path: babyai-go-to-red-ball-no-dists/train-*
- split: test
path: babyai-go-to-red-ball-no-dists/test-*
- config_name: babyai-go-to-red-blue-ball
data_files:
- split: train
path: babyai-go-to-red-blue-ball/train-*
- split: test
path: babyai-go-to-red-blue-ball/test-*
- config_name: babyai-go-to-seq
data_files:
- split: train
path: babyai-go-to-seq/train-*
- split: test
path: babyai-go-to-seq/test-*
- config_name: babyai-key-corridor
data_files:
- split: test
path: babyai-key-corridor/test-*
- split: train
path: babyai-key-corridor/train-*
- config_name: babyai-key-in-box
data_files:
- split: test
path: babyai-key-in-box/test-*
- split: train
path: babyai-key-in-box/train-*
- config_name: babyai-mini-boss-level
data_files:
- split: test
path: babyai-mini-boss-level/test-*
- split: train
path: babyai-mini-boss-level/train-*
- config_name: babyai-move-two-across-s8n9
data_files:
- split: test
path: babyai-move-two-across-s8n9/test-*
- split: train
path: babyai-move-two-across-s8n9/train-*
- config_name: babyai-one-room-s8
data_files:
- split: test
path: babyai-one-room-s8/test-*
- split: train
path: babyai-one-room-s8/train-*
- config_name: babyai-open
data_files:
- split: test
path: babyai-open/test-*
- split: train
path: babyai-open/train-*
- config_name: babyai-open-door
data_files:
- split: test
path: babyai-open-door/test-*
- split: train
path: babyai-open-door/train-*
- config_name: babyai-open-doors-order-n4
data_files:
- split: test
path: babyai-open-doors-order-n4/test-*
- split: train
path: babyai-open-doors-order-n4/train-*
- config_name: babyai-open-red-door
data_files:
- split: test
path: babyai-open-red-door/test-*
- split: train
path: babyai-open-red-door/train-*
- config_name: babyai-open-two-doors
data_files:
- split: test
path: babyai-open-two-doors/test-*
- split: train
path: babyai-open-two-doors/train-*
- config_name: babyai-pickup
data_files:
- split: test
path: babyai-pickup/test-*
- split: train
path: babyai-pickup/train-*
- config_name: babyai-pickup-above
data_files:
- split: test
path: babyai-pickup-above/test-*
- split: train
path: babyai-pickup-above/train-*
- config_name: babyai-pickup-dist
data_files:
- split: test
path: babyai-pickup-dist/test-*
- split: train
path: babyai-pickup-dist/train-*
- config_name: babyai-pickup-loc
data_files:
- split: test
path: babyai-pickup-loc/test-*
- split: train
path: babyai-pickup-loc/train-*
- config_name: babyai-synth
data_files:
- split: test
path: babyai-synth/test-*
- split: train
path: babyai-synth/train-*
- config_name: babyai-synth-loc
data_files:
- split: test
path: babyai-synth-loc/test-*
- split: train
path: babyai-synth-loc/train-*
- config_name: babyai-synth-seq
data_files:
- split: test
path: babyai-synth-seq/test-*
- split: train
path: babyai-synth-seq/train-*
- config_name: babyai-unblock-pickup
data_files:
- split: test
path: babyai-unblock-pickup/test-*
- split: train
path: babyai-unblock-pickup/train-*
- config_name: babyai-unlock-local
data_files:
- split: test
path: babyai-unlock-local/test-*
- split: train
path: babyai-unlock-local/train-*
- config_name: babyai-unlock-pickup
data_files:
- split: test
path: babyai-unlock-pickup/test-*
- split: train
path: babyai-unlock-pickup/train-*
- config_name: conceptual-captions
data_files:
- split: test
path: conceptual-captions/test-*
- split: train
path: conceptual-captions/train-*
- config_name: metaworld-assembly
data_files:
- split: train
path: metaworld-assembly/train-*
- split: test
path: metaworld-assembly/test-*
- config_name: metaworld-basketball
data_files:
- split: train
path: metaworld-basketball/train-*
- split: test
path: metaworld-basketball/test-*
- config_name: metaworld-bin-picking
data_files:
- split: train
path: metaworld-bin-picking/train-*
- split: test
path: metaworld-bin-picking/test-*
- config_name: metaworld-box-close
data_files:
- split: train
path: metaworld-box-close/train-*
- split: test
path: metaworld-box-close/test-*
- config_name: metaworld-button-press
data_files:
- split: train
path: metaworld-button-press/train-*
- split: test
path: metaworld-button-press/test-*
- config_name: metaworld-button-press-topdown
data_files:
- split: train
path: metaworld-button-press-topdown/train-*
- split: test
path: metaworld-button-press-topdown/test-*
- config_name: metaworld-button-press-topdown-wall
data_files:
- split: train
path: metaworld-button-press-topdown-wall/train-*
- split: test
path: metaworld-button-press-topdown-wall/test-*
- config_name: metaworld-button-press-wall
data_files:
- split: train
path: metaworld-button-press-wall/train-*
- split: test
path: metaworld-button-press-wall/test-*
- config_name: metaworld-coffee-button
data_files:
- split: train
path: metaworld-coffee-button/train-*
- split: test
path: metaworld-coffee-button/test-*
- config_name: metaworld-coffee-pull
data_files:
- split: train
path: metaworld-coffee-pull/train-*
- split: test
path: metaworld-coffee-pull/test-*
- config_name: metaworld-coffee-push
data_files:
- split: train
path: metaworld-coffee-push/train-*
- split: test
path: metaworld-coffee-push/test-*
- config_name: metaworld-dial-turn
data_files:
- split: train
path: metaworld-dial-turn/train-*
- split: test
path: metaworld-dial-turn/test-*
- config_name: metaworld-disassemble
data_files:
- split: train
path: metaworld-disassemble/train-*
- split: test
path: metaworld-disassemble/test-*
- config_name: metaworld-door-close
data_files:
- split: train
path: metaworld-door-close/train-*
- split: test
path: metaworld-door-close/test-*
- config_name: metaworld-door-lock
data_files:
- split: train
path: metaworld-door-lock/train-*
- split: test
path: metaworld-door-lock/test-*
- config_name: metaworld-door-open
data_files:
- split: train
path: metaworld-door-open/train-*
- split: test
path: metaworld-door-open/test-*
- config_name: metaworld-door-unlock
data_files:
- split: train
path: metaworld-door-unlock/train-*
- split: test
path: metaworld-door-unlock/test-*
- config_name: metaworld-drawer-close
data_files:
- split: train
path: metaworld-drawer-close/train-*
- split: test
path: metaworld-drawer-close/test-*
- config_name: metaworld-drawer-open
data_files:
- split: train
path: metaworld-drawer-open/train-*
- split: test
path: metaworld-drawer-open/test-*
- config_name: metaworld-faucet-close
data_files:
- split: train
path: metaworld-faucet-close/train-*
- split: test
path: metaworld-faucet-close/test-*
- config_name: metaworld-faucet-open
data_files:
- split: train
path: metaworld-faucet-open/train-*
- split: test
path: metaworld-faucet-open/test-*
- config_name: metaworld-hammer
data_files:
- split: train
path: metaworld-hammer/train-*
- split: test
path: metaworld-hammer/test-*
- config_name: metaworld-hand-insert
data_files:
- split: train
path: metaworld-hand-insert/train-*
- split: test
path: metaworld-hand-insert/test-*
- config_name: metaworld-handle-press
data_files:
- split: train
path: metaworld-handle-press/train-*
- split: test
path: metaworld-handle-press/test-*
- config_name: metaworld-handle-press-side
data_files:
- split: train
path: metaworld-handle-press-side/train-*
- split: test
path: metaworld-handle-press-side/test-*
- config_name: metaworld-handle-pull
data_files:
- split: train
path: metaworld-handle-pull/train-*
- split: test
path: metaworld-handle-pull/test-*
- config_name: metaworld-handle-pull-side
data_files:
- split: train
path: metaworld-handle-pull-side/train-*
- split: test
path: metaworld-handle-pull-side/test-*
- config_name: metaworld-lever-pull
data_files:
- split: train
path: metaworld-lever-pull/train-*
- split: test
path: metaworld-lever-pull/test-*
- config_name: metaworld-peg-insert-side
data_files:
- split: train
path: metaworld-peg-insert-side/train-*
- split: test
path: metaworld-peg-insert-side/test-*
- config_name: metaworld-peg-unplug-side
data_files:
- split: train
path: metaworld-peg-unplug-side/train-*
- split: test
path: metaworld-peg-unplug-side/test-*
- config_name: metaworld-pick-out-of-hole
data_files:
- split: train
path: metaworld-pick-out-of-hole/train-*
- split: test
path: metaworld-pick-out-of-hole/test-*
- config_name: metaworld-pick-place
data_files:
- split: train
path: metaworld-pick-place/train-*
- split: test
path: metaworld-pick-place/test-*
- config_name: metaworld-pick-place-wall
data_files:
- split: train
path: metaworld-pick-place-wall/train-*
- split: test
path: metaworld-pick-place-wall/test-*
- config_name: metaworld-plate-slide
data_files:
- split: train
path: metaworld-plate-slide/train-*
- split: test
path: metaworld-plate-slide/test-*
- config_name: metaworld-plate-slide-back
data_files:
- split: train
path: metaworld-plate-slide-back/train-*
- split: test
path: metaworld-plate-slide-back/test-*
- config_name: metaworld-plate-slide-back-side
data_files:
- split: train
path: metaworld-plate-slide-back-side/train-*
- split: test
path: metaworld-plate-slide-back-side/test-*
- config_name: metaworld-plate-slide-side
data_files:
- split: train
path: metaworld-plate-slide-side/train-*
- split: test
path: metaworld-plate-slide-side/test-*
- config_name: metaworld-push
data_files:
- split: train
path: metaworld-push/train-*
- split: test
path: metaworld-push/test-*
- config_name: metaworld-push-back
data_files:
- split: train
path: metaworld-push-back/train-*
- split: test
path: metaworld-push-back/test-*
- config_name: metaworld-push-wall
data_files:
- split: train
path: metaworld-push-wall/train-*
- split: test
path: metaworld-push-wall/test-*
- config_name: metaworld-reach
data_files:
- split: train
path: metaworld-reach/train-*
- split: test
path: metaworld-reach/test-*
- config_name: metaworld-reach-wall
data_files:
- split: train
path: metaworld-reach-wall/train-*
- split: test
path: metaworld-reach-wall/test-*
- config_name: metaworld-shelf-place
data_files:
- split: train
path: metaworld-shelf-place/train-*
- split: test
path: metaworld-shelf-place/test-*
- config_name: metaworld-soccer
data_files:
- split: train
path: metaworld-soccer/train-*
- split: test
path: metaworld-soccer/test-*
- config_name: metaworld-stick-pull
data_files:
- split: train
path: metaworld-stick-pull/train-*
- split: test
path: metaworld-stick-pull/test-*
- config_name: metaworld-stick-push
data_files:
- split: train
path: metaworld-stick-push/train-*
- split: test
path: metaworld-stick-push/test-*
- config_name: metaworld-sweep
data_files:
- split: train
path: metaworld-sweep/train-*
- split: test
path: metaworld-sweep/test-*
- config_name: metaworld-sweep-into
data_files:
- split: train
path: metaworld-sweep-into/train-*
- split: test
path: metaworld-sweep-into/test-*
- config_name: metaworld-window-close
data_files:
- split: train
path: metaworld-window-close/train-*
- split: test
path: metaworld-window-close/test-*
- config_name: metaworld-window-open
data_files:
- split: train
path: metaworld-window-open/train-*
- split: test
path: metaworld-window-open/test-*
- config_name: mujoco-ant
data_files:
- split: train
path: mujoco-ant/train-*
- split: test
path: mujoco-ant/test-*
- config_name: mujoco-doublependulum
data_files:
- split: train
path: mujoco-doublependulum/train-*
- split: test
path: mujoco-doublependulum/test-*
- config_name: mujoco-halfcheetah
data_files:
- split: train
path: mujoco-halfcheetah/train-*
- split: test
path: mujoco-halfcheetah/test-*
- config_name: mujoco-hopper
data_files:
- split: train
path: mujoco-hopper/train-*
- split: test
path: mujoco-hopper/test-*
- config_name: mujoco-humanoid
data_files:
- split: train
path: mujoco-humanoid/train-*
- split: test
path: mujoco-humanoid/test-*
- config_name: mujoco-pendulum
data_files:
- split: train
path: mujoco-pendulum/train-*
- split: test
path: mujoco-pendulum/test-*
- config_name: mujoco-pusher
data_files:
- split: train
path: mujoco-pusher/train-*
- split: test
path: mujoco-pusher/test-*
- config_name: mujoco-reacher
data_files:
- split: train
path: mujoco-reacher/train-*
- split: test
path: mujoco-reacher/test-*
- config_name: mujoco-standup
data_files:
- split: train
path: mujoco-standup/train-*
- split: test
path: mujoco-standup/test-*
- config_name: mujoco-swimmer
data_files:
- split: train
path: mujoco-swimmer/train-*
- split: test
path: mujoco-swimmer/test-*
- config_name: mujoco-walker
data_files:
- split: train
path: mujoco-walker/train-*
- split: test
path: mujoco-walker/test-*
- config_name: ok-vqa
data_files:
- split: train
path: ok-vqa/train-*
- split: test
path: ok-vqa/test-*
tags:
- imitation-learning
- reinforcement-learning
- text-generation
- question-answering
- generalist-agent
dataset_info:
- config_name: atari-alien
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1340568536.0
num_examples: 97
- name: test
num_bytes: 140147997.0
num_examples: 11
download_size: 139482052
dataset_size: 1480716533.0
- config_name: atari-amidar
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 839195896.0
num_examples: 146
- name: test
num_bytes: 76328889.0
num_examples: 17
download_size: 849996308
dataset_size: 915524785.0
- config_name: atari-assault
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 798961431.0
num_examples: 53
- name: test
num_bytes: 70630737.0
num_examples: 6
download_size: 856465142
dataset_size: 869592168.0
- config_name: atari-asterix
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 981904668.0
num_examples: 470
- name: test
num_bytes: 94826831.0
num_examples: 53
download_size: 1025083959
dataset_size: 1076731499.0
- config_name: atari-asteroids
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 774344616.0
num_examples: 17
- name: test
num_bytes: 52617462.0
num_examples: 2
download_size: 815573512
dataset_size: 826962078.0
- config_name: atari-atlantis
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 915242786.0
num_examples: 44
- name: test
num_bytes: 68743372.0
num_examples: 5
download_size: 969604640
dataset_size: 983986158.0
- config_name: atari-bankheist
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1623230516.0
num_examples: 222
- name: test
num_bytes: 182769923.0
num_examples: 25
download_size: 1743163262
dataset_size: 1806000439.0
- config_name: atari-battlezone
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1406320758.0
num_examples: 97
- name: test
num_bytes: 167008797.0
num_examples: 11
download_size: 640049534
dataset_size: 1573329555.0
- config_name: atari-beamrider
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1028942918.0
num_examples: 46
- name: test
num_bytes: 165781602.0
num_examples: 6
download_size: 1190822803
dataset_size: 1194724520.0
- config_name: atari-berzerk
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 599497245.0
num_examples: 17
- name: test
num_bytes: 75010244.0
num_examples: 2
download_size: 652845047
dataset_size: 674507489.0
- config_name: atari-bowling
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 546770697.0
num_examples: 193
- name: test
num_bytes: 62611921.0
num_examples: 22
download_size: 534548773
dataset_size: 609382618.0
- config_name: atari-boxing
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1081525678.975
num_examples: 1025
- name: test
num_bytes: 119411032.0
num_examples: 114
download_size: 1196687855
dataset_size: 1200936710.975
- config_name: atari-breakout
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 449338850.0
num_examples: 32
- name: test
num_bytes: 57704753.0
num_examples: 4
download_size: 355232930
dataset_size: 507043603.0
- config_name: atari-centipede
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 740721041.0
num_examples: 460
- name: test
num_bytes: 85208346.0
num_examples: 52
download_size: 819207107
dataset_size: 825929387.0
- config_name: atari-choppercommand
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 989964507.0
num_examples: 144
- name: test
num_bytes: 147199310.0
num_examples: 16
download_size: 1131175930
dataset_size: 1137163817.0
- config_name: atari-crazyclimber
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1246068403.0
num_examples: 88
- name: test
num_bytes: 139541935.0
num_examples: 10
download_size: 1294452085
dataset_size: 1385610338.0
- config_name: atari-defender
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 631539225.0
num_examples: 16
- name: test
num_bytes: 78383287.0
num_examples: 2
download_size: 620482245
dataset_size: 709922512.0
- config_name: atari-demonattack
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 624524718.0
num_examples: 18
- name: test
num_bytes: 77648737.0
num_examples: 2
download_size: 692930877
dataset_size: 702173455.0
- config_name: atari-doubledunk
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: test
num_bytes: 123241754.0
num_examples: 51
- name: train
num_bytes: 1109840257.0
num_examples: 456
download_size: 1208221748
dataset_size: 1233082011.0
- config_name: atari-enduro
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1341529954.0
num_examples: 16
- name: test
num_bytes: 170147714.0
num_examples: 2
download_size: 1506759932
dataset_size: 1511677668.0
- config_name: atari-fishingderby
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1515746411.0
num_examples: 275
- name: test
num_bytes: 179086977.0
num_examples: 31
download_size: 1692400820
dataset_size: 1694833388.0
- config_name: atari-freeway
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1109519748.0
num_examples: 219
- name: test
num_bytes: 126516219.0
num_examples: 25
download_size: 1232267662
dataset_size: 1236035967.0
- config_name: atari-frostbite
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1461470198.0
num_examples: 188
- name: test
num_bytes: 168294758.0
num_examples: 21
download_size: 1623699715
dataset_size: 1629764956.0
- config_name: atari-gopher
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 838220280.0
num_examples: 23
- name: test
num_bytes: 112043092.0
num_examples: 3
download_size: 942000464
dataset_size: 950263372.0
- config_name: atari-gravitar
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 795642642.0
num_examples: 750
- name: test
num_bytes: 88650726.0
num_examples: 84
download_size: 877506629
dataset_size: 884293368.0
- config_name: atari-hero
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1093415256.0
num_examples: 166
- name: test
num_bytes: 125418914.0
num_examples: 19
download_size: 1203346008
dataset_size: 1218834170.0
- config_name: atari-icehockey
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 764843072.0
num_examples: 118
- name: test
num_bytes: 87267657.0
num_examples: 14
download_size: 778055672
dataset_size: 852110729.0
- config_name: atari-jamesbond
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 735033584.0
num_examples: 54
- name: test
num_bytes: 168937080.0
num_examples: 7
download_size: 899088453
dataset_size: 903970664.0
- config_name: atari-kangaroo
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1040140729.0
num_examples: 495
- name: test
num_bytes: 112177810.0
num_examples: 56
download_size: 1148401746
dataset_size: 1152318539.0
- config_name: atari-krull
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 2283525995.0
num_examples: 318
- name: test
num_bytes: 253656157.0
num_examples: 36
download_size: 2526820904
dataset_size: 2537182152.0
- config_name: atari-kungfumaster
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1459405811.0
num_examples: 150
- name: test
num_bytes: 175710328.0
num_examples: 17
download_size: 1609871392
dataset_size: 1635116139.0
- config_name: atari-montezumarevenge
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1358041617.0
num_examples: 389
- name: test
num_bytes: 151969510.0
num_examples: 44
download_size: 1496389769
dataset_size: 1510011127.0
- config_name: atari-mspacman
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1450638504.0
num_examples: 179
- name: test
num_bytes: 158188150.0
num_examples: 20
download_size: 157083760
dataset_size: 1608826654.0
- config_name: atari-namethisgame
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1303134716.0
num_examples: 45
- name: test
num_bytes: 180906060.0
num_examples: 6
download_size: 1480907677
dataset_size: 1484040776.0
- config_name: atari-phoenix
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 710710054.0
num_examples: 17
- name: test
num_bytes: 90041382.0
num_examples: 2
download_size: 789132045
dataset_size: 800751436.0
- config_name: atari-pitfall
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1038921456.0
num_examples: 42
- name: test
num_bytes: 95477942.0
num_examples: 5
download_size: 563920504
dataset_size: 1134399398.0
- config_name: atari-pong
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: test
num_bytes: 42460330.0
num_examples: 31
- name: train
num_bytes: 372438874.0
num_examples: 272
download_size: 340157509
dataset_size: 414899204.0
- config_name: atari-privateeye
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: test
num_bytes: 188566614.0
num_examples: 19
- name: train
num_bytes: 1646331664.0
num_examples: 166
download_size: 999585816
dataset_size: 1834898278.0
- config_name: atari-qbert
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: test
num_bytes: 212314952.0
num_examples: 12
- name: train
num_bytes: 1906885976.0
num_examples: 105
download_size: 2114236276
dataset_size: 2119200928.0
- config_name: atari-riverraid
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: test
num_bytes: 138639529.0
num_examples: 31
- name: train
num_bytes: 1336041601.0
num_examples: 277
download_size: 1451357887
dataset_size: 1474681130.0
- config_name: atari-roadrunner
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: test
num_bytes: 102119437.0
num_examples: 24
- name: train
num_bytes: 913351876.0
num_examples: 212
download_size: 1001454818
dataset_size: 1015471313.0
- config_name: atari-robotank
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: test
num_bytes: 128435803.0
num_examples: 7
- name: train
num_bytes: 1292214032.0
num_examples: 63
download_size: 1388205947
dataset_size: 1420649835.0
- config_name: atari-seaquest
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: test
num_bytes: 91834003.0
num_examples: 24
- name: train
num_bytes: 828174074.0
num_examples: 209
download_size: 908365754
dataset_size: 920008077.0
- config_name: atari-skiing
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1141286076.0
num_examples: 917
- name: test
num_bytes: 127551492.0
num_examples: 102
download_size: 1265105500
dataset_size: 1268837568.0
- config_name: atari-solaris
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 1146266482.0
num_examples: 34
- name: test
num_bytes: 122871787.0
num_examples: 4
download_size: 1257863864
dataset_size: 1269138269.0
- config_name: atari-spaceinvaders
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 888515140.0
num_examples: 30
- name: test
num_bytes: 183628032.0
num_examples: 4
download_size: 1044841686
dataset_size: 1072143172.0
- config_name: atari-stargunner
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 615092285.0
num_examples: 31
- name: test
num_bytes: 71315788.0
num_examples: 4
download_size: 677077474
dataset_size: 686408073.0
- config_name: atari-surround
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 526004197.0
num_examples: 144
- name: test
num_bytes: 67282927.0
num_examples: 17
download_size: 532120267
dataset_size: 593287124.0
- config_name: atari-tennis
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 709632525.0
num_examples: 49
- name: test
num_bytes: 76212648.0
num_examples: 6
download_size: 539956655
dataset_size: 785845173.0
- config_name: atari-timepilot
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 849962378.0
num_examples: 48
- name: test
num_bytes: 95939303.0
num_examples: 6
download_size: 919663541
dataset_size: 945901681.0
- config_name: atari-tutankham
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 833317180.0
num_examples: 27
- name: test
num_bytes: 137596199.0
num_examples: 4
download_size: 528781594
dataset_size: 970913379.0
- config_name: atari-upndown
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: train
num_bytes: 2963452811.0
num_examples: 16
- name: test
num_bytes: 371856958.0
num_examples: 2
download_size: 3320647022
dataset_size: 3335309769.0
- config_name: atari-venture
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: test
num_bytes: 88888187.0
num_examples: 25
- name: train
num_bytes: 884080096.0
num_examples: 216
download_size: 869134091
dataset_size: 972968283.0
- config_name: atari-videopinball
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: test
num_bytes: 50315326.0
num_examples: 3
- name: train
num_bytes: 1330483745.0
num_examples: 22
download_size: 1377534468
dataset_size: 1380799071.0
- config_name: atari-wizardofwor
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: test
num_bytes: 121295756.0
num_examples: 14
- name: train
num_bytes: 1015986420.0
num_examples: 124
download_size: 1082615829
dataset_size: 1137282176.0
- config_name: atari-yarsrevenge
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: test
num_bytes: 278195918.0
num_examples: 4
- name: train
num_bytes: 2348309471.0
num_examples: 31
download_size: 1988218999
dataset_size: 2626505389.0
- config_name: atari-zaxxon
features:
- name: image_observations
sequence: image
- name: rewards
sequence: float32
- name: discrete_actions
sequence: int64
splits:
- name: test
num_bytes: 117311384.0
num_examples: 8
- name: train
num_bytes: 982507552.0
num_examples: 64
download_size: 1093792295
dataset_size: 1099818936.0
- config_name: babyai-action-obj-door
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 730102581
num_examples: 95000
- name: test
num_bytes: 38820823
num_examples: 5000
download_size: 15937785
dataset_size: 768923404
- config_name: babyai-blocked-unlock-pickup
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 207846215
num_examples: 5000
- name: train
num_bytes: 3944315285
num_examples: 95000
download_size: 47671576
dataset_size: 4152161500
- config_name: babyai-boss-level
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 524421727
num_examples: 5000
- name: train
num_bytes: 10122220692
num_examples: 95000
download_size: 171013846
dataset_size: 10646642419
- config_name: babyai-boss-level-no-unlock
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 512206014
num_examples: 5000
- name: train
num_bytes: 9951813143
num_examples: 95000
download_size: 166637143
dataset_size: 10464019157
- config_name: babyai-find-obj-s5
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 3525778032
num_examples: 95000
- name: test
num_bytes: 183685740
num_examples: 5000
download_size: 49738428
dataset_size: 3709463772
- config_name: babyai-go-to
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 6152451450
num_examples: 95000
- name: test
num_bytes: 319842603
num_examples: 5000
download_size: 101378644
dataset_size: 6472294053
- config_name: babyai-go-to-door
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 615768109
num_examples: 95000
- name: test
num_bytes: 32599120
num_examples: 5000
download_size: 8940753
dataset_size: 648367229
- config_name: babyai-go-to-local
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 618625078
num_examples: 95000
- name: test
num_bytes: 32783633
num_examples: 5000
download_size: 14568281
dataset_size: 651408711
- config_name: babyai-go-to-obj
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 576503446
num_examples: 95000
- name: test
num_bytes: 30207684
num_examples: 5000
download_size: 8102560
dataset_size: 606711130
- config_name: babyai-go-to-obj-door
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 698247097
num_examples: 95000
- name: test
num_bytes: 36554007
num_examples: 5000
download_size: 18138758
dataset_size: 734801104
- config_name: babyai-go-to-red-ball
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 617255758
num_examples: 95000
- name: test
num_bytes: 32552614
num_examples: 5000
download_size: 14101801
dataset_size: 649808372
- config_name: babyai-go-to-red-ball-grey
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 685059164
num_examples: 95000
- name: test
num_bytes: 36316718
num_examples: 5000
download_size: 14234379
dataset_size: 721375882
- config_name: babyai-go-to-red-ball-no-dists
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 575338070
num_examples: 95000
- name: test
num_bytes: 30355826
num_examples: 5000
download_size: 7108473
dataset_size: 605693896
- config_name: babyai-go-to-red-blue-ball
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 684110113
num_examples: 95000
- name: test
num_bytes: 36050340
num_examples: 5000
download_size: 15617708
dataset_size: 720160453
- config_name: babyai-go-to-seq
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 8659717841
num_examples: 95000
- name: test
num_bytes: 457950086
num_examples: 5000
download_size: 142792284
dataset_size: 9117667927
- config_name: babyai-key-corridor
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 673861952
num_examples: 5000
- name: train
num_bytes: 12830544960
num_examples: 95000
download_size: 192785385
dataset_size: 13504406912
- config_name: babyai-key-in-box
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 9114424
num_examples: 5000
- name: train
num_bytes: 170766251
num_examples: 95000
download_size: 2402474
dataset_size: 179880675
- config_name: babyai-mini-boss-level
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 165697671
num_examples: 5000
- name: train
num_bytes: 3160839261
num_examples: 95000
download_size: 49046590
dataset_size: 3326536932
- config_name: babyai-move-two-across-s8n9
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 263104296
num_examples: 5000
- name: train
num_bytes: 5010029188
num_examples: 95000
download_size: 67260892
dataset_size: 5273133484
- config_name: babyai-one-room-s8
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 35849856
num_examples: 5000
- name: train
num_bytes: 678323712
num_examples: 95000
download_size: 8726372
dataset_size: 714173568
- config_name: babyai-open
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 184341054
num_examples: 5000
- name: train
num_bytes: 3552284018
num_examples: 95000
download_size: 2850718
dataset_size: 3736625072
- config_name: babyai-open-door
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 44954852
num_examples: 5000
- name: train
num_bytes: 857776914
num_examples: 95000
download_size: 11397484
dataset_size: 902731766
- config_name: babyai-open-doors-order-n4
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 65109790
num_examples: 5000
- name: train
num_bytes: 1224959587
num_examples: 95000
download_size: 14918459
dataset_size: 1290069377
- config_name: babyai-open-red-door
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 28865701
num_examples: 5000
- name: train
num_bytes: 547345717
num_examples: 95000
download_size: 2723624
dataset_size: 576211418
- config_name: babyai-open-two-doors
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 85096451
num_examples: 5000
- name: train
num_bytes: 1614499890
num_examples: 95000
download_size: 12535076
dataset_size: 1699596341
- config_name: babyai-pickup
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 324751988
num_examples: 5000
- name: train
num_bytes: 6247776138
num_examples: 95000
download_size: 103094535
dataset_size: 6572528126
- config_name: babyai-pickup-above
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 181653115
num_examples: 5000
- name: train
num_bytes: 3399366642
num_examples: 95000
download_size: 47780316
dataset_size: 3581019757
- config_name: babyai-pickup-dist
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 29384140
num_examples: 5000
- name: train
num_bytes: 555920169
num_examples: 95000
download_size: 10606303
dataset_size: 585304309
- config_name: babyai-pickup-loc
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 36556968
num_examples: 5000
- name: train
num_bytes: 709012750
num_examples: 95000
download_size: 15292435
dataset_size: 745569718
- config_name: babyai-synth
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 307405687
num_examples: 5000
- name: train
num_bytes: 5948279603
num_examples: 95000
download_size: 100838075
dataset_size: 6255685290
- config_name: babyai-synth-loc
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 290016584
num_examples: 5000
- name: train
num_bytes: 5488393137
num_examples: 95000
download_size: 93570653
dataset_size: 5778409721
- config_name: babyai-synth-seq
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 489211184
num_examples: 5000
- name: train
num_bytes: 9238807765
num_examples: 95000
download_size: 140373267
dataset_size: 9728018949
- config_name: babyai-unblock-pickup
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 349148205
num_examples: 5000
- name: train
num_bytes: 6483599187
num_examples: 95000
download_size: 109831237
dataset_size: 6832747392
- config_name: babyai-unlock-local
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 85036094
num_examples: 5000
- name: train
num_bytes: 1620777960
num_examples: 95000
download_size: 21461309
dataset_size: 1705814054
- config_name: babyai-unlock-pickup
features:
- name: text_observations
sequence: string
- name: discrete_observations
sequence:
sequence: int64
length: 148
- name: discrete_actions
sequence: int64
- name: rewards
sequence: float32
splits:
- name: test
num_bytes: 120199548
num_examples: 5000
- name: train
num_bytes: 2279983679
num_examples: 95000
download_size: 26099013
dataset_size: 2400183227
- config_name: conceptual-captions
features:
- name: images
dtype: image
- name: text
dtype: string
splits:
- name: test
num_bytes: 1564922274.875
num_examples: 12465
- name: train
num_bytes: 321742591779.0
num_examples: 2620472
download_size: 7559495686
dataset_size: 323307514053.875
- config_name: metaworld-assembly
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 31556512
dataset_size: 309971200
- config_name: metaworld-basketball
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 13457975
dataset_size: 309971200
- config_name: metaworld-bin-picking
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 148239551
dataset_size: 309971200
- config_name: metaworld-box-close
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 155046141
dataset_size: 309971200
- config_name: metaworld-button-press
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 92407404
dataset_size: 309971200
- config_name: metaworld-button-press-topdown
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 99643997
dataset_size: 309971200
- config_name: metaworld-button-press-topdown-wall
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 102330609
dataset_size: 309971200
- config_name: metaworld-button-press-wall
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 98686929
dataset_size: 309971200
- config_name: metaworld-coffee-button
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 98541376
dataset_size: 309971200
- config_name: metaworld-coffee-pull
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 141657803
dataset_size: 309971200
- config_name: metaworld-coffee-push
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 153493123
dataset_size: 309971200
- config_name: metaworld-dial-turn
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 90092180
dataset_size: 309971200
- config_name: metaworld-disassemble
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 55699141
dataset_size: 309971200
- config_name: metaworld-door-close
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 132047898
dataset_size: 309971200
- config_name: metaworld-door-lock
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 108135090
dataset_size: 309971200
- config_name: metaworld-door-open
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 123463142
dataset_size: 309971200
- config_name: metaworld-door-unlock
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 107047389
dataset_size: 309971200
- config_name: metaworld-drawer-close
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 86742866
dataset_size: 309971200
- config_name: metaworld-drawer-open
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 87426230
dataset_size: 309971200
- config_name: metaworld-faucet-close
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 75525957
dataset_size: 309971200
- config_name: metaworld-faucet-open
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 82798110
dataset_size: 309971200
- config_name: metaworld-hammer
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 156766229
dataset_size: 309971200
- config_name: metaworld-hand-insert
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 115425570
dataset_size: 309971200
- config_name: metaworld-handle-press
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 88721833
dataset_size: 309971200
- config_name: metaworld-handle-press-side
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 90271855
dataset_size: 309971200
- config_name: metaworld-handle-pull
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 106520317
dataset_size: 309971200
- config_name: metaworld-handle-pull-side
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 104725703
dataset_size: 309971200
- config_name: metaworld-lever-pull
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 147893313
dataset_size: 309971200
- config_name: metaworld-peg-insert-side
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 133765390
dataset_size: 309971200
- config_name: metaworld-peg-unplug-side
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 152488362
dataset_size: 309971200
- config_name: metaworld-pick-out-of-hole
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 15063825
dataset_size: 309971200
- config_name: metaworld-pick-place
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 156685126
dataset_size: 309971200
- config_name: metaworld-pick-place-wall
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 152697114
dataset_size: 309971200
- config_name: metaworld-plate-slide
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 91689118
dataset_size: 309971200
- config_name: metaworld-plate-slide-back
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 17682663
dataset_size: 309971200
- config_name: metaworld-plate-slide-back-side
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 16397415
dataset_size: 309971200
- config_name: metaworld-plate-slide-side
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 88672818
dataset_size: 309971200
- config_name: metaworld-push
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 146425498
dataset_size: 309971200
- config_name: metaworld-push-back
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 115758693
dataset_size: 309971200
- config_name: metaworld-push-wall
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 138978942
dataset_size: 309971200
- config_name: metaworld-reach
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 151264193
dataset_size: 309971200
- config_name: metaworld-reach-wall
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 153008204
dataset_size: 309971200
- config_name: metaworld-shelf-place
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 126421788
dataset_size: 309971200
- config_name: metaworld-soccer
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 139325515
dataset_size: 309971200
- config_name: metaworld-stick-pull
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 150611675
dataset_size: 309971200
- config_name: metaworld-stick-push
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 145549289
dataset_size: 309971200
- config_name: metaworld-sweep
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 144411349
dataset_size: 309971200
- config_name: metaworld-sweep-into
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 116977226
dataset_size: 309971200
- config_name: metaworld-window-close
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 82738762
dataset_size: 309971200
- config_name: metaworld-window-open
features:
- name: continuous_observations
sequence:
sequence: float32
length: 39
- name: continuous_actions
sequence:
sequence: float32
length: 4
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 281792000
num_examples: 16000
- name: test
num_bytes: 28179200
num_examples: 1600
download_size: 82547802
dataset_size: 309971200
- config_name: mujoco-ant
features:
- name: continuous_observations
sequence:
sequence: float32
- name: continuous_actions
sequence:
sequence: float32
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 1334666176
num_examples: 9000
- name: test
num_bytes: 149007264
num_examples: 1000
download_size: 1427489194
dataset_size: 1483673440
- config_name: mujoco-doublependulum
features:
- name: continuous_observations
sequence:
sequence: float32
- name: continuous_actions
sequence:
sequence: float32
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 539380200
num_examples: 9000
- name: test
num_bytes: 59838360
num_examples: 1000
download_size: 423057943
dataset_size: 599218560
- config_name: mujoco-halfcheetah
features:
- name: continuous_observations
sequence:
sequence: float32
- name: continuous_actions
sequence:
sequence: float32
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 936108000
num_examples: 9000
- name: test
num_bytes: 104012000
num_examples: 1000
download_size: 983767586
dataset_size: 1040120000
- config_name: mujoco-hopper
features:
- name: continuous_observations
sequence:
sequence: float32
- name: continuous_actions
sequence:
sequence: float32
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 277504480
num_examples: 9000
- name: test
num_bytes: 30493476
num_examples: 1000
download_size: 291016996
dataset_size: 307997956
- config_name: mujoco-humanoid
features:
- name: continuous_observations
sequence:
sequence: float32
- name: rewards
sequence: float32
- name: continuous_actions
sequence:
sequence: float32
splits:
- name: train
num_bytes: 12855318192
num_examples: 9000
- name: test
num_bytes: 1436554272
num_examples: 1000
download_size: 10321727430
dataset_size: 14291872464
- config_name: mujoco-pendulum
features:
- name: continuous_observations
sequence:
sequence: float32
- name: continuous_actions
sequence:
sequence: float32
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 137118592
num_examples: 9000
- name: test
num_bytes: 15128704
num_examples: 1000
download_size: 107926228
dataset_size: 152247296
- config_name: mujoco-pusher
features:
- name: continuous_observations
sequence:
sequence: float32
- name: continuous_actions
sequence:
sequence: float32
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 118908000
num_examples: 9000
- name: test
num_bytes: 13212000
num_examples: 1000
download_size: 124763158
dataset_size: 132120000
- config_name: mujoco-reacher
features:
- name: continuous_observations
sequence:
sequence: float32
- name: continuous_actions
sequence:
sequence: float32
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 28908000
num_examples: 9000
- name: test
num_bytes: 3212000
num_examples: 1000
download_size: 34000959
dataset_size: 32120000
- config_name: mujoco-standup
features:
- name: rewards
sequence: float32
- name: continuous_observations
sequence:
sequence: float32
- name: continuous_actions
sequence:
sequence: float32
splits:
- name: train
num_bytes: 14256108000
num_examples: 9000
- name: test
num_bytes: 1584012000
num_examples: 1000
download_size: 1163281621
dataset_size: 15840120000
- config_name: mujoco-swimmer
features:
- name: continuous_observations
sequence:
sequence: float32
- name: continuous_actions
sequence:
sequence: float32
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 468108000
num_examples: 9000
- name: test
num_bytes: 52012000
num_examples: 1000
download_size: 459798751
dataset_size: 520120000
- config_name: mujoco-walker
features:
- name: continuous_observations
sequence:
sequence: float32
- name: continuous_actions
sequence:
sequence: float32
- name: rewards
sequence: float32
splits:
- name: train
num_bytes: 858590040
num_examples: 9000
- name: test
num_bytes: 95183024
num_examples: 1000
download_size: 892883623
dataset_size: 953773064
- config_name: ok-vqa
features:
- name: images
dtype: image
- name: text
dtype: string
splits:
- name: train
num_bytes: 149757863.0
num_examples: 9009
- name: test
num_bytes: 84544434.0
num_examples: 5046
download_size: 233832618
dataset_size: 234302297.0
---
# GIA Dataset
## Dataset Description
The GIA dataset combines a wide range of individual datasets. It includes expert demonstrations by expert RL agents, image and caption pairs, textual data and more. The GIA dataset is part of the GIA project, which aims to build a multimodal generalist agent.
### Usage
```python
>>> from datasets import load_dataset
>>> dataset = load_dataset("gia-project/gia-dataset", "metaworld-assembly")
>>> first_episode = dataset["train"][0]
>>> first_episode.keys()
dict_keys(['continuous_observations', 'continuous_actions', 'rewards'])
>>> len(first_episode["rewards"])
500
>>> first_episode["continuous_actions"][0]
[6.459120273590088, 2.2422609329223633, -5.914587020874023, -19.799840927124023]
```
## Dataset Structure
### Data Instances
<details>
<summary>Click to expand the score information for each task</summary>
The following table presents a comparative analysis of scores across various domains and tasks. The scores highlight the performance difference between a random agent and the episodes recorded in our dataset.
| Task | Random Agent Score | Dataset Episode Score |
| ----------------------------------- | :-----------------: | :-------------------: |
| **Atari** | | |
| atari-alien | 205.50 ± 111.97 | 16912.50 ± 7087.42 |
| atari-amidar | 2.38 ± 2.50 | 2164.71 ± 1229.47 |
| atari-assault | 262.50 ± 89.61 | 15699.12 ± 9572.12 |
| atari-asterix | 213.50 ± 110.87 | 3699.62 ± 2421.30 |
| atari-asteroids | 856.40 ± 434.32 | 177011.05 ± 35334.20 |
| atari-atlantis | 17764.00 ± 6662.43 | 320679.59 ± 418247.37 |
| atari-bankheist | 13.40 ± 11.07 | 1322.43 ± 60.84 |
| atari-battlezone | 2170.00 ± 2121.58 | 295592.59 ± 161960.96 |
| atari-beamrider | 357.28 ± 143.97 | 29589.35 ± 16132.96 |
| atari-berzerk | 160.10 ± 118.87 | 57085.26 ± 13104.53 |
| atari-bowling | 23.81 ± 6.07 | 20.40 ± 7.29 |
| atari-boxing | 0.52 ± 4.37 | 97.97 ± 3.77 |
| atari-breakout | 1.24 ± 1.30 | 702.97 ± 203.62 |
| atari-centipede | 2150.06 ± 1113.28 | 11624.29 ± 4918.34 |
| atari-choppercommand | 875.00 ± 416.98 | 90990.62 ± 270876.93 |
| atari-crazyclimber | 7376.00 ± 2253.09 | 179296.94 ± 39862.06 |
| atari-defender | 3417.50 ± 1443.41 | 351958.33 ± 40466.82 |
| atari-demonattack | 165.55 ± 92.93 | 92195.25 ± 26174.79 |
| atari-doubledunk | -18.54 ± 3.07 | 20.94 ± 3.65 |
| atari-enduro | 0.00 ± 0.00 | 2292.22 ± 147.54 |
| atari-fishingderby | -93.90 ± 3.51 | 7.18 ± 25.06 |
| atari-freeway | 0.01 ± 0.10 | 33.88 ± 0.35 |
| atari-frostbite | 67.60 ± 37.61 | 13196.12 ± 4341.00 |
| atari-gopher | 319.40 ± 228.24 | 81676.15 ± 46329.48 |
| atari-gravitar | 188.50 ± 203.33 | 3986.57 ± 1729.05 |
| atari-hero | 475.25 ± 894.95 | 44677.35 ± 1754.42 |
| atari-icehockey | -9.83 ± 3.24 | 25.17 ± 5.79 |
| atari-jamesbond | 28.50 ± 45.42 | 27786.89 ± 33819.20 |
| atari-kangaroo | 52.00 ± 108.15 | 574.05 ± 636.94 |
| atari-krull | 1754.00 ± 583.56 | 11439.83 ± 1218.34 |
| atari-kungfumaster | 390.00 ± 359.03 | 32392.81 ± 10006.55 |
| atari-montezumarevenge | 0.00 ± 0.00 | 393.53 ± 50.45 |
| atari-mspacman | 246.40 ± 121.22 | 6896.08 ± 2031.99 |
| atari-namethisgame | 2447.40 ± 888.97 | 22991.18 ± 2473.15 |
| atari-phoenix | 776.80 ± 635.86 | 424583.16 ± 97649.17 |
| atari-pitfall | -259.75 ± 384.26 | -1.45 ± 4.50 |
| atari-pong | -20.22 ± 0.95 | 20.99 ± 0.18 |
| atari-privateeye | 41.65 ± 191.83 | 100.00 ± 0.00 |
| atari-qbert | 164.25 ± 151.79 | 42971.37 ± 85070.72 |
| atari-riverraid | 1474.40 ± 314.59 | 14800.94 ± 7924.56 |
| atari-roadrunner | 11.00 ± 42.18 | 77942.80 ± 6088.62 |
| atari-robotank | 1.87 ± 1.59 | 80.51 ± 13.28 |
| atari-seaquest | 73.20 ± 57.91 | 2597.34 ± 386.09 |
| atari-skiing | -16299.52 ± 1850.70 | -10738.06 ± 111.13 |
| atari-solaris | 2360.40 ± 1852.03 | 1353.68 ± 516.96 |
| atari-spaceinvaders | 137.20 ± 95.82 | 29425.29 ± 23623.89 |
| atari-stargunner | 652.00 ± 312.24 | 360588.57 ± 49207.71 |
| atari-surround | -9.99 ± 0.10 | 9.39 ± 0.85 |
| atari-tennis | -23.95 ± 0.22 | 11.11 ± 7.57 |
| atari-timepilot | 3396.00 ± 2128.85 | 69583.33 ± 29838.67 |
| atari-tutankham | 12.73 ± 17.40 | 291.16 ± 30.37 |
| atari-upndown | 358.90 ± 380.11 | 429418.33 ± 7187.43 |
| atari-venture | 0.00 ± 0.00 | 0.00 ± 0.00 |
| atari-videopinball | 23917.17 ± 19449.59 | 441507.92 ± 283264.62 |
| atari-wizardofwor | 620.00 ± 837.85 | 49333.33 ± 16157.08 |
| atari-yarsrevenge | 3503.91 ± 906.14 | 270262.86 ± 161815.96 |
| atari-zaxxon | 21.00 ± 102.27 | 73097.22 ± 14825.77 |
| **BabyAI** | | |
| babyai-action-obj-door | 0.37 ± 0.39 | 0.99 ± 0.01 |
| babyai-blocked-unlock-pickup | 0.00 ± 0.02 | 0.95 ± 0.01 |
| babyai-boss-level | 0.06 ± 0.21 | 0.94 ± 0.05 |
| babyai-boss-level-no-unlock | 0.06 ± 0.19 | 0.94 ± 0.05 |
| babyai-find-obj-s5 | 0.08 ± 0.23 | 0.95 ± 0.04 |
| babyai-go-to | 0.13 ± 0.29 | 0.92 ± 0.07 |
| babyai-go-to-door | 0.45 ± 0.38 | 0.99 ± 0.00 |
| babyai-go-to-imp-unlock | 0.08 ± 0.23 | |
| babyai-go-to-local | 0.16 ± 0.30 | 0.93 ± 0.04 |
| babyai-go-to-obj | 0.13 ± 0.27 | 0.93 ± 0.03 |
| babyai-go-to-obj-door | 0.53 ± 0.39 | 0.99 ± 0.01 |
| babyai-go-to-red-ball | 0.17 ± 0.30 | 0.93 ± 0.04 |
| babyai-go-to-red-ball-grey | 0.12 ± 0.27 | 0.92 ± 0.05 |
| babyai-go-to-red-ball-no-dists | 0.14 ± 0.28 | 0.93 ± 0.03 |
| babyai-go-to-red-blue-ball | 0.12 ± 0.27 | 0.92 ± 0.05 |
| babyai-go-to-seq | 0.08 ± 0.23 | 0.94 ± 0.05 |
| babyai-key-corridor | 0.00 ± 0.00 | 0.91 ± 0.01 |
| babyai-key-in-box | 0.01 ± 0.06 | |
| babyai-mini-boss-level | 0.07 ± 0.21 | 0.89 ± 0.10 |
| babyai-move-two-across-s8n9 | 0.00 ± 0.00 | 0.96 ± 0.01 |
| babyai-one-room-s8 | 0.08 ± 0.21 | 0.92 ± 0.03 |
| babyai-open | 0.10 ± 0.24 | 0.95 ± 0.05 |
| babyai-open-door | 0.23 ± 0.34 | 0.99 ± 0.00 |
| babyai-open-doors-order-n4 | 0.16 ± 0.30 | 0.99 ± 0.01 |
| babyai-open-red-door | 0.08 ± 0.21 | 0.92 ± 0.03 |
| babyai-open-two-doors | 0.08 ± 0.20 | 0.98 ± 0.00 |
| babyai-pickup | 0.08 ± 0.22 | 0.92 ± 0.07 |
| babyai-pickup-above | 0.02 ± 0.09 | 0.91 ± 0.07 |
| babyai-pickup-dist | 0.10 ± 0.24 | 0.86 ± 0.21 |
| babyai-pickup-loc | 0.08 ± 0.23 | 0.91 ± 0.04 |
| babyai-synth | 0.11 ± 0.26 | 0.93 ± 0.06 |
| babyai-synth-loc | 0.13 ± 0.29 | 0.94 ± 0.06 |
| babyai-synth-seq | 0.07 ± 0.20 | 0.95 ± 0.04 |
| babyai-unblock-pickup | 0.08 ± 0.22 | 0.91 ± 0.08 |
| babyai-unlock | 0.03 ± 0.15 | |
| babyai-unlock-local | 0.01 ± 0.09 | 0.98 ± 0.01 |
| babyai-unlock-pickup | 0.00 ± 0.00 | 0.75 ± 0.04 |
| babyai-unlock-to-unlock | 0.00 ± 0.00 | |
| **MetaWorld** | | |
| metaworld-assembly | 45.30 ± 4.13 | 245.99 ± 3.50 |
| metaworld-basketball | 2.81 ± 1.24 | 627.99 ± 1.98 |
| metaworld-bin-picking | 1.89 ± 0.45 | 425.58 ± 101.86 |
| metaworld-box-close | 76.39 ± 17.91 | 512.49 ± 107.81 |
| metaworld-button-press | 31.73 ± 5.20 | 643.10 ± 12.85 |
| metaworld-button-press-topdown | 28.97 ± 10.37 | 490.18 ± 27.21 |
| metaworld-button-press-topdown-wall | 29.04 ± 10.52 | 497.19 ± 31.37 |
| metaworld-button-press-wall | 8.98 ± 3.99 | 675.41 ± 15.04 |
| metaworld-coffee-button | 31.72 ± 6.36 | 731.08 ± 29.34 |
| metaworld-coffee-pull | 4.09 ± 0.38 | 259.86 ± 88.48 |
| metaworld-coffee-push | 4.17 ± 0.76 | 496.78 ± 118.20 |
| metaworld-dial-turn | 29.64 ± 16.67 | 793.56 ± 80.06 |
| metaworld-disassemble | 40.31 ± 7.53 | 42.83 ± 6.30 |
| metaworld-door-close | 5.30 ± 1.33 | 529.75 ± 27.24 |
| metaworld-door-lock | 112.35 ± 28.63 | 811.52 ± 34.07 |
| metaworld-door-open | 56.37 ± 11.23 | 581.94 ± 19.67 |
| metaworld-door-unlock | 94.17 ± 15.56 | 802.88 ± 17.05 |
| metaworld-drawer-close | 116.73 ± 253.11 | 867.92 ± 4.48 |
| metaworld-drawer-open | 126.85 ± 25.22 | 492.99 ± 2.52 |
| metaworld-faucet-close | 253.12 ± 22.94 | 753.92 ± 13.42 |
| metaworld-faucet-open | 244.10 ± 23.25 | 705.76 ± 7.15 |
| metaworld-hammer | 95.33 ± 9.02 | 693.17 ± 34.62 |
| metaworld-hand-insert | 2.75 ± 3.53 | 740.53 ± 36.69 |
| metaworld-handle-press | 80.41 ± 110.19 | 855.91 ± 72.75 |
| metaworld-handle-press-side | 57.00 ± 39.47 | 861.12 ± 20.01 |
| metaworld-handle-pull | 10.34 ± 13.54 | 669.35 ± 24.81 |
| metaworld-handle-pull-side | 2.13 ± 2.76 | 384.65 ± 102.89 |
| metaworld-lever-pull | 60.31 ± 15.77 | 612.04 ± 38.85 |
| metaworld-peg-insert-side | 1.71 ± 0.36 | 315.23 ± 140.07 |
| metaworld-peg-unplug-side | 4.75 ± 2.83 | 456.12 ± 81.65 |
| metaworld-pick-out-of-hole | 1.51 ± 0.24 | 219.61 ± 88.85 |
| metaworld-pick-place | 1.61 ± 0.99 | 419.10 ± 98.19 |
| metaworld-pick-place-wall | 0.00 ± 0.01 | 450.57 ± 64.10 |
| metaworld-plate-slide | 74.64 ± 13.84 | 527.01 ± 155.34 |
| metaworld-plate-slide-back | 33.47 ± 11.22 | 718.22 ± 87.41 |
| metaworld-plate-slide-back-side | 34.34 ± 11.53 | 729.61 ± 69.15 |
| metaworld-plate-slide-side | 22.61 ± 17.36 | 662.81 ± 102.81 |
| metaworld-push | 5.51 ± 2.43 | 750.57 ± 43.98 |
| metaworld-push-back | 1.21 ± 0.16 | 85.05 ± 107.12 |
| metaworld-push-wall | 6.13 ± 3.17 | 748.87 ± 10.62 |
| metaworld-reach | 149.67 ± 44.70 | 681.37 ± 133.68 |
| metaworld-reach-wall | 143.26 ± 36.56 | 746.12 ± 104.19 |
| metaworld-shelf-place | 0.00 ± 0.01 | 241.34 ± 24.60 |
| metaworld-soccer | 5.66 ± 4.61 | 375.15 ± 140.24 |
| metaworld-stick-pull | 2.64 ± 1.41 | 523.55 ± 18.94 |
| metaworld-stick-push | 2.81 ± 1.04 | 627.95 ± 10.20 |
| metaworld-sweep | 11.23 ± 7.28 | 494.85 ± 43.29 |
| metaworld-sweep-into | 12.55 ± 10.72 | 799.21 ± 19.07 |
| metaworld-window-close | 57.46 ± 7.11 | 591.30 ± 38.63 |
| metaworld-window-open | 43.36 ± 2.09 | 590.82 ± 57.08 |
| **MuJoCo** | | |
| mujoco-ant | -59.95 ± 99.62 | 5846.42 ± 942.55 |
| mujoco-doublependulum | 57.46 ± 17.54 | 9338.69 ± 352.61 |
| mujoco-halfcheetah | -284.97 ± 79.83 | 7437.77 ± 173.30 |
| mujoco-hopper | 18.38 ± 17.09 | 1858.73 ± 534.07 |
| mujoco-humanoid | 122.02 ± 35.28 | 6281.02 ± 1795.84 |
| mujoco-pendulum | 6.07 ± 3.47 | 475.40 ± 178.96 |
| mujoco-pusher | -149.69 ± 7.41 | -25.21 ± 6.66 |
| mujoco-reacher | -43.00 ± 3.91 | -5.68 ± 2.53 |
| mujoco-standup | 33135.75 ± 2481.89 | 273574.16 ± 85253.26 |
| mujoco-swimmer | 0.80 ± 10.71 | 92.18 ± 4.44 |
| mujoco-walker | 2.68 ± 6.06 | 4631.22 ± 1059.01 |
</details>
### Data Fields
- `text`: a `string` feature
- `images`: a `image` feature
- `image_observations` : a `Sequence(image)` feature
- `text_observations` : a `Sequence(string)` feature
- `discrete_observations`: a `Sequence(Sequence(int64))` feature
- `continuous_observations`: a `Sequence(Sequence(float32))` feature
- `continuous_actions`: a `Sequence(Sequence(float32))` feature
- `discrete_actions`: a `Sequence(int64)` feature
- `rewards`: a `Sequence(float32)` feature
### Data Splits
- `train`: `` examples
- `test`: `` examples
## Dataset Creation
This section describes how our dataset was created. We specifically detail how data for each domain and task were generated. The generation scripts are available in the [GIA repository](https://github.com/huggingface/gia). For RL tasks, we trained one agent per task using the [Sample Factory](https://www.samplefactory.dev). Then we used the trained agent to generate episodes.
### Atari
We used the 57 [ALE/Atari](https://github.com/Farama-Foundation/Arcade-Learning-Environment) games as our environment, configuring the following parameters for our experiments. We rendered the images in grayscale with an 84x84 pixel resolution. The agent interacted with the environment every 4 frames. Sticky actions were not used, and the raw reward (no clipping) was reported. Episodes were stored as complete, i.e. with no termination on life loss.
### BabyAI
We used BabyAI's implementation from [Minigrid](https://github.com/Farama-Foundation/Minigrid).
We reused the [bot agent](https://github.com/mila-iqia/babyai) provided with BabyAI's paper and adapted it to the new Minigrid API.
Using the bot, we generated 100.000 episodes for each of the 39 tasks of [Minigrid's BabyAI](https://minigrid.farama.org/environments/babyai/) and stored for each step:
- the mission: str
- the concatenation of the symbolic observation flattened and the direction: Array of integers of size (147,)
- the action: integer
- the reward: float
### Conceptual Captions
The [Conceptual Captions](https://github.com/google-research-datasets/conceptual-captions/tree/master) dataset, offered by Google LLC, comprises pairs of image links and their corresponding captions. Each image has been downloaded and, when required, resized to ensure the maximum dimension does not exceed 352 pixels.
### MetaWorld
We used the 50 tasks from [MetaWorld v2](https://github.com/Farama-Foundation/Metaworld). We constrained the episode to a duration of 100 timesteps, which is always sufficient to solve the task.
### MuJoCo
We used the 11 environments of Gymnasium MuJoCo.
### OK-VQA
The [OK-VQA](https://okvqa.allenai.org/index.html) dataset released by Kenneth Marino, Mohammad Rastegari, Ali Farhadi, Roozbeh Mottaghi was used.
The data were formatted to match Hugging Face dataset's requirements and images were resized such that the largest dimension is at most 352.
### OSCAR
We modified the "unshuffled_deduplicated_en" split of [OSCAR 2019](https://huggingface.co/datasets/oscar) dataset, initially put together by Pedro J. Ortiz, Benoît Sagot, and Laurent Romary and licensed under [CC BY 4.0](https://oscar-project.github.io/documentation/versions/oscar-2019/#license).
We cleaned and deduplicated the dataset using [the methods](https://github.com/bigscience-workshop/data-preparation/tree/main/preprocessing/training/01b_oscar_cleaning_and_filtering) and parameters used for the [ROOTS dataset](https://arxiv.org/abs/2303.03915) (Lurençon et al., 2023).
The dataset was splitted into 30 even shards each cleaned and deduplicated independently before being concatenated again.
## Considerations for Using the Data
### Known Issues
- Some BabyAI tasks are missing due to incompatibility with the training bot:
- `babyai-key-in-box`
- `babyai-go-to-imp-unlock`
- `babyai-unlock-to-unlock`
- `babyai-unlock`
- For some atari tasks, the episode is too long, causing an `OverflowError` when loading the dataset:
- `atari-enduro`
- For some tasks, although the score can be higher than the random agent, we can't consider the task as solved:
- `atari-bowling`
- `atari-privateeye`
- `atari-solaris`
- `atari-venture`
- `metaworld-bin-picking`
- `metaworld-disassemble`
- `metaworld-peg-insert-side`
- `metaworld-plate-slide`
- `metaworld-push-back`
### Future Developments
We plan to expand the dataset to include the following additional domains:
- [ ] DM Lab
- [ ] Sokoban
- [ ] Procgen
- [ ] DM Control Suite (w and w/o pixels)
## Additional Information
### Licensing Information
This dataset is release under the Apache 2.0 license.
### Citation Information
```bibtex
@misc{gallouedec2023giadataset,
title={GIA Dataset: A Multi-Modal, Multi-Task Learning Resource},
author={Gallouédec, Quentin and Beeching, Edward and Romac, Clément},
year={2023},
howpublished={\url{https://huggingface.co/datasets/gia-project/gia-dataset}},
note={Part of the GIA Project}
}
```
## Acknowledgment
We would like to extend our sincere gratitude to:
- [Shengyi Costa Huang](https://huggingface.co/vwxyzjn) for his invaluable assistance with the pretrained models used in this research |
seamew/THUCNewsText | 2021-08-19T00:04:34.000Z | [
"region:us"
] | seamew | null | null | null | 4 | 8,285 | Entry not found |
hf-internal-testing/cats_vs_dogs_sample | 2023-04-11T17:04:37.000Z | [
"region:us"
] | hf-internal-testing | null | \\n@Inproceedings (Conference){asirra-a-captcha-that-exploits-interest-aligned-manual-image-categorization,
author = {Elson, Jeremy and Douceur, John (JD) and Howell, Jon and Saul, Jared},
title = {Asirra: A CAPTCHA that Exploits Interest-Aligned Manual Image Categorization},
booktitle = {Proceedings of 14th ACM Conference on Computer and Communications Security (CCS)},
year = {2007},
month = {October},
publisher = {Association for Computing Machinery, Inc.},
url = {https://www.microsoft.com/en-us/research/publication/asirra-a-captcha-that-exploits-interest-aligned-manual-image-categorization/},
edition = {Proceedings of 14th ACM Conference on Computer and Communications Security (CCS)},
} | null | 0 | 8,128 | Entry not found |
BigScienceBiasEval/crows_pairs_multilingual | 2022-04-26T16:26:28.000Z | [
"license:cc-by-sa-4.0",
"arxiv:2010.00133",
"region:us"
] | BigScienceBiasEval | This is a revised version of CrowS-Pairs that measures stereotypes in language modelling in both English and French. | @inproceedings{neveol2022french,
title={French CrowS-Pairs: Extending a challenge dataset for measuring social bias in masked language models to a language other than English},
author={N{\'e}v{\'e}ol, Aur{\'e}lie and Dupont, Yoann and Bezan{\c{c}}on, Julien and Fort, Kar{\"e}n},
booktitle={ACL 2022-60th Annual Meeting of the Association for Computational Linguistics},
year={2022}
} | null | 2 | 8,099 | ---
license: cc-by-sa-4.0
---
Original from https://gitlab.inria.fr/french-crows-pairs/acl-2022-paper-data-and-code/-/tree/main/.
# Data Statement for CrowS-Pairs-fr
> **How to use this document:**
> Fill in each section according to the instructions. Give as much detail as you can, but there's no need to extrapolate. The goal is to help people understand your data when they approach it. This could be someone looking at it in ten years, or it could be you yourself looking back at the data in two years.
> For full details, the best source is the original Data Statements paper, here: https://www.aclweb.org/anthology/Q18-1041/ .
> Instruction fields are given as blockquotes; delete the instructions when you're done, and provide the file with your data, for example as "DATASTATEMENT.md". The lists in some blocks are designed to be filled in, but it's good to also leave a written description of what's happening, as well as the list. It's fine to skip some fields if the information isn't known.
> Only blockquoted content should be deleted; the final about statement should be left intact.
Data set name: Crows-Pairs-fr
Citation (if available): Névéol A, Dupont Y, Bezançon J, Fort K. French CrowS-Pairs: Extending a challenge dataset for measuring social bias in masked language models to a language other than English. Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics - ACL 2022
Data set developer(s): Aurélie Névéol, Yoann Dupont, Julien Bezançon, Karën Fort
Data statement author(s): Aurélie Névéol, Yoann Dupont
Others who contributed to this document: N/A
License: Creative Commons Attribution-ShareAlike 4.0 (CC BY-SA 4.0).
## A. CURATION RATIONALE
> *Explanation.* Which texts were included and what were the goals in selecting texts, both in the original collection and in any further sub-selection? This can be especially important in datasets too large to thoroughly inspect by hand. An explicit statement of the curation rationale can help dataset users make inferences about what other kinds of texts systems trained with them could conceivably generalize to.
The French part of the corpus was built by first translating the original 1,508 sentence pairs of the English corpus into French.
We then adapted the crowdsourcing method described by [Nangia et al. (2020)](https://arxiv.org/pdf/2010.00133) to collect additional sentences expressing a stereotype relevant to the French socio-cultural environment. Data collection is implemented through LanguageARC [(Fiumara et al., 2020)](https://www.aclweb.org/anthology/2020.cllrd-1.1.pdf), a citizen science platform supporting the development of language resources dedicated to social improvement. We created a LanguageARC project (https://languagearc.com/projects/19) to collect these additional sentences. Participants were asked to submit a statement that expressed a stereotype in French along with a selection of ten bias types: the nine bias types offered in CrowS-Pairs and the additional category _other_. We collected 210 additional sentences this way.
## B. LANGUAGE VARIETY/VARIETIES
> *Explanation.* Languages differ from each other in structural ways that can interact with NLP algorithms. Within a language, regional or social dialects can also show great variation (Chambers and Trudgill, 1998). The language and language variety should be described with a language tag from BCP-47 identifying the language variety (e.g., en-US or yue-Hant-HK), and a prose description of the language variety, glossing the BCP-47 tag and also providing further information (e.g., "English as spoken in Palo Alto, California", or "Cantonese written with traditional characters by speakers in Hong Kong who are bilingual in Mandarin").
* BCP-47 language tags: fr-FR
* Language variety description: French spoken by native French people from metropolitan France.
## C. CONTRIBUTOR DEMOGRAPHIC
> ## C. SPEAKER DEMOGRAPHIC
> *Explanation.* Sociolinguistics has found that variation (in pronunciation, prosody, word choice, and grammar) correlates with speaker demographic characteristics (Labov, 1966), as speakers use linguistic variation to construct and project identities (Eckert and Rickford, 2001). Transfer from native languages (L1) can affect the language produced by non-native (L2) speakers (Ellis, 1994, Ch. 8). A further important type of variation is disordered speech (e.g., dysarthria). Specifications include:
N/A
## D. ANNOTATOR DEMOGRAPHIC
> *Explanation.* What are the demographic characteristics of the annotators and annotation guideline developers? Their own “social address” influences their experience with language and thus their perception of what they are annotating. Specifications include:
Participants to the collection project were recruited through calls for volunteers posted to social media and mailing lists in the French research community.
## E. SPEECH SITUATION
N/A
## F. TEXT CHARACTERISTICS
> *Explanation.* Both genre and topic influence the vocabulary and structural characteristics of texts (Biber, 1995), and should be specified.
Collected data is a collection of offensive stereotyped statements in French, they might be upsetting.
Along these stereotyped statements are paired anti-stereotyped statements.
## G. RECORDING QUALITY
N/A
## H. OTHER
> *Explanation.* There may be other information of relevance as well. Please use this space to develop any further categories that are relevant for your dataset.
## I. PROVENANCE APPENDIX
Examples were gathered using the LanguageArc site and by creating a dedicated project: https://languagearc.com/projects/19
## About this document
A data statement is a characterization of a dataset that provides context to allow developers and users to better understand how experimental results might generalize, how software might be appropriately deployed, and what biases might be reflected in systems built on the software.
Data Statements are from the University of Washington. Contact: [datastatements@uw.edu](mailto:datastatements@uw.edu). This document template is licensed as [CC0](https://creativecommons.org/share-your-work/public-domain/cc0/).
This version of the markdown Data Statement is from June 4th 2020. The Data Statement template is based on worksheets distributed at the [2020 LREC workshop on Data Statements](https://sites.google.com/uw.edu/data-statements-for-nlp/), by Emily M. Bender, Batya Friedman, and Angelina McMillan-Major. Adapted to community Markdown template by Leon Dercyznski.
|
opus_books | 2022-11-03T16:47:07.000Z | [
"task_categories:translation",
"annotations_creators:found",
"language_creators:found",
"multilinguality:multilingual",
"size_categories:1K<n<10K",
"source_datasets:original",
"language:ca",
"language:de",
"language:el",
"language:en",
"language:eo",
"language:es",
"language:fi",
"language:fr",
"language:hu",
"language:it",
"language:nl",
"language:no",
"language:pl",
"language:pt",
"language:ru",
"language:sv",
"license:unknown",
"region:us"
] | null | This is a collection of copyright free books aligned by Andras Farkas, which are available from http://www.farkastranslations.com/bilingual_books.php
Note that the texts are rather dated due to copyright issues and that some of them are manually reviewed (check the meta-data at the top of the corpus files in XML). The source is multilingually aligned, which is available from http://www.farkastranslations.com/bilingual_books.php. In OPUS, the alignment is formally bilingual but the multilingual alignment can be recovered from the XCES sentence alignment files. Note also that the alignment units from the original source may include multi-sentence paragraphs, which are split and sentence-aligned in OPUS.
All texts are freely available for personal, educational and research use. Commercial use (e.g. reselling as parallel books) and mass redistribution without explicit permission are not granted. Please acknowledge the source when using the data!
16 languages, 64 bitexts
total number of files: 158
total number of tokens: 19.50M
total number of sentence fragments: 0.91M | @InProceedings{TIEDEMANN12.463,
author = {J�rg Tiedemann},
title = {Parallel Data, Tools and Interfaces in OPUS},
booktitle = {Proceedings of the Eight International Conference on Language Resources and Evaluation (LREC'12)},
year = {2012},
month = {may},
date = {23-25},
address = {Istanbul, Turkey},
editor = {Nicoletta Calzolari (Conference Chair) and Khalid Choukri and Thierry Declerck and Mehmet Ugur Dogan and Bente Maegaard and Joseph Mariani and Jan Odijk and Stelios Piperidis},
publisher = {European Language Resources Association (ELRA)},
isbn = {978-2-9517408-7-7},
language = {english}
} | null | 15 | 7,964 | ---
annotations_creators:
- found
language_creators:
- found
language:
- ca
- de
- el
- en
- eo
- es
- fi
- fr
- hu
- it
- nl
- 'no'
- pl
- pt
- ru
- sv
license:
- unknown
multilinguality:
- multilingual
size_categories:
- 1K<n<10K
source_datasets:
- original
task_categories:
- translation
task_ids: []
paperswithcode_id: null
pretty_name: OpusBooks
dataset_info:
- config_name: ca-de
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- ca
- de
splits:
- name: train
num_bytes: 899565
num_examples: 4445
download_size: 349126
dataset_size: 899565
- config_name: ca-en
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- ca
- en
splits:
- name: train
num_bytes: 863174
num_examples: 4605
download_size: 336276
dataset_size: 863174
- config_name: de-en
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- de
- en
splits:
- name: train
num_bytes: 13739047
num_examples: 51467
download_size: 5124458
dataset_size: 13739047
- config_name: el-en
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- el
- en
splits:
- name: train
num_bytes: 552579
num_examples: 1285
download_size: 175537
dataset_size: 552579
- config_name: de-eo
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- de
- eo
splits:
- name: train
num_bytes: 398885
num_examples: 1363
download_size: 150822
dataset_size: 398885
- config_name: en-eo
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- en
- eo
splits:
- name: train
num_bytes: 386231
num_examples: 1562
download_size: 145339
dataset_size: 386231
- config_name: de-es
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- de
- es
splits:
- name: train
num_bytes: 7592487
num_examples: 27526
download_size: 2802010
dataset_size: 7592487
- config_name: el-es
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- el
- es
splits:
- name: train
num_bytes: 527991
num_examples: 1096
download_size: 168306
dataset_size: 527991
- config_name: en-es
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- en
- es
splits:
- name: train
num_bytes: 25291783
num_examples: 93470
download_size: 9257150
dataset_size: 25291783
- config_name: eo-es
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- eo
- es
splits:
- name: train
num_bytes: 409591
num_examples: 1677
download_size: 154950
dataset_size: 409591
- config_name: en-fi
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- en
- fi
splits:
- name: train
num_bytes: 715039
num_examples: 3645
download_size: 266714
dataset_size: 715039
- config_name: es-fi
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- es
- fi
splits:
- name: train
num_bytes: 710462
num_examples: 3344
download_size: 264316
dataset_size: 710462
- config_name: de-fr
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- de
- fr
splits:
- name: train
num_bytes: 9544399
num_examples: 34916
download_size: 3556168
dataset_size: 9544399
- config_name: el-fr
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- el
- fr
splits:
- name: train
num_bytes: 539933
num_examples: 1237
download_size: 169241
dataset_size: 539933
- config_name: en-fr
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- en
- fr
splits:
- name: train
num_bytes: 32997199
num_examples: 127085
download_size: 12009501
dataset_size: 32997199
- config_name: eo-fr
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- eo
- fr
splits:
- name: train
num_bytes: 412999
num_examples: 1588
download_size: 152040
dataset_size: 412999
- config_name: es-fr
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- es
- fr
splits:
- name: train
num_bytes: 14382198
num_examples: 56319
download_size: 5203099
dataset_size: 14382198
- config_name: fi-fr
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- fi
- fr
splits:
- name: train
num_bytes: 746097
num_examples: 3537
download_size: 276633
dataset_size: 746097
- config_name: ca-hu
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- ca
- hu
splits:
- name: train
num_bytes: 886162
num_examples: 4463
download_size: 346425
dataset_size: 886162
- config_name: de-hu
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- de
- hu
splits:
- name: train
num_bytes: 13515043
num_examples: 51780
download_size: 5069455
dataset_size: 13515043
- config_name: el-hu
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- el
- hu
splits:
- name: train
num_bytes: 546290
num_examples: 1090
download_size: 176715
dataset_size: 546290
- config_name: en-hu
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- en
- hu
splits:
- name: train
num_bytes: 35256934
num_examples: 137151
download_size: 13232578
dataset_size: 35256934
- config_name: eo-hu
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- eo
- hu
splits:
- name: train
num_bytes: 389112
num_examples: 1636
download_size: 151332
dataset_size: 389112
- config_name: fr-hu
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- fr
- hu
splits:
- name: train
num_bytes: 22483133
num_examples: 89337
download_size: 8328639
dataset_size: 22483133
- config_name: de-it
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- de
- it
splits:
- name: train
num_bytes: 7760020
num_examples: 27381
download_size: 2811066
dataset_size: 7760020
- config_name: en-it
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- en
- it
splits:
- name: train
num_bytes: 8993803
num_examples: 32332
download_size: 3295251
dataset_size: 8993803
- config_name: eo-it
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- eo
- it
splits:
- name: train
num_bytes: 387606
num_examples: 1453
download_size: 146899
dataset_size: 387606
- config_name: es-it
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- es
- it
splits:
- name: train
num_bytes: 7837703
num_examples: 28868
download_size: 2864028
dataset_size: 7837703
- config_name: fr-it
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- fr
- it
splits:
- name: train
num_bytes: 4752171
num_examples: 14692
download_size: 1737670
dataset_size: 4752171
- config_name: hu-it
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- hu
- it
splits:
- name: train
num_bytes: 8445585
num_examples: 30949
download_size: 3101681
dataset_size: 8445585
- config_name: ca-nl
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- ca
- nl
splits:
- name: train
num_bytes: 884823
num_examples: 4329
download_size: 340308
dataset_size: 884823
- config_name: de-nl
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- de
- nl
splits:
- name: train
num_bytes: 3561764
num_examples: 15622
download_size: 1325189
dataset_size: 3561764
- config_name: en-nl
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- en
- nl
splits:
- name: train
num_bytes: 10278038
num_examples: 38652
download_size: 3727995
dataset_size: 10278038
- config_name: es-nl
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- es
- nl
splits:
- name: train
num_bytes: 9062389
num_examples: 32247
download_size: 3245558
dataset_size: 9062389
- config_name: fr-nl
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- fr
- nl
splits:
- name: train
num_bytes: 10408148
num_examples: 40017
download_size: 3720151
dataset_size: 10408148
- config_name: hu-nl
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- hu
- nl
splits:
- name: train
num_bytes: 10814173
num_examples: 43428
download_size: 3998988
dataset_size: 10814173
- config_name: it-nl
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- it
- nl
splits:
- name: train
num_bytes: 1328305
num_examples: 2359
download_size: 476875
dataset_size: 1328305
- config_name: en-no
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- en
- 'no'
splits:
- name: train
num_bytes: 661978
num_examples: 3499
download_size: 246977
dataset_size: 661978
- config_name: es-no
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- es
- 'no'
splits:
- name: train
num_bytes: 729125
num_examples: 3585
download_size: 270796
dataset_size: 729125
- config_name: fi-no
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- fi
- 'no'
splits:
- name: train
num_bytes: 691181
num_examples: 3414
download_size: 256267
dataset_size: 691181
- config_name: fr-no
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- fr
- 'no'
splits:
- name: train
num_bytes: 692786
num_examples: 3449
download_size: 256501
dataset_size: 692786
- config_name: hu-no
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- hu
- 'no'
splits:
- name: train
num_bytes: 695497
num_examples: 3410
download_size: 267047
dataset_size: 695497
- config_name: en-pl
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- en
- pl
splits:
- name: train
num_bytes: 583091
num_examples: 2831
download_size: 226855
dataset_size: 583091
- config_name: fi-pl
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- fi
- pl
splits:
- name: train
num_bytes: 613791
num_examples: 2814
download_size: 236123
dataset_size: 613791
- config_name: fr-pl
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- fr
- pl
splits:
- name: train
num_bytes: 614248
num_examples: 2825
download_size: 235905
dataset_size: 614248
- config_name: hu-pl
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- hu
- pl
splits:
- name: train
num_bytes: 616161
num_examples: 2859
download_size: 245670
dataset_size: 616161
- config_name: de-pt
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- de
- pt
splits:
- name: train
num_bytes: 317155
num_examples: 1102
download_size: 116319
dataset_size: 317155
- config_name: en-pt
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- en
- pt
splits:
- name: train
num_bytes: 309689
num_examples: 1404
download_size: 111837
dataset_size: 309689
- config_name: eo-pt
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- eo
- pt
splits:
- name: train
num_bytes: 311079
num_examples: 1259
download_size: 116157
dataset_size: 311079
- config_name: es-pt
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- es
- pt
splits:
- name: train
num_bytes: 326884
num_examples: 1327
download_size: 120549
dataset_size: 326884
- config_name: fr-pt
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- fr
- pt
splits:
- name: train
num_bytes: 324616
num_examples: 1263
download_size: 115920
dataset_size: 324616
- config_name: hu-pt
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- hu
- pt
splits:
- name: train
num_bytes: 302972
num_examples: 1184
download_size: 115002
dataset_size: 302972
- config_name: it-pt
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- it
- pt
splits:
- name: train
num_bytes: 301428
num_examples: 1163
download_size: 111050
dataset_size: 301428
- config_name: de-ru
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- de
- ru
splits:
- name: train
num_bytes: 5764673
num_examples: 17373
download_size: 1799371
dataset_size: 5764673
- config_name: en-ru
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- en
- ru
splits:
- name: train
num_bytes: 5190880
num_examples: 17496
download_size: 1613419
dataset_size: 5190880
- config_name: es-ru
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- es
- ru
splits:
- name: train
num_bytes: 5281130
num_examples: 16793
download_size: 1648606
dataset_size: 5281130
- config_name: fr-ru
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- fr
- ru
splits:
- name: train
num_bytes: 2474210
num_examples: 8197
download_size: 790541
dataset_size: 2474210
- config_name: hu-ru
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- hu
- ru
splits:
- name: train
num_bytes: 7818688
num_examples: 26127
download_size: 2469765
dataset_size: 7818688
- config_name: it-ru
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- it
- ru
splits:
- name: train
num_bytes: 5316952
num_examples: 17906
download_size: 1620478
dataset_size: 5316952
- config_name: en-sv
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- en
- sv
splits:
- name: train
num_bytes: 790785
num_examples: 3095
download_size: 304975
dataset_size: 790785
- config_name: fr-sv
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- fr
- sv
splits:
- name: train
num_bytes: 833553
num_examples: 3002
download_size: 321660
dataset_size: 833553
- config_name: it-sv
features:
- name: id
dtype: string
- name: translation
dtype:
translation:
languages:
- it
- sv
splits:
- name: train
num_bytes: 811413
num_examples: 2998
download_size: 307821
dataset_size: 811413
---
# Dataset Card for OpusBooks
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** http://opus.nlpl.eu/Books.php
- **Repository:** None
- **Paper:** http://www.lrec-conf.org/proceedings/lrec2012/pdf/463_Paper.pdf
- **Leaderboard:** [More Information Needed]
- **Point of Contact:** [More Information Needed]
### Dataset Summary
[More Information Needed]
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
Here are some examples of questions and facts:
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
[More Information Needed]
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
[More Information Needed]
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
Thanks to [@abhishekkrthakur](https://github.com/abhishekkrthakur) for adding this dataset. |
lamini/lamini_docs | 2023-07-23T23:48:57.000Z | [
"region:us"
] | lamini | null | null | null | 6 | 7,760 | ---
dataset_info:
features:
- name: question
dtype: string
- name: answer
dtype: string
- name: input_ids
sequence: int32
- name: attention_mask
sequence: int8
- name: labels
sequence: int64
splits:
- name: train
num_bytes: 1846734.3
num_examples: 1260
- name: test
num_bytes: 205192.7
num_examples: 140
download_size: 698607
dataset_size: 2051927.0
---
# Dataset Card for "lamini_docs"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
open-llm-leaderboard/details_Charlie911__vicuna-7b-v1.5-lora-mctaco | 2023-09-17T20:27:35.000Z | [
"region:us"
] | open-llm-leaderboard | null | null | null | 0 | 7,747 | ---
pretty_name: Evaluation run of Charlie911/vicuna-7b-v1.5-lora-mctaco
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [Charlie911/vicuna-7b-v1.5-lora-mctaco](https://huggingface.co/Charlie911/vicuna-7b-v1.5-lora-mctaco)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 64 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 3 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the agregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_Charlie911__vicuna-7b-v1.5-lora-mctaco\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-09-17T20:27:23.554125](https://huggingface.co/datasets/open-llm-leaderboard/details_Charlie911__vicuna-7b-v1.5-lora-mctaco/blob/main/results_2023-09-17T20-27-23.554125.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"em\": 0.2488464765100671,\n\
\ \"em_stderr\": 0.004427614016278926,\n \"f1\": 0.28849937080536914,\n\
\ \"f1_stderr\": 0.00442953185165108,\n \"acc\": 0.372010258662628,\n\
\ \"acc_stderr\": 0.00929094831305589\n },\n \"harness|drop|3\": {\n\
\ \"em\": 0.2488464765100671,\n \"em_stderr\": 0.004427614016278926,\n\
\ \"f1\": 0.28849937080536914,\n \"f1_stderr\": 0.00442953185165108\n\
\ },\n \"harness|gsm8k|5\": {\n \"acc\": 0.04473085670962851,\n \
\ \"acc_stderr\": 0.005693886131407047\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.6992896606156275,\n \"acc_stderr\": 0.012888010494704732\n\
\ }\n}\n```"
repo_url: https://huggingface.co/Charlie911/vicuna-7b-v1.5-lora-mctaco
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|arc:challenge|25_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|arc:challenge|25_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_drop_3
data_files:
- split: 2023_09_17T20_27_23.554125
path:
- '**/details_harness|drop|3_2023-09-17T20-27-23.554125.parquet'
- split: latest
path:
- '**/details_harness|drop|3_2023-09-17T20-27-23.554125.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_09_17T20_27_23.554125
path:
- '**/details_harness|gsm8k|5_2023-09-17T20-27-23.554125.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-09-17T20-27-23.554125.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hellaswag|10_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hellaswag|10_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-09-01T09:00:53.100273.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-09-01T09:03:24.370765.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-management|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-management|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-virology|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-virology|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- '**/details_harness|truthfulqa:mc|0_2023-09-01T09:00:53.100273.parquet'
- split: 2023_09_01T09_03_24.370765
path:
- '**/details_harness|truthfulqa:mc|0_2023-09-01T09:03:24.370765.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2023-09-01T09:03:24.370765.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_09_17T20_27_23.554125
path:
- '**/details_harness|winogrande|5_2023-09-17T20-27-23.554125.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-09-17T20-27-23.554125.parquet'
- config_name: results
data_files:
- split: 2023_09_01T09_00_53.100273
path:
- results_2023-09-01T09:00:53.100273.parquet
- split: 2023_09_01T09_03_24.370765
path:
- results_2023-09-01T09:03:24.370765.parquet
- split: 2023_09_17T20_27_23.554125
path:
- results_2023-09-17T20-27-23.554125.parquet
- split: latest
path:
- results_2023-09-17T20-27-23.554125.parquet
---
# Dataset Card for Evaluation run of Charlie911/vicuna-7b-v1.5-lora-mctaco
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/Charlie911/vicuna-7b-v1.5-lora-mctaco
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [Charlie911/vicuna-7b-v1.5-lora-mctaco](https://huggingface.co/Charlie911/vicuna-7b-v1.5-lora-mctaco) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 64 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 3 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_Charlie911__vicuna-7b-v1.5-lora-mctaco",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-09-17T20:27:23.554125](https://huggingface.co/datasets/open-llm-leaderboard/details_Charlie911__vicuna-7b-v1.5-lora-mctaco/blob/main/results_2023-09-17T20-27-23.554125.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"em": 0.2488464765100671,
"em_stderr": 0.004427614016278926,
"f1": 0.28849937080536914,
"f1_stderr": 0.00442953185165108,
"acc": 0.372010258662628,
"acc_stderr": 0.00929094831305589
},
"harness|drop|3": {
"em": 0.2488464765100671,
"em_stderr": 0.004427614016278926,
"f1": 0.28849937080536914,
"f1_stderr": 0.00442953185165108
},
"harness|gsm8k|5": {
"acc": 0.04473085670962851,
"acc_stderr": 0.005693886131407047
},
"harness|winogrande|5": {
"acc": 0.6992896606156275,
"acc_stderr": 0.012888010494704732
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] |
yahma/alpaca-cleaned | 2023-04-10T20:29:06.000Z | [
"task_categories:text-generation",
"language:en",
"license:cc-by-4.0",
"instruction-finetuning",
"region:us"
] | yahma | null | null | null | 246 | 7,741 | ---
license: cc-by-4.0
language:
- en
tags:
- instruction-finetuning
pretty_name: Alpaca-Cleaned
task_categories:
- text-generation
---
# Dataset Card for Alpaca-Cleaned
- **Repository:** https://github.com/gururise/AlpacaDataCleaned
## Dataset Description
This is a cleaned version of the original Alpaca Dataset released by Stanford. The following issues have been identified in the original release and fixed in this dataset:
1. **Hallucinations:** Many instructions in the original dataset had instructions referencing data on the internet, which just caused GPT3 to hallucinate an answer.
```
"instruction":"Summarize the given article in 200 Words.",
"input": "https://www.bbc.com/news/world-51461830",
"output": "The recent protests in Belarus are part of a growing wave of discontent against the rule of Alexander Lukashenko, who has led the country since 1994. Demonstrations against his government have been ongoing since August, and come in the wake of an election widely seen as rigged. As police and security forces have cracked down on protesters, the movement has only grown, with hundreds of thousands of people taking to the streets to call for new leadership. Human rights activists have accused the government of using extreme violence in order prosecute protesters, leading to an international outcry and sanctions from the European Union. As the situation in Belarus continues to unfold, the international community is monitoring the situation closely - hoping for a peaceful resolution, while also standing firm against the oppressive tactics employed by the Belarusian state."
"instruction": "Classify the following shape according to simple geometry.",
"input": "<Picture Attached>",
"output": "Triangle"
"instruction": "Given a recent news article, write a headline for it.",
"input": "https://www.news.example.com/article-1",
"output": "\"NASA Launches a Mars 2020 Mission to Search for Signs of Ancient Life\""
```
2. **Merged Instructions:** There were many instructions that were merged together in the original dataset for some reason.
```
"instruction": "Should I invest in stocks?",
"input": "",
"output": "It depends on your individual situation. Investing in stocks carries a certain amount of risk but can also offer potential rewards. Make sure to research the different types of stocks, understand the risks and rewards, and develop a plan that works for you.\n##\n18. Instruction: Identify the noun in the following sentence:\n18. Input:\nThe large apple was delicious.\n18. Output:\nApple"
```
3. **Empty outputs:** Some entries in the original dataset had empty outputs.
```
"instruction": "Calculate the area of the following shape in square centimeters.",
"input": "rectangle of size 4 cm x 5 cm",
"output": ""
```
4. **Empty code examples:** Some descriptions in the original dataset were missing code examples, making it difficult to understand the intended behavior of the code.
5. **Instructions to generate images:** Some descriptions in the original dataset included instructions to generate images, something obviously not possible.
```
"instruction": "Create a graphic or logo that visually represents the word \"courage\".",
"input": "",
"output": "<No Output>"
```
6. **N/A outputs:** Some code snippets in the original dataset had N/A outputs.
7. **Inconsistent input field:** The original dataset had inconsistent usage of the input field when it was supposed to be empty.
```
"input":"<no input>"
"input":"No input"
"input":"noinput"
"input":"<noinput>"
```
8. **Wrong answers:** Some instructions/questions in the original dataset had incorrect answers. About 80% of the math problems are estimated to have incorrect answers.
```
"instruction": "Calculate the median of the following data set.",
"input": "1, 2, 4, 5, 8, 9",
"output": "5"
"instruction": "Convert 25m to km.",
"input": "",
"output": "25km"
```
9. **Non-Sensical/Unclear instructions:** Many instructions are unclear, we try to clarify (or re-write) if instructions are non-sensical. Instructions that are slightly unclear, but where one could deduce the meaning are not altered.
```
"instruction": "Freeze the following sample of yogurt for 10 minutes.",
"input": "Yogurt sample",
"output": "<noinput>"
"instruction": "Increase the font size to 12 points.",
"input": "",
"output": "The font size has been increased to 12 points."
```
10. **Extraneous escape and control characters:** The original dataset had several entries with extraneous escape and control characters.
### Original Alpaca Dataset Summary
Alpaca is a dataset of 52,000 instructions and demonstrations generated by OpenAI's `text-davinci-003` engine. This instruction data can be used to conduct instruction-tuning for language models and make the language model follow instruction better.
The authors built on the data generation pipeline from [Self-Instruct framework](https://github.com/yizhongw/self-instruct) and made the following modifications:
- The `text-davinci-003` engine to generate the instruction data instead of `davinci`.
- A [new prompt](https://github.com/tatsu-lab/stanford_alpaca/blob/main/prompt.txt) was written that explicitly gave the requirement of instruction generation to `text-davinci-003`.
- Much more aggressive batch decoding was used, i.e., generating 20 instructions at once, which significantly reduced the cost of data generation.
- The data generation pipeline was simplified by discarding the difference between classification and non-classification instructions.
- Only a single instance was generated for each instruction, instead of 2 to 3 instances as in Self-Instruct.
This produced an instruction-following dataset with 52K examples obtained at a much lower cost (less than $500).
In a preliminary study, the authors also found that the 52K generated data to be much more diverse than the data released by [Self-Instruct](https://github.com/yizhongw/self-instruct/blob/main/data/seed_tasks.jsonl).
### Supported Tasks and Leaderboards
The Alpaca dataset designed for instruction training pretrained language models.
### Languages
The data in Alpaca are in English (BCP-47 en).
## Dataset Structure
### Data Instances
An example of "train" looks as follows:
```json
{
"instruction": "Create a classification task by clustering the given list of items.",
"input": "Apples, oranges, bananas, strawberries, pineapples",
"output": "Class 1: Apples, Oranges\nClass 2: Bananas, Strawberries\nClass 3: Pineapples",
"text": "Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.\n\n### Instruction:\nCreate a classification task by clustering the given list of items.\n\n### Input:\nApples, oranges, bananas, strawberries, pineapples\n\n### Response:\nClass 1: Apples, Oranges\nClass 2: Bananas, Strawberries\nClass 3: Pineapples",
}
```
### Data Fields
The data fields are as follows:
* `instruction`: describes the task the model should perform. Each of the 52K instructions is unique.
* `input`: optional context or input for the task. For example, when the instruction is "Summarize the following article", the input is the article. Around 40% of the examples have an input.
* `output`: the answer to the instruction as generated by `text-davinci-003`.
* `text`: the `instruction`, `input` and `output` formatted with the [prompt template](https://github.com/tatsu-lab/stanford_alpaca#data-release) used by the authors for fine-tuning their models.
### Data Splits
| | train |
|---------------|------:|
| alpaca | 52002 |
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
Excerpt the [blog post](https://crfm.stanford.edu/2023/03/13/alpaca.html) accompanying the release of this dataset:
> We believe that releasing the above assets will enable the academic community to perform controlled scientific studies on instruction-following language models, resulting in better science and ultimately new techniques to address the existing deficiencies with these models. At the same time, any release carries some risk. First, we recognize that releasing our training recipe reveals the feasibility of certain capabilities. On one hand, this enables more people (including bad actors) to create models that could cause harm (either intentionally or not). On the other hand, this awareness might incentivize swift defensive action, especially from the academic community, now empowered by the means to perform deeper safety research on such models. Overall, we believe that the benefits for the research community outweigh the risks of this particular release. Given that we are releasing the training recipe, we believe that releasing the data, model weights, and training code incur minimal further risk, given the simplicity of the recipe. At the same time, releasing these assets has enormous benefits for reproducible science, so that the academic community can use standard datasets, models, and code to perform controlled comparisons and to explore extensions. Deploying an interactive demo for Alpaca also poses potential risks, such as more widely disseminating harmful content and lowering the barrier for spam, fraud, or disinformation. We have put into place two risk mitigation strategies. First, we have implemented a content filter using OpenAI’s content moderation API, which filters out harmful content as defined by OpenAI’s usage policies. Second, we watermark all the model outputs using the method described in Kirchenbauer et al. 2023, so that others can detect (with some probability) whether an output comes from Alpaca 7B. Finally, we have strict terms and conditions for using the demo; it is restricted to non-commercial uses and to uses that follow LLaMA’s license agreement. We understand that these mitigation measures can be circumvented once we release the model weights or if users train their own instruction-following models. However, by installing these mitigations, we hope to advance the best practices and ultimately develop community norms for the responsible deployment of foundation models.
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
The `alpaca` data is generated by a language model (`text-davinci-003`) and inevitably contains some errors or biases. We encourage users to use this data with caution and propose new methods to filter or improve the imperfections.
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
The dataset is available under the [Creative Commons NonCommercial (CC BY-NC 4.0)](https://creativecommons.org/licenses/by-nc/4.0/legalcode).
### Citation Information
```
@misc{alpaca,
author = {Rohan Taori and Ishaan Gulrajani and Tianyi Zhang and Yann Dubois and Xuechen Li and Carlos Guestrin and Percy Liang and Tatsunori B. Hashimoto },
title = {Stanford Alpaca: An Instruction-following LLaMA model},
year = {2023},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/tatsu-lab/stanford_alpaca}},
}
```
### Contributions
[More Information Needed] |
garage-bAInd/Open-Platypus | 2023-09-17T16:56:19.000Z | [
"size_categories:10K<n<100K",
"language:en",
"arxiv:2308.07317",
"region:us"
] | garage-bAInd | null | null | null | 225 | 7,535 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
dataset_info:
features:
- name: input
dtype: string
- name: output
dtype: string
- name: instruction
dtype: string
- name: data_source
dtype: string
splits:
- name: train
num_bytes: 30776452
num_examples: 24926
download_size: 15565850
dataset_size: 30776452
language:
- en
size_categories:
- 10K<n<100K
---
# OpenPlatypus
This dataset is focused on improving LLM logical reasoning skills and was used to train the Platypus2 models. It is comprised of the following datasets, which were filtered using keyword search and then Sentence Transformers to remove questions with a similarity above 80%:
| Dataset Name | License Type |
|--------------------------------------------------------------|--------------|
| [PRM800K](https://github.com/openai/prm800k) | MIT |
| [ScienceQA](https://github.com/lupantech/ScienceQA) | [Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International](https://creativecommons.org/licenses/by-nc-sa/4.0/) |
| [SciBench](https://github.com/mandyyyyii/scibench) | MIT |
| [ReClor](https://whyu.me/reclor/) | Non-commercial |
| [TheoremQA](https://huggingface.co/datasets/wenhu/TheoremQA) | MIT |
| [`nuprl/leetcode-solutions-python-testgen-gpt4`](https://huggingface.co/datasets/nuprl/leetcode-solutions-python-testgen-gpt4/viewer/nuprl--leetcode-solutions-python-testgen-gpt4/train?p=1) | None listed |
| [`jondurbin/airoboros-gpt4-1.4.1`](https://huggingface.co/datasets/jondurbin/airoboros-gpt4-1.4.1) | other |
| [`TigerResearch/tigerbot-kaggle-leetcodesolutions-en-2k`](https://huggingface.co/datasets/TigerResearch/tigerbot-kaggle-leetcodesolutions-en-2k/viewer/TigerResearch--tigerbot-kaggle-leetcodesolutions-en-2k/train?p=2) | apache-2.0 |
| [openbookQA](https://huggingface.co/datasets/openbookqa/viewer/additional/train?row=35) | apache-2.0 |
| [ARB](https://arb.duckai.org) | MIT |
| [`timdettmers/openassistant-guanaco`](https://huggingface.co/datasets/timdettmers/openassistant-guanaco) | apache-2.0 |
## Data Contamination Check
We've removed approximately 200 questions that appear in the Hugging Face benchmark test sets. Please see our [paper](https://arxiv.org/abs/2308.07317) and [project webpage](https://platypus-llm.github.io) for additional information.
## Model Info
Please see models at [`garage-bAInd`](https://huggingface.co/garage-bAInd).
## Training and filtering code
Please see the [Platypus GitHub repo](https://github.com/arielnlee/Platypus).
## Citations
```bibtex
@article{platypus2023,
title={Platypus: Quick, Cheap, and Powerful Refinement of LLMs},
author={Ariel N. Lee and Cole J. Hunter and Nataniel Ruiz},
booktitle={arXiv preprint arxiv:2308.07317},
year={2023}
}
```
```bibtex
@article{lightman2023lets,
title={Let's Verify Step by Step},
author={Lightman, Hunter and Kosaraju, Vineet and Burda, Yura and Edwards, Harri and Baker, Bowen and Lee, Teddy and Leike, Jan and Schulman, John and Sutskever, Ilya and Cobbe, Karl},
journal={preprint arXiv:2305.20050},
year={2023}
}
```
```bibtex
@inproceedings{lu2022learn,
title={Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering},
author={Lu, Pan and Mishra, Swaroop and Xia, Tony and Qiu, Liang and Chang, Kai-Wei and Zhu, Song-Chun and Tafjord, Oyvind and Clark, Peter and Ashwin Kalyan},
booktitle={The 36th Conference on Neural Information Processing Systems (NeurIPS)},
year={2022}
}
```
```bibtex
@misc{wang2023scibench,
title={SciBench: Evaluating College-Level Scientific Problem-Solving Abilities of Large Language Models},
author={Xiaoxuan Wang and Ziniu Hu and Pan Lu and Yanqiao Zhu and Jieyu Zhang and Satyen Subramaniam and Arjun R. Loomba and Shichang Zhang and Yizhou Sun and Wei Wang},
year={2023},
arXiv eprint 2307.10635
}
```
```bibtex
@inproceedings{yu2020reclor,
author = {Yu, Weihao and Jiang, Zihang and Dong, Yanfei and Feng, Jiashi},
title = {ReClor: A Reading Comprehension Dataset Requiring Logical Reasoning},
booktitle = {International Conference on Learning Representations (ICLR)},
month = {April},
year = {2020}
}
```
```bibtex
@article{chen2023theoremqa,
title={TheoremQA: A Theorem-driven Question Answering dataset},
author={Chen, Wenhu and Ming Yin, Max Ku, Elaine Wan, Xueguang Ma, Jianyu Xu, Tony Xia, Xinyi Wang, Pan Lu},
journal={preprint arXiv:2305.12524},
year={2023}
}
```
```bibtex
@inproceedings{OpenBookQA2018,
title={Can a Suit of Armor Conduct Electricity? A New Dataset for Open Book Question Answering},
author={Todor Mihaylov and Peter Clark and Tushar Khot and Ashish Sabharwal},
booktitle={EMNLP},
year={2018}
}
```
```bibtex
@misc{sawada2023arb,
title={ARB: Advanced Reasoning Benchmark for Large Language Models},
author={Tomohiro Sawada and Daniel Paleka and Alexander Havrilla and Pranav Tadepalli and Paula Vidas and Alexander Kranias and John J. Nay and Kshitij Gupta and Aran Komatsuzaki},
arXiv eprint 2307.13692,
year={2023}
}
``` |
wiki_dpr | 2023-04-05T13:43:12.000Z | [
"task_categories:fill-mask",
"task_categories:text-generation",
"task_ids:language-modeling",
"task_ids:masked-language-modeling",
"annotations_creators:no-annotation",
"language_creators:crowdsourced",
"multilinguality:multilingual",
"size_categories:10M<n<100M",
"source_datasets:original",
"language:en",
"license:cc-by-sa-3.0",
"license:gfdl",
"text-search",
"arxiv:2004.04906",
"region:us"
] | null | This is the wikipedia split used to evaluate the Dense Passage Retrieval (DPR) model.
It contains 21M passages from wikipedia along with their DPR embeddings.
The wikipedia articles were split into multiple, disjoint text blocks of 100 words as passages. | @misc{karpukhin2020dense,
title={Dense Passage Retrieval for Open-Domain Question Answering},
author={Vladimir Karpukhin and Barlas Oğuz and Sewon Min and Patrick Lewis and Ledell Wu and Sergey Edunov and Danqi Chen and Wen-tau Yih},
year={2020},
eprint={2004.04906},
archivePrefix={arXiv},
primaryClass={cs.CL}
} | null | 18 | 7,498 | ---
annotations_creators:
- no-annotation
language_creators:
- crowdsourced
language:
- en
license:
- cc-by-sa-3.0
- gfdl
multilinguality:
- multilingual
size_categories:
- 10M<n<100M
source_datasets:
- original
task_categories:
- fill-mask
- text-generation
task_ids:
- language-modeling
- masked-language-modeling
pretty_name: Wiki-DPR
tags:
- text-search
dataset_info:
- config_name: psgs_w100.nq.exact
features:
- name: id
dtype: string
- name: text
dtype: string
- name: title
dtype: string
- name: embeddings
sequence: float32
splits:
- name: train
num_bytes: 78419281788
num_examples: 21015300
download_size: 70965697456
dataset_size: 78419281788
- config_name: psgs_w100.nq.compressed
features:
- name: id
dtype: string
- name: text
dtype: string
- name: title
dtype: string
- name: embeddings
sequence: float32
splits:
- name: train
num_bytes: 78419281788
num_examples: 21015300
download_size: 70965697456
dataset_size: 78419281788
- config_name: psgs_w100.nq.no_index
features:
- name: id
dtype: string
- name: text
dtype: string
- name: title
dtype: string
- name: embeddings
sequence: float32
splits:
- name: train
num_bytes: 78419281788
num_examples: 21015300
download_size: 70965697456
dataset_size: 78419281788
- config_name: psgs_w100.multiset.exact
features:
- name: id
dtype: string
- name: text
dtype: string
- name: title
dtype: string
- name: embeddings
sequence: float32
splits:
- name: train
num_bytes: 78419281788
num_examples: 21015300
download_size: 70965697456
dataset_size: 78419281788
- config_name: psgs_w100.multiset.compressed
features:
- name: id
dtype: string
- name: text
dtype: string
- name: title
dtype: string
- name: embeddings
sequence: float32
splits:
- name: train
num_bytes: 78419281788
num_examples: 21015300
download_size: 70965697456
dataset_size: 78419281788
- config_name: psgs_w100.multiset.no_index
features:
- name: id
dtype: string
- name: text
dtype: string
- name: title
dtype: string
- name: embeddings
sequence: float32
splits:
- name: train
num_bytes: 78419281788
num_examples: 21015300
download_size: 70965697456
dataset_size: 78419281788
---
# Dataset Card for "wiki_dpr"
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://github.com/facebookresearch/DPR](https://github.com/facebookresearch/DPR)
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 425.79 GB
- **Size of the generated dataset:** 470.52 GB
- **Total amount of disk used:** 978.05 GB
### Dataset Summary
This is the wikipedia split used to evaluate the Dense Passage Retrieval (DPR) model.
It contains 21M passages from wikipedia along with their DPR embeddings.
The wikipedia articles were split into multiple, disjoint text blocks of 100 words as passages.
The wikipedia dump is the one from Dec. 20, 2018.
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Dataset Structure
### Data Instances
Each instance contains a paragraph of at most 100 words, as well as the title of the wikipedia page it comes from, and the DPR embedding (a 768-d vector).
#### psgs_w100.multiset.compressed
- **Size of downloaded dataset files:** 70.97 GB
- **Size of the generated dataset:** 78.42 GB
- **Total amount of disk used:** 152.26 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{'id': '1',
'text': 'Aaron Aaron ( or ; "Ahärôn") is a prophet, high priest, and the brother of Moses in the Abrahamic religions. Knowledge of Aaron, along with his brother Moses, comes exclusively from religious texts, such as the Bible and Quran. The Hebrew Bible relates that, unlike Moses, who grew up in the Egyptian royal court, Aaron and his elder sister Miriam remained with their kinsmen in the eastern border-land of Egypt (Goshen). When Moses first confronted the Egyptian king about the Israelites, Aaron served as his brother\'s spokesman ("prophet") to the Pharaoh. Part of the Law (Torah) that Moses received from'],
'title': 'Aaron',
'embeddings': [-0.07233893871307373,
0.48035329580307007,
0.18650995194911957,
-0.5287084579467773,
-0.37329429388046265,
0.37622880935668945,
0.25524479150772095,
...
-0.336689829826355,
0.6313082575798035,
-0.7025573253631592]}
```
#### psgs_w100.multiset.exact
- **Size of downloaded dataset files:** 70.97 GB
- **Size of the generated dataset:** 78.42 GB
- **Total amount of disk used:** 187.38 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{'id': '1',
'text': 'Aaron Aaron ( or ; "Ahärôn") is a prophet, high priest, and the brother of Moses in the Abrahamic religions. Knowledge of Aaron, along with his brother Moses, comes exclusively from religious texts, such as the Bible and Quran. The Hebrew Bible relates that, unlike Moses, who grew up in the Egyptian royal court, Aaron and his elder sister Miriam remained with their kinsmen in the eastern border-land of Egypt (Goshen). When Moses first confronted the Egyptian king about the Israelites, Aaron served as his brother\'s spokesman ("prophet") to the Pharaoh. Part of the Law (Torah) that Moses received from'],
'title': 'Aaron',
'embeddings': [-0.07233893871307373,
0.48035329580307007,
0.18650995194911957,
-0.5287084579467773,
-0.37329429388046265,
0.37622880935668945,
0.25524479150772095,
...
-0.336689829826355,
0.6313082575798035,
-0.7025573253631592]}
```
#### psgs_w100.multiset.no_index
- **Size of downloaded dataset files:** 70.97 GB
- **Size of the generated dataset:** 78.42 GB
- **Total amount of disk used:** 149.38 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{'id': '1',
'text': 'Aaron Aaron ( or ; "Ahärôn") is a prophet, high priest, and the brother of Moses in the Abrahamic religions. Knowledge of Aaron, along with his brother Moses, comes exclusively from religious texts, such as the Bible and Quran. The Hebrew Bible relates that, unlike Moses, who grew up in the Egyptian royal court, Aaron and his elder sister Miriam remained with their kinsmen in the eastern border-land of Egypt (Goshen). When Moses first confronted the Egyptian king about the Israelites, Aaron served as his brother\'s spokesman ("prophet") to the Pharaoh. Part of the Law (Torah) that Moses received from'],
'title': 'Aaron',
'embeddings': [-0.07233893871307373,
0.48035329580307007,
0.18650995194911957,
-0.5287084579467773,
-0.37329429388046265,
0.37622880935668945,
0.25524479150772095,
...
-0.336689829826355,
0.6313082575798035,
-0.7025573253631592]}
```
#### psgs_w100.nq.compressed
- **Size of downloaded dataset files:** 70.97 GB
- **Size of the generated dataset:** 78.42 GB
- **Total amount of disk used:** 152.26 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{'id': '1',
'text': 'Aaron Aaron ( or ; "Ahärôn") is a prophet, high priest, and the brother of Moses in the Abrahamic religions. Knowledge of Aaron, along with his brother Moses, comes exclusively from religious texts, such as the Bible and Quran. The Hebrew Bible relates that, unlike Moses, who grew up in the Egyptian royal court, Aaron and his elder sister Miriam remained with their kinsmen in the eastern border-land of Egypt (Goshen). When Moses first confronted the Egyptian king about the Israelites, Aaron served as his brother\'s spokesman ("prophet") to the Pharaoh. Part of the Law (Torah) that Moses received from'],
'title': 'Aaron',
'embeddings': [0.013342111371457577,
0.582173764705658,
-0.31309744715690613,
-0.6991612911224365,
-0.5583199858665466,
0.5187504887580872,
0.7152731418609619,
...
-0.5385938286781311,
0.8093984127044678,
-0.4741983711719513]}
```
#### psgs_w100.nq.exact
- **Size of downloaded dataset files:** 70.97 GB
- **Size of the generated dataset:** 78.42 GB
- **Total amount of disk used:** 187.38 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{'id': '1',
'text': 'Aaron Aaron ( or ; "Ahärôn") is a prophet, high priest, and the brother of Moses in the Abrahamic religions. Knowledge of Aaron, along with his brother Moses, comes exclusively from religious texts, such as the Bible and Quran. The Hebrew Bible relates that, unlike Moses, who grew up in the Egyptian royal court, Aaron and his elder sister Miriam remained with their kinsmen in the eastern border-land of Egypt (Goshen). When Moses first confronted the Egyptian king about the Israelites, Aaron served as his brother\'s spokesman ("prophet") to the Pharaoh. Part of the Law (Torah) that Moses received from'],
'title': 'Aaron',
'embeddings': [0.013342111371457577,
0.582173764705658,
-0.31309744715690613,
-0.6991612911224365,
-0.5583199858665466,
0.5187504887580872,
0.7152731418609619,
...
-0.5385938286781311,
0.8093984127044678,
-0.4741983711719513]}
```
### Data Fields
The data fields are the same among all splits.
#### psgs_w100.multiset.compressed
- `id`: a `string` feature.
- `text`: a `string` feature.
- `title`: a `string` feature.
- `embeddings`: a `list` of `float32` features.
#### psgs_w100.multiset.exact
- `id`: a `string` feature.
- `text`: a `string` feature.
- `title`: a `string` feature.
- `embeddings`: a `list` of `float32` features.
#### psgs_w100.multiset.no_index
- `id`: a `string` feature.
- `text`: a `string` feature.
- `title`: a `string` feature.
- `embeddings`: a `list` of `float32` features.
#### psgs_w100.nq.compressed
- `id`: a `string` feature.
- `text`: a `string` feature.
- `title`: a `string` feature.
- `embeddings`: a `list` of `float32` features.
#### psgs_w100.nq.exact
- `id`: a `string` feature.
- `text`: a `string` feature.
- `title`: a `string` feature.
- `embeddings`: a `list` of `float32` features.
### Data Splits
| name | train |
|-----------------------------|-------:|
|psgs_w100.multiset.compressed|21015300|
|psgs_w100.multiset.exact |21015300|
|psgs_w100.multiset.no_index |21015300|
|psgs_w100.nq.compressed |21015300|
|psgs_w100.nq.exact |21015300|
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Citation Information
```
@misc{karpukhin2020dense,
title={Dense Passage Retrieval for Open-Domain Question Answering},
author={Vladimir Karpukhin and Barlas Oğuz and Sewon Min and Patrick Lewis and Ledell Wu and Sergey Edunov and Danqi Chen and Wen-tau Yih},
year={2020},
eprint={2004.04906},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
### Contributions
Thanks to [@thomwolf](https://github.com/thomwolf), [@lewtun](https://github.com/lewtun), [@lhoestq](https://github.com/lhoestq) for adding this dataset. |
roneneldan/TinyStories | 2023-08-16T16:54:12.000Z | [
"arxiv:2305.07759",
"region:us"
] | roneneldan | null | null | null | 251 | 7,470 | License: CDLA-Sharing-1.0
-------------
Dataset containing synthetically generated (by GPT-3.5 and GPT-4) short stories that only use a small vocabulary.
Described in the following paper: https://arxiv.org/abs/2305.07759.
The models referred to in the paper were trained on TinyStories-train.txt (the file tinystories-valid.txt can be used for validation loss). These models can be found on Huggingface, at roneneldan/TinyStories-1M/3M/8M/28M/33M/1Layer-21M.
There are two more resources:
tinystories_all_data.tar.gz - contains a superset of the stories together with metadata and the prompt that was used to create each story.
TinyStoriesV2-GPT4-train.txt - Is a new version of the dataset that is based on generations by GPT-4 only (the original dataset also has generations by GPT-3.5 which are of lesser quality). It contains all the examples in TinyStories.txt which were GPT-4 generated as a subset (but is significantly larger).
|
bigcode/the-stack-dedup | 2023-08-17T08:21:58.000Z | [
"task_categories:text-generation",
"language_creators:crowdsourced",
"language_creators:expert-generated",
"multilinguality:multilingual",
"size_categories:unknown",
"language:code",
"license:other",
"arxiv:2211.15533",
"arxiv:2107.03374",
"arxiv:2207.14157",
"region:us"
] | bigcode | null | null | null | 234 | 7,452 | ---
annotations_creators: []
language_creators:
- crowdsourced
- expert-generated
language:
- code
license:
- other
multilinguality:
- multilingual
pretty_name: The-Stack
size_categories:
- unknown
source_datasets: []
task_categories:
- text-generation
task_ids: []
extra_gated_prompt: |-
## Terms of Use for The Stack
The Stack dataset is a collection of source code in over 300 programming languages. We ask that you read and acknowledge the following points before using the dataset:
1. The Stack is a collection of source code from repositories with various licenses. Any use of all or part of the code gathered in The Stack must abide by the terms of the original licenses, including attribution clauses when relevant. We facilitate this by providing provenance information for each data point.
2. The Stack is regularly updated to enact validated data removal requests. By clicking on "Access repository", you agree to update your own version of The Stack to the most recent usable version specified by the maintainers in [the following thread](https://huggingface.co/datasets/bigcode/the-stack/discussions/7). If you have questions about dataset versions and allowed uses, please also ask them in the dataset’s [community discussions](https://huggingface.co/datasets/bigcode/the-stack/discussions/new). We will also notify users via email when the latest usable version changes.
3. To host, share, or otherwise provide access to The Stack dataset, you must include [these Terms of Use](https://huggingface.co/datasets/bigcode/the-stack#terms-of-use-for-the-stack) and require users to agree to it.
By clicking on "Access repository" below, you accept that your contact information (email address and username) can be shared with the dataset maintainers as well.
extra_gated_fields:
Email: text
I have read the License and agree with its terms: checkbox
---
# Dataset Card for The Stack

## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Changelog](#changelog)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [How to use it](#how-to-use-it)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
- [Terms of Use for The Stack](#terms-of-use-for-the-stack)
## Dataset Description
- **Homepage:** https://www.bigcode-project.org/
- **Repository:** https://github.com/bigcode-project
- **Paper:** https://arxiv.org/abs/2211.15533
- **Leaderboard:** N/A
- **Point of Contact:** contact@bigcode-project.org
### Changelog
|Release|Description|
|-|-|
|v1.0| Initial release of the Stack. Included 30 programming languages and 18 permissive licenses. **Note:** Three included licenses (MPL/EPL/LGPL) are considered weak copyleft licenses. The resulting near-deduplicated dataset is 1.5TB in size. |
|v1.1| The three copyleft licenses ((MPL/EPL/LGPL) were excluded and the list of permissive licenses extended to 193 licenses in total. The list of programming languages was increased from 30 to 358 languages. Also opt-out request submitted by 15.11.2022 were excluded from this version of the dataset. The resulting near-deduplicated dataset is 3TB in size.|
|v1.2| Opt-out request submitted by 09.02.2022 were excluded from this version of the dataset. A stronger near-deduplication strategy was applied resulting leading to 2.7TB in size.|
### Dataset Summary
The Stack contains over 6TB of permissively-licensed source code files covering 358 programming languages. The dataset was created as part of the [BigCode Project](https://www.bigcode-project.org/), an open scientific collaboration working on the responsible development of Large Language Models for Code (Code LLMs). The Stack serves as a pre-training dataset for Code LLMs, i.e., code-generating AI systems which enable the synthesis of programs from natural language descriptions as well as other from code snippets. **This is the near-deduplicated version with 3TB data.**
### Supported Tasks and Leaderboards
The Stack is a pre-training dataset for creating code LLMs. Code LLMs can be used for a wide variety of downstream tasks such as code completion from natural language descriptions ([HumanEval](https://huggingface.co/datasets/openai_humaneval), [MBPP](https://huggingface.co/datasets/mbpp)), documentation generation for individual functions ([CodeSearchNet](https://huggingface.co/datasets/code_search_net)), and auto-completion of code snippets ([HumanEval-Infilling](https://github.com/openai/human-eval-infilling)). However, these downstream evaluation benchmarks are outside the scope of The Stack.
### Languages
The following natural languages appear in the comments and docstrings from files in the dataset: EN, ZH, FR, PT, ES, RU, DE, KO, JA, UZ, IT, ID, RO, AR, FA, CA, HU, ML, NL, TR, TE, EL, EO, BN, LV, GL, PL, GU, CEB, IA, KN, SH, MK, UR, SV, LA, JKA, MY, SU, CS, MN. This kind of data is essential for applications such as documentation generation and natural-language-to-code translation.
The dataset contains **358 programming languages**. The full list can be found [here](https://huggingface.co/datasets/bigcode/the-stack-dedup/blob/main/programming-languages.json).
### How to use it
```python
from datasets import load_dataset
# full dataset (3TB of data)
ds = load_dataset("bigcode/the-stack-dedup", split="train")
# specific language (e.g. Dockerfiles)
ds = load_dataset("bigcode/the-stack-dedup", data_dir="data/dockerfile", split="train")
# dataset streaming (will only download the data as needed)
ds = load_dataset("bigcode/the-stack-dedup", streaming=True, split="train")
for sample in iter(ds): print(sample["content"])
```
## Dataset Structure
### Data Instances
Each data instance corresponds to one file. The content of the file is in the `content` feature, and other features (`repository_name`, `licenses`, etc.) provide some metadata. Note that a given file can appear in several different repositories that satisfy our safe-license criterion. If that is the case, only the first – in alphabetical order -- of these repositories is shown for simplicity.
### Data Fields
- `content` (string): the content of the file.
- `size` (integer): size of the uncompressed file.
- `lang` (string): the programming language.
- `ext` (string): file extension
- `avg_line_length` (float): the average line-length of the file.
- `max_line_length` (integer): the maximum line-length of the file.
- `alphanum_fraction` (float): the fraction of characters in the file that are alphabetical or numerical characters.
- `hexsha` (string): unique git hash of file
- `max_{stars|forks|issues}_repo_path` (string): path to file in repo containing this file with maximum number of `{stars|forks|issues}`
- `max_{stars|forks|issues}_repo_name` (string): name of repo containing this file with maximum number of `{stars|forks|issues}`
- `max_{stars|forks|issues}_repo_head_hexsha` (string): hexsha of repository head
- `max_{stars|forks|issues}_repo_licenses` (string): licenses in repository
- `max_{stars|forks|issues}_count` (integer): number of `{stars|forks|issues}` in repository
- `max_{stars|forks|issues}_repo_{stars|forks|issues}_min_datetime` (string): first timestamp of a `{stars|forks|issues}` event
- `max_{stars|forks|issues}_repo_{stars|forks|issues}_max_datetime` (string): last timestamp of a `{stars|forks|issues}` event
### Data Splits
The dataset has no splits and all data is loaded as train split by default. If you want to setup a custom train-test split beware that dataset contains a lot of near-duplicates which can cause leakage into the test split.
## Dataset Creation
### Curation Rationale
One of the challenges faced by researchers working on code LLMs is the lack of openness and transparency around the development of these systems. Most prior works described the high-level data collection process but did not release the training data. It is therefore difficult for other researchers to fully reproduce these models and understand what kind of pre-training data leads to high-performing code LLMs. By releasing an open large-scale code dataset we hope to make training of code LLMs more reproducible. **This is the near-deduplicated version with 3TB data.**
### Source Data
#### Initial Data Collection and Normalization
220.92M active GitHub repository names were collected from the event archives published between January 1st, 2015 and March 31st, 2022 on [GHArchive](https://gharchive.org/). Only 137.36M of these repositories were public and accessible on GitHub – others were not accessible as they had been deleted by their owners. 51.76B files were downloaded from the public repositories on GitHub between November 2021 and June 2022. 5.28B files were unique. The uncompressed size of all stored files is 92.36TB.
The list of programming language extensions is taken from this [list](https://gist.github.com/ppisarczyk/43962d06686722d26d176fad46879d41) (also provided in Appendix C of the paper).
Near-deduplication was implemented in the pre-processing pipeline on top of exact deduplication. To find near-duplicates, MinHash with 256 permutations of all documents was computed in linear time. Locality Sensitive Hashing was used to find the clusters of duplicates. Jaccard Similarities were computed inside these clusters to remove any false positives and with a similarity threshold of 0.85. Roughly 40% of permissively licensed files were (near-)duplicates. See section 3 of the paper for further details.
The following are not stored:
- Files that cannot contribute to training code: binary, empty, could not be decoded
- Files larger than 1MB
- The excluded file extensions are listed in Appendix B of the paper.
##### License detection
Permissive licenses have minimal restrictions on how the software can be copied, modified, and redistributed. The full list of licenses can be found [here](https://huggingface.co/datasets/bigcode/the-stack-dedup/blob/main/licenses.json)
GHArchive contained the license information for approximately 12% of the collected repositories. For the remaining repositories, [go-license-detector](https://github.com/src-d/go-license-detector) was run to detect the most likely SPDX license identifier. The detector did not detect a license for ~81% of the repositories, in which case the repository was excluded from the dataset.
A file was in included in the safe license dataset if at least one of the repositories containing the file had a permissive license.
#### Who are the source language producers?
The source (code) language producers are users of GitHub that created unique repository names between January 1st, 2015, and March 31st, 2022.
### Personal and Sensitive Information
The released dataset may contain sensitive information such as emails, IP addresses, and API/ssh keys that have previously been published to public repositories on GitHub. Deduplication has helped to reduce the amount of sensitive data that may exist. In the event that the dataset contains personal information, researchers should only use public, non-personal information in support of conducting and publishing their [open-access](https://en.wikipedia.org/wiki/Open_access) research. Personal information should not be used for spamming purposes, including sending unsolicited emails or selling of personal information. Complaints, removal requests, and "do not contact" requests can be sent to contact@bigcode-project.org.
The PII pipeline for this dataset is still a work in progress (see this [issue](https://github.com/bigcode-project/admin/issues/9) for updates). Researchers that wish to contribute to the anonymization pipeline of the project can apply to join [here](https://www.bigcode-project.org/docs/about/join/). Developers with source code in the dataset can request to have it removed [here](https://www.bigcode-project.org/docs/about/ip/) (proof of code contribution is required).
### Opting out of The Stack
We are giving developers the ability to have their code removed from the dataset upon request. The process for submitting and enacting removal requests will keep evolving throughout the project as we receive feedback and build up more data governance tools.
You can check if your code is in The Stack with the following ["Am I In The Stack?" Space](https://huggingface.co/spaces/bigcode/in-the-stack). If you'd like to have your data removed from the dataset follow the [instructions on GitHub](https://github.com/bigcode-project/opt-out-v2).
## Considerations for Using the Data
### Social Impact of Dataset
The Stack is an output of the BigCode Project. BigCode aims to be responsible by design and by default. The project is conducted in the spirit of Open Science, focused on the responsible development of LLMs for code.
With the release of The Stack, we aim to increase access, reproducibility, and transparency of code LLMs in the research community. Work to de-risk and improve on the implementation of ethical best practices of code LLMs is conducted in various BigCode working groups. The Legal, Ethics, and Governance working group has explored topics such as licensing (including copyleft and the intended use of permissively licensed code), attribution of generated code to original code, rights to restrict processing, the inclusion of Personally Identifiable Information (PII), and risks of malicious code, among other topics. This work is ongoing as of October 25th, 2022.
We expect code LLMs to enable people from diverse backgrounds to write higher quality code and develop low-code applications. Mission-critical software could become easier to maintain as professional developers are guided by code-generating systems on how to write more robust and efficient code. While the social impact is intended to be positive, the increased accessibility of code LLMs comes with certain risks such as over-reliance on the generated code and long-term effects on the software development job market.
A broader impact analysis relating to Code LLMs can be found in section 7 of this [paper](https://arxiv.org/abs/2107.03374). An in-depth risk assessments for Code LLMs can be found in section 4 of this [paper](https://arxiv.org/abs/2207.14157).
### Discussion of Biases
The code collected from GitHub does not contain demographic information or proxy information about the demographics. However, it is not without risks,
as the comments within the code may contain harmful or offensive language, which could be learned by the models.
Widely adopted programming languages like C and Javascript are overrepresented compared to niche programming languages like Julia and Scala. Some programming languages such as SQL, Batchfile, TypeScript are less likely to be permissively licensed (4% vs the average 10%). This may result in a biased representation of those languages. Permissively licensed files also tend to be longer.
Roughly 40 natural languages are present in docstrings and comments with English being the most prevalent. In python files, it makes up ~96% of the dataset.
For further information on data analysis of the Stack, see this [repo](https://github.com/bigcode-project/bigcode-analysis).
### Other Known Limitations
One of the current limitations of The Stack is that scraped HTML for websites may not be compliant with Web Content Accessibility Guidelines ([WCAG](https://www.w3.org/WAI/standards-guidelines/wcag/)). This could have an impact on HTML-generated code that may introduce web accessibility issues.
The training dataset could contain malicious code and/or the model could be used to generate malware or ransomware.
To the best of our knowledge, all files contained in the dataset are licensed with one of the permissive licenses (see list in [Licensing information](#licensing-information)). The accuracy of license attribution is limited by the accuracy of GHArchive and go-license-detector. Any mistakes should be reported to BigCode Project for review and follow-up as needed.
## Additional Information
### Dataset Curators
1. Harm de Vries, ServiceNow Research, harm.devries@servicenow.com
2. Leandro von Werra, Hugging Face, leandro@huggingface.co
### Licensing Information
The Stack is a collection of source code from repositories with various licenses. Any use of all or part of the code gathered in The Stack must abide by the terms of the original licenses, including attribution clauses when relevant. We facilitate this by providing provenance information for each data point.
The list of [SPDX license identifiers](https://spdx.org/licenses/) included in the dataset can be found [here](https://huggingface.co/datasets/bigcode/the-stack-dedup/blob/main/licenses.json).
### Citation Information
```
@article{Kocetkov2022TheStack,
title={The Stack: 3 TB of permissively licensed source code},
author={Kocetkov, Denis and Li, Raymond and Ben Allal, Loubna and Li, Jia and Mou,Chenghao and Muñoz Ferrandis, Carlos and Jernite, Yacine and Mitchell, Margaret and Hughes, Sean and Wolf, Thomas and Bahdanau, Dzmitry and von Werra, Leandro and de Vries, Harm},
journal={Preprint},
year={2022}
}
```
### Contributions
[More Information Needed]
## Terms of Use for The Stack
The Stack dataset is a collection of source code in over 300 programming languages. We ask that you read and acknowledge the following points before using the dataset:
1. The Stack is a collection of source code from repositories with various licenses. Any use of all or part of the code gathered in The Stack must abide by the terms of the original licenses, including attribution clauses when relevant. We facilitate this by providing provenance information for each data point.
2. The Stack is regularly updated to enact validated data removal requests. By clicking on "Access repository", you agree to update your own version of The Stack to the most recent usable version specified by the maintainers in [the following thread](https://huggingface.co/datasets/bigcode/the-stack/discussions/7). If you have questions about dataset versions and allowed uses, please also ask them in the dataset’s [community discussions](https://huggingface.co/datasets/bigcode/the-stack/discussions/new). We will also notify users via email when the latest usable version changes.
3. To host, share, or otherwise provide access to The Stack dataset, you must include these Terms of Use and require users to agree to it.
|
e2e_nlg | 2022-11-18T19:59:40.000Z | [
"task_categories:text2text-generation",
"annotations_creators:crowdsourced",
"language_creators:crowdsourced",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:cc-by-sa-4.0",
"meaning-representation-to-text",
"arxiv:1706.09254",
"arxiv:1901.11528",
"region:us"
] | null | The E2E dataset is used for training end-to-end, data-driven natural language generation systems in the restaurant domain, which is ten times bigger than existing, frequently used datasets in this area.
The E2E dataset poses new challenges:
(1) its human reference texts show more lexical richness and syntactic variation, including discourse phenomena;
(2) generating from this set requires content selection. As such, learning from this dataset promises more natural, varied and less template-like system utterances.
E2E is released in the following paper where you can find more details and baseline results:
https://arxiv.org/abs/1706.09254 | @article{dusek.etal2020:csl,
title = {Evaluating the {{State}}-of-the-{{Art}} of {{End}}-to-{{End Natural Language Generation}}: {{The E2E NLG Challenge}}},
author = {Du{\v{s}}ek, Ond\v{r}ej and Novikova, Jekaterina and Rieser, Verena},
year = {2020},
month = jan,
volume = {59},
pages = {123--156},
doi = {10.1016/j.csl.2019.06.009},
archivePrefix = {arXiv},
eprint = {1901.11528},
eprinttype = {arxiv},
journal = {Computer Speech & Language}
} | null | 7 | 7,392 | ---
annotations_creators:
- crowdsourced
language_creators:
- crowdsourced
language:
- en
license:
- cc-by-sa-4.0
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- text2text-generation
task_ids: []
paperswithcode_id: e2e
pretty_name: End-to-End NLG Challenge
tags:
- meaning-representation-to-text
dataset_info:
features:
- name: meaning_representation
dtype: string
- name: human_reference
dtype: string
splits:
- name: train
num_bytes: 9435824
num_examples: 42061
- name: validation
num_bytes: 1171723
num_examples: 4672
- name: test
num_bytes: 1320205
num_examples: 4693
download_size: 11812316
dataset_size: 11927752
---
# Dataset Card for End-to-End NLG Challenge
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [homepage](http://www.macs.hw.ac.uk/InteractionLab/E2E/)
- **Repository:** [repository](https://github.com/tuetschek/e2e-dataset/)
- **Paper:** [paper](https://arxiv.org/abs/1706.09254)
- **Leaderboard:** [leaderboard](http://www.macs.hw.ac.uk/InteractionLab/E2E/)
### Dataset Summary
The E2E dataset is used for training end-to-end, data-driven natural language generation systems in the restaurant domain, which is ten times bigger than existing, frequently used datasets in this area.
The E2E dataset poses new challenges:
(1) its human reference texts show more lexical richness and syntactic variation, including discourse phenomena;
(2) generating from this set requires content selection. As such, learning from this dataset promises more natural, varied and less template-like system utterances.
E2E is released in the following paper where you can find more details and baseline results:
https://arxiv.org/abs/1706.09254
### Supported Tasks and Leaderboards
- `text2text-generation-other-meaning-representation-to-text`: The dataset can be used to train a model to generate descriptions in the restaurant domain from meaning representations, which consists in taking as input some data about a restaurant and generate a sentence in natural language that presents the different aspects of the data about the restaurant.. Success on this task is typically measured by achieving a *high* [BLEU](https://huggingface.co/metrics/bleu), [NIST](https://huggingface.co/metrics/nist), [METEOR](https://huggingface.co/metrics/meteor), [Rouge-L](https://huggingface.co/metrics/rouge), [CIDEr](https://huggingface.co/metrics/cider). The TGen model (Dusek and Jurcıcek, 2016a) was used a baseline, had the following scores:
| | BLEU | NIST | METEOR | ROUGE_L | CIDEr |
| -------- | ------ | ------ | ------ | ------- | ------ |
| BASELINE | 0.6593 | 8.6094 | 0.4483 | 0.6850 | 2.2338 |
This task has an inactive leaderboard which can be found [here](http://www.macs.hw.ac.uk/InteractionLab/E2E/) and ranks models based on the metrics above.
### Languages
The dataset is in english (en).
## Dataset Structure
### Data Instances
Example of one instance:
```
{'human_reference': 'The Vaults pub near Café Adriatic has a 5 star rating. Prices start at £30.',
'meaning_representation': 'name[The Vaults], eatType[pub], priceRange[more than £30], customer rating[5 out of 5], near[Café Adriatic]'}
```
### Data Fields
- `human_reference`: string, the text is natural language that describes the different characteristics in the meaning representation
- `meaning_representation`: list of slots and values to generate a description from
Each MR consists of 3–8 attributes (slots), such as name, food or area, and their values.
### Data Splits
The dataset is split into training, validation and testing sets (in a 76.5-8.5-15 ratio), keeping a similar distribution of MR and reference text lengths and ensuring that MRs in different sets are distinct.
| | train | validation | test |
| ----- |-------:|------------:|------:|
| N. Instances | 42061 | 4672 | 4693 |
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
[More Information Needed]
#### Initial Data Collection and Normalization
The data was collected using the CrowdFlower platform and quality-controlled following Novikova et al. (2016).
#### Who are the source language producers?
[More Information Needed]
### Annotations
Following Novikova et al. (2016), the E2E data was collected using pictures as stimuli, which was shown to elicit significantly more natural, more informative, and better phrased human references than textual MRs.
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
```
@article{dusek.etal2020:csl,
title = {Evaluating the {{State}}-of-the-{{Art}} of {{End}}-to-{{End Natural Language Generation}}: {{The E2E NLG Challenge}}},
author = {Du{\v{s}}ek, Ond\v{r}ej and Novikova, Jekaterina and Rieser, Verena},
year = {2020},
month = jan,
volume = {59},
pages = {123--156},
doi = {10.1016/j.csl.2019.06.009},
archivePrefix = {arXiv},
eprint = {1901.11528},
eprinttype = {arxiv},
journal = {Computer Speech \& Language}
```
### Contributions
Thanks to [@lhoestq](https://github.com/lhoestq) for adding this dataset. |
codeparrot/github-code-clean | 2022-07-05T09:35:14.000Z | [
"license:apache-2.0",
"region:us"
] | codeparrot | The GitHub Code clean dataset in a more filtered version of codeparrot/github-code dataset, it consists of 115M code files from GitHub in 32 programming languages with 60 extensions totaling in almost 1TB of text data. | null | null | 55 | 7,339 | ---
license: apache-2.0
---
This is a cleaner version of [Github-code dataset](https://huggingface.co/datasets/codeparrot/github-code), we add the following filters:
* Average line length < 100
* Alpha numeric characters fraction > 0.25
* Remove auto-generated files (keyword search)
3.39M files are removed making up 2.94% of the dataset. |
financial_phrasebank | 2023-07-26T06:27:17.000Z | [
"task_categories:text-classification",
"task_ids:multi-class-classification",
"task_ids:sentiment-classification",
"annotations_creators:expert-generated",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:1K<n<10K",
"source_datasets:original",
"language:en",
"license:cc-by-nc-sa-3.0",
"finance",
"arxiv:1307.5336",
"region:us"
] | null | The key arguments for the low utilization of statistical techniques in
financial sentiment analysis have been the difficulty of implementation for
practical applications and the lack of high quality training data for building
such models. Especially in the case of finance and economic texts, annotated
collections are a scarce resource and many are reserved for proprietary use
only. To resolve the missing training data problem, we present a collection of
∼ 5000 sentences to establish human-annotated standards for benchmarking
alternative modeling techniques.
The objective of the phrase level annotation task was to classify each example
sentence into a positive, negative or neutral category by considering only the
information explicitly available in the given sentence. Since the study is
focused only on financial and economic domains, the annotators were asked to
consider the sentences from the view point of an investor only; i.e. whether
the news may have positive, negative or neutral influence on the stock price.
As a result, sentences which have a sentiment that is not relevant from an
economic or financial perspective are considered neutral.
This release of the financial phrase bank covers a collection of 4840
sentences. The selected collection of phrases was annotated by 16 people with
adequate background knowledge on financial markets. Three of the annotators
were researchers and the remaining 13 annotators were master’s students at
Aalto University School of Business with majors primarily in finance,
accounting, and economics.
Given the large number of overlapping annotations (5 to 8 annotations per
sentence), there are several ways to define a majority vote based gold
standard. To provide an objective comparison, we have formed 4 alternative
reference datasets based on the strength of majority agreement: all annotators
agree, >=75% of annotators agree, >=66% of annotators agree and >=50% of
annotators agree. | @article{Malo2014GoodDO,
title={Good debt or bad debt: Detecting semantic orientations in economic texts},
author={P. Malo and A. Sinha and P. Korhonen and J. Wallenius and P. Takala},
journal={Journal of the Association for Information Science and Technology},
year={2014},
volume={65}
} | null | 104 | 7,304 | ---
annotations_creators:
- expert-generated
language_creators:
- found
language:
- en
license:
- cc-by-nc-sa-3.0
multilinguality:
- monolingual
size_categories:
- 1K<n<10K
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- multi-class-classification
- sentiment-classification
pretty_name: FinancialPhrasebank
dataset_info:
- config_name: sentences_allagree
features:
- name: sentence
dtype: string
- name: label
dtype:
class_label:
names:
'0': negative
'1': neutral
'2': positive
splits:
- name: train
num_bytes: 303371
num_examples: 2264
download_size: 681890
dataset_size: 303371
- config_name: sentences_75agree
features:
- name: sentence
dtype: string
- name: label
dtype:
class_label:
names:
'0': negative
'1': neutral
'2': positive
splits:
- name: train
num_bytes: 472703
num_examples: 3453
download_size: 681890
dataset_size: 472703
- config_name: sentences_66agree
features:
- name: sentence
dtype: string
- name: label
dtype:
class_label:
names:
'0': negative
'1': neutral
'2': positive
splits:
- name: train
num_bytes: 587152
num_examples: 4217
download_size: 681890
dataset_size: 587152
- config_name: sentences_50agree
features:
- name: sentence
dtype: string
- name: label
dtype:
class_label:
names:
'0': negative
'1': neutral
'2': positive
splits:
- name: train
num_bytes: 679240
num_examples: 4846
download_size: 681890
dataset_size: 679240
tags:
- finance
---
# Dataset Card for financial_phrasebank
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [Kaggle](https://www.kaggle.com/ankurzing/sentiment-analysis-for-financial-news) [ResearchGate](https://www.researchgate.net/publication/251231364_FinancialPhraseBank-v10)
- **Repository:**
- **Paper:** [Arxiv](https://arxiv.org/abs/1307.5336)
- **Leaderboard:** [Kaggle](https://www.kaggle.com/ankurzing/sentiment-analysis-for-financial-news/code) [PapersWithCode](https://paperswithcode.com/sota/sentiment-analysis-on-financial-phrasebank) =
- **Point of Contact:** [Pekka Malo](mailto:pekka.malo@aalto.fi) [Ankur Sinha](mailto:ankur.sinha@aalto.fi)
### Dataset Summary
Polar sentiment dataset of sentences from financial news. The dataset consists of 4840 sentences from English language financial news categorised by sentiment. The dataset is divided by agreement rate of 5-8 annotators.
### Supported Tasks and Leaderboards
Sentiment Classification
### Languages
English
## Dataset Structure
### Data Instances
```
{ "sentence": "Pharmaceuticals group Orion Corp reported a fall in its third-quarter earnings that were hit by larger expenditures on R&D and marketing .",
"label": "negative"
}
```
### Data Fields
- sentence: a tokenized line from the dataset
- label: a label corresponding to the class as a string: 'positive', 'negative' or 'neutral'
### Data Splits
There's no train/validation/test split.
However the dataset is available in four possible configurations depending on the percentage of agreement of annotators:
`sentences_50agree`; Number of instances with >=50% annotator agreement: 4846
`sentences_66agree`: Number of instances with >=66% annotator agreement: 4217
`sentences_75agree`: Number of instances with >=75% annotator agreement: 3453
`sentences_allagree`: Number of instances with 100% annotator agreement: 2264
## Dataset Creation
### Curation Rationale
The key arguments for the low utilization of statistical techniques in
financial sentiment analysis have been the difficulty of implementation for
practical applications and the lack of high quality training data for building
such models. Especially in the case of finance and economic texts, annotated
collections are a scarce resource and many are reserved for proprietary use
only. To resolve the missing training data problem, we present a collection of
∼ 5000 sentences to establish human-annotated standards for benchmarking
alternative modeling techniques.
The objective of the phrase level annotation task was to classify each example
sentence into a positive, negative or neutral category by considering only the
information explicitly available in the given sentence. Since the study is
focused only on financial and economic domains, the annotators were asked to
consider the sentences from the view point of an investor only; i.e. whether
the news may have positive, negative or neutral influence on the stock price.
As a result, sentences which have a sentiment that is not relevant from an
economic or financial perspective are considered neutral.
### Source Data
#### Initial Data Collection and Normalization
The corpus used in this paper is made out of English news on all listed
companies in OMX Helsinki. The news has been downloaded from the LexisNexis
database using an automated web scraper. Out of this news database, a random
subset of 10,000 articles was selected to obtain good coverage across small and
large companies, companies in different industries, as well as different news
sources. Following the approach taken by Maks and Vossen (2010), we excluded
all sentences which did not contain any of the lexicon entities. This reduced
the overall sample to 53,400 sentences, where each has at least one or more
recognized lexicon entity. The sentences were then classified according to the
types of entity sequences detected. Finally, a random sample of ∼5000 sentences
was chosen to represent the overall news database.
#### Who are the source language producers?
The source data was written by various financial journalists.
### Annotations
#### Annotation process
This release of the financial phrase bank covers a collection of 4840
sentences. The selected collection of phrases was annotated by 16 people with
adequate background knowledge on financial markets.
Given the large number of overlapping annotations (5 to 8 annotations per
sentence), there are several ways to define a majority vote based gold
standard. To provide an objective comparison, we have formed 4 alternative
reference datasets based on the strength of majority agreement:
#### Who are the annotators?
Three of the annotators were researchers and the remaining 13 annotators were
master's students at Aalto University School of Business with majors primarily
in finance, accounting, and economics.
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
All annotators were from the same institution and so interannotator agreement
should be understood with this taken into account.
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
This work is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported License. To view a copy of this license, visit http://creativecommons.org/licenses/by-nc-sa/3.0/.
If you are interested in commercial use of the data, please contact the following authors for an appropriate license:
- [Pekka Malo](mailto:pekka.malo@aalto.fi)
- [Ankur Sinha](mailto:ankur.sinha@aalto.fi)
### Citation Information
```
@article{Malo2014GoodDO,
title={Good debt or bad debt: Detecting semantic orientations in economic texts},
author={P. Malo and A. Sinha and P. Korhonen and J. Wallenius and P. Takala},
journal={Journal of the Association for Information Science and Technology},
year={2014},
volume={65}
}
```
### Contributions
Thanks to [@frankier](https://github.com/frankier) for adding this dataset. |
agemagician/uniref50 | 2023-10-07T23:04:56.000Z | [
"region:us"
] | agemagician | null | null | null | 2 | 7,299 | Entry not found |
opus100 | 2023-06-01T14:59:58.000Z | [
"task_categories:translation",
"annotations_creators:no-annotation",
"language_creators:found",
"multilinguality:translation",
"size_categories:100K<n<1M",
"size_categories:10K<n<100K",
"size_categories:1K<n<10K",
"size_categories:1M<n<10M",
"size_categories:n<1K",
"source_datasets:extended",
"language:af",
"language:am",
"language:an",
"language:ar",
"language:as",
"language:az",
"language:be",
"language:bg",
"language:bn",
"language:br",
"language:bs",
"language:ca",
"language:cs",
"language:cy",
"language:da",
"language:de",
"language:dz",
"language:el",
"language:en",
"language:eo",
"language:es",
"language:et",
"language:eu",
"language:fa",
"language:fi",
"language:fr",
"language:fy",
"language:ga",
"language:gd",
"language:gl",
"language:gu",
"language:ha",
"language:he",
"language:hi",
"language:hr",
"language:hu",
"language:hy",
"language:id",
"language:ig",
"language:is",
"language:it",
"language:ja",
"language:ka",
"language:kk",
"language:km",
"language:kn",
"language:ko",
"language:ku",
"language:ky",
"language:li",
"language:lt",
"language:lv",
"language:mg",
"language:mk",
"language:ml",
"language:mn",
"language:mr",
"language:ms",
"language:mt",
"language:my",
"language:nb",
"language:ne",
"language:nl",
"language:nn",
"language:no",
"language:oc",
"language:or",
"language:pa",
"language:pl",
"language:ps",
"language:pt",
"language:ro",
"language:ru",
"language:rw",
"language:se",
"language:sh",
"language:si",
"language:sk",
"language:sl",
"language:sq",
"language:sr",
"language:sv",
"language:ta",
"language:te",
"language:tg",
"language:th",
"language:tk",
"language:tr",
"language:tt",
"language:ug",
"language:uk",
"language:ur",
"language:uz",
"language:vi",
"language:wa",
"language:xh",
"language:yi",
"language:yo",
"language:zh",
"language:zu",
"license:unknown",
"arxiv:2004.11867",
"region:us"
] | null | OPUS-100 is English-centric, meaning that all training pairs include English on either the source or target side.
The corpus covers 100 languages (including English).OPUS-100 contains approximately 55M sentence pairs.
Of the 99 language pairs, 44 have 1M sentence pairs of training data, 73 have at least 100k, and 95 have at least 10k. | @misc{zhang2020improving,
title={Improving Massively Multilingual Neural Machine Translation and Zero-Shot Translation},
author={Biao Zhang and Philip Williams and Ivan Titov and Rico Sennrich},
year={2020},
eprint={2004.11867},
archivePrefix={arXiv},
primaryClass={cs.CL}
} | null | 56 | 7,186 | ---
pretty_name: Opus100
task_categories:
- translation
multilinguality:
- translation
task_ids: []
language:
- af
- am
- an
- ar
- as
- az
- be
- bg
- bn
- br
- bs
- ca
- cs
- cy
- da
- de
- dz
- el
- en
- eo
- es
- et
- eu
- fa
- fi
- fr
- fy
- ga
- gd
- gl
- gu
- ha
- he
- hi
- hr
- hu
- hy
- id
- ig
- is
- it
- ja
- ka
- kk
- km
- kn
- ko
- ku
- ky
- li
- lt
- lv
- mg
- mk
- ml
- mn
- mr
- ms
- mt
- my
- nb
- ne
- nl
- nn
- 'no'
- oc
- or
- pa
- pl
- ps
- pt
- ro
- ru
- rw
- se
- sh
- si
- sk
- sl
- sq
- sr
- sv
- ta
- te
- tg
- th
- tk
- tr
- tt
- ug
- uk
- ur
- uz
- vi
- wa
- xh
- yi
- yo
- zh
- zu
annotations_creators:
- no-annotation
language_creators:
- found
source_datasets:
- extended
size_categories:
- 100K<n<1M
- 10K<n<100K
- 1K<n<10K
- 1M<n<10M
- n<1K
license:
- unknown
paperswithcode_id: opus-100
dataset_info:
- config_name: af-en
features:
- name: translation
dtype:
translation:
languages:
- af
- en
splits:
- name: test
num_bytes: 135916
num_examples: 2000
- name: train
num_bytes: 18726471
num_examples: 275512
- name: validation
num_bytes: 132777
num_examples: 2000
download_size: 7505036
dataset_size: 18995164
- config_name: am-en
features:
- name: translation
dtype:
translation:
languages:
- am
- en
splits:
- name: test
num_bytes: 588029
num_examples: 2000
- name: train
num_bytes: 21950644
num_examples: 89027
- name: validation
num_bytes: 566077
num_examples: 2000
download_size: 7004193
dataset_size: 23104750
- config_name: an-en
features:
- name: translation
dtype:
translation:
languages:
- an
- en
splits:
- name: train
num_bytes: 438332
num_examples: 6961
download_size: 96148
dataset_size: 438332
- config_name: ar-en
features:
- name: translation
dtype:
translation:
languages:
- ar
- en
splits:
- name: test
num_bytes: 331648
num_examples: 2000
- name: train
num_bytes: 152766484
num_examples: 1000000
- name: validation
num_bytes: 2272106
num_examples: 2000
download_size: 55286865
dataset_size: 155370238
- config_name: as-en
features:
- name: translation
dtype:
translation:
languages:
- as
- en
splits:
- name: test
num_bytes: 261466
num_examples: 2000
- name: train
num_bytes: 15634648
num_examples: 138479
- name: validation
num_bytes: 248139
num_examples: 2000
download_size: 4183517
dataset_size: 16144253
- config_name: az-en
features:
- name: translation
dtype:
translation:
languages:
- az
- en
splits:
- name: test
num_bytes: 393109
num_examples: 2000
- name: train
num_bytes: 56431259
num_examples: 262089
- name: validation
num_bytes: 407109
num_examples: 2000
download_size: 18897341
dataset_size: 57231477
- config_name: be-en
features:
- name: translation
dtype:
translation:
languages:
- be
- en
splits:
- name: test
num_bytes: 166858
num_examples: 2000
- name: train
num_bytes: 5298500
num_examples: 67312
- name: validation
num_bytes: 175205
num_examples: 2000
download_size: 1906088
dataset_size: 5640563
- config_name: bg-en
features:
- name: translation
dtype:
translation:
languages:
- bg
- en
splits:
- name: test
num_bytes: 243751
num_examples: 2000
- name: train
num_bytes: 108930347
num_examples: 1000000
- name: validation
num_bytes: 234848
num_examples: 2000
download_size: 36980744
dataset_size: 109408946
- config_name: bn-en
features:
- name: translation
dtype:
translation:
languages:
- bn
- en
splits:
- name: test
num_bytes: 510101
num_examples: 2000
- name: train
num_bytes: 249906846
num_examples: 1000000
- name: validation
num_bytes: 498414
num_examples: 2000
download_size: 72999655
dataset_size: 250915361
- config_name: br-en
features:
- name: translation
dtype:
translation:
languages:
- br
- en
splits:
- name: test
num_bytes: 127925
num_examples: 2000
- name: train
num_bytes: 8539006
num_examples: 153447
- name: validation
num_bytes: 133772
num_examples: 2000
download_size: 3323458
dataset_size: 8800703
- config_name: bs-en
features:
- name: translation
dtype:
translation:
languages:
- bs
- en
splits:
- name: test
num_bytes: 168622
num_examples: 2000
- name: train
num_bytes: 75082948
num_examples: 1000000
- name: validation
num_bytes: 172481
num_examples: 2000
download_size: 30746956
dataset_size: 75424051
- config_name: ca-en
features:
- name: translation
dtype:
translation:
languages:
- ca
- en
splits:
- name: test
num_bytes: 205666
num_examples: 2000
- name: train
num_bytes: 88405510
num_examples: 1000000
- name: validation
num_bytes: 212637
num_examples: 2000
download_size: 36267794
dataset_size: 88823813
- config_name: cs-en
features:
- name: translation
dtype:
translation:
languages:
- cs
- en
splits:
- name: test
num_bytes: 205274
num_examples: 2000
- name: train
num_bytes: 91897719
num_examples: 1000000
- name: validation
num_bytes: 219084
num_examples: 2000
download_size: 39673827
dataset_size: 92322077
- config_name: cy-en
features:
- name: translation
dtype:
translation:
languages:
- cy
- en
splits:
- name: test
num_bytes: 124289
num_examples: 2000
- name: train
num_bytes: 17244980
num_examples: 289521
- name: validation
num_bytes: 118856
num_examples: 2000
download_size: 6487005
dataset_size: 17488125
- config_name: da-en
features:
- name: translation
dtype:
translation:
languages:
- da
- en
splits:
- name: test
num_bytes: 298123
num_examples: 2000
- name: train
num_bytes: 126425274
num_examples: 1000000
- name: validation
num_bytes: 300624
num_examples: 2000
download_size: 50404122
dataset_size: 127024021
- config_name: de-en
features:
- name: translation
dtype:
translation:
languages:
- de
- en
splits:
- name: test
num_bytes: 330959
num_examples: 2000
- name: train
num_bytes: 152246756
num_examples: 1000000
- name: validation
num_bytes: 332350
num_examples: 2000
download_size: 67205361
dataset_size: 152910065
- config_name: dz-en
features:
- name: translation
dtype:
translation:
languages:
- dz
- en
splits:
- name: train
num_bytes: 81162
num_examples: 624
download_size: 17814
dataset_size: 81162
- config_name: el-en
features:
- name: translation
dtype:
translation:
languages:
- el
- en
splits:
- name: test
num_bytes: 302393
num_examples: 2000
- name: train
num_bytes: 127964703
num_examples: 1000000
- name: validation
num_bytes: 291234
num_examples: 2000
download_size: 43973686
dataset_size: 128558330
- config_name: en-eo
features:
- name: translation
dtype:
translation:
languages:
- en
- eo
splits:
- name: test
num_bytes: 167386
num_examples: 2000
- name: train
num_bytes: 24431953
num_examples: 337106
- name: validation
num_bytes: 168838
num_examples: 2000
download_size: 9999313
dataset_size: 24768177
- config_name: en-es
features:
- name: translation
dtype:
translation:
languages:
- en
- es
splits:
- name: test
num_bytes: 326270
num_examples: 2000
- name: train
num_bytes: 136643904
num_examples: 1000000
- name: validation
num_bytes: 326735
num_examples: 2000
download_size: 55534068
dataset_size: 137296909
- config_name: en-et
features:
- name: translation
dtype:
translation:
languages:
- en
- et
splits:
- name: test
num_bytes: 272171
num_examples: 2000
- name: train
num_bytes: 112299053
num_examples: 1000000
- name: validation
num_bytes: 276962
num_examples: 2000
download_size: 46235623
dataset_size: 112848186
- config_name: en-eu
features:
- name: translation
dtype:
translation:
languages:
- en
- eu
splits:
- name: test
num_bytes: 280885
num_examples: 2000
- name: train
num_bytes: 112330085
num_examples: 1000000
- name: validation
num_bytes: 281503
num_examples: 2000
download_size: 46389313
dataset_size: 112892473
- config_name: en-fa
features:
- name: translation
dtype:
translation:
languages:
- en
- fa
splits:
- name: test
num_bytes: 296556
num_examples: 2000
- name: train
num_bytes: 125401335
num_examples: 1000000
- name: validation
num_bytes: 291129
num_examples: 2000
download_size: 44568447
dataset_size: 125989020
- config_name: en-fi
features:
- name: translation
dtype:
translation:
languages:
- en
- fi
splits:
- name: test
num_bytes: 245822
num_examples: 2000
- name: train
num_bytes: 106025790
num_examples: 1000000
- name: validation
num_bytes: 247227
num_examples: 2000
download_size: 42563103
dataset_size: 106518839
- config_name: en-fr
features:
- name: translation
dtype:
translation:
languages:
- en
- fr
splits:
- name: test
num_bytes: 469731
num_examples: 2000
- name: train
num_bytes: 201441250
num_examples: 1000000
- name: validation
num_bytes: 481484
num_examples: 2000
download_size: 81009778
dataset_size: 202392465
- config_name: en-fy
features:
- name: translation
dtype:
translation:
languages:
- en
- fy
splits:
- name: test
num_bytes: 101246
num_examples: 2000
- name: train
num_bytes: 3895688
num_examples: 54342
- name: validation
num_bytes: 100129
num_examples: 2000
download_size: 1522187
dataset_size: 4097063
- config_name: en-ga
features:
- name: translation
dtype:
translation:
languages:
- en
- ga
splits:
- name: test
num_bytes: 503317
num_examples: 2000
- name: train
num_bytes: 42132742
num_examples: 289524
- name: validation
num_bytes: 503217
num_examples: 2000
download_size: 14998873
dataset_size: 43139276
- config_name: en-gd
features:
- name: translation
dtype:
translation:
languages:
- en
- gd
splits:
- name: test
num_bytes: 218362
num_examples: 1606
- name: train
num_bytes: 1254795
num_examples: 16316
- name: validation
num_bytes: 203885
num_examples: 1605
download_size: 564053
dataset_size: 1677042
- config_name: en-gl
features:
- name: translation
dtype:
translation:
languages:
- en
- gl
splits:
- name: test
num_bytes: 190699
num_examples: 2000
- name: train
num_bytes: 43327444
num_examples: 515344
- name: validation
num_bytes: 193606
num_examples: 2000
download_size: 18056665
dataset_size: 43711749
- config_name: en-gu
features:
- name: translation
dtype:
translation:
languages:
- en
- gu
splits:
- name: test
num_bytes: 199733
num_examples: 2000
- name: train
num_bytes: 33641975
num_examples: 318306
- name: validation
num_bytes: 205550
num_examples: 2000
download_size: 9407543
dataset_size: 34047258
- config_name: en-ha
features:
- name: translation
dtype:
translation:
languages:
- en
- ha
splits:
- name: test
num_bytes: 407352
num_examples: 2000
- name: train
num_bytes: 20391964
num_examples: 97983
- name: validation
num_bytes: 411526
num_examples: 2000
download_size: 6898482
dataset_size: 21210842
- config_name: en-he
features:
- name: translation
dtype:
translation:
languages:
- en
- he
splits:
- name: test
num_bytes: 208475
num_examples: 2000
- name: train
num_bytes: 91160431
num_examples: 1000000
- name: validation
num_bytes: 209446
num_examples: 2000
download_size: 31214136
dataset_size: 91578352
- config_name: en-hi
features:
- name: translation
dtype:
translation:
languages:
- en
- hi
splits:
- name: test
num_bytes: 496578
num_examples: 2000
- name: train
num_bytes: 124923977
num_examples: 534319
- name: validation
num_bytes: 474087
num_examples: 2000
download_size: 35993452
dataset_size: 125894642
- config_name: en-hr
features:
- name: translation
dtype:
translation:
languages:
- en
- hr
splits:
- name: test
num_bytes: 179644
num_examples: 2000
- name: train
num_bytes: 75310316
num_examples: 1000000
- name: validation
num_bytes: 179623
num_examples: 2000
download_size: 30728154
dataset_size: 75669583
- config_name: en-hu
features:
- name: translation
dtype:
translation:
languages:
- en
- hu
splits:
- name: test
num_bytes: 206047
num_examples: 2000
- name: train
num_bytes: 87484262
num_examples: 1000000
- name: validation
num_bytes: 208315
num_examples: 2000
download_size: 35696235
dataset_size: 87898624
- config_name: en-hy
features:
- name: translation
dtype:
translation:
languages:
- en
- hy
splits:
- name: train
num_bytes: 652631
num_examples: 7059
download_size: 215246
dataset_size: 652631
- config_name: en-id
features:
- name: translation
dtype:
translation:
languages:
- en
- id
splits:
- name: test
num_bytes: 177693
num_examples: 2000
- name: train
num_bytes: 78699773
num_examples: 1000000
- name: validation
num_bytes: 180032
num_examples: 2000
download_size: 29914089
dataset_size: 79057498
- config_name: en-ig
features:
- name: translation
dtype:
translation:
languages:
- en
- ig
splits:
- name: test
num_bytes: 137332
num_examples: 1843
- name: train
num_bytes: 1612539
num_examples: 18415
- name: validation
num_bytes: 135995
num_examples: 1843
download_size: 391849
dataset_size: 1885866
- config_name: en-is
features:
- name: translation
dtype:
translation:
languages:
- en
- is
splits:
- name: test
num_bytes: 170887
num_examples: 2000
- name: train
num_bytes: 73964915
num_examples: 1000000
- name: validation
num_bytes: 170640
num_examples: 2000
download_size: 28831218
dataset_size: 74306442
- config_name: en-it
features:
- name: translation
dtype:
translation:
languages:
- en
- it
splits:
- name: test
num_bytes: 299037
num_examples: 2000
- name: train
num_bytes: 123655086
num_examples: 1000000
- name: validation
num_bytes: 294362
num_examples: 2000
download_size: 50903618
dataset_size: 124248485
- config_name: en-ja
features:
- name: translation
dtype:
translation:
languages:
- en
- ja
splits:
- name: test
num_bytes: 190999
num_examples: 2000
- name: train
num_bytes: 88349369
num_examples: 1000000
- name: validation
num_bytes: 191419
num_examples: 2000
download_size: 34452575
dataset_size: 88731787
- config_name: en-ka
features:
- name: translation
dtype:
translation:
languages:
- en
- ka
splits:
- name: test
num_bytes: 256227
num_examples: 2000
- name: train
num_bytes: 42465706
num_examples: 377306
- name: validation
num_bytes: 260416
num_examples: 2000
download_size: 12743188
dataset_size: 42982349
- config_name: en-kk
features:
- name: translation
dtype:
translation:
languages:
- en
- kk
splits:
- name: test
num_bytes: 137664
num_examples: 2000
- name: train
num_bytes: 7124378
num_examples: 79927
- name: validation
num_bytes: 139665
num_examples: 2000
download_size: 2425372
dataset_size: 7401707
- config_name: en-km
features:
- name: translation
dtype:
translation:
languages:
- en
- km
splits:
- name: test
num_bytes: 289027
num_examples: 2000
- name: train
num_bytes: 19680611
num_examples: 111483
- name: validation
num_bytes: 302527
num_examples: 2000
download_size: 5193620
dataset_size: 20272165
- config_name: en-ko
features:
- name: translation
dtype:
translation:
languages:
- en
- ko
splits:
- name: test
num_bytes: 190696
num_examples: 2000
- name: train
num_bytes: 93665332
num_examples: 1000000
- name: validation
num_bytes: 189368
num_examples: 2000
download_size: 37602794
dataset_size: 94045396
- config_name: en-kn
features:
- name: translation
dtype:
translation:
languages:
- en
- kn
splits:
- name: test
num_bytes: 77205
num_examples: 918
- name: train
num_bytes: 1833334
num_examples: 14537
- name: validation
num_bytes: 77607
num_examples: 917
download_size: 525449
dataset_size: 1988146
- config_name: en-ku
features:
- name: translation
dtype:
translation:
languages:
- en
- ku
splits:
- name: test
num_bytes: 247847
num_examples: 2000
- name: train
num_bytes: 49107864
num_examples: 144844
- name: validation
num_bytes: 239325
num_examples: 2000
download_size: 14252198
dataset_size: 49595036
- config_name: en-ky
features:
- name: translation
dtype:
translation:
languages:
- en
- ky
splits:
- name: test
num_bytes: 142530
num_examples: 2000
- name: train
num_bytes: 1879298
num_examples: 27215
- name: validation
num_bytes: 138487
num_examples: 2000
download_size: 616902
dataset_size: 2160315
- config_name: en-li
features:
- name: translation
dtype:
translation:
languages:
- en
- li
splits:
- name: test
num_bytes: 93350
num_examples: 2000
- name: train
num_bytes: 1628601
num_examples: 25535
- name: validation
num_bytes: 92906
num_examples: 2000
download_size: 450092
dataset_size: 1814857
- config_name: en-lt
features:
- name: translation
dtype:
translation:
languages:
- en
- lt
splits:
- name: test
num_bytes: 482615
num_examples: 2000
- name: train
num_bytes: 177061044
num_examples: 1000000
- name: validation
num_bytes: 469117
num_examples: 2000
download_size: 69388131
dataset_size: 178012776
- config_name: en-lv
features:
- name: translation
dtype:
translation:
languages:
- en
- lv
splits:
- name: test
num_bytes: 536576
num_examples: 2000
- name: train
num_bytes: 206051849
num_examples: 1000000
- name: validation
num_bytes: 522072
num_examples: 2000
download_size: 78952903
dataset_size: 207110497
- config_name: en-mg
features:
- name: translation
dtype:
translation:
languages:
- en
- mg
splits:
- name: test
num_bytes: 525067
num_examples: 2000
- name: train
num_bytes: 130865649
num_examples: 590771
- name: validation
num_bytes: 511171
num_examples: 2000
download_size: 52470504
dataset_size: 131901887
- config_name: en-mk
features:
- name: translation
dtype:
translation:
languages:
- en
- mk
splits:
- name: test
num_bytes: 308934
num_examples: 2000
- name: train
num_bytes: 117069489
num_examples: 1000000
- name: validation
num_bytes: 305498
num_examples: 2000
download_size: 39517761
dataset_size: 117683921
- config_name: en-ml
features:
- name: translation
dtype:
translation:
languages:
- en
- ml
splits:
- name: test
num_bytes: 340626
num_examples: 2000
- name: train
num_bytes: 199971743
num_examples: 822746
- name: validation
num_bytes: 334459
num_examples: 2000
download_size: 48654808
dataset_size: 200646828
- config_name: en-mn
features:
- name: translation
dtype:
translation:
languages:
- en
- mn
splits:
- name: train
num_bytes: 250778
num_examples: 4294
download_size: 42039
dataset_size: 250778
- config_name: en-mr
features:
- name: translation
dtype:
translation:
languages:
- en
- mr
splits:
- name: test
num_bytes: 238612
num_examples: 2000
- name: train
num_bytes: 2724131
num_examples: 27007
- name: validation
num_bytes: 235540
num_examples: 2000
download_size: 910211
dataset_size: 3198283
- config_name: en-ms
features:
- name: translation
dtype:
translation:
languages:
- en
- ms
splits:
- name: test
num_bytes: 179705
num_examples: 2000
- name: train
num_bytes: 76829645
num_examples: 1000000
- name: validation
num_bytes: 180183
num_examples: 2000
download_size: 29807607
dataset_size: 77189533
- config_name: en-mt
features:
- name: translation
dtype:
translation:
languages:
- en
- mt
splits:
- name: test
num_bytes: 566134
num_examples: 2000
- name: train
num_bytes: 222222396
num_examples: 1000000
- name: validation
num_bytes: 594386
num_examples: 2000
download_size: 84757608
dataset_size: 223382916
- config_name: en-my
features:
- name: translation
dtype:
translation:
languages:
- en
- my
splits:
- name: test
num_bytes: 337351
num_examples: 2000
- name: train
num_bytes: 3673501
num_examples: 24594
- name: validation
num_bytes: 336155
num_examples: 2000
download_size: 1038600
dataset_size: 4347007
- config_name: en-nb
features:
- name: translation
dtype:
translation:
languages:
- en
- nb
splits:
- name: test
num_bytes: 334117
num_examples: 2000
- name: train
num_bytes: 13611709
num_examples: 142906
- name: validation
num_bytes: 324400
num_examples: 2000
download_size: 5706626
dataset_size: 14270226
- config_name: en-ne
features:
- name: translation
dtype:
translation:
languages:
- en
- ne
splits:
- name: test
num_bytes: 186527
num_examples: 2000
- name: train
num_bytes: 44136280
num_examples: 406381
- name: validation
num_bytes: 204920
num_examples: 2000
download_size: 11711988
dataset_size: 44527727
- config_name: en-nl
features:
- name: translation
dtype:
translation:
languages:
- en
- nl
splits:
- name: test
num_bytes: 282755
num_examples: 2000
- name: train
num_bytes: 112327073
num_examples: 1000000
- name: validation
num_bytes: 270940
num_examples: 2000
download_size: 45374708
dataset_size: 112880768
- config_name: en-nn
features:
- name: translation
dtype:
translation:
languages:
- en
- nn
splits:
- name: test
num_bytes: 179007
num_examples: 2000
- name: train
num_bytes: 32924821
num_examples: 486055
- name: validation
num_bytes: 187650
num_examples: 2000
download_size: 12742134
dataset_size: 33291478
- config_name: en-no
features:
- name: translation
dtype:
translation:
languages:
- en
- 'no'
splits:
- name: test
num_bytes: 173328
num_examples: 2000
- name: train
num_bytes: 74106283
num_examples: 1000000
- name: validation
num_bytes: 178013
num_examples: 2000
download_size: 28851262
dataset_size: 74457624
- config_name: en-oc
features:
- name: translation
dtype:
translation:
languages:
- en
- oc
splits:
- name: test
num_bytes: 82350
num_examples: 2000
- name: train
num_bytes: 1627206
num_examples: 35791
- name: validation
num_bytes: 81650
num_examples: 2000
download_size: 607192
dataset_size: 1791206
- config_name: en-or
features:
- name: translation
dtype:
translation:
languages:
- en
- or
splits:
- name: test
num_bytes: 163947
num_examples: 1318
- name: train
num_bytes: 1500749
num_examples: 14273
- name: validation
num_bytes: 155331
num_examples: 1317
download_size: 499401
dataset_size: 1820027
- config_name: en-pa
features:
- name: translation
dtype:
translation:
languages:
- en
- pa
splits:
- name: test
num_bytes: 133909
num_examples: 2000
- name: train
num_bytes: 8509228
num_examples: 107296
- name: validation
num_bytes: 136196
num_examples: 2000
download_size: 2589682
dataset_size: 8779333
- config_name: en-pl
features:
- name: translation
dtype:
translation:
languages:
- en
- pl
splits:
- name: test
num_bytes: 212503
num_examples: 2000
- name: train
num_bytes: 95248523
num_examples: 1000000
- name: validation
num_bytes: 218216
num_examples: 2000
download_size: 39320454
dataset_size: 95679242
- config_name: en-ps
features:
- name: translation
dtype:
translation:
languages:
- en
- ps
splits:
- name: test
num_bytes: 93003
num_examples: 2000
- name: train
num_bytes: 4436576
num_examples: 79127
- name: validation
num_bytes: 95164
num_examples: 2000
download_size: 1223087
dataset_size: 4624743
- config_name: en-pt
features:
- name: translation
dtype:
translation:
languages:
- en
- pt
splits:
- name: test
num_bytes: 296122
num_examples: 2000
- name: train
num_bytes: 118243649
num_examples: 1000000
- name: validation
num_bytes: 292082
num_examples: 2000
download_size: 48087550
dataset_size: 118831853
- config_name: en-ro
features:
- name: translation
dtype:
translation:
languages:
- en
- ro
splits:
- name: test
num_bytes: 198647
num_examples: 2000
- name: train
num_bytes: 85249851
num_examples: 1000000
- name: validation
num_bytes: 199172
num_examples: 2000
download_size: 35032743
dataset_size: 85647670
- config_name: en-ru
features:
- name: translation
dtype:
translation:
languages:
- en
- ru
splits:
- name: test
num_bytes: 490984
num_examples: 2000
- name: train
num_bytes: 195101737
num_examples: 1000000
- name: validation
num_bytes: 490246
num_examples: 2000
download_size: 68501634
dataset_size: 196082967
- config_name: en-rw
features:
- name: translation
dtype:
translation:
languages:
- en
- rw
splits:
- name: test
num_bytes: 136197
num_examples: 2000
- name: train
num_bytes: 15286303
num_examples: 173823
- name: validation
num_bytes: 134965
num_examples: 2000
download_size: 5233241
dataset_size: 15557465
- config_name: en-se
features:
- name: translation
dtype:
translation:
languages:
- en
- se
splits:
- name: test
num_bytes: 85705
num_examples: 2000
- name: train
num_bytes: 2047412
num_examples: 35907
- name: validation
num_bytes: 83672
num_examples: 2000
download_size: 806982
dataset_size: 2216789
- config_name: en-sh
features:
- name: translation
dtype:
translation:
languages:
- en
- sh
splits:
- name: test
num_bytes: 569487
num_examples: 2000
- name: train
num_bytes: 60900239
num_examples: 267211
- name: validation
num_bytes: 555602
num_examples: 2000
download_size: 22357505
dataset_size: 62025328
- config_name: en-si
features:
- name: translation
dtype:
translation:
languages:
- en
- si
splits:
- name: test
num_bytes: 271743
num_examples: 2000
- name: train
num_bytes: 114951675
num_examples: 979109
- name: validation
num_bytes: 271244
num_examples: 2000
download_size: 33247484
dataset_size: 115494662
- config_name: en-sk
features:
- name: translation
dtype:
translation:
languages:
- en
- sk
splits:
- name: test
num_bytes: 258042
num_examples: 2000
- name: train
num_bytes: 111743868
num_examples: 1000000
- name: validation
num_bytes: 255470
num_examples: 2000
download_size: 46618395
dataset_size: 112257380
- config_name: en-sl
features:
- name: translation
dtype:
translation:
languages:
- en
- sl
splits:
- name: test
num_bytes: 205478
num_examples: 2000
- name: train
num_bytes: 90270957
num_examples: 1000000
- name: validation
num_bytes: 198662
num_examples: 2000
download_size: 37536724
dataset_size: 90675097
- config_name: en-sq
features:
- name: translation
dtype:
translation:
languages:
- en
- sq
splits:
- name: test
num_bytes: 275379
num_examples: 2000
- name: train
num_bytes: 105745981
num_examples: 1000000
- name: validation
num_bytes: 267312
num_examples: 2000
download_size: 42697338
dataset_size: 106288672
- config_name: en-sr
features:
- name: translation
dtype:
translation:
languages:
- en
- sr
splits:
- name: test
num_bytes: 180232
num_examples: 2000
- name: train
num_bytes: 75726835
num_examples: 1000000
- name: validation
num_bytes: 184246
num_examples: 2000
download_size: 31260575
dataset_size: 76091313
- config_name: en-sv
features:
- name: translation
dtype:
translation:
languages:
- en
- sv
splits:
- name: test
num_bytes: 271014
num_examples: 2000
- name: train
num_bytes: 116985953
num_examples: 1000000
- name: validation
num_bytes: 279994
num_examples: 2000
download_size: 46694960
dataset_size: 117536961
- config_name: en-ta
features:
- name: translation
dtype:
translation:
languages:
- en
- ta
splits:
- name: test
num_bytes: 351990
num_examples: 2000
- name: train
num_bytes: 74044524
num_examples: 227014
- name: validation
num_bytes: 335557
num_examples: 2000
download_size: 17652443
dataset_size: 74732071
- config_name: en-te
features:
- name: translation
dtype:
translation:
languages:
- en
- te
splits:
- name: test
num_bytes: 190595
num_examples: 2000
- name: train
num_bytes: 6688625
num_examples: 64352
- name: validation
num_bytes: 193666
num_examples: 2000
download_size: 2011832
dataset_size: 7072886
- config_name: en-tg
features:
- name: translation
dtype:
translation:
languages:
- en
- tg
splits:
- name: test
num_bytes: 372120
num_examples: 2000
- name: train
num_bytes: 35477177
num_examples: 193882
- name: validation
num_bytes: 371728
num_examples: 2000
download_size: 11389877
dataset_size: 36221025
- config_name: en-th
features:
- name: translation
dtype:
translation:
languages:
- en
- th
splits:
- name: test
num_bytes: 290581
num_examples: 2000
- name: train
num_bytes: 132821031
num_examples: 1000000
- name: validation
num_bytes: 288366
num_examples: 2000
download_size: 38147204
dataset_size: 133399978
- config_name: en-tk
features:
- name: translation
dtype:
translation:
languages:
- en
- tk
splits:
- name: test
num_bytes: 83886
num_examples: 1852
- name: train
num_bytes: 719633
num_examples: 13110
- name: validation
num_bytes: 81014
num_examples: 1852
download_size: 157481
dataset_size: 884533
- config_name: en-tr
features:
- name: translation
dtype:
translation:
languages:
- en
- tr
splits:
- name: test
num_bytes: 183833
num_examples: 2000
- name: train
num_bytes: 78946365
num_examples: 1000000
- name: validation
num_bytes: 181917
num_examples: 2000
download_size: 30892429
dataset_size: 79312115
- config_name: en-tt
features:
- name: translation
dtype:
translation:
languages:
- en
- tt
splits:
- name: test
num_bytes: 693276
num_examples: 2000
- name: train
num_bytes: 35313258
num_examples: 100843
- name: validation
num_bytes: 701670
num_examples: 2000
download_size: 9940523
dataset_size: 36708204
- config_name: en-ug
features:
- name: translation
dtype:
translation:
languages:
- en
- ug
splits:
- name: test
num_bytes: 620881
num_examples: 2000
- name: train
num_bytes: 31576580
num_examples: 72170
- name: validation
num_bytes: 631236
num_examples: 2000
download_size: 8687743
dataset_size: 32828697
- config_name: en-uk
features:
- name: translation
dtype:
translation:
languages:
- en
- uk
splits:
- name: test
num_bytes: 249750
num_examples: 2000
- name: train
num_bytes: 104230356
num_examples: 1000000
- name: validation
num_bytes: 247131
num_examples: 2000
download_size: 37415496
dataset_size: 104727237
- config_name: en-ur
features:
- name: translation
dtype:
translation:
languages:
- en
- ur
splits:
- name: test
num_bytes: 538564
num_examples: 2000
- name: train
num_bytes: 268961304
num_examples: 753913
- name: validation
num_bytes: 529316
num_examples: 2000
download_size: 81092186
dataset_size: 270029184
- config_name: en-uz
features:
- name: translation
dtype:
translation:
languages:
- en
- uz
splits:
- name: test
num_bytes: 408683
num_examples: 2000
- name: train
num_bytes: 38375434
num_examples: 173157
- name: validation
num_bytes: 398861
num_examples: 2000
download_size: 11791643
dataset_size: 39182978
- config_name: en-vi
features:
- name: translation
dtype:
translation:
languages:
- en
- vi
splits:
- name: test
num_bytes: 192752
num_examples: 2000
- name: train
num_bytes: 82615270
num_examples: 1000000
- name: validation
num_bytes: 194729
num_examples: 2000
download_size: 30647296
dataset_size: 83002751
- config_name: en-wa
features:
- name: translation
dtype:
translation:
languages:
- en
- wa
splits:
- name: test
num_bytes: 87099
num_examples: 2000
- name: train
num_bytes: 6085948
num_examples: 104496
- name: validation
num_bytes: 87726
num_examples: 2000
download_size: 2119821
dataset_size: 6260773
- config_name: en-xh
features:
- name: translation
dtype:
translation:
languages:
- en
- xh
splits:
- name: test
num_bytes: 318660
num_examples: 2000
- name: train
num_bytes: 50607248
num_examples: 439671
- name: validation
num_bytes: 315839
num_examples: 2000
download_size: 20503199
dataset_size: 51241747
- config_name: en-yi
features:
- name: translation
dtype:
translation:
languages:
- en
- yi
splits:
- name: test
num_bytes: 96490
num_examples: 2000
- name: train
num_bytes: 1275143
num_examples: 15010
- name: validation
num_bytes: 99826
num_examples: 2000
download_size: 284031
dataset_size: 1471459
- config_name: en-yo
features:
- name: translation
dtype:
translation:
languages:
- en
- yo
splits:
- name: train
num_bytes: 979769
num_examples: 10375
download_size: 177540
dataset_size: 979769
- config_name: en-zh
features:
- name: translation
dtype:
translation:
languages:
- en
- zh
splits:
- name: test
num_bytes: 511372
num_examples: 2000
- name: train
num_bytes: 200062983
num_examples: 1000000
- name: validation
num_bytes: 512364
num_examples: 2000
download_size: 83265500
dataset_size: 201086719
- config_name: en-zu
features:
- name: translation
dtype:
translation:
languages:
- en
- zu
splits:
- name: test
num_bytes: 117518
num_examples: 2000
- name: train
num_bytes: 2799590
num_examples: 38616
- name: validation
num_bytes: 120141
num_examples: 2000
download_size: 889951
dataset_size: 3037249
- config_name: ar-de
features:
- name: translation
dtype:
translation:
languages:
- ar
- de
splits:
- name: test
num_bytes: 238599
num_examples: 2000
download_size: 2556791
dataset_size: 238599
- config_name: ar-fr
features:
- name: translation
dtype:
translation:
languages:
- ar
- fr
splits:
- name: test
num_bytes: 547382
num_examples: 2000
download_size: 2556791
dataset_size: 547382
- config_name: ar-nl
features:
- name: translation
dtype:
translation:
languages:
- ar
- nl
splits:
- name: test
num_bytes: 212936
num_examples: 2000
download_size: 2556791
dataset_size: 212936
- config_name: ar-ru
features:
- name: translation
dtype:
translation:
languages:
- ar
- ru
splits:
- name: test
num_bytes: 808270
num_examples: 2000
download_size: 2556791
dataset_size: 808270
- config_name: ar-zh
features:
- name: translation
dtype:
translation:
languages:
- ar
- zh
splits:
- name: test
num_bytes: 713412
num_examples: 2000
download_size: 2556791
dataset_size: 713412
- config_name: de-fr
features:
- name: translation
dtype:
translation:
languages:
- de
- fr
splits:
- name: test
num_bytes: 458746
num_examples: 2000
download_size: 2556791
dataset_size: 458746
- config_name: de-nl
features:
- name: translation
dtype:
translation:
languages:
- de
- nl
splits:
- name: test
num_bytes: 403886
num_examples: 2000
download_size: 2556791
dataset_size: 403886
- config_name: de-ru
features:
- name: translation
dtype:
translation:
languages:
- de
- ru
splits:
- name: test
num_bytes: 315779
num_examples: 2000
download_size: 2556791
dataset_size: 315779
- config_name: de-zh
features:
- name: translation
dtype:
translation:
languages:
- de
- zh
splits:
- name: test
num_bytes: 280397
num_examples: 2000
download_size: 2556791
dataset_size: 280397
- config_name: fr-nl
features:
- name: translation
dtype:
translation:
languages:
- fr
- nl
splits:
- name: test
num_bytes: 368646
num_examples: 2000
download_size: 2556791
dataset_size: 368646
- config_name: fr-ru
features:
- name: translation
dtype:
translation:
languages:
- fr
- ru
splits:
- name: test
num_bytes: 732724
num_examples: 2000
download_size: 2556791
dataset_size: 732724
- config_name: fr-zh
features:
- name: translation
dtype:
translation:
languages:
- fr
- zh
splits:
- name: test
num_bytes: 619394
num_examples: 2000
download_size: 2556791
dataset_size: 619394
- config_name: nl-ru
features:
- name: translation
dtype:
translation:
languages:
- nl
- ru
splits:
- name: test
num_bytes: 256067
num_examples: 2000
download_size: 2556791
dataset_size: 256067
- config_name: nl-zh
features:
- name: translation
dtype:
translation:
languages:
- nl
- zh
splits:
- name: test
num_bytes: 183641
num_examples: 2000
download_size: 2556791
dataset_size: 183641
- config_name: ru-zh
features:
- name: translation
dtype:
translation:
languages:
- ru
- zh
splits:
- name: test
num_bytes: 916114
num_examples: 2000
download_size: 2556791
dataset_size: 916114
config_names:
- af-en
- am-en
- an-en
- ar-de
- ar-en
- ar-fr
- ar-nl
- ar-ru
- ar-zh
- as-en
- az-en
- be-en
- bg-en
- bn-en
- br-en
- bs-en
- ca-en
- cs-en
- cy-en
- da-en
- de-en
- de-fr
- de-nl
- de-ru
- de-zh
- dz-en
- el-en
- en-eo
- en-es
- en-et
- en-eu
- en-fa
- en-fi
- en-fr
- en-fy
- en-ga
- en-gd
- en-gl
- en-gu
- en-ha
- en-he
- en-hi
- en-hr
- en-hu
- en-hy
- en-id
- en-ig
- en-is
- en-it
- en-ja
- en-ka
- en-kk
- en-km
- en-kn
- en-ko
- en-ku
- en-ky
- en-li
- en-lt
- en-lv
- en-mg
- en-mk
- en-ml
- en-mn
- en-mr
- en-ms
- en-mt
- en-my
- en-nb
- en-ne
- en-nl
- en-nn
- en-no
- en-oc
- en-or
- en-pa
- en-pl
- en-ps
- en-pt
- en-ro
- en-ru
- en-rw
- en-se
- en-sh
- en-si
- en-sk
- en-sl
- en-sq
- en-sr
- en-sv
- en-ta
- en-te
- en-tg
- en-th
- en-tk
- en-tr
- en-tt
- en-ug
- en-uk
- en-ur
- en-uz
- en-vi
- en-wa
- en-xh
- en-yi
- en-yo
- en-zh
- en-zu
- fr-nl
- fr-ru
- fr-zh
- nl-ru
- nl-zh
- ru-zh
---
# Dataset Card for Opus100
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [Link](http://opus.nlpl.eu/opus-100.php)
- **Repository:** [GitHub](https://github.com/EdinburghNLP/opus-100-corpus)
- **Paper:** [ARXIV](https://arxiv.org/abs/2004.11867)
- **Leaderboard:**
- **Point of Contact:**
### Dataset Summary
OPUS-100 is English-centric, meaning that all training pairs include English on either the source or target side. The corpus covers 100 languages (including English). Selected the languages based on the volume of parallel data available in OPUS.
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
OPUS-100 contains approximately 55M sentence pairs. Of the 99 language pairs, 44 have 1M sentence pairs of training data, 73 have at least 100k, and 95 have at least 10k.
## Dataset Structure
### Data Instances
```
{
"ca": "El departament de bombers té el seu propi equip d'investigació.",
"en": "Well, the fire department has its own investigative unit."
}
```
### Data Fields
- `src_tag`: `string` text in source language
- `tgt_tag`: `string` translation of source language in target language
### Data Splits
The dataset is split into training, development, and test portions. Data was prepared by randomly sampled up to 1M sentence pairs per language pair for training and up to 2000 each for development and test. To ensure that there was no overlap (at the monolingual sentence level) between the training and development/test data, they applied a filter during sampling to exclude sentences that had already been sampled. Note that this was done cross-lingually so that, for instance, an English sentence in the Portuguese-English portion of the training data could not occur in the Hindi-English test set.
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
[More Information Needed]
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
```
@misc{zhang2020improving,
title={Improving Massively Multilingual Neural Machine Translation and Zero-Shot Translation},
author={Biao Zhang and Philip Williams and Ivan Titov and Rico Sennrich},
year={2020},
eprint={2004.11867},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
### Contributions
Thanks to [@vasudevgupta7](https://github.com/vasudevgupta7) for adding this dataset. |
wics/strategy-qa | 2023-05-10T06:12:13.000Z | [
"license:other",
"region:us"
] | wics | null | 3 | 7,054 | ---
license: other
---
| ||
xquad | 2023-04-05T13:45:22.000Z | [
"task_categories:question-answering",
"task_ids:extractive-qa",
"annotations_creators:expert-generated",
"language_creators:expert-generated",
"multilinguality:multilingual",
"size_categories:unknown",
"source_datasets:extended|squad",
"language:ar",
"language:de",
"language:el",
"language:en",
"language:es",
"language:hi",
"language:ro",
"language:ru",
"language:th",
"language:tr",
"language:vi",
"language:zh",
"license:cc-by-sa-4.0",
"arxiv:1910.11856",
"region:us"
] | null | XQuAD (Cross-lingual Question Answering Dataset) is a benchmark dataset for evaluating cross-lingual question answering
performance. The dataset consists of a subset of 240 paragraphs and 1190 question-answer pairs from the development set
of SQuAD v1.1 (Rajpurkar et al., 2016) together with their professional translations into ten languages: Spanish, German,
Greek, Russian, Turkish, Arabic, Vietnamese, Thai, Chinese, Hindi and Romanian. Consequently, the dataset is entirely parallel
across 12 languages. | @article{Artetxe:etal:2019,
author = {Mikel Artetxe and Sebastian Ruder and Dani Yogatama},
title = {On the cross-lingual transferability of monolingual representations},
journal = {CoRR},
volume = {abs/1910.11856},
year = {2019},
archivePrefix = {arXiv},
eprint = {1910.11856}
} | null | 11 | 7,024 | ---
pretty_name: XQuAD
annotations_creators:
- expert-generated
language_creators:
- expert-generated
language:
- ar
- de
- el
- en
- es
- hi
- ro
- ru
- th
- tr
- vi
- zh
license:
- cc-by-sa-4.0
multilinguality:
- multilingual
size_categories:
- unknown
source_datasets:
- extended|squad
task_categories:
- question-answering
task_ids:
- extractive-qa
paperswithcode_id: xquad
dataset_info:
- config_name: xquad.ar
features:
- name: id
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
sequence:
- name: text
dtype: string
- name: answer_start
dtype: int32
splits:
- name: validation
num_bytes: 1722799
num_examples: 1190
download_size: 13962158
dataset_size: 1722799
- config_name: xquad.de
features:
- name: id
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
sequence:
- name: text
dtype: string
- name: answer_start
dtype: int32
splits:
- name: validation
num_bytes: 1283301
num_examples: 1190
download_size: 13962158
dataset_size: 1283301
- config_name: xquad.zh
features:
- name: id
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
sequence:
- name: text
dtype: string
- name: answer_start
dtype: int32
splits:
- name: validation
num_bytes: 984241
num_examples: 1190
download_size: 13962158
dataset_size: 984241
- config_name: xquad.vi
features:
- name: id
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
sequence:
- name: text
dtype: string
- name: answer_start
dtype: int32
splits:
- name: validation
num_bytes: 1477239
num_examples: 1190
download_size: 13962158
dataset_size: 1477239
- config_name: xquad.en
features:
- name: id
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
sequence:
- name: text
dtype: string
- name: answer_start
dtype: int32
splits:
- name: validation
num_bytes: 1116123
num_examples: 1190
download_size: 13962158
dataset_size: 1116123
- config_name: xquad.es
features:
- name: id
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
sequence:
- name: text
dtype: string
- name: answer_start
dtype: int32
splits:
- name: validation
num_bytes: 1273499
num_examples: 1190
download_size: 13962158
dataset_size: 1273499
- config_name: xquad.hi
features:
- name: id
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
sequence:
- name: text
dtype: string
- name: answer_start
dtype: int32
splits:
- name: validation
num_bytes: 2682975
num_examples: 1190
download_size: 13962158
dataset_size: 2682975
- config_name: xquad.el
features:
- name: id
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
sequence:
- name: text
dtype: string
- name: answer_start
dtype: int32
splits:
- name: validation
num_bytes: 2206690
num_examples: 1190
download_size: 13962158
dataset_size: 2206690
- config_name: xquad.th
features:
- name: id
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
sequence:
- name: text
dtype: string
- name: answer_start
dtype: int32
splits:
- name: validation
num_bytes: 2854959
num_examples: 1190
download_size: 13962158
dataset_size: 2854959
- config_name: xquad.tr
features:
- name: id
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
sequence:
- name: text
dtype: string
- name: answer_start
dtype: int32
splits:
- name: validation
num_bytes: 1210763
num_examples: 1190
download_size: 13962158
dataset_size: 1210763
- config_name: xquad.ru
features:
- name: id
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
sequence:
- name: text
dtype: string
- name: answer_start
dtype: int32
splits:
- name: validation
num_bytes: 2136990
num_examples: 1190
download_size: 13962158
dataset_size: 2136990
- config_name: xquad.ro
features:
- name: id
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
sequence:
- name: text
dtype: string
- name: answer_start
dtype: int32
splits:
- name: validation
num_bytes: 1299450
num_examples: 1190
download_size: 13962158
dataset_size: 1299450
---
# Dataset Card for "xquad"
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://github.com/deepmind/xquad](https://github.com/deepmind/xquad)
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 146.31 MB
- **Size of the generated dataset:** 18.97 MB
- **Total amount of disk used:** 165.28 MB
### Dataset Summary
XQuAD (Cross-lingual Question Answering Dataset) is a benchmark dataset for evaluating cross-lingual question answering
performance. The dataset consists of a subset of 240 paragraphs and 1190 question-answer pairs from the development set
of SQuAD v1.1 (Rajpurkar et al., 2016) together with their professional translations into ten languages: Spanish, German,
Greek, Russian, Turkish, Arabic, Vietnamese, Thai, Chinese, and Hindi. Consequently, the dataset is entirely parallel
across 11 languages.
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Dataset Structure
### Data Instances
#### xquad.ar
- **Size of downloaded dataset files:** 13.30 MB
- **Size of the generated dataset:** 1.72 MB
- **Total amount of disk used:** 15.03 MB
An example of 'validation' looks as follows.
```
This example was too long and was cropped:
{
"answers": {
"answer_start": [527],
"text": ["136"]
},
"context": "\"Die Verteidigung der Panthers gab nur 308 Punkte ab und belegte den sechsten Platz in der Liga, während sie die NFL mit 24 Inte...",
"id": "56beb4343aeaaa14008c925c",
"question": "Wie viele Sacks erzielte Jared Allen in seiner Karriere?"
}
```
#### xquad.de
- **Size of downloaded dataset files:** 13.30 MB
- **Size of the generated dataset:** 1.29 MB
- **Total amount of disk used:** 14.59 MB
An example of 'validation' looks as follows.
```
This example was too long and was cropped:
{
"answers": {
"answer_start": [527],
"text": ["136"]
},
"context": "\"Die Verteidigung der Panthers gab nur 308 Punkte ab und belegte den sechsten Platz in der Liga, während sie die NFL mit 24 Inte...",
"id": "56beb4343aeaaa14008c925c",
"question": "Wie viele Sacks erzielte Jared Allen in seiner Karriere?"
}
```
#### xquad.el
- **Size of downloaded dataset files:** 13.30 MB
- **Size of the generated dataset:** 2.21 MB
- **Total amount of disk used:** 15.51 MB
An example of 'validation' looks as follows.
```
This example was too long and was cropped:
{
"answers": {
"answer_start": [527],
"text": ["136"]
},
"context": "\"Die Verteidigung der Panthers gab nur 308 Punkte ab und belegte den sechsten Platz in der Liga, während sie die NFL mit 24 Inte...",
"id": "56beb4343aeaaa14008c925c",
"question": "Wie viele Sacks erzielte Jared Allen in seiner Karriere?"
}
```
#### xquad.en
- **Size of downloaded dataset files:** 13.30 MB
- **Size of the generated dataset:** 1.12 MB
- **Total amount of disk used:** 14.42 MB
An example of 'validation' looks as follows.
```
This example was too long and was cropped:
{
"answers": {
"answer_start": [527],
"text": ["136"]
},
"context": "\"Die Verteidigung der Panthers gab nur 308 Punkte ab und belegte den sechsten Platz in der Liga, während sie die NFL mit 24 Inte...",
"id": "56beb4343aeaaa14008c925c",
"question": "Wie viele Sacks erzielte Jared Allen in seiner Karriere?"
}
```
#### xquad.es
- **Size of downloaded dataset files:** 13.30 MB
- **Size of the generated dataset:** 1.28 MB
- **Total amount of disk used:** 14.58 MB
An example of 'validation' looks as follows.
```
This example was too long and was cropped:
{
"answers": {
"answer_start": [527],
"text": ["136"]
},
"context": "\"Die Verteidigung der Panthers gab nur 308 Punkte ab und belegte den sechsten Platz in der Liga, während sie die NFL mit 24 Inte...",
"id": "56beb4343aeaaa14008c925c",
"question": "Wie viele Sacks erzielte Jared Allen in seiner Karriere?"
}
```
### Data Fields
The data fields are the same among all splits.
#### xquad.ar
- `id`: a `string` feature.
- `context`: a `string` feature.
- `question`: a `string` feature.
- `answers`: a dictionary feature containing:
- `text`: a `string` feature.
- `answer_start`: a `int32` feature.
#### xquad.de
- `id`: a `string` feature.
- `context`: a `string` feature.
- `question`: a `string` feature.
- `answers`: a dictionary feature containing:
- `text`: a `string` feature.
- `answer_start`: a `int32` feature.
#### xquad.el
- `id`: a `string` feature.
- `context`: a `string` feature.
- `question`: a `string` feature.
- `answers`: a dictionary feature containing:
- `text`: a `string` feature.
- `answer_start`: a `int32` feature.
#### xquad.en
- `id`: a `string` feature.
- `context`: a `string` feature.
- `question`: a `string` feature.
- `answers`: a dictionary feature containing:
- `text`: a `string` feature.
- `answer_start`: a `int32` feature.
#### xquad.es
- `id`: a `string` feature.
- `context`: a `string` feature.
- `question`: a `string` feature.
- `answers`: a dictionary feature containing:
- `text`: a `string` feature.
- `answer_start`: a `int32` feature.
### Data Splits
| name | validation |
| -------- | ---------: |
| xquad.ar | 1190 |
| xquad.de | 1190 |
| xquad.el | 1190 |
| xquad.en | 1190 |
| xquad.es | 1190 |
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Citation Information
```
@article{Artetxe:etal:2019,
author = {Mikel Artetxe and Sebastian Ruder and Dani Yogatama},
title = {On the cross-lingual transferability of monolingual representations},
journal = {CoRR},
volume = {abs/1910.11856},
year = {2019},
archivePrefix = {arXiv},
eprint = {1910.11856}
}
```
### Contributions
Thanks to [@lewtun](https://github.com/lewtun), [@patrickvonplaten](https://github.com/patrickvonplaten), [@thomwolf](https://github.com/thomwolf) for adding this dataset. |
paws-x | 2023-01-25T14:42:16.000Z | [
"task_categories:text-classification",
"task_ids:semantic-similarity-classification",
"task_ids:semantic-similarity-scoring",
"task_ids:text-scoring",
"task_ids:multi-input-text-classification",
"annotations_creators:expert-generated",
"annotations_creators:machine-generated",
"language_creators:expert-generated",
"language_creators:machine-generated",
"multilinguality:multilingual",
"size_categories:10K<n<100K",
"source_datasets:extended|other-paws",
"language:de",
"language:en",
"language:es",
"language:fr",
"language:ja",
"language:ko",
"language:zh",
"license:other",
"paraphrase-identification",
"arxiv:1908.11828",
"region:us"
] | null | PAWS-X, a multilingual version of PAWS (Paraphrase Adversaries from Word Scrambling) for six languages.
This dataset contains 23,659 human translated PAWS evaluation pairs and 296,406 machine
translated training pairs in six typologically distinct languages: French, Spanish, German,
Chinese, Japanese, and Korean. English language is available by default. All translated
pairs are sourced from examples in PAWS-Wiki.
For further details, see the accompanying paper: PAWS-X: A Cross-lingual Adversarial Dataset
for Paraphrase Identification (https://arxiv.org/abs/1908.11828)
NOTE: There might be some missing or wrong labels in the dataset and we have replaced them with -1. | @InProceedings{pawsx2019emnlp,
title = {{PAWS-X: A Cross-lingual Adversarial Dataset for Paraphrase Identification}},
author = {Yang, Yinfei and Zhang, Yuan and Tar, Chris and Baldridge, Jason},
booktitle = {Proc. of EMNLP},
year = {2019}
} | null | 17 | 7,002 | ---
annotations_creators:
- expert-generated
- machine-generated
language_creators:
- expert-generated
- machine-generated
language:
- de
- en
- es
- fr
- ja
- ko
- zh
license:
- other
multilinguality:
- multilingual
size_categories:
- 10K<n<100K
source_datasets:
- extended|other-paws
task_categories:
- text-classification
task_ids:
- semantic-similarity-classification
- semantic-similarity-scoring
- text-scoring
- multi-input-text-classification
paperswithcode_id: paws-x
pretty_name: 'PAWS-X: A Cross-lingual Adversarial Dataset for Paraphrase Identification'
tags:
- paraphrase-identification
dataset_info:
- config_name: en
features:
- name: id
dtype: int32
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': '0'
'1': '1'
splits:
- name: train
num_bytes: 12215953
num_examples: 49401
- name: test
num_bytes: 494734
num_examples: 2000
- name: validation
num_bytes: 492287
num_examples: 2000
download_size: 30282057
dataset_size: 13202974
- config_name: de
features:
- name: id
dtype: int32
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': '0'
'1': '1'
splits:
- name: train
num_bytes: 12801824
num_examples: 49401
- name: test
num_bytes: 524214
num_examples: 2000
- name: validation
num_bytes: 514009
num_examples: 2000
download_size: 30282057
dataset_size: 13840047
- config_name: es
features:
- name: id
dtype: int32
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': '0'
'1': '1'
splits:
- name: train
num_bytes: 12808486
num_examples: 49401
- name: test
num_bytes: 519111
num_examples: 2000
- name: validation
num_bytes: 513888
num_examples: 2000
download_size: 30282057
dataset_size: 13841485
- config_name: fr
features:
- name: id
dtype: int32
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': '0'
'1': '1'
splits:
- name: train
num_bytes: 13295597
num_examples: 49401
- name: test
num_bytes: 535101
num_examples: 2000
- name: validation
num_bytes: 533031
num_examples: 2000
download_size: 30282057
dataset_size: 14363729
- config_name: ja
features:
- name: id
dtype: int32
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': '0'
'1': '1'
splits:
- name: train
num_bytes: 15041632
num_examples: 49401
- name: test
num_bytes: 668636
num_examples: 2000
- name: validation
num_bytes: 661778
num_examples: 2000
download_size: 30282057
dataset_size: 16372046
- config_name: ko
features:
- name: id
dtype: int32
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': '0'
'1': '1'
splits:
- name: train
num_bytes: 13934221
num_examples: 49401
- name: test
num_bytes: 562300
num_examples: 2000
- name: validation
num_bytes: 554875
num_examples: 2000
download_size: 30282057
dataset_size: 15051396
- config_name: zh
features:
- name: id
dtype: int32
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': '0'
'1': '1'
splits:
- name: train
num_bytes: 10815499
num_examples: 49401
- name: test
num_bytes: 474644
num_examples: 2000
- name: validation
num_bytes: 473118
num_examples: 2000
download_size: 30282057
dataset_size: 11763261
---
# Dataset Card for PAWS-X: A Cross-lingual Adversarial Dataset for Paraphrase Identification
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [PAWS-X](https://github.com/google-research-datasets/paws/tree/master/pawsx)
- **Repository:** [PAWS-X](https://github.com/google-research-datasets/paws/tree/master/pawsx)
- **Paper:** [PAWS-X: A Cross-lingual Adversarial Dataset for Paraphrase Identification](https://arxiv.org/abs/1908.11828)
- **Point of Contact:** [Yinfei Yang](yinfeiy@google.com)
### Dataset Summary
This dataset contains 23,659 **human** translated PAWS evaluation pairs and
296,406 **machine** translated training pairs in six typologically distinct
languages: French, Spanish, German, Chinese, Japanese, and Korean. All
translated pairs are sourced from examples in
[PAWS-Wiki](https://github.com/google-research-datasets/paws#paws-wiki).
For further details, see the accompanying paper:
[PAWS-X: A Cross-lingual Adversarial Dataset for Paraphrase
Identification](https://arxiv.org/abs/1908.11828)
### Supported Tasks and Leaderboards
It has been majorly used for paraphrase identification for English and other 6 languages namely French, Spanish, German, Chinese, Japanese, and Korean
### Languages
The dataset is in English, French, Spanish, German, Chinese, Japanese, and Korean
## Dataset Structure
### Data Instances
For en:
```
id : 1
sentence1 : In Paris , in October 1560 , he secretly met the English ambassador , Nicolas Throckmorton , asking him for a passport to return to England through Scotland .
sentence2 : In October 1560 , he secretly met with the English ambassador , Nicolas Throckmorton , in Paris , and asked him for a passport to return to Scotland through England .
label : 0
```
For fr:
```
id : 1
sentence1 : À Paris, en octobre 1560, il rencontra secrètement l'ambassadeur d'Angleterre, Nicolas Throckmorton, lui demandant un passeport pour retourner en Angleterre en passant par l'Écosse.
sentence2 : En octobre 1560, il rencontra secrètement l'ambassadeur d'Angleterre, Nicolas Throckmorton, à Paris, et lui demanda un passeport pour retourner en Écosse par l'Angleterre.
label : 0
```
### Data Fields
All files are in tsv format with four columns:
Column Name | Data
:---------- | :--------------------------------------------------------
id | An ID that matches the ID of the source pair in PAWS-Wiki
sentence1 | The first sentence
sentence2 | The second sentence
label | Label for each pair
The source text of each translation can be retrieved by looking up the ID in the
corresponding file in PAWS-Wiki.
### Data Splits
The numbers of examples for each of the seven languages are shown below:
Language | Train | Dev | Test
:------- | ------: | -----: | -----:
en | 49,401 | 2,000 | 2,000
fr | 49,401 | 2,000 | 2,000
es | 49,401 | 2,000 | 2,000
de | 49,401 | 2,000 | 2,000
zh | 49,401 | 2,000 | 2,000
ja | 49,401 | 2,000 | 2,000
ko | 49,401 | 2,000 | 2,000
> **Caveat**: please note that the dev and test sets of PAWS-X are both sourced
> from the dev set of PAWS-Wiki. As a consequence, the same `sentence 1` may
> appear in both the dev and test sets. Nevertheless our data split guarantees
> that there is no overlap on sentence pairs (`sentence 1` + `sentence 2`)
> between dev and test.
## Dataset Creation
### Curation Rationale
Most existing work on adversarial data generation focuses on English. For example, PAWS (Paraphrase Adversaries from Word Scrambling) (Zhang et al., 2019) consists of challenging English paraphrase identification pairs from Wikipedia and Quora. They remedy this gap with PAWS-X, a new dataset of 23,659 human translated PAWS evaluation pairs in six typologically distinct languages: French, Spanish, German, Chinese, Japanese, and Korean. They provide baseline numbers for three models with different capacity to capture non-local context and sentence structure, and using different multilingual training and evaluation regimes. Multilingual BERT (Devlin et al., 2019) fine-tuned on PAWS English plus machine-translated data performs the best, with a range of 83.1-90.8 accuracy across the non-English languages and an average accuracy gain of 23% over the next best model. PAWS-X shows the effectiveness of deep, multilingual pre-training while also leaving considerable headroom as a new challenge to drive multilingual research that better captures structure and contextual information.
### Source Data
PAWS (Paraphrase Adversaries from Word Scrambling)
#### Initial Data Collection and Normalization
All translated pairs are sourced from examples in [PAWS-Wiki](https://github.com/google-research-datasets/paws#paws-wiki)
#### Who are the source language producers?
This dataset contains 23,659 human translated PAWS evaluation pairs and 296,406 machine translated training pairs in six typologically distinct languages: French, Spanish, German, Chinese, Japanese, and Korean.
### Annotations
#### Annotation process
If applicable, describe the annotation process and any tools used, or state otherwise. Describe the amount of data annotated, if not all. Describe or reference annotation guidelines provided to the annotators. If available, provide interannotator statistics. Describe any annotation validation processes.
#### Who are the annotators?
The paper mentions the translate team, especially Mengmeng Niu, for the help with the annotations.
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
List the people involved in collecting the dataset and their affiliation(s). If funding information is known, include it here.
### Licensing Information
The dataset may be freely used for any purpose, although acknowledgement of Google LLC ("Google") as the data source would be appreciated. The dataset is provided "AS IS" without any warranty, express or implied. Google disclaims all liability for any damages, direct or indirect, resulting from the use of the dataset.
### Citation Information
```
@InProceedings{pawsx2019emnlp,
title = {{PAWS-X: A Cross-lingual Adversarial Dataset for Paraphrase Identification}},
author = {Yang, Yinfei and Zhang, Yuan and Tar, Chris and Baldridge, Jason},
booktitle = {Proc. of EMNLP},
year = {2019}
}
```
### Contributions
Thanks to [@bhavitvyamalik](https://github.com/bhavitvyamalik), [@gowtham1997](https://github.com/gowtham1997) for adding this dataset. |
multi_woz_v22 | 2023-01-25T14:41:08.000Z | [
"task_categories:text-generation",
"task_categories:fill-mask",
"task_categories:token-classification",
"task_categories:text-classification",
"task_ids:dialogue-modeling",
"task_ids:multi-class-classification",
"task_ids:parsing",
"annotations_creators:machine-generated",
"language_creators:crowdsourced",
"language_creators:machine-generated",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:apache-2.0",
"arxiv:1810.00278",
"region:us"
] | null | Multi-Domain Wizard-of-Oz dataset (MultiWOZ), a fully-labeled collection of human-human written conversations spanning over multiple domains and topics.
MultiWOZ 2.1 (Eric et al., 2019) identified and fixed many erroneous annotations and user utterances in the original version, resulting in an
improved version of the dataset. MultiWOZ 2.2 is a yet another improved version of this dataset, which identifies and fizes dialogue state annotation errors
across 17.3% of the utterances on top of MultiWOZ 2.1 and redefines the ontology by disallowing vocabularies of slots with a large number of possible values
(e.g., restaurant name, time of booking) and introducing standardized slot span annotations for these slots. | @article{corr/abs-2007-12720,
author = {Xiaoxue Zang and
Abhinav Rastogi and
Srinivas Sunkara and
Raghav Gupta and
Jianguo Zhang and
Jindong Chen},
title = {MultiWOZ 2.2 : {A} Dialogue Dataset with Additional Annotation Corrections
and State Tracking Baselines},
journal = {CoRR},
volume = {abs/2007.12720},
year = {2020},
url = {https://arxiv.org/abs/2007.12720},
archivePrefix = {arXiv},
eprint = {2007.12720}
} | null | 14 | 6,967 | ---
annotations_creators:
- machine-generated
language_creators:
- crowdsourced
- machine-generated
language:
- en
license:
- apache-2.0
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- text-generation
- fill-mask
- token-classification
- text-classification
task_ids:
- dialogue-modeling
- multi-class-classification
- parsing
paperswithcode_id: multiwoz
pretty_name: Multi-domain Wizard-of-Oz
dataset_info:
- config_name: v2.2
features:
- name: dialogue_id
dtype: string
- name: services
sequence: string
- name: turns
sequence:
- name: turn_id
dtype: string
- name: speaker
dtype:
class_label:
names:
'0': USER
'1': SYSTEM
- name: utterance
dtype: string
- name: frames
sequence:
- name: service
dtype: string
- name: state
struct:
- name: active_intent
dtype: string
- name: requested_slots
sequence: string
- name: slots_values
sequence:
- name: slots_values_name
dtype: string
- name: slots_values_list
sequence: string
- name: slots
sequence:
- name: slot
dtype: string
- name: value
dtype: string
- name: start
dtype: int32
- name: exclusive_end
dtype: int32
- name: copy_from
dtype: string
- name: copy_from_value
sequence: string
- name: dialogue_acts
struct:
- name: dialog_act
sequence:
- name: act_type
dtype: string
- name: act_slots
sequence:
- name: slot_name
dtype: string
- name: slot_value
dtype: string
- name: span_info
sequence:
- name: act_type
dtype: string
- name: act_slot_name
dtype: string
- name: act_slot_value
dtype: string
- name: span_start
dtype: int32
- name: span_end
dtype: int32
splits:
- name: train
num_bytes: 68222649
num_examples: 8437
- name: validation
num_bytes: 8990945
num_examples: 1000
- name: test
num_bytes: 9027095
num_examples: 1000
download_size: 276592909
dataset_size: 86240689
- config_name: v2.2_active_only
features:
- name: dialogue_id
dtype: string
- name: services
sequence: string
- name: turns
sequence:
- name: turn_id
dtype: string
- name: speaker
dtype:
class_label:
names:
'0': USER
'1': SYSTEM
- name: utterance
dtype: string
- name: frames
sequence:
- name: service
dtype: string
- name: state
struct:
- name: active_intent
dtype: string
- name: requested_slots
sequence: string
- name: slots_values
sequence:
- name: slots_values_name
dtype: string
- name: slots_values_list
sequence: string
- name: slots
sequence:
- name: slot
dtype: string
- name: value
dtype: string
- name: start
dtype: int32
- name: exclusive_end
dtype: int32
- name: copy_from
dtype: string
- name: copy_from_value
sequence: string
- name: dialogue_acts
struct:
- name: dialog_act
sequence:
- name: act_type
dtype: string
- name: act_slots
sequence:
- name: slot_name
dtype: string
- name: slot_value
dtype: string
- name: span_info
sequence:
- name: act_type
dtype: string
- name: act_slot_name
dtype: string
- name: act_slot_value
dtype: string
- name: span_start
dtype: int32
- name: span_end
dtype: int32
splits:
- name: train
num_bytes: 40937577
num_examples: 8437
- name: validation
num_bytes: 5377939
num_examples: 1000
- name: test
num_bytes: 5410819
num_examples: 1000
download_size: 276592909
dataset_size: 51726335
---
# Dataset Card for MultiWOZ
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Repository:** [MultiWOZ 2.2 github repository](https://github.com/budzianowski/multiwoz/tree/master/data/MultiWOZ_2.2)
- **Paper:** [MultiWOZ v2](https://arxiv.org/abs/1810.00278), and [MultiWOZ v2.2](https://www.aclweb.org/anthology/2020.nlp4convai-1.13.pdf)
- **Point of Contact:** [Paweł Budzianowski](pfb30@cam.ac.uk)
### Dataset Summary
Multi-Domain Wizard-of-Oz dataset (MultiWOZ), a fully-labeled collection of human-human written conversations spanning over multiple domains and topics.
MultiWOZ 2.1 (Eric et al., 2019) identified and fixed many erroneous annotations and user utterances in the original version, resulting in an
improved version of the dataset. MultiWOZ 2.2 is a yet another improved version of this dataset, which identifies and fixes dialogue state annotation errors
across 17.3% of the utterances on top of MultiWOZ 2.1 and redefines the ontology by disallowing vocabularies of slots with a large number of possible values
(e.g., restaurant name, time of booking) and introducing standardized slot span annotations for these slots.
### Supported Tasks and Leaderboards
This dataset supports a range of task.
- **Generative dialogue modeling** or `dialogue-modeling`: the text of the dialogues can be used to train a sequence model on the utterances. Performance on this task is typically evaluated with delexicalized-[BLEU](https://huggingface.co/metrics/bleu), inform rate and request success.
- **Intent state tracking**, a `multi-class-classification` task: predict the belief state of the user side of the conversation, performance is measured by [F1](https://huggingface.co/metrics/f1).
- **Dialog act prediction**, a `parsing` task: parse an utterance into the corresponding dialog acts for the system to use. [F1](https://huggingface.co/metrics/f1) is typically reported.
### Languages
The text in the dataset is in English (`en`).
## Dataset Structure
### Data Instances
A data instance is a full multi-turn dialogue between a `USER` and a `SYSTEM`. Each turn has a single utterance, e.g.:
```
['What fun places can I visit in the East?',
'We have five spots which include boating, museums and entertainment. Any preferences that you have?']
```
The utterances of the `USER` are also annotated with frames denoting their intent and believe state:
```
[{'service': ['attraction'],
'slots': [{'copy_from': [],
'copy_from_value': [],
'exclusive_end': [],
'slot': [],
'start': [],
'value': []}],
'state': [{'active_intent': 'find_attraction',
'requested_slots': [],
'slots_values': {'slots_values_list': [['east']],
'slots_values_name': ['attraction-area']}}]},
{'service': [], 'slots': [], 'state': []}]
```
Finally, each of the utterances is annotated with dialog acts which provide a structured representation of what the `USER` or `SYSTEM` is inquiring or giving information about.
```
[{'dialog_act': {'act_slots': [{'slot_name': ['east'],
'slot_value': ['area']}],
'act_type': ['Attraction-Inform']},
'span_info': {'act_slot_name': ['area'],
'act_slot_value': ['east'],
'act_type': ['Attraction-Inform'],
'span_end': [39],
'span_start': [35]}},
{'dialog_act': {'act_slots': [{'slot_name': ['none'], 'slot_value': ['none']},
{'slot_name': ['boating', 'museums', 'entertainment', 'five'],
'slot_value': ['type', 'type', 'type', 'choice']}],
'act_type': ['Attraction-Select', 'Attraction-Inform']},
'span_info': {'act_slot_name': ['type', 'type', 'type', 'choice'],
'act_slot_value': ['boating', 'museums', 'entertainment', 'five'],
'act_type': ['Attraction-Inform',
'Attraction-Inform',
'Attraction-Inform',
'Attraction-Inform'],
'span_end': [40, 49, 67, 12],
'span_start': [33, 42, 54, 8]}}]
```
### Data Fields
Each dialogue instance has the following fields:
- `dialogue_id`: a unique ID identifying the dialog. The MUL and PMUL names refer to strictly multi domain dialogues (at least 2 main domains are involved) while the SNG, SSNG and WOZ names refer to single domain dialogues with potentially sub-domains like booking.
- `services`: a list of services mentioned in the dialog, such as `train` or `hospitals`.
- `turns`: the sequence of utterances with their annotations, including:
- `turn_id`: a turn identifier, unique per dialog.
- `speaker`: either the `USER` or `SYSTEM`.
- `utterance`: the text of the utterance.
- `dialogue_acts`: The structured parse of the utterance into dialog acts in the system's grammar
- `act_type`: Such as e.g. `Attraction-Inform` to seek or provide information about an `attraction`
- `act_slots`: provide more details about the action
- `span_info`: maps these `act_slots` to the `utterance` text.
- `frames`: only for `USER` utterances, track the user's belief state, i.e. a structured representation of what they are trying to achieve in the fialog. This decomposes into:
- `service`: the service they are interested in
- `state`: their belief state including their `active_intent` and further information expressed in `requested_slots`
- `slots`: a mapping of the `requested_slots` to where they are mentioned in the text. It takes one of two forms, detailed next:
The first type are span annotations that identify the location where slot values have been mentioned in the utterances for non-categorical slots. These span annotations are represented as follows:
```
{
"slots": [
{
"slot": String of slot name.
"start": Int denoting the index of the starting character in the utterance corresponding to the slot value.
"exclusive_end": Int denoting the index of the character just after the last character corresponding to the slot value in the utterance. In python, utterance[start:exclusive_end] gives the slot value.
"value": String of value. It equals to utterance[start:exclusive_end], where utterance is the current utterance in string.
}
]
}
```
There are also some non-categorical slots whose values are carried over from another slot in the dialogue state. Their values don"t explicitly appear in the utterances. For example, a user utterance can be "I also need a taxi from the restaurant to the hotel.", in which the state values of "taxi-departure" and "taxi-destination" are respectively carried over from that of "restaurant-name" and "hotel-name". For these slots, instead of annotating them as spans, a "copy from" annotation identifies the slot it copies the value from. This annotation is formatted as follows,
```
{
"slots": [
{
"slot": Slot name string.
"copy_from": The slot to copy from.
"value": A list of slot values being . It corresponds to the state values of the "copy_from" slot.
}
]
}
```
### Data Splits
The dataset is split into a `train`, `validation`, and `test` split with the following sizes:
| | train | validation | test |
|---------------------|------:|-----------:|-----:|
| Number of dialogues | 8438 | 1000 | 1000 |
| Number of turns | 42190 | 5000 | 5000 |
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
The initial dataset (Versions 1.0 and 2.0) was created by a team of researchers from the [Cambridge Dialogue Systems Group](https://mi.eng.cam.ac.uk/research/dialogue/corpora/). Version 2.1 was developed on top of v2.0 by a team from Amazon, and v2.2 was developed by a team of Google researchers.
### Licensing Information
The dataset is released under the Apache License 2.0.
### Citation Information
You can cite the following for the various versions of MultiWOZ:
Version 1.0
```
@inproceedings{ramadan2018large,
title={Large-Scale Multi-Domain Belief Tracking with Knowledge Sharing},
author={Ramadan, Osman and Budzianowski, Pawe{\l} and Gasic, Milica},
booktitle={Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics},
volume={2},
pages={432--437},
year={2018}
}
```
Version 2.0
```
@inproceedings{budzianowski2018large,
Author = {Budzianowski, Pawe{\l} and Wen, Tsung-Hsien and Tseng, Bo-Hsiang and Casanueva, I{\~n}igo and Ultes Stefan and Ramadan Osman and Ga{\v{s}}i\'c, Milica},
title={MultiWOZ - A Large-Scale Multi-Domain Wizard-of-Oz Dataset for Task-Oriented Dialogue Modelling},
booktitle={Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (EMNLP)},
year={2018}
}
```
Version 2.1
```
@article{eric2019multiwoz,
title={MultiWOZ 2.1: Multi-Domain Dialogue State Corrections and State Tracking Baselines},
author={Eric, Mihail and Goel, Rahul and Paul, Shachi and Sethi, Abhishek and Agarwal, Sanchit and Gao, Shuyag and Hakkani-Tur, Dilek},
journal={arXiv preprint arXiv:1907.01669},
year={2019}
}
```
Version 2.2
```
@inproceedings{zang2020multiwoz,
title={MultiWOZ 2.2: A Dialogue Dataset with Additional Annotation Corrections and State Tracking Baselines},
author={Zang, Xiaoxue and Rastogi, Abhinav and Sunkara, Srinivas and Gupta, Raghav and Zhang, Jianguo and Chen, Jindong},
booktitle={Proceedings of the 2nd Workshop on Natural Language Processing for Conversational AI, ACL 2020},
pages={109--117},
year={2020}
}
```
### Contributions
Thanks to [@yjernite](https://github.com/yjernite) for adding this dataset. |
angelolab/ark_example | 2023-08-31T19:36:09.000Z | [
"task_categories:image-segmentation",
"task_ids:instance-segmentation",
"annotations_creators:no-annotation",
"size_categories:n<1K",
"source_datasets:original",
"license:apache-2.0",
"MIBI",
"Multiplexed-Imaging",
"region:us"
] | angelolab | This dataset contains 11 Field of Views (FOVs), each with 22 channels. | @InProceedings{huggingface:dataset,
title = {Ark Analysis Example Dataset},
author={Angelo Lab},
year={2022}
} | null | 0 | 6,953 | ---
annotations_creators:
- no-annotation
language: []
language_creators: []
license:
- apache-2.0
multilinguality: []
pretty_name: An example dataset for analyzing multiplexed imaging data.
size_categories:
- n<1K
source_datasets:
- original
tags:
- MIBI
- Multiplexed-Imaging
task_categories:
- image-segmentation
task_ids:
- instance-segmentation
---
# Dataset Card for [Dataset Name]
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:**
- **Repository:**
- **Paper:**
- **Leaderboard:**
- **Point of Contact:**
### Dataset Summary
[More Information Needed]
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
Thanks to [@angelolab](https://github.com/angelolab) for adding this dataset. |
universal_dependencies | 2023-06-01T14:59:56.000Z | [
"task_categories:token-classification",
"task_ids:parsing",
"annotations_creators:expert-generated",
"language_creators:crowdsourced",
"multilinguality:multilingual",
"size_categories:1K<n<10K",
"source_datasets:original",
"language:af",
"language:aii",
"language:ajp",
"language:akk",
"language:am",
"language:apu",
"language:aqz",
"language:ar",
"language:be",
"language:bg",
"language:bho",
"language:bm",
"language:br",
"language:bxr",
"language:ca",
"language:ckt",
"language:cop",
"language:cs",
"language:cu",
"language:cy",
"language:da",
"language:de",
"language:el",
"language:en",
"language:es",
"language:et",
"language:eu",
"language:fa",
"language:fi",
"language:fo",
"language:fr",
"language:fro",
"language:ga",
"language:gd",
"language:gl",
"language:got",
"language:grc",
"language:gsw",
"language:gun",
"language:gv",
"language:he",
"language:hi",
"language:hr",
"language:hsb",
"language:hu",
"language:hy",
"language:id",
"language:is",
"language:it",
"language:ja",
"language:kfm",
"language:kk",
"language:kmr",
"language:ko",
"language:koi",
"language:kpv",
"language:krl",
"language:la",
"language:lt",
"language:lv",
"language:lzh",
"language:mdf",
"language:mr",
"language:mt",
"language:myu",
"language:myv",
"language:nl",
"language:no",
"language:nyq",
"language:olo",
"language:orv",
"language:otk",
"language:pcm",
"language:pl",
"language:pt",
"language:ro",
"language:ru",
"language:sa",
"language:sk",
"language:sl",
"language:sme",
"language:sms",
"language:soj",
"language:sq",
"language:sr",
"language:sv",
"language:swl",
"language:ta",
"language:te",
"language:th",
"language:tl",
"language:tpn",
"language:tr",
"language:ug",
"language:uk",
"language:ur",
"language:vi",
"language:wbp",
"language:wo",
"language:yo",
"language:yue",
"language:zh",
"license:unknown",
"constituency-parsing",
"dependency-parsing",
"region:us"
] | null | Universal Dependencies is a project that seeks to develop cross-linguistically consistent treebank annotation for many languages, with the goal of facilitating multilingual parser development, cross-lingual learning, and parsing research from a language typology perspective. The annotation scheme is based on (universal) Stanford dependencies (de Marneffe et al., 2006, 2008, 2014), Google universal part-of-speech tags (Petrov et al., 2012), and the Interset interlingua for morphosyntactic tagsets (Zeman, 2008). | null | null | 13 | 6,927 | ---
annotations_creators:
- expert-generated
language_creators:
- crowdsourced
language:
- af
- aii
- ajp
- akk
- am
- apu
- aqz
- ar
- be
- bg
- bho
- bm
- br
- bxr
- ca
- ckt
- cop
- cs
- cu
- cy
- da
- de
- el
- en
- es
- et
- eu
- fa
- fi
- fo
- fr
- fro
- ga
- gd
- gl
- got
- grc
- gsw
- gun
- gv
- he
- hi
- hr
- hsb
- hu
- hy
- id
- is
- it
- ja
- kfm
- kk
- kmr
- ko
- koi
- kpv
- krl
- la
- lt
- lv
- lzh
- mdf
- mr
- mt
- myu
- myv
- nl
- 'no'
- nyq
- olo
- orv
- otk
- pcm
- pl
- pt
- ro
- ru
- sa
- sk
- sl
- sme
- sms
- soj
- sq
- sr
- sv
- swl
- ta
- te
- th
- tl
- tpn
- tr
- ug
- uk
- ur
- vi
- wbp
- wo
- yo
- yue
- zh
license:
- unknown
multilinguality:
- multilingual
size_categories:
- 1K<n<10K
source_datasets:
- original
task_categories:
- token-classification
task_ids:
- parsing
paperswithcode_id: universal-dependencies
pretty_name: Universal Dependencies Treebank
tags:
- constituency-parsing
- dependency-parsing
dataset_info:
- config_name: af_afribooms
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 3523113
num_examples: 1315
- name: validation
num_bytes: 547285
num_examples: 194
- name: test
num_bytes: 1050299
num_examples: 425
download_size: 3088237
dataset_size: 5120697
- config_name: akk_pisandub
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 153470
num_examples: 101
download_size: 101789
dataset_size: 153470
- config_name: akk_riao
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 3374577
num_examples: 1804
download_size: 2022357
dataset_size: 3374577
- config_name: aqz_tudet
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 8286
num_examples: 24
download_size: 5683
dataset_size: 8286
- config_name: sq_tsa
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 116034
num_examples: 60
download_size: 68875
dataset_size: 116034
- config_name: am_att
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 1554859
num_examples: 1074
download_size: 1019607
dataset_size: 1554859
- config_name: grc_perseus
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 22611612
num_examples: 11476
- name: validation
num_bytes: 3152233
num_examples: 1137
- name: test
num_bytes: 3004502
num_examples: 1306
download_size: 18898313
dataset_size: 28768347
- config_name: grc_proiel
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 30938089
num_examples: 15014
- name: validation
num_bytes: 2264551
num_examples: 1019
- name: test
num_bytes: 2192289
num_examples: 1047
download_size: 23715831
dataset_size: 35394929
- config_name: apu_ufpa
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 75578
num_examples: 76
download_size: 69565
dataset_size: 75578
- config_name: ar_nyuad
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 79064476
num_examples: 15789
- name: validation
num_bytes: 9859912
num_examples: 1986
- name: test
num_bytes: 9880240
num_examples: 1963
download_size: 58583673
dataset_size: 98804628
- config_name: ar_padt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 58537298
num_examples: 6075
- name: validation
num_bytes: 7787253
num_examples: 909
- name: test
num_bytes: 7428063
num_examples: 680
download_size: 51208169
dataset_size: 73752614
- config_name: ar_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2816625
num_examples: 1000
download_size: 2084082
dataset_size: 2816625
- config_name: hy_armtdp
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 7697891
num_examples: 1975
- name: validation
num_bytes: 988849
num_examples: 249
- name: test
num_bytes: 947287
num_examples: 278
download_size: 6886567
dataset_size: 9634027
- config_name: aii_as
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 52540
num_examples: 57
download_size: 32639
dataset_size: 52540
- config_name: bm_crb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 1502886
num_examples: 1026
download_size: 892924
dataset_size: 1502886
- config_name: eu_bdt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 8199861
num_examples: 5396
- name: validation
num_bytes: 2701073
num_examples: 1798
- name: test
num_bytes: 2734601
num_examples: 1799
download_size: 8213576
dataset_size: 13635535
- config_name: be_hse
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 34880663
num_examples: 21555
- name: validation
num_bytes: 1745668
num_examples: 1090
- name: test
num_bytes: 1818113
num_examples: 889
download_size: 26433402
dataset_size: 38444444
- config_name: bho_bhtb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 947740
num_examples: 357
download_size: 614159
dataset_size: 947740
- config_name: br_keb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 1026257
num_examples: 888
download_size: 679680
dataset_size: 1026257
- config_name: bg_btb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 18545312
num_examples: 8907
- name: validation
num_bytes: 2393174
num_examples: 1115
- name: test
num_bytes: 2344136
num_examples: 1116
download_size: 14910603
dataset_size: 23282622
- config_name: bxr_bdt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 17364
num_examples: 19
- name: test
num_bytes: 1116630
num_examples: 908
download_size: 726053
dataset_size: 1133994
- config_name: yue_hk
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 1242850
num_examples: 1004
download_size: 710060
dataset_size: 1242850
- config_name: ca_ancora
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 46502842
num_examples: 13123
- name: validation
num_bytes: 6282364
num_examples: 1709
- name: test
num_bytes: 6441038
num_examples: 1846
download_size: 35924146
dataset_size: 59226244
- config_name: zh_cfl
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 660584
num_examples: 451
download_size: 384725
dataset_size: 660584
- config_name: zh_gsd
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 9268661
num_examples: 3997
- name: validation
num_bytes: 1188371
num_examples: 500
- name: test
num_bytes: 1130467
num_examples: 500
download_size: 6828367
dataset_size: 11587499
- config_name: zh_gsdsimp
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 9268663
num_examples: 3997
- name: validation
num_bytes: 1188383
num_examples: 500
- name: test
num_bytes: 1130459
num_examples: 500
download_size: 6828419
dataset_size: 11587505
- config_name: zh_hk
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 880193
num_examples: 1004
download_size: 494447
dataset_size: 880193
- config_name: zh_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2425817
num_examples: 1000
download_size: 1606982
dataset_size: 2425817
- config_name: ckt_hse
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 808669
num_examples: 1004
download_size: 771943
dataset_size: 808669
- config_name: lzh_kyoto
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 26615708
num_examples: 38669
- name: validation
num_bytes: 3770507
num_examples: 5296
- name: test
num_bytes: 3155207
num_examples: 4469
download_size: 22658287
dataset_size: 33541422
- config_name: cop_scriptorium
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 3944468
num_examples: 1089
- name: validation
num_bytes: 1566786
num_examples: 381
- name: test
num_bytes: 1487709
num_examples: 403
download_size: 4502996
dataset_size: 6998963
- config_name: hr_set
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 19104315
num_examples: 6914
- name: validation
num_bytes: 2787184
num_examples: 960
- name: test
num_bytes: 3035797
num_examples: 1136
download_size: 15103034
dataset_size: 24927296
- config_name: cs_cac
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 81527862
num_examples: 23478
- name: validation
num_bytes: 1898678
num_examples: 603
- name: test
num_bytes: 1878841
num_examples: 628
download_size: 55990235
dataset_size: 85305381
- config_name: cs_cltt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 4277239
num_examples: 860
- name: validation
num_bytes: 752253
num_examples: 129
- name: test
num_bytes: 646103
num_examples: 136
download_size: 3745656
dataset_size: 5675595
- config_name: cs_fictree
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 21490020
num_examples: 10160
- name: validation
num_bytes: 2677727
num_examples: 1309
- name: test
num_bytes: 2679930
num_examples: 1291
download_size: 17464342
dataset_size: 26847677
- config_name: cs_pdt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 201356662
num_examples: 68495
- name: validation
num_bytes: 27366981
num_examples: 9270
- name: test
num_bytes: 29817339
num_examples: 10148
download_size: 171506068
dataset_size: 258540982
- config_name: cs_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 3195818
num_examples: 1000
download_size: 2231853
dataset_size: 3195818
- config_name: da_ddt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 8689809
num_examples: 4383
- name: validation
num_bytes: 1117939
num_examples: 564
- name: test
num_bytes: 1082651
num_examples: 565
download_size: 6425281
dataset_size: 10890399
- config_name: nl_alpino
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 22503950
num_examples: 12264
- name: validation
num_bytes: 1411253
num_examples: 718
- name: test
num_bytes: 1354908
num_examples: 596
download_size: 16858557
dataset_size: 25270111
- config_name: nl_lassysmall
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 9001614
num_examples: 5787
- name: validation
num_bytes: 1361552
num_examples: 676
- name: test
num_bytes: 1391136
num_examples: 875
download_size: 8034396
dataset_size: 11754302
- config_name: en_esl
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 5335977
num_examples: 4124
- name: validation
num_bytes: 648562
num_examples: 500
- name: test
num_bytes: 651829
num_examples: 500
download_size: 3351548
dataset_size: 6636368
- config_name: en_ewt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 22755753
num_examples: 12543
- name: validation
num_bytes: 2829889
num_examples: 2002
- name: test
num_bytes: 2820398
num_examples: 2077
download_size: 16893922
dataset_size: 28406040
- config_name: en_gum
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 8999554
num_examples: 4287
- name: validation
num_bytes: 1704949
num_examples: 784
- name: test
num_bytes: 1743317
num_examples: 890
download_size: 7702761
dataset_size: 12447820
- config_name: en_gumreddit
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 1365930
num_examples: 587
- name: validation
num_bytes: 317546
num_examples: 150
- name: test
num_bytes: 374707
num_examples: 158
download_size: 1195979
dataset_size: 2058183
- config_name: en_lines
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 5728898
num_examples: 3176
- name: validation
num_bytes: 1911762
num_examples: 1032
- name: test
num_bytes: 1766797
num_examples: 1035
download_size: 5522254
dataset_size: 9407457
- config_name: en_partut
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 4133445
num_examples: 1781
- name: validation
num_bytes: 265039
num_examples: 156
- name: test
num_bytes: 326834
num_examples: 153
download_size: 2720286
dataset_size: 4725318
- config_name: en_pronouns
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 207364
num_examples: 285
download_size: 147181
dataset_size: 207364
- config_name: en_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2282027
num_examples: 1000
download_size: 1340563
dataset_size: 2282027
- config_name: myv_jr
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2763297
num_examples: 1690
download_size: 1945981
dataset_size: 2763297
- config_name: et_edt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 42901059
num_examples: 24633
- name: validation
num_bytes: 5551620
num_examples: 3125
- name: test
num_bytes: 5994421
num_examples: 3214
download_size: 32393618
dataset_size: 54447100
- config_name: et_ewt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 4199896
num_examples: 2837
- name: validation
num_bytes: 1089459
num_examples: 743
- name: test
num_bytes: 1600116
num_examples: 913
download_size: 4044147
dataset_size: 6889471
- config_name: fo_farpahc
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 2114958
num_examples: 1020
- name: validation
num_bytes: 809707
num_examples: 300
- name: test
num_bytes: 798245
num_examples: 301
download_size: 2186706
dataset_size: 3722910
- config_name: fo_oft
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 1220792
num_examples: 1208
download_size: 802681
dataset_size: 1220792
- config_name: fi_ftb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 16800109
num_examples: 14981
- name: validation
num_bytes: 2074201
num_examples: 1875
- name: test
num_bytes: 2144908
num_examples: 1867
download_size: 13132466
dataset_size: 21019218
- config_name: fi_ood
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2366923
num_examples: 2122
download_size: 1480506
dataset_size: 2366923
- config_name: fi_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2086421
num_examples: 1000
download_size: 1411514
dataset_size: 2086421
- config_name: fi_tdt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 22065448
num_examples: 12217
- name: validation
num_bytes: 2483303
num_examples: 1364
- name: test
num_bytes: 2855263
num_examples: 1555
download_size: 16692242
dataset_size: 27404014
- config_name: fr_fqb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2674644
num_examples: 2289
download_size: 1556235
dataset_size: 2674644
- config_name: fr_ftb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 44714315
num_examples: 14759
- name: validation
num_bytes: 3929428
num_examples: 1235
- name: test
num_bytes: 7583038
num_examples: 2541
download_size: 30926802
dataset_size: 56226781
- config_name: fr_gsd
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 38329902
num_examples: 14449
- name: validation
num_bytes: 3861548
num_examples: 1476
- name: test
num_bytes: 1086926
num_examples: 416
download_size: 25492044
dataset_size: 43278376
- config_name: fr_partut
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 2620477
num_examples: 803
- name: validation
num_bytes: 205839
num_examples: 107
- name: test
num_bytes: 288829
num_examples: 110
download_size: 1817897
dataset_size: 3115145
- config_name: fr_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2660405
num_examples: 1000
download_size: 1685033
dataset_size: 2660405
- config_name: fr_sequoia
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 5370647
num_examples: 2231
- name: validation
num_bytes: 1065411
num_examples: 412
- name: test
num_bytes: 1067676
num_examples: 456
download_size: 4415282
dataset_size: 7503734
- config_name: fr_spoken
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 1625626
num_examples: 1167
- name: validation
num_bytes: 1091750
num_examples: 909
- name: test
num_bytes: 1078438
num_examples: 730
download_size: 2483341
dataset_size: 3795814
- config_name: gl_ctg
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 8157432
num_examples: 2272
- name: validation
num_bytes: 3057483
num_examples: 860
- name: test
num_bytes: 3053764
num_examples: 861
download_size: 8230649
dataset_size: 14268679
- config_name: gl_treegal
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 1804389
num_examples: 600
- name: test
num_bytes: 1174023
num_examples: 400
download_size: 1741471
dataset_size: 2978412
- config_name: de_gsd
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 32297384
num_examples: 13814
- name: validation
num_bytes: 1504189
num_examples: 799
- name: test
num_bytes: 2000117
num_examples: 977
download_size: 21507364
dataset_size: 35801690
- config_name: de_hdt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 334214761
num_examples: 153035
- name: validation
num_bytes: 39099013
num_examples: 18434
- name: test
num_bytes: 39519143
num_examples: 18459
download_size: 249243037
dataset_size: 412832917
- config_name: de_lit
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 3327891
num_examples: 1922
download_size: 2060988
dataset_size: 3327891
- config_name: de_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2684407
num_examples: 1000
download_size: 1731875
dataset_size: 2684407
- config_name: got_proiel
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 5175361
num_examples: 3387
- name: validation
num_bytes: 1498101
num_examples: 985
- name: test
num_bytes: 1518642
num_examples: 1029
download_size: 5225655
dataset_size: 8192104
- config_name: el_gdt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 6028077
num_examples: 1662
- name: validation
num_bytes: 1492610
num_examples: 403
- name: test
num_bytes: 1521094
num_examples: 456
download_size: 5788161
dataset_size: 9041781
- config_name: he_htb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 17324640
num_examples: 5241
- name: validation
num_bytes: 1440985
num_examples: 484
- name: test
num_bytes: 1550465
num_examples: 491
download_size: 12054025
dataset_size: 20316090
- config_name: qhe_hiencs
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 1510145
num_examples: 1448
- name: validation
num_bytes: 244129
num_examples: 225
- name: test
num_bytes: 236291
num_examples: 225
download_size: 914584
dataset_size: 1990565
- config_name: hi_hdtb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 61893814
num_examples: 13304
- name: validation
num_bytes: 7748544
num_examples: 1659
- name: test
num_bytes: 7786343
num_examples: 1684
download_size: 51589681
dataset_size: 77428701
- config_name: hi_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 3384789
num_examples: 1000
download_size: 2303495
dataset_size: 3384789
- config_name: hu_szeged
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 2822934
num_examples: 910
- name: validation
num_bytes: 1584932
num_examples: 441
- name: test
num_bytes: 1419130
num_examples: 449
download_size: 3687905
dataset_size: 5826996
- config_name: is_icepahc
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 97197159
num_examples: 34007
- name: validation
num_bytes: 18931295
num_examples: 4865
- name: test
num_bytes: 19039838
num_examples: 5157
download_size: 85106126
dataset_size: 135168292
- config_name: is_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2304432
num_examples: 1000
download_size: 1525635
dataset_size: 2304432
- config_name: id_csui
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 1611334
num_examples: 656
- name: test
num_bytes: 888832
num_examples: 374
download_size: 1448601
dataset_size: 2500166
- config_name: id_gsd
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 11728948
num_examples: 4477
- name: validation
num_bytes: 1513894
num_examples: 559
- name: test
num_bytes: 1417208
num_examples: 557
download_size: 9487349
dataset_size: 14660050
- config_name: id_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 1768596
num_examples: 1000
download_size: 1149692
dataset_size: 1768596
- config_name: ga_idt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 10327215
num_examples: 4005
- name: validation
num_bytes: 1057313
num_examples: 451
- name: test
num_bytes: 1109028
num_examples: 454
download_size: 7417728
dataset_size: 12493556
- config_name: it_isdt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 33510781
num_examples: 13121
- name: validation
num_bytes: 1439348
num_examples: 564
- name: test
num_bytes: 1267932
num_examples: 482
download_size: 20998527
dataset_size: 36218061
- config_name: it_partut
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 5428686
num_examples: 1781
- name: validation
num_bytes: 335085
num_examples: 156
- name: test
num_bytes: 413752
num_examples: 153
download_size: 3582155
dataset_size: 6177523
- config_name: it_postwita
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 10523322
num_examples: 5368
- name: validation
num_bytes: 1299818
num_examples: 671
- name: test
num_bytes: 1344079
num_examples: 674
download_size: 7611319
dataset_size: 13167219
- config_name: it_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2612838
num_examples: 1000
download_size: 1641073
dataset_size: 2612838
- config_name: it_twittiro
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 2536429
num_examples: 1138
- name: validation
num_bytes: 323504
num_examples: 144
- name: test
num_bytes: 316211
num_examples: 142
download_size: 1894686
dataset_size: 3176144
- config_name: it_vit
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 24536095
num_examples: 8277
- name: validation
num_bytes: 3144507
num_examples: 743
- name: test
num_bytes: 2870355
num_examples: 1067
download_size: 17605311
dataset_size: 30550957
- config_name: ja_bccwj
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 119164443
num_examples: 40740
- name: validation
num_bytes: 23390188
num_examples: 8417
- name: test
num_bytes: 21904413
num_examples: 7871
download_size: 87340125
dataset_size: 164459044
- config_name: ja_gsd
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 36905139
num_examples: 7027
- name: validation
num_bytes: 2662999
num_examples: 501
- name: test
num_bytes: 2858141
num_examples: 543
download_size: 30397358
dataset_size: 42426279
- config_name: ja_modern
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 3062149
num_examples: 822
download_size: 2163988
dataset_size: 3062149
- config_name: ja_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 6322307
num_examples: 1000
download_size: 4661525
dataset_size: 6322307
- config_name: krl_kkpp
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 370378
num_examples: 228
download_size: 226103
dataset_size: 370378
- config_name: kk_ktb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 64737
num_examples: 31
- name: test
num_bytes: 1263246
num_examples: 1047
download_size: 849300
dataset_size: 1327983
- config_name: kfm_aha
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 8464
num_examples: 10
download_size: 6290
dataset_size: 8464
- config_name: koi_uh
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 117629
num_examples: 81
download_size: 91509
dataset_size: 117629
- config_name: kpv_ikdp
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 182189
num_examples: 132
download_size: 121684
dataset_size: 182189
- config_name: kpv_lattice
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 685683
num_examples: 435
download_size: 467085
dataset_size: 685683
- config_name: ko_gsd
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 5480313
num_examples: 4400
- name: validation
num_bytes: 1156603
num_examples: 950
- name: test
num_bytes: 1129555
num_examples: 989
download_size: 4882238
dataset_size: 7766471
- config_name: ko_kaist
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 29037654
num_examples: 23010
- name: validation
num_bytes: 2511880
num_examples: 2066
- name: test
num_bytes: 2792215
num_examples: 2287
download_size: 21855177
dataset_size: 34341749
- config_name: ko_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2511856
num_examples: 1000
download_size: 2024810
dataset_size: 2511856
- config_name: kmr_mg
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 30374
num_examples: 20
- name: test
num_bytes: 1248564
num_examples: 734
download_size: 765158
dataset_size: 1278938
- config_name: la_ittb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 54306304
num_examples: 22775
- name: validation
num_bytes: 4236222
num_examples: 2101
- name: test
num_bytes: 4221459
num_examples: 2101
download_size: 40247546
dataset_size: 62763985
- config_name: la_llct
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 26885433
num_examples: 7289
- name: validation
num_bytes: 3363915
num_examples: 850
- name: test
num_bytes: 3352500
num_examples: 884
download_size: 21975884
dataset_size: 33601848
- config_name: la_perseus
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 2542043
num_examples: 1334
- name: test
num_bytes: 1575350
num_examples: 939
download_size: 2573703
dataset_size: 4117393
- config_name: la_proiel
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 24956038
num_examples: 15917
- name: validation
num_bytes: 2020476
num_examples: 1234
- name: test
num_bytes: 2029828
num_examples: 1260
download_size: 18434442
dataset_size: 29006342
- config_name: lv_lvtb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 29167529
num_examples: 10156
- name: validation
num_bytes: 4501172
num_examples: 1664
- name: test
num_bytes: 4565919
num_examples: 1823
download_size: 25227301
dataset_size: 38234620
- config_name: lt_alksnis
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 7272501
num_examples: 2341
- name: validation
num_bytes: 1763901
num_examples: 617
- name: test
num_bytes: 1648521
num_examples: 684
download_size: 7008248
dataset_size: 10684923
- config_name: lt_hse
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 433214
num_examples: 153
- name: validation
num_bytes: 433214
num_examples: 153
- name: test
num_bytes: 433214
num_examples: 153
download_size: 265619
dataset_size: 1299642
- config_name: olo_kkpp
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 18096
num_examples: 19
- name: test
num_bytes: 175355
num_examples: 106
download_size: 121837
dataset_size: 193451
- config_name: mt_mudt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 1858001
num_examples: 1123
- name: validation
num_bytes: 826004
num_examples: 433
- name: test
num_bytes: 892629
num_examples: 518
download_size: 2011753
dataset_size: 3576634
- config_name: gv_cadhan
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 483042
num_examples: 291
download_size: 287206
dataset_size: 483042
- config_name: mr_ufal
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 420345
num_examples: 373
- name: validation
num_bytes: 60791
num_examples: 46
- name: test
num_bytes: 56582
num_examples: 47
download_size: 339354
dataset_size: 537718
- config_name: gun_dooley
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 1037858
num_examples: 1046
download_size: 571571
dataset_size: 1037858
- config_name: gun_thomas
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 143111
num_examples: 98
download_size: 92963
dataset_size: 143111
- config_name: mdf_jr
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 234147
num_examples: 167
download_size: 162330
dataset_size: 234147
- config_name: myu_tudet
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 26202
num_examples: 62
download_size: 20315
dataset_size: 26202
- config_name: pcm_nsc
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 16079391
num_examples: 7279
- name: validation
num_bytes: 2099571
num_examples: 991
- name: test
num_bytes: 2063685
num_examples: 972
download_size: 14907410
dataset_size: 20242647
- config_name: nyq_aha
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 8723
num_examples: 10
download_size: 6387
dataset_size: 8723
- config_name: sme_giella
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 1987666
num_examples: 2257
- name: test
num_bytes: 1142396
num_examples: 865
download_size: 1862302
dataset_size: 3130062
- config_name: no_bokmaal
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 25647647
num_examples: 15696
- name: validation
num_bytes: 3828310
num_examples: 2409
- name: test
num_bytes: 3151638
num_examples: 1939
download_size: 19177350
dataset_size: 32627595
- config_name: no_nynorsk
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 25630539
num_examples: 14174
- name: validation
num_bytes: 3277649
num_examples: 1890
- name: test
num_bytes: 2601676
num_examples: 1511
download_size: 18532495
dataset_size: 31509864
- config_name: no_nynorsklia
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 3500907
num_examples: 3412
- name: validation
num_bytes: 1003845
num_examples: 881
- name: test
num_bytes: 999943
num_examples: 957
download_size: 3349676
dataset_size: 5504695
- config_name: cu_proiel
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 6106144
num_examples: 4124
- name: validation
num_bytes: 1639912
num_examples: 1073
- name: test
num_bytes: 1648459
num_examples: 1141
download_size: 6239839
dataset_size: 9394515
- config_name: fro_srcmf
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 11959859
num_examples: 13909
- name: validation
num_bytes: 1526574
num_examples: 1842
- name: test
num_bytes: 1535923
num_examples: 1927
download_size: 9043098
dataset_size: 15022356
- config_name: orv_rnc
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 1527306
num_examples: 320
- name: test
num_bytes: 2552216
num_examples: 637
download_size: 2627398
dataset_size: 4079522
- config_name: orv_torot
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 18077991
num_examples: 13336
- name: validation
num_bytes: 2408313
num_examples: 1852
- name: test
num_bytes: 2347934
num_examples: 1756
download_size: 15296362
dataset_size: 22834238
- config_name: otk_tonqq
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 22829
num_examples: 18
download_size: 14389
dataset_size: 22829
- config_name: fa_perdt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 48654947
num_examples: 26196
- name: validation
num_bytes: 2687750
num_examples: 1456
- name: test
num_bytes: 2600303
num_examples: 1455
download_size: 33606395
dataset_size: 53943000
- config_name: fa_seraji
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 12627691
num_examples: 4798
- name: validation
num_bytes: 1634327
num_examples: 599
- name: test
num_bytes: 1675134
num_examples: 600
download_size: 9890107
dataset_size: 15937152
- config_name: pl_lfg
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 16810910
num_examples: 13774
- name: validation
num_bytes: 2093712
num_examples: 1745
- name: test
num_bytes: 2100915
num_examples: 1727
download_size: 14865541
dataset_size: 21005537
- config_name: pl_pdb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 44652289
num_examples: 17722
- name: validation
num_bytes: 5494883
num_examples: 2215
- name: test
num_bytes: 5322608
num_examples: 2215
download_size: 36340919
dataset_size: 55469780
- config_name: pl_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2943603
num_examples: 1000
download_size: 1943983
dataset_size: 2943603
- config_name: pt_bosque
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 22808617
num_examples: 8328
- name: validation
num_bytes: 1201577
num_examples: 560
- name: test
num_bytes: 1131511
num_examples: 476
download_size: 15201503
dataset_size: 25141705
- config_name: pt_gsd
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 22208385
num_examples: 9664
- name: validation
num_bytes: 2805628
num_examples: 1210
- name: test
num_bytes: 2732063
num_examples: 1204
download_size: 15300844
dataset_size: 27746076
- config_name: pt_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2431942
num_examples: 1000
download_size: 1516883
dataset_size: 2431942
- config_name: ro_nonstandard
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 74489083
num_examples: 24121
- name: validation
num_bytes: 2663152
num_examples: 1052
- name: test
num_bytes: 3017162
num_examples: 1052
download_size: 50345748
dataset_size: 80169397
- config_name: ro_rrt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 23695399
num_examples: 8043
- name: validation
num_bytes: 2190973
num_examples: 752
- name: test
num_bytes: 2092520
num_examples: 729
download_size: 17187956
dataset_size: 27978892
- config_name: ro_simonero
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 15390734
num_examples: 3747
- name: validation
num_bytes: 1926639
num_examples: 443
- name: test
num_bytes: 1940787
num_examples: 491
download_size: 11409378
dataset_size: 19258160
- config_name: ru_gsd
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 10504099
num_examples: 3850
- name: validation
num_bytes: 1635884
num_examples: 579
- name: test
num_bytes: 1597603
num_examples: 601
download_size: 8830986
dataset_size: 13737586
- config_name: ru_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2695958
num_examples: 1000
download_size: 1869304
dataset_size: 2695958
- config_name: ru_syntagrus
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 126305584
num_examples: 48814
- name: validation
num_bytes: 17043673
num_examples: 6584
- name: test
num_bytes: 16880203
num_examples: 6491
download_size: 102745164
dataset_size: 160229460
- config_name: ru_taiga
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 5802733
num_examples: 3138
- name: validation
num_bytes: 1382140
num_examples: 945
- name: test
num_bytes: 1314084
num_examples: 881
download_size: 5491427
dataset_size: 8498957
- config_name: sa_ufal
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 431697
num_examples: 230
download_size: 424675
dataset_size: 431697
- config_name: sa_vedic
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 2179608
num_examples: 2524
- name: test
num_bytes: 1209605
num_examples: 1473
download_size: 2041583
dataset_size: 3389213
- config_name: gd_arcosg
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 3952356
num_examples: 1990
- name: validation
num_bytes: 1038211
num_examples: 645
- name: test
num_bytes: 1034788
num_examples: 538
download_size: 3474087
dataset_size: 6025355
- config_name: sr_set
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 9309552
num_examples: 3328
- name: validation
num_bytes: 1503953
num_examples: 536
- name: test
num_bytes: 1432672
num_examples: 520
download_size: 7414381
dataset_size: 12246177
- config_name: sms_giellagas
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 174744
num_examples: 104
download_size: 116491
dataset_size: 174744
- config_name: sk_snk
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 12017312
num_examples: 8483
- name: validation
num_bytes: 1863926
num_examples: 1060
- name: test
num_bytes: 1943012
num_examples: 1061
download_size: 10013420
dataset_size: 15824250
- config_name: sl_ssj
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 16713639
num_examples: 6478
- name: validation
num_bytes: 2070847
num_examples: 734
- name: test
num_bytes: 2083062
num_examples: 788
download_size: 12455962
dataset_size: 20867548
- config_name: sl_sst
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 2903675
num_examples: 2078
- name: test
num_bytes: 1493885
num_examples: 1110
download_size: 2655777
dataset_size: 4397560
- config_name: soj_aha
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 6218
num_examples: 8
download_size: 4577
dataset_size: 6218
- config_name: ajp_madar
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 71956
num_examples: 100
download_size: 43174
dataset_size: 71956
- config_name: es_ancora
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 50101327
num_examples: 14305
- name: validation
num_bytes: 5883940
num_examples: 1654
- name: test
num_bytes: 5928986
num_examples: 1721
download_size: 37668083
dataset_size: 61914253
- config_name: es_gsd
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 39582074
num_examples: 14187
- name: validation
num_bytes: 3834443
num_examples: 1400
- name: test
num_bytes: 1253720
num_examples: 426
download_size: 26073760
dataset_size: 44670237
- config_name: es_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2595946
num_examples: 1000
download_size: 1628475
dataset_size: 2595946
- config_name: swl_sslc
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 57443
num_examples: 87
- name: validation
num_bytes: 59002
num_examples: 82
- name: test
num_bytes: 24542
num_examples: 34
download_size: 81699
dataset_size: 140987
- config_name: sv_lines
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 6731662
num_examples: 3176
- name: validation
num_bytes: 2239951
num_examples: 1032
- name: test
num_bytes: 2070626
num_examples: 1035
download_size: 7245283
dataset_size: 11042239
- config_name: sv_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2554725
num_examples: 1000
download_size: 1722516
dataset_size: 2554725
- config_name: sv_talbanken
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 9287256
num_examples: 4303
- name: validation
num_bytes: 1361535
num_examples: 504
- name: test
num_bytes: 2835742
num_examples: 1219
download_size: 8476012
dataset_size: 13484533
- config_name: gsw_uzh
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 111357
num_examples: 100
download_size: 59675
dataset_size: 111357
- config_name: tl_trg
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 86696
num_examples: 128
download_size: 61344
dataset_size: 86696
- config_name: tl_ugnayan
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 90863
num_examples: 94
download_size: 55207
dataset_size: 90863
- config_name: ta_mwtt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 522349
num_examples: 534
download_size: 414263
dataset_size: 522349
- config_name: ta_ttb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 1538780
num_examples: 400
- name: validation
num_bytes: 305206
num_examples: 80
- name: test
num_bytes: 478941
num_examples: 120
download_size: 1753448
dataset_size: 2322927
- config_name: te_mtg
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 703512
num_examples: 1051
- name: validation
num_bytes: 91547
num_examples: 131
- name: test
num_bytes: 99757
num_examples: 146
download_size: 643764
dataset_size: 894816
- config_name: th_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2341697
num_examples: 1000
download_size: 1606517
dataset_size: 2341697
- config_name: tpn_tudet
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 8089
num_examples: 8
download_size: 5447
dataset_size: 8089
- config_name: qtd_sagt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 583697
num_examples: 285
- name: validation
num_bytes: 1564765
num_examples: 801
- name: test
num_bytes: 1710777
num_examples: 805
download_size: 2299611
dataset_size: 3859239
- config_name: tr_boun
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 12827173
num_examples: 7803
- name: validation
num_bytes: 1577760
num_examples: 979
- name: test
num_bytes: 1580727
num_examples: 979
download_size: 9742035
dataset_size: 15985660
- config_name: tr_gb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2146729
num_examples: 2880
download_size: 1474083
dataset_size: 2146729
- config_name: tr_imst
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 5063905
num_examples: 3664
- name: validation
num_bytes: 1342351
num_examples: 988
- name: test
num_bytes: 1347524
num_examples: 983
download_size: 4711018
dataset_size: 7753780
- config_name: tr_pud
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 2021772
num_examples: 1000
download_size: 1359487
dataset_size: 2021772
- config_name: uk_iu
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 18886802
num_examples: 5496
- name: validation
num_bytes: 2592721
num_examples: 672
- name: test
num_bytes: 3561164
num_examples: 892
download_size: 17344586
dataset_size: 25040687
- config_name: hsb_ufal
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 54257
num_examples: 23
- name: test
num_bytes: 1246592
num_examples: 623
download_size: 781067
dataset_size: 1300849
- config_name: ur_udtb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 19808745
num_examples: 4043
- name: validation
num_bytes: 2652349
num_examples: 552
- name: test
num_bytes: 2702596
num_examples: 535
download_size: 15901007
dataset_size: 25163690
- config_name: ug_udt
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 2570856
num_examples: 1656
- name: validation
num_bytes: 1406032
num_examples: 900
- name: test
num_bytes: 1371993
num_examples: 900
download_size: 3455092
dataset_size: 5348881
- config_name: vi_vtb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 1689772
num_examples: 1400
- name: validation
num_bytes: 948019
num_examples: 800
- name: test
num_bytes: 987207
num_examples: 800
download_size: 2055529
dataset_size: 3624998
- config_name: wbp_ufal
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 48533
num_examples: 55
download_size: 38326
dataset_size: 48533
- config_name: cy_ccg
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 1629465
num_examples: 704
- name: test
num_bytes: 1779002
num_examples: 953
download_size: 1984759
dataset_size: 3408467
- config_name: wo_wtb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: train
num_bytes: 2781883
num_examples: 1188
- name: validation
num_bytes: 1204839
num_examples: 449
- name: test
num_bytes: 1227124
num_examples: 470
download_size: 3042699
dataset_size: 5213846
- config_name: yo_ytb
features:
- name: idx
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: lemmas
sequence: string
- name: upos
sequence:
class_label:
names:
'0': NOUN
'1': PUNCT
'2': ADP
'3': NUM
'4': SYM
'5': SCONJ
'6': ADJ
'7': PART
'8': DET
'9': CCONJ
'10': PROPN
'11': PRON
'12': X
'13': _
'14': ADV
'15': INTJ
'16': VERB
'17': AUX
- name: xpos
sequence: string
- name: feats
sequence: string
- name: head
sequence: string
- name: deprel
sequence: string
- name: deps
sequence: string
- name: misc
sequence: string
splits:
- name: test
num_bytes: 905766
num_examples: 318
download_size: 567955
dataset_size: 905766
config_names:
- af_afribooms
- aii_as
- ajp_madar
- akk_pisandub
- akk_riao
- am_att
- apu_ufpa
- aqz_tudet
- ar_nyuad
- ar_padt
- ar_pud
- be_hse
- bg_btb
- bho_bhtb
- bm_crb
- br_keb
- bxr_bdt
- ca_ancora
- ckt_hse
- cop_scriptorium
- cs_cac
- cs_cltt
- cs_fictree
- cs_pdt
- cs_pud
- cu_proiel
- cy_ccg
- da_ddt
- de_gsd
- de_hdt
- de_lit
- de_pud
- el_gdt
- en_esl
- en_ewt
- en_gum
- en_gumreddit
- en_lines
- en_partut
- en_pronouns
- en_pud
- es_ancora
- es_gsd
- es_pud
- et_edt
- et_ewt
- eu_bdt
- fa_perdt
- fa_seraji
- fi_ftb
- fi_ood
- fi_pud
- fi_tdt
- fo_farpahc
- fo_oft
- fr_fqb
- fr_ftb
- fr_gsd
- fr_partut
- fr_pud
- fr_sequoia
- fr_spoken
- fro_srcmf
- ga_idt
- gd_arcosg
- gl_ctg
- gl_treegal
- got_proiel
- grc_perseus
- grc_proiel
- gsw_uzh
- gun_dooley
- gun_thomas
- gv_cadhan
- he_htb
- hi_hdtb
- hi_pud
- hr_set
- hsb_ufal
- hu_szeged
- hy_armtdp
- id_csui
- id_gsd
- id_pud
- is_icepahc
- is_pud
- it_isdt
- it_partut
- it_postwita
- it_pud
- it_twittiro
- it_vit
- ja_bccwj
- ja_gsd
- ja_modern
- ja_pud
- kfm_aha
- kk_ktb
- kmr_mg
- ko_gsd
- ko_kaist
- ko_pud
- koi_uh
- kpv_ikdp
- kpv_lattice
- krl_kkpp
- la_ittb
- la_llct
- la_perseus
- la_proiel
- lt_alksnis
- lt_hse
- lv_lvtb
- lzh_kyoto
- mdf_jr
- mr_ufal
- mt_mudt
- myu_tudet
- myv_jr
- nl_alpino
- nl_lassysmall
- no_bokmaal
- no_nynorsk
- no_nynorsklia
- nyq_aha
- olo_kkpp
- orv_rnc
- orv_torot
- otk_tonqq
- pcm_nsc
- pl_lfg
- pl_pdb
- pl_pud
- pt_bosque
- pt_gsd
- pt_pud
- qhe_hiencs
- qtd_sagt
- ro_nonstandard
- ro_rrt
- ro_simonero
- ru_gsd
- ru_pud
- ru_syntagrus
- ru_taiga
- sa_ufal
- sa_vedic
- sk_snk
- sl_ssj
- sl_sst
- sme_giella
- sms_giellagas
- soj_aha
- sq_tsa
- sr_set
- sv_lines
- sv_pud
- sv_talbanken
- swl_sslc
- ta_mwtt
- ta_ttb
- te_mtg
- th_pud
- tl_trg
- tl_ugnayan
- tpn_tudet
- tr_boun
- tr_gb
- tr_imst
- tr_pud
- ug_udt
- uk_iu
- ur_udtb
- vi_vtb
- wbp_ufal
- wo_wtb
- yo_ytb
- yue_hk
- zh_cfl
- zh_gsd
- zh_gsdsimp
- zh_hk
- zh_pud
---
# Dataset Card for Universal Dependencies Treebank
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [Universal Dependencies](https://universaldependencies.org/)
- **Repository:**
- **Paper:**
- **Leaderboard:**
- **Point of Contact:**
### Dataset Summary
[More Information Needed]
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
Thanks to [@patrickvonplaten](https://github.com/patrickvonplaten), [@jplu](https://github.com/jplu) for adding this dataset. |
common_voice | 2023-06-27T07:46:51.000Z | [
"task_categories:automatic-speech-recognition",
"annotations_creators:crowdsourced",
"language_creators:crowdsourced",
"multilinguality:multilingual",
"size_categories:100K<n<1M",
"size_categories:10K<n<100K",
"size_categories:1K<n<10K",
"size_categories:n<1K",
"source_datasets:extended|common_voice",
"language:ab",
"language:ar",
"language:as",
"language:br",
"language:ca",
"language:cnh",
"language:cs",
"language:cv",
"language:cy",
"language:de",
"language:dv",
"language:el",
"language:en",
"language:eo",
"language:es",
"language:et",
"language:eu",
"language:fa",
"language:fi",
"language:fr",
"language:fy",
"language:ga",
"language:hi",
"language:hsb",
"language:hu",
"language:ia",
"language:id",
"language:it",
"language:ja",
"language:ka",
"language:kab",
"language:ky",
"language:lg",
"language:lt",
"language:lv",
"language:mn",
"language:mt",
"language:nl",
"language:or",
"language:pa",
"language:pl",
"language:pt",
"language:rm",
"language:ro",
"language:ru",
"language:rw",
"language:sah",
"language:sl",
"language:sv",
"language:ta",
"language:th",
"language:tr",
"language:tt",
"language:uk",
"language:vi",
"language:vot",
"language:zh",
"license:cc0-1.0",
"region:us"
] | null | Common Voice is Mozilla's initiative to help teach machines how real people speak.
The dataset currently consists of 7,335 validated hours of speech in 60 languages, but we’re always adding more voices and languages. | @inproceedings{commonvoice:2020,
author = {Ardila, R. and Branson, M. and Davis, K. and Henretty, M. and Kohler, M. and Meyer, J. and Morais, R. and Saunders, L. and Tyers, F. M. and Weber, G.},
title = {Common Voice: A Massively-Multilingual Speech Corpus},
booktitle = {Proceedings of the 12th Conference on Language Resources and Evaluation (LREC 2020)},
pages = {4211--4215},
year = 2020
} | null | 99 | 6,810 | ---
pretty_name: Common Voice
annotations_creators:
- crowdsourced
language_creators:
- crowdsourced
language:
- ab
- ar
- as
- br
- ca
- cnh
- cs
- cv
- cy
- de
- dv
- el
- en
- eo
- es
- et
- eu
- fa
- fi
- fr
- fy
- ga
- hi
- hsb
- hu
- ia
- id
- it
- ja
- ka
- kab
- ky
- lg
- lt
- lv
- mn
- mt
- nl
- or
- pa
- pl
- pt
- rm
- ro
- ru
- rw
- sah
- sl
- sv
- ta
- th
- tr
- tt
- uk
- vi
- vot
- zh
language_bcp47:
- fy-NL
- ga-IE
- pa-IN
- rm-sursilv
- rm-vallader
- sv-SE
- zh-CN
- zh-HK
- zh-TW
license:
- cc0-1.0
multilinguality:
- multilingual
size_categories:
- 100K<n<1M
- 10K<n<100K
- 1K<n<10K
- n<1K
source_datasets:
- extended|common_voice
task_categories:
- automatic-speech-recognition
task_ids: []
paperswithcode_id: common-voice
dataset_info:
- config_name: ab
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 1295622
num_examples: 22
- name: test
num_bytes: 411844
num_examples: 9
- name: validation
- name: other
num_bytes: 40023390
num_examples: 752
- name: validated
num_bytes: 1707426
num_examples: 31
- name: invalidated
num_bytes: 361626
num_examples: 8
download_size: 41038412
dataset_size: 43799908
- config_name: ar
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 359335168
num_examples: 14227
- name: test
num_bytes: 237546641
num_examples: 7622
- name: validation
num_bytes: 209606861
num_examples: 7517
- name: other
num_bytes: 515822404
num_examples: 18283
- name: validated
num_bytes: 1182522872
num_examples: 43291
- name: invalidated
num_bytes: 194805036
num_examples: 6333
download_size: 1756264615
dataset_size: 2699638982
- config_name: as
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 11442279
num_examples: 270
- name: test
num_bytes: 5071343
num_examples: 110
- name: validation
num_bytes: 5480156
num_examples: 124
- name: other
- name: validated
num_bytes: 21993698
num_examples: 504
- name: invalidated
num_bytes: 886145
num_examples: 31
download_size: 22226465
dataset_size: 44873621
- config_name: br
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 62238289
num_examples: 2780
- name: test
num_bytes: 54461339
num_examples: 2087
- name: validation
num_bytes: 46995570
num_examples: 1997
- name: other
num_bytes: 269858143
num_examples: 10912
- name: validated
num_bytes: 203503622
num_examples: 8560
- name: invalidated
num_bytes: 20861017
num_examples: 623
download_size: 465276982
dataset_size: 657917980
- config_name: ca
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 12966939466
num_examples: 285584
- name: test
num_bytes: 745761890
num_examples: 15724
- name: validation
num_bytes: 716442038
num_examples: 15724
- name: other
num_bytes: 2693542910
num_examples: 64446
- name: validated
num_bytes: 18115833966
num_examples: 416701
- name: invalidated
num_bytes: 850402888
num_examples: 18846
download_size: 20743110341
dataset_size: 36088923158
- config_name: cnh
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 18866674
num_examples: 807
- name: test
num_bytes: 24675321
num_examples: 752
- name: validation
num_bytes: 22162315
num_examples: 756
- name: other
num_bytes: 84878963
num_examples: 2934
- name: validated
num_bytes: 69330148
num_examples: 2432
- name: invalidated
num_bytes: 13642724
num_examples: 433
download_size: 161331331
dataset_size: 233556145
- config_name: cs
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 215205282
num_examples: 5655
- name: test
num_bytes: 148499476
num_examples: 4144
- name: validation
num_bytes: 148312130
num_examples: 4118
- name: other
num_bytes: 282225475
num_examples: 7475
- name: validated
num_bytes: 1019817024
num_examples: 30431
- name: invalidated
num_bytes: 24717823
num_examples: 685
download_size: 1271909933
dataset_size: 1838777210
- config_name: cv
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 31649510
num_examples: 931
- name: test
num_bytes: 32513061
num_examples: 788
- name: validation
num_bytes: 28429779
num_examples: 818
- name: other
num_bytes: 288294623
num_examples: 6927
- name: validated
num_bytes: 126717875
num_examples: 3496
- name: invalidated
num_bytes: 57923138
num_examples: 1282
download_size: 439329081
dataset_size: 565527986
- config_name: cy
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 271642649
num_examples: 6839
- name: test
num_bytes: 206865596
num_examples: 4820
- name: validation
num_bytes: 201813388
num_examples: 4776
- name: other
num_bytes: 688469886
num_examples: 17919
- name: validated
num_bytes: 2763112391
num_examples: 72984
- name: invalidated
num_bytes: 146874576
num_examples: 3648
download_size: 3434474658
dataset_size: 4278778486
- config_name: de
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 11463160619
num_examples: 246525
- name: test
num_bytes: 744617681
num_examples: 15588
- name: validation
num_bytes: 729559862
num_examples: 15588
- name: other
num_bytes: 464513461
num_examples: 10095
- name: validated
num_bytes: 22402489041
num_examples: 565186
- name: invalidated
num_bytes: 1440604803
num_examples: 32789
download_size: 23283812097
dataset_size: 37244945467
- config_name: dv
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 118576140
num_examples: 2680
- name: test
num_bytes: 94281409
num_examples: 2202
- name: validation
num_bytes: 94117088
num_examples: 2077
- name: other
- name: validated
num_bytes: 528571107
num_examples: 11866
- name: invalidated
num_bytes: 37694847
num_examples: 840
download_size: 540488041
dataset_size: 873240591
- config_name: el
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 80759076
num_examples: 2316
- name: test
num_bytes: 53820491
num_examples: 1522
- name: validation
num_bytes: 44818565
num_examples: 1401
- name: other
num_bytes: 186861175
num_examples: 5659
- name: validated
num_bytes: 204446790
num_examples: 5996
- name: invalidated
num_bytes: 6023769
num_examples: 185
download_size: 381570611
dataset_size: 576729866
- config_name: en
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 26088826658
num_examples: 564337
- name: test
num_bytes: 758718688
num_examples: 16164
- name: validation
num_bytes: 795638801
num_examples: 16164
- name: other
num_bytes: 5796244022
num_examples: 169895
- name: validated
num_bytes: 48425872575
num_examples: 1224864
- name: invalidated
num_bytes: 9122973965
num_examples: 189562
download_size: 60613063630
dataset_size: 90988274709
- config_name: eo
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 993655930
num_examples: 19587
- name: test
num_bytes: 420153812
num_examples: 8969
- name: validation
num_bytes: 391427586
num_examples: 8987
- name: other
num_bytes: 142476819
num_examples: 2946
- name: validated
num_bytes: 2603249289
num_examples: 58094
- name: invalidated
num_bytes: 238105462
num_examples: 4736
download_size: 2883560869
dataset_size: 4789068898
- config_name: es
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 6918333205
num_examples: 161813
- name: test
num_bytes: 754049291
num_examples: 15089
- name: validation
num_bytes: 735558084
num_examples: 15089
- name: other
num_bytes: 5528972205
num_examples: 144791
- name: validated
num_bytes: 9623788388
num_examples: 236314
- name: invalidated
num_bytes: 1664876264
num_examples: 40640
download_size: 16188844718
dataset_size: 25225577437
- config_name: et
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 161124199
num_examples: 2966
- name: test
num_bytes: 133183135
num_examples: 2509
- name: validation
num_bytes: 137604813
num_examples: 2507
- name: other
num_bytes: 30339130
num_examples: 569
- name: validated
num_bytes: 573417188
num_examples: 10683
- name: invalidated
num_bytes: 193019544
num_examples: 3557
download_size: 767174465
dataset_size: 1228688009
- config_name: eu
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 317322801
num_examples: 7505
- name: test
num_bytes: 238866501
num_examples: 5172
- name: validation
num_bytes: 228150083
num_examples: 5172
- name: other
num_bytes: 988079897
num_examples: 23570
- name: validated
num_bytes: 2621488299
num_examples: 63009
- name: invalidated
num_bytes: 208553909
num_examples: 5387
download_size: 3664586106
dataset_size: 4602461490
- config_name: fa
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 239255087
num_examples: 7593
- name: test
num_bytes: 217939210
num_examples: 5213
- name: validation
num_bytes: 196558067
num_examples: 5213
- name: other
num_bytes: 737017546
num_examples: 22510
- name: validated
num_bytes: 8120181903
num_examples: 251659
- name: invalidated
num_bytes: 499570226
num_examples: 11698
download_size: 8884585819
dataset_size: 10010522039
- config_name: fi
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 16017393
num_examples: 460
- name: test
num_bytes: 16117529
num_examples: 428
- name: validation
num_bytes: 15471757
num_examples: 415
- name: other
num_bytes: 5836400
num_examples: 149
- name: validated
num_bytes: 47669391
num_examples: 1305
- name: invalidated
num_bytes: 2228215
num_examples: 59
download_size: 49882909
dataset_size: 103340685
- config_name: fr
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 12439892070
num_examples: 298982
- name: test
num_bytes: 733943163
num_examples: 15763
- name: validation
num_bytes: 703801114
num_examples: 15763
- name: other
num_bytes: 117998889
num_examples: 3222
- name: validated
num_bytes: 17921836252
num_examples: 461004
- name: invalidated
num_bytes: 1794149368
num_examples: 40351
download_size: 19130141984
dataset_size: 33711620856
- config_name: fy-NL
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 159116360
num_examples: 3927
- name: test
num_bytes: 126913262
num_examples: 3020
- name: validation
num_bytes: 112288554
num_examples: 2790
- name: other
num_bytes: 893887467
num_examples: 21569
- name: validated
num_bytes: 429651922
num_examples: 10495
- name: invalidated
num_bytes: 38985422
num_examples: 1031
download_size: 1237743070
dataset_size: 1760842987
- config_name: ga-IE
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 15396820
num_examples: 541
- name: test
num_bytes: 16611739
num_examples: 506
- name: validation
num_bytes: 14897739
num_examples: 497
- name: other
num_bytes: 61948768
num_examples: 2130
- name: validated
num_bytes: 93371649
num_examples: 3352
- name: invalidated
num_bytes: 10993268
num_examples: 409
download_size: 156553447
dataset_size: 213219983
- config_name: hi
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 4860737
num_examples: 157
- name: test
num_bytes: 4728043
num_examples: 127
- name: validation
num_bytes: 5569352
num_examples: 135
- name: other
num_bytes: 4176110
num_examples: 139
- name: validated
num_bytes: 15158052
num_examples: 419
- name: invalidated
num_bytes: 2801051
num_examples: 60
download_size: 21424045
dataset_size: 37293345
- config_name: hsb
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 43049910
num_examples: 808
- name: test
num_bytes: 20929094
num_examples: 387
- name: validation
num_bytes: 8769458
num_examples: 172
- name: other
num_bytes: 3173841
num_examples: 62
- name: validated
num_bytes: 72748422
num_examples: 1367
- name: invalidated
num_bytes: 5589972
num_examples: 227
download_size: 79362060
dataset_size: 154260697
- config_name: hu
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 126163153
num_examples: 3348
- name: test
num_bytes: 57056435
num_examples: 1649
- name: validation
num_bytes: 50306925
num_examples: 1434
- name: other
num_bytes: 12051094
num_examples: 295
- name: validated
num_bytes: 234307671
num_examples: 6457
- name: invalidated
num_bytes: 5881521
num_examples: 169
download_size: 242758708
dataset_size: 485766799
- config_name: ia
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 96577153
num_examples: 3477
- name: test
num_bytes: 33204678
num_examples: 899
- name: validation
num_bytes: 67436779
num_examples: 1601
- name: other
num_bytes: 30937041
num_examples: 1095
- name: validated
num_bytes: 197248304
num_examples: 5978
- name: invalidated
num_bytes: 6769573
num_examples: 192
download_size: 226499645
dataset_size: 432173528
- config_name: id
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 63515863
num_examples: 2130
- name: test
num_bytes: 60711104
num_examples: 1844
- name: validation
num_bytes: 56963520
num_examples: 1835
- name: other
num_bytes: 206578628
num_examples: 6782
- name: validated
num_bytes: 272570942
num_examples: 8696
- name: invalidated
num_bytes: 16566129
num_examples: 470
download_size: 475918233
dataset_size: 676906186
- config_name: it
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 2555546829
num_examples: 58015
- name: test
num_bytes: 656285877
num_examples: 12928
- name: validation
num_bytes: 621955330
num_examples: 12928
- name: other
num_bytes: 671213467
num_examples: 14549
- name: validated
num_bytes: 4552252754
num_examples: 102579
- name: invalidated
num_bytes: 564610354
num_examples: 12189
download_size: 5585781573
dataset_size: 9621864611
- config_name: ja
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 27600264
num_examples: 722
- name: test
num_bytes: 26475556
num_examples: 632
- name: validation
num_bytes: 22098940
num_examples: 586
- name: other
num_bytes: 34588931
num_examples: 885
- name: validated
num_bytes: 106916400
num_examples: 3072
- name: invalidated
num_bytes: 17819020
num_examples: 504
download_size: 152879796
dataset_size: 235499111
- config_name: ka
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 47790695
num_examples: 1058
- name: test
num_bytes: 30301524
num_examples: 656
- name: validation
num_bytes: 24951079
num_examples: 527
- name: other
num_bytes: 2144603
num_examples: 44
- name: validated
num_bytes: 104135978
num_examples: 2275
- name: invalidated
num_bytes: 7004160
num_examples: 139
download_size: 104280554
dataset_size: 216328039
- config_name: kab
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 3219289101
num_examples: 120530
- name: test
num_bytes: 446453041
num_examples: 14622
- name: validation
num_bytes: 414159937
num_examples: 14622
- name: other
num_bytes: 2282481767
num_examples: 88021
- name: validated
num_bytes: 15310455176
num_examples: 573718
- name: invalidated
num_bytes: 581587104
num_examples: 18134
download_size: 17171606918
dataset_size: 22254426126
- config_name: ky
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 75460488
num_examples: 1955
- name: test
num_bytes: 57116561
num_examples: 1503
- name: validation
num_bytes: 61393867
num_examples: 1511
- name: other
num_bytes: 258081579
num_examples: 7223
- name: validated
num_bytes: 355742823
num_examples: 9236
- name: invalidated
num_bytes: 41007711
num_examples: 926
download_size: 579440853
dataset_size: 848803029
- config_name: lg
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 46910479
num_examples: 1250
- name: test
num_bytes: 26951803
num_examples: 584
- name: validation
num_bytes: 16709367
num_examples: 384
- name: other
num_bytes: 111180838
num_examples: 3110
- name: validated
num_bytes: 90606863
num_examples: 2220
- name: invalidated
num_bytes: 14069959
num_examples: 290
download_size: 208197149
dataset_size: 306429309
- config_name: lt
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 34605356
num_examples: 931
- name: test
num_bytes: 19940391
num_examples: 466
- name: validation
num_bytes: 10462851
num_examples: 244
- name: other
num_bytes: 71150206
num_examples: 1629
- name: validated
num_bytes: 65138550
num_examples: 1644
- name: invalidated
num_bytes: 4414780
num_examples: 102
download_size: 135299706
dataset_size: 205712134
- config_name: lv
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 67269173
num_examples: 2552
- name: test
num_bytes: 56937435
num_examples: 1882
- name: validation
num_bytes: 55289058
num_examples: 2002
- name: other
num_bytes: 40259801
num_examples: 1560
- name: validated
num_bytes: 179726893
num_examples: 6444
- name: invalidated
num_bytes: 4383319
num_examples: 143
download_size: 208307691
dataset_size: 403865679
- config_name: mn
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 89913910
num_examples: 2183
- name: test
num_bytes: 86737041
num_examples: 1862
- name: validation
num_bytes: 82343275
num_examples: 1837
- name: other
num_bytes: 146365394
num_examples: 3272
- name: validated
num_bytes: 327264827
num_examples: 7487
- name: invalidated
num_bytes: 31764232
num_examples: 667
download_size: 486369317
dataset_size: 764388679
- config_name: mt
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 73850815
num_examples: 2036
- name: test
num_bytes: 66520195
num_examples: 1617
- name: validation
num_bytes: 56412066
num_examples: 1516
- name: other
num_bytes: 220666971
num_examples: 5714
- name: validated
num_bytes: 218212969
num_examples: 5747
- name: invalidated
num_bytes: 12328068
num_examples: 314
download_size: 425114242
dataset_size: 647991084
- config_name: nl
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 321946148
num_examples: 9460
- name: test
num_bytes: 205287443
num_examples: 5708
- name: validation
num_bytes: 186095353
num_examples: 4938
- name: other
num_bytes: 801418
num_examples: 27
- name: validated
num_bytes: 1710636990
num_examples: 52488
- name: invalidated
num_bytes: 115133112
num_examples: 3308
download_size: 1741827548
dataset_size: 2539900464
- config_name: or
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 16067910
num_examples: 388
- name: test
num_bytes: 4270651
num_examples: 98
- name: validation
num_bytes: 5485937
num_examples: 129
- name: other
num_bytes: 177775963
num_examples: 4302
- name: validated
num_bytes: 25824418
num_examples: 615
- name: invalidated
num_bytes: 2701922
num_examples: 62
download_size: 199077358
dataset_size: 232126801
- config_name: pa-IN
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 7572499
num_examples: 211
- name: test
num_bytes: 4375532
num_examples: 116
- name: validation
num_bytes: 1702492
num_examples: 44
- name: other
num_bytes: 56683312
num_examples: 1411
- name: validated
num_bytes: 13650443
num_examples: 371
- name: invalidated
num_bytes: 1690766
num_examples: 43
download_size: 69748265
dataset_size: 85675044
- config_name: pl
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 273394509
num_examples: 7468
- name: test
num_bytes: 205047541
num_examples: 5153
- name: validation
num_bytes: 195917307
num_examples: 5153
- name: other
num_bytes: 442144781
num_examples: 12848
- name: validated
num_bytes: 3150860197
num_examples: 90791
- name: invalidated
num_bytes: 180801918
num_examples: 4601
download_size: 3537012341
dataset_size: 4448166253
- config_name: pt
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 231451724
num_examples: 6514
- name: test
num_bytes: 180108694
num_examples: 4641
- name: validation
num_bytes: 165966139
num_examples: 4592
- name: other
num_bytes: 283497435
num_examples: 8390
- name: validated
num_bytes: 1480529669
num_examples: 41584
- name: invalidated
num_bytes: 67948392
num_examples: 1740
download_size: 1704252567
dataset_size: 2409502053
- config_name: rm-sursilv
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 62396326
num_examples: 1384
- name: test
num_bytes: 51707733
num_examples: 1194
- name: validation
num_bytes: 52114252
num_examples: 1205
- name: other
num_bytes: 93351293
num_examples: 2102
- name: validated
num_bytes: 166218231
num_examples: 3783
- name: invalidated
num_bytes: 30593270
num_examples: 639
download_size: 275950479
dataset_size: 456381105
- config_name: rm-vallader
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 29528457
num_examples: 574
- name: test
num_bytes: 18805466
num_examples: 378
- name: validation
num_bytes: 17012341
num_examples: 357
- name: other
num_bytes: 36890435
num_examples: 727
- name: validated
num_bytes: 65711922
num_examples: 1316
- name: invalidated
num_bytes: 9356204
num_examples: 374
download_size: 108113989
dataset_size: 177304825
- config_name: ro
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 107235430
num_examples: 3399
- name: test
num_bytes: 60106568
num_examples: 1778
- name: validation
num_bytes: 30358457
num_examples: 858
- name: other
num_bytes: 65805210
num_examples: 1945
- name: validated
num_bytes: 197820619
num_examples: 6039
- name: invalidated
num_bytes: 11108104
num_examples: 485
download_size: 261978702
dataset_size: 472434388
- config_name: ru
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 686168722
num_examples: 15481
- name: test
num_bytes: 385349488
num_examples: 8007
- name: validation
num_bytes: 361164462
num_examples: 7963
- name: other
num_bytes: 450644862
num_examples: 10247
- name: validated
num_bytes: 3212213931
num_examples: 74256
- name: invalidated
num_bytes: 145739451
num_examples: 3056
download_size: 3655676916
dataset_size: 5241280916
- config_name: rw
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 21645788973
num_examples: 515197
- name: test
num_bytes: 707959382
num_examples: 15724
- name: validation
num_bytes: 698662384
num_examples: 15032
- name: other
num_bytes: 923146896
num_examples: 22923
- name: validated
num_bytes: 35011249432
num_examples: 832929
- name: invalidated
num_bytes: 7969286423
num_examples: 206790
download_size: 42545189583
dataset_size: 66956093490
- config_name: sah
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 68286985
num_examples: 1442
- name: test
num_bytes: 38534020
num_examples: 757
- name: validation
num_bytes: 17900397
num_examples: 405
- name: other
num_bytes: 62594222
num_examples: 1275
- name: validated
num_bytes: 124800352
num_examples: 2606
- name: invalidated
num_bytes: 3594160
num_examples: 66
download_size: 181245626
dataset_size: 315710136
- config_name: sl
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 66122967
num_examples: 2038
- name: test
num_bytes: 26872195
num_examples: 881
- name: validation
num_bytes: 16353097
num_examples: 556
- name: other
num_bytes: 79268518
num_examples: 2502
- name: validated
num_bytes: 148371273
num_examples: 4669
- name: invalidated
num_bytes: 3048301
num_examples: 92
download_size: 222751292
dataset_size: 340036351
- config_name: sv-SE
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 62727263
num_examples: 2331
- name: test
num_bytes: 59127381
num_examples: 2027
- name: validation
num_bytes: 53846355
num_examples: 2019
- name: other
num_bytes: 109970049
num_examples: 3043
- name: validated
num_bytes: 327049001
num_examples: 12552
- name: invalidated
num_bytes: 13462567
num_examples: 462
download_size: 421434184
dataset_size: 626182616
- config_name: ta
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 69052658
num_examples: 2009
- name: test
num_bytes: 67616865
num_examples: 1781
- name: validation
num_bytes: 63248009
num_examples: 1779
- name: other
num_bytes: 246650792
num_examples: 7428
- name: validated
num_bytes: 438961956
num_examples: 12652
- name: invalidated
num_bytes: 23587453
num_examples: 594
download_size: 679766097
dataset_size: 909117733
- config_name: th
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 100435725
num_examples: 2917
- name: test
num_bytes: 82030679
num_examples: 2188
- name: validation
num_bytes: 63237632
num_examples: 1922
- name: other
num_bytes: 95235301
num_examples: 2671
- name: validated
num_bytes: 245734783
num_examples: 7028
- name: invalidated
num_bytes: 18247080
num_examples: 467
download_size: 341305736
dataset_size: 604921200
- config_name: tr
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 57879052
num_examples: 1831
- name: test
num_bytes: 60268059
num_examples: 1647
- name: validation
num_bytes: 54914798
num_examples: 1647
- name: other
num_bytes: 10954154
num_examples: 325
- name: validated
num_bytes: 585777527
num_examples: 18685
- name: invalidated
num_bytes: 59288266
num_examples: 1726
download_size: 620848700
dataset_size: 829081856
- config_name: tt
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 348132697
num_examples: 11211
- name: test
num_bytes: 135120057
num_examples: 4485
- name: validation
num_bytes: 61690964
num_examples: 2127
- name: other
num_bytes: 62158038
num_examples: 1798
- name: validated
num_bytes: 767791517
num_examples: 25781
- name: invalidated
num_bytes: 10403128
num_examples: 287
download_size: 777153207
dataset_size: 1385296401
- config_name: uk
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 161925063
num_examples: 4035
- name: test
num_bytes: 138422211
num_examples: 3235
- name: validation
num_bytes: 135483169
num_examples: 3236
- name: other
num_bytes: 327979131
num_examples: 8161
- name: validated
num_bytes: 889863965
num_examples: 22337
- name: invalidated
num_bytes: 55745301
num_examples: 1255
download_size: 1218559031
dataset_size: 1709418840
- config_name: vi
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 6244454
num_examples: 221
- name: test
num_bytes: 6656365
num_examples: 198
- name: validation
num_bytes: 6531856
num_examples: 200
- name: other
num_bytes: 31315434
num_examples: 870
- name: validated
num_bytes: 19432595
num_examples: 619
- name: invalidated
num_bytes: 2981661
num_examples: 78
download_size: 51929480
dataset_size: 73162365
- config_name: vot
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 146467
num_examples: 3
- name: test
- name: validation
- name: other
num_bytes: 7963322
num_examples: 411
- name: validated
num_bytes: 146467
num_examples: 3
- name: invalidated
num_bytes: 107949
num_examples: 6
download_size: 7792602
dataset_size: 8364205
- config_name: zh-CN
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 793667379
num_examples: 18541
- name: test
num_bytes: 420202544
num_examples: 8760
- name: validation
num_bytes: 396096323
num_examples: 8743
- name: other
num_bytes: 381264783
num_examples: 8948
- name: validated
num_bytes: 1618113625
num_examples: 36405
- name: invalidated
num_bytes: 266234479
num_examples: 5305
download_size: 2184602350
dataset_size: 3875579133
- config_name: zh-HK
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 221459521
num_examples: 7506
- name: test
num_bytes: 217627041
num_examples: 5172
- name: validation
num_bytes: 196071110
num_examples: 5172
- name: other
num_bytes: 1319233252
num_examples: 38830
- name: validated
num_bytes: 1482087591
num_examples: 41835
- name: invalidated
num_bytes: 124170969
num_examples: 2999
download_size: 2774145806
dataset_size: 3560649484
- config_name: zh-TW
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
splits:
- name: train
num_bytes: 97323787
num_examples: 3507
- name: test
num_bytes: 85512325
num_examples: 2895
- name: validation
num_bytes: 80402637
num_examples: 2895
- name: other
num_bytes: 623801957
num_examples: 22477
- name: validated
num_bytes: 1568842090
num_examples: 61232
- name: invalidated
num_bytes: 100241443
num_examples: 3584
download_size: 2182836295
dataset_size: 2556124239
config_names:
- ab
- ar
- as
- br
- ca
- cnh
- cs
- cv
- cy
- de
- dv
- el
- en
- eo
- es
- et
- eu
- fa
- fi
- fr
- fy-NL
- ga-IE
- hi
- hsb
- hu
- ia
- id
- it
- ja
- ka
- kab
- ky
- lg
- lt
- lv
- mn
- mt
- nl
- or
- pa-IN
- pl
- pt
- rm-sursilv
- rm-vallader
- ro
- ru
- rw
- sah
- sl
- sv-SE
- ta
- th
- tr
- tt
- uk
- vi
- vot
- zh-CN
- zh-HK
- zh-TW
---
# Dataset Card for common_voice
<div class="course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400">
<p><b>Deprecated:</b> Dataset "common_voice" is deprecated and will soon be deleted. Use datasets under <a href="https://huggingface.co/mozilla-foundation">mozilla-foundation</a> organisation instead. For example, you can load <a href="https://huggingface.co/datasets/mozilla-foundation/common_voice_13_0">Common Voice 13</a> dataset via <code>load_dataset("mozilla-foundation/common_voice_13_0", "en")</code></p>
</div>
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** https://commonvoice.mozilla.org/en/datasets
- **Repository:** https://github.com/common-voice/common-voice
- **Paper:** https://commonvoice.mozilla.org/en/datasets
- **Leaderboard:** [Needs More Information]
- **Point of Contact:** [Needs More Information]
### Dataset Summary
The Common Voice dataset consists of a unique MP3 and corresponding text file. Many of the 9,283 recorded hours in the dataset also include demographic metadata like age, sex, and accent that can help train the accuracy of speech recognition engines.
The dataset currently consists of 7,335 validated hours in 60 languages, but were always adding more voices and languages. Take a look at our Languages page to request a language or start contributing.
### Supported Tasks and Leaderboards
[Needs More Information]
### Languages
English
## Dataset Structure
### Data Instances
A typical data point comprises the path to the audio file, called path and its sentence. Additional fields include accent, age, client_id, up_votes down_votes, gender, locale and segment.
`
{'accent': 'netherlands', 'age': 'fourties', 'client_id': 'bbbcb732e0f422150c30ff3654bbab572e2a617da107bca22ff8b89ab2e4f124d03b6a92c48322862f60bd0179ae07baf0f9b4f9c4e11d581e0cec70f703ba54', 'down_votes': 0, 'gender': 'male', 'locale': 'nl', 'path': 'nl/clips/common_voice_nl_23522441.mp3', 'segment': "''", 'sentence': 'Ik vind dat een dubieuze procedure.', 'up_votes': 2, 'audio': {'path': `nl/clips/common_voice_nl_23522441.mp3', 'array': array([-0.00048828, -0.00018311, -0.00137329, ..., 0.00079346, 0.00091553, 0.00085449], dtype=float32), 'sampling_rate': 48000}
`
### Data Fields
client_id: An id for which client (voice) made the recording
path: The path to the audio file
audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: `dataset[0]["audio"]` the audio file is automatically decoded and resampled to `dataset.features["audio"].sampling_rate`. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the `"audio"` column, *i.e.* `dataset[0]["audio"]` should **always** be preferred over `dataset["audio"][0]`.
sentence: The sentence the user was prompted to speak
up_votes: How many upvotes the audio file has received from reviewers
down_votes: How many downvotes the audio file has received from reviewers
age: The age of the speaker.
gender: The gender of the speaker
accent: Accent of the speaker
locale: The locale of the speaker
segment: Usually empty field
### Data Splits
The speech material has been subdivided into portions for dev, train, test, validated, invalidated, reported and other.
The validated data is data that has been validated with reviewers and recieved upvotes that the data is of high quality.
The invalidated data is data has been invalidated by reviewers
and recieved downvotes that the data is of low quality.
The reported data is data that has been reported, for different reasons.
The other data is data that has not yet been reviewed.
The dev, test, train are all data that has been reviewed, deemed of high quality and split into dev, test and train.
## Dataset Creation
### Curation Rationale
[Needs More Information]
### Source Data
#### Initial Data Collection and Normalization
[Needs More Information]
#### Who are the source language producers?
[Needs More Information]
### Annotations
#### Annotation process
[Needs More Information]
#### Who are the annotators?
[Needs More Information]
### Personal and Sensitive Information
The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in the Common Voice dataset.
## Considerations for Using the Data
### Social Impact of Dataset
The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in the Common Voice dataset.
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
Public Domain, [CC-0](https://creativecommons.org/share-your-work/public-domain/cc0/)
### Citation Information
```
@inproceedings{commonvoice:2020,
author = {Ardila, R. and Branson, M. and Davis, K. and Henretty, M. and Kohler, M. and Meyer, J. and Morais, R. and Saunders, L. and Tyers, F. M. and Weber, G.},
title = {Common Voice: A Massively-Multilingual Speech Corpus},
booktitle = {Proceedings of the 12th Conference on Language Resources and Evaluation (LREC 2020)},
pages = {4211--4215},
year = 2020
}
```
### Contributions
Thanks to [@BirgerMoell](https://github.com/BirgerMoell) for adding this dataset. |
mteb/mtop_domain | 2022-11-21T19:59:05.000Z | [
"task_categories:text-classification",
"language:de",
"language:en",
"language:es",
"language:fr",
"language:hi",
"language:th",
"region:us"
] | mteb | null | null | null | 2 | 6,786 | ---
task_categories:
- text-classification
language:
- de
- en
- es
- fr
- hi
- th
--- |
sick | 2023-01-25T14:44:16.000Z | [
"task_categories:text-classification",
"task_ids:natural-language-inference",
"annotations_creators:crowdsourced",
"language_creators:crowdsourced",
"multilinguality:monolingual",
"size_categories:1K<n<10K",
"source_datasets:extended|image-flickr-8k",
"source_datasets:extended|semeval2012-sts-msr-video",
"language:en",
"license:cc-by-nc-sa-3.0",
"region:us"
] | null | Shared and internationally recognized benchmarks are fundamental for the development of any computational system.
We aim to help the research community working on compositional distributional semantic models (CDSMs) by providing SICK (Sentences Involving Compositional Knowldedge), a large size English benchmark tailored for them.
SICK consists of about 10,000 English sentence pairs that include many examples of the lexical, syntactic and semantic phenomena that CDSMs are expected to account for, but do not require dealing with other aspects of existing sentential data sets (idiomatic multiword expressions, named entities, telegraphic language) that are not within the scope of CDSMs.
By means of crowdsourcing techniques, each pair was annotated for two crucial semantic tasks: relatedness in meaning (with a 5-point rating scale as gold score) and entailment relation between the two elements (with three possible gold labels: entailment, contradiction, and neutral).
The SICK data set was used in SemEval-2014 Task 1, and it freely available for research purposes. | @inproceedings{marelli-etal-2014-sick,
title = "A {SICK} cure for the evaluation of compositional distributional semantic models",
author = "Marelli, Marco and
Menini, Stefano and
Baroni, Marco and
Bentivogli, Luisa and
Bernardi, Raffaella and
Zamparelli, Roberto",
booktitle = "Proceedings of the Ninth International Conference on Language Resources and Evaluation ({LREC}'14)",
month = may,
year = "2014",
address = "Reykjavik, Iceland",
publisher = "European Language Resources Association (ELRA)",
url = "http://www.lrec-conf.org/proceedings/lrec2014/pdf/363_Paper.pdf",
pages = "216--223",
} | null | 5 | 6,754 | ---
annotations_creators:
- crowdsourced
language_creators:
- crowdsourced
language:
- en
license:
- cc-by-nc-sa-3.0
multilinguality:
- monolingual
size_categories:
- 1K<n<10K
source_datasets:
- extended|image-flickr-8k
- extended|semeval2012-sts-msr-video
task_categories:
- text-classification
task_ids:
- natural-language-inference
paperswithcode_id: sick
pretty_name: Sentences Involving Compositional Knowledge
dataset_info:
features:
- name: id
dtype: string
- name: sentence_A
dtype: string
- name: sentence_B
dtype: string
- name: label
dtype:
class_label:
names:
'0': entailment
'1': neutral
'2': contradiction
- name: relatedness_score
dtype: float32
- name: entailment_AB
dtype: string
- name: entailment_BA
dtype: string
- name: sentence_A_original
dtype: string
- name: sentence_B_original
dtype: string
- name: sentence_A_dataset
dtype: string
- name: sentence_B_dataset
dtype: string
splits:
- name: train
num_bytes: 1180530
num_examples: 4439
- name: validation
num_bytes: 132913
num_examples: 495
- name: test
num_bytes: 1305846
num_examples: 4906
download_size: 217584
dataset_size: 2619289
---
# Dataset Card for sick
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** http://marcobaroni.org/composes/sick.html
- **Repository:** [Needs More Information]
- **Paper:** https://www.aclweb.org/anthology/L14-1314/
- **Leaderboard:** [Needs More Information]
- **Point of Contact:** [Needs More Information]
### Dataset Summary
Shared and internationally recognized benchmarks are fundamental for the development of any computational system. We aim to help the research community working on compositional distributional semantic models (CDSMs) by providing SICK (Sentences Involving Compositional Knowldedge), a large size English benchmark tailored for them. SICK consists of about 10,000 English sentence pairs that include many examples of the lexical, syntactic and semantic phenomena that CDSMs are expected to account for, but do not require dealing with other aspects of existing sentential data sets (idiomatic multiword expressions, named entities, telegraphic language) that are not within the scope of CDSMs. By means of crowdsourcing techniques, each pair was annotated for two crucial semantic tasks: relatedness in meaning (with a 5-point rating scale as gold score) and entailment relation between the two elements (with three possible gold labels: entailment, contradiction, and neutral). The SICK data set was used in SemEval-2014 Task 1, and it freely available for research purposes.
### Supported Tasks and Leaderboards
[Needs More Information]
### Languages
The dataset is in English.
## Dataset Structure
### Data Instances
Example instance:
```
{
"entailment_AB": "A_neutral_B",
"entailment_BA": "B_neutral_A",
"label": 1,
"id": "1",
"relatedness_score": 4.5,
"sentence_A": "A group of kids is playing in a yard and an old man is standing in the background",
"sentence_A_dataset": "FLICKR",
"sentence_A_original": "A group of children playing in a yard, a man in the background.",
"sentence_B": "A group of boys in a yard is playing and a man is standing in the background",
"sentence_B_dataset": "FLICKR",
"sentence_B_original": "A group of children playing in a yard, a man in the background."
}
```
### Data Fields
- pair_ID: sentence pair ID
- sentence_A: sentence A
- sentence_B: sentence B
- label: textual entailment gold label: entailment (0), neutral (1) or contradiction (2)
- relatedness_score: semantic relatedness gold score (on a 1-5 continuous scale)
- entailment_AB: entailment for the A-B order (A_neutral_B, A_entails_B, or A_contradicts_B)
- entailment_BA: entailment for the B-A order (B_neutral_A, B_entails_A, or B_contradicts_A)
- sentence_A_original: original sentence from which sentence A is derived
- sentence_B_original: original sentence from which sentence B is derived
- sentence_A_dataset: dataset from which the original sentence A was extracted (FLICKR vs. SEMEVAL)
- sentence_B_dataset: dataset from which the original sentence B was extracted (FLICKR vs. SEMEVAL)
### Data Splits
Train Trial Test
4439 495 4906
## Dataset Creation
### Curation Rationale
[Needs More Information]
### Source Data
#### Initial Data Collection and Normalization
[Needs More Information]
#### Who are the source language producers?
[Needs More Information]
### Annotations
#### Annotation process
[Needs More Information]
#### Who are the annotators?
[Needs More Information]
### Personal and Sensitive Information
[Needs More Information]
## Considerations for Using the Data
### Social Impact of Dataset
[Needs More Information]
### Discussion of Biases
[Needs More Information]
### Other Known Limitations
[Needs More Information]
## Additional Information
### Dataset Curators
[Needs More Information]
### Licensing Information
[Needs More Information]
### Citation Information
```
@inproceedings{marelli-etal-2014-sick,
title = "A {SICK} cure for the evaluation of compositional distributional semantic models",
author = "Marelli, Marco and
Menini, Stefano and
Baroni, Marco and
Bentivogli, Luisa and
Bernardi, Raffaella and
Zamparelli, Roberto",
booktitle = "Proceedings of the Ninth International Conference on Language Resources and Evaluation ({LREC}'14)",
month = may,
year = "2014",
address = "Reykjavik, Iceland",
publisher = "European Language Resources Association (ELRA)",
url = "http://www.lrec-conf.org/proceedings/lrec2014/pdf/363_Paper.pdf",
pages = "216--223",
}
```
### Contributions
Thanks to [@calpt](https://github.com/calpt) for adding this dataset. |
llm-lens/vocab_tags | 2023-06-29T02:50:09.000Z | [
"region:us"
] | llm-lens | null | null | null | 1 | 6,697 | ---
dataset_info:
features:
- name: prompt_descriptions
dtype: string
splits:
- name: train
num_bytes: 346971
num_examples: 22131
download_size: 298971
dataset_size: 346971
---
# Dataset Card for "vocab_tags"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
dair-ai/emotion | 2023-04-20T08:08:15.000Z | [
"task_categories:text-classification",
"task_ids:multi-class-classification",
"annotations_creators:machine-generated",
"language_creators:machine-generated",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:other",
"emotion-classification",
"region:us"
] | dair-ai | Emotion is a dataset of English Twitter messages with six basic emotions: anger, fear, joy, love, sadness, and surprise. For more detailed information please refer to the paper. | @inproceedings{saravia-etal-2018-carer,
title = "{CARER}: Contextualized Affect Representations for Emotion Recognition",
author = "Saravia, Elvis and
Liu, Hsien-Chi Toby and
Huang, Yen-Hao and
Wu, Junlin and
Chen, Yi-Shin",
booktitle = "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing",
month = oct # "-" # nov,
year = "2018",
address = "Brussels, Belgium",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/D18-1404",
doi = "10.18653/v1/D18-1404",
pages = "3687--3697",
abstract = "Emotions are expressed in nuanced ways, which varies by collective or individual experiences, knowledge, and beliefs. Therefore, to understand emotion, as conveyed through text, a robust mechanism capable of capturing and modeling different linguistic nuances and phenomena is needed. We propose a semi-supervised, graph-based algorithm to produce rich structural descriptors which serve as the building blocks for constructing contextualized affect representations from text. The pattern-based representations are further enriched with word embeddings and evaluated through several emotion recognition tasks. Our experimental results demonstrate that the proposed method outperforms state-of-the-art techniques on emotion recognition tasks.",
} | null | 127 | 6,650 | ---
annotations_creators:
- machine-generated
language_creators:
- machine-generated
language:
- en
license:
- other
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- multi-class-classification
paperswithcode_id: emotion
pretty_name: Emotion
tags:
- emotion-classification
dataset_info:
- config_name: split
features:
- name: text
dtype: string
- name: label
dtype:
class_label:
names:
'0': sadness
'1': joy
'2': love
'3': anger
'4': fear
'5': surprise
splits:
- name: train
num_bytes: 1741597
num_examples: 16000
- name: validation
num_bytes: 214703
num_examples: 2000
- name: test
num_bytes: 217181
num_examples: 2000
download_size: 740883
dataset_size: 2173481
- config_name: unsplit
features:
- name: text
dtype: string
- name: label
dtype:
class_label:
names:
'0': sadness
'1': joy
'2': love
'3': anger
'4': fear
'5': surprise
splits:
- name: train
num_bytes: 45445685
num_examples: 416809
download_size: 15388281
dataset_size: 45445685
train-eval-index:
- config: default
task: text-classification
task_id: multi_class_classification
splits:
train_split: train
eval_split: test
col_mapping:
text: text
label: target
metrics:
- type: accuracy
name: Accuracy
- type: f1
name: F1 macro
args:
average: macro
- type: f1
name: F1 micro
args:
average: micro
- type: f1
name: F1 weighted
args:
average: weighted
- type: precision
name: Precision macro
args:
average: macro
- type: precision
name: Precision micro
args:
average: micro
- type: precision
name: Precision weighted
args:
average: weighted
- type: recall
name: Recall macro
args:
average: macro
- type: recall
name: Recall micro
args:
average: micro
- type: recall
name: Recall weighted
args:
average: weighted
---
# Dataset Card for "emotion"
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://github.com/dair-ai/emotion_dataset](https://github.com/dair-ai/emotion_dataset)
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 16.13 MB
- **Size of the generated dataset:** 47.62 MB
- **Total amount of disk used:** 63.75 MB
### Dataset Summary
Emotion is a dataset of English Twitter messages with six basic emotions: anger, fear, joy, love, sadness, and surprise. For more detailed information please refer to the paper.
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Dataset Structure
### Data Instances
An example looks as follows.
```
{
"text": "im feeling quite sad and sorry for myself but ill snap out of it soon",
"label": 0
}
```
### Data Fields
The data fields are:
- `text`: a `string` feature.
- `label`: a classification label, with possible values including `sadness` (0), `joy` (1), `love` (2), `anger` (3), `fear` (4), `surprise` (5).
### Data Splits
The dataset has 2 configurations:
- split: with a total of 20_000 examples split into train, validation and split
- unsplit: with a total of 416_809 examples in a single train split
| name | train | validation | test |
|---------|-------:|-----------:|-----:|
| split | 16000 | 2000 | 2000 |
| unsplit | 416809 | n/a | n/a |
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
The dataset should be used for educational and research purposes only.
### Citation Information
If you use this dataset, please cite:
```
@inproceedings{saravia-etal-2018-carer,
title = "{CARER}: Contextualized Affect Representations for Emotion Recognition",
author = "Saravia, Elvis and
Liu, Hsien-Chi Toby and
Huang, Yen-Hao and
Wu, Junlin and
Chen, Yi-Shin",
booktitle = "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing",
month = oct # "-" # nov,
year = "2018",
address = "Brussels, Belgium",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/D18-1404",
doi = "10.18653/v1/D18-1404",
pages = "3687--3697",
abstract = "Emotions are expressed in nuanced ways, which varies by collective or individual experiences, knowledge, and beliefs. Therefore, to understand emotion, as conveyed through text, a robust mechanism capable of capturing and modeling different linguistic nuances and phenomena is needed. We propose a semi-supervised, graph-based algorithm to produce rich structural descriptors which serve as the building blocks for constructing contextualized affect representations from text. The pattern-based representations are further enriched with word embeddings and evaluated through several emotion recognition tasks. Our experimental results demonstrate that the proposed method outperforms state-of-the-art techniques on emotion recognition tasks.",
}
```
### Contributions
Thanks to [@lhoestq](https://github.com/lhoestq), [@thomwolf](https://github.com/thomwolf), [@lewtun](https://github.com/lewtun) for adding this dataset.
|
enwik8 | 2023-04-06T14:14:17.000Z | [
"task_categories:fill-mask",
"task_categories:text-generation",
"task_ids:language-modeling",
"task_ids:masked-language-modeling",
"annotations_creators:no-annotation",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:mit",
"region:us"
] | null | The dataset is based on the Hutter Prize (http://prize.hutter1.net) and contains the first 10^8 bytes of English Wikipedia in 2006 in XML | null | null | 4 | 6,627 | ---
annotations_creators:
- no-annotation
language_creators:
- found
language:
- en
license:
- mit
multilinguality:
- monolingual
pretty_name: enwik8
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- fill-mask
- text-generation
task_ids:
- language-modeling
- masked-language-modeling
dataset_info:
- config_name: enwik8
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 104299244
num_examples: 1128024
download_size: 36445475
dataset_size: 102383126
- config_name: enwik8-raw
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 100000008
num_examples: 1
download_size: 36445475
dataset_size: 100000008
---
# Dataset Card for enwik8
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-instances)
- [Data Splits](#data-instances)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
## Dataset Description
- **Homepage:** http://mattmahoney.net/dc/textdata.html
- **Repository:** [Needs More Information]
- **Paper:** [Needs More Information]
- **Leaderboard:** https://paperswithcode.com/sota/language-modelling-on-enwiki8
- **Point of Contact:** [Needs More Information]
- **Size of downloaded dataset files:** 36.45 MB
- **Size of the generated dataset:** 102.38 MB
- **Total amount of disk used:** 138.83 MB
### Dataset Summary
The enwik8 dataset is the first 100,000,000 (100M) bytes of the English Wikipedia XML dump on Mar. 3, 2006 and is typically used to measure a model's ability to compress data.
### Supported Tasks and Leaderboards
A leaderboard for byte-level causal language modelling can be found on [paperswithcode](https://paperswithcode.com/sota/language-modelling-on-enwiki8)
### Languages
en
## Dataset Structure
### Data Instances
- **Size of downloaded dataset files:** 36.45 MB
- **Size of the generated dataset:** 102.38 MB
- **Total amount of disk used:** 138.83 MB
```
{
"text": "In [[Denmark]], the [[Freetown Christiania]] was created in downtown [[Copenhagen]]....",
}
```
### Data Fields
The data fields are the same among all sets.
#### enwik8
- `text`: a `string` feature.
#### enwik8-raw
- `text`: a `string` feature.
### Data Splits
| dataset | train |
| --- | --- |
| enwik8 | 1128024 |
| enwik8- raw | 1 |
## Dataset Creation
### Curation Rationale
[Needs More Information]
### Source Data
#### Initial Data Collection and Normalization
The data is just English Wikipedia XML dump on Mar. 3, 2006 split by line for enwik8 and not split by line for enwik8-raw.
#### Who are the source language producers?
[Needs More Information]
### Annotations
#### Annotation process
[Needs More Information]
#### Who are the annotators?
[Needs More Information]
### Personal and Sensitive Information
[Needs More Information]
## Considerations for Using the Data
### Social Impact of Dataset
[Needs More Information]
### Discussion of Biases
[Needs More Information]
### Other Known Limitations
[Needs More Information]
## Additional Information
### Dataset Curators
[Needs More Information]
### Licensing Information
[Needs More Information]
### Citation Information
Dataset is not part of a publication, and can therefore not be cited.
### Contributions
Thanks to [@HallerPatrick](https://github.com/HallerPatrick) for adding this dataset and [@mtanghu](https://github.com/mtanghu) for updating it. |
oscar | 2023-06-01T14:59:59.000Z | [
"task_categories:text-generation",
"task_categories:fill-mask",
"task_ids:language-modeling",
"task_ids:masked-language-modeling",
"annotations_creators:no-annotation",
"language_creators:found",
"multilinguality:multilingual",
"size_categories:100K<n<1M",
"size_categories:100M<n<1B",
"size_categories:10K<n<100K",
"size_categories:10M<n<100M",
"size_categories:1K<n<10K",
"size_categories:1M<n<10M",
"size_categories:n<1K",
"source_datasets:original",
"language:af",
"language:als",
"language:am",
"language:an",
"language:ar",
"language:arz",
"language:as",
"language:ast",
"language:av",
"language:az",
"language:azb",
"language:ba",
"language:bar",
"language:bcl",
"language:be",
"language:bg",
"language:bh",
"language:bn",
"language:bo",
"language:bpy",
"language:br",
"language:bs",
"language:bxr",
"language:ca",
"language:cbk",
"language:ce",
"language:ceb",
"language:ckb",
"language:cs",
"language:cv",
"language:cy",
"language:da",
"language:de",
"language:diq",
"language:dsb",
"language:dv",
"language:el",
"language:eml",
"language:en",
"language:eo",
"language:es",
"language:et",
"language:eu",
"language:fa",
"language:fi",
"language:fr",
"language:frr",
"language:fy",
"language:ga",
"language:gd",
"language:gl",
"language:gn",
"language:gom",
"language:gu",
"language:he",
"language:hi",
"language:hr",
"language:hsb",
"language:ht",
"language:hu",
"language:hy",
"language:ia",
"language:id",
"language:ie",
"language:ilo",
"language:io",
"language:is",
"language:it",
"language:ja",
"language:jbo",
"language:jv",
"language:ka",
"language:kk",
"language:km",
"language:kn",
"language:ko",
"language:krc",
"language:ku",
"language:kv",
"language:kw",
"language:ky",
"language:la",
"language:lb",
"language:lez",
"language:li",
"language:lmo",
"language:lo",
"language:lrc",
"language:lt",
"language:lv",
"language:mai",
"language:mg",
"language:mhr",
"language:min",
"language:mk",
"language:ml",
"language:mn",
"language:mr",
"language:mrj",
"language:ms",
"language:mt",
"language:mwl",
"language:my",
"language:myv",
"language:mzn",
"language:nah",
"language:nap",
"language:nds",
"language:ne",
"language:new",
"language:nl",
"language:nn",
"language:no",
"language:oc",
"language:or",
"language:os",
"language:pa",
"language:pam",
"language:pl",
"language:pms",
"language:pnb",
"language:ps",
"language:pt",
"language:qu",
"language:rm",
"language:ro",
"language:ru",
"language:sa",
"language:sah",
"language:scn",
"language:sd",
"language:sh",
"language:si",
"language:sk",
"language:sl",
"language:so",
"language:sq",
"language:sr",
"language:su",
"language:sv",
"language:sw",
"language:ta",
"language:te",
"language:tg",
"language:th",
"language:tk",
"language:tl",
"language:tr",
"language:tt",
"language:tyv",
"language:ug",
"language:uk",
"language:ur",
"language:uz",
"language:vec",
"language:vi",
"language:vo",
"language:wa",
"language:war",
"language:wuu",
"language:xal",
"language:xmf",
"language:yi",
"language:yo",
"language:yue",
"language:zh",
"license:cc0-1.0",
"arxiv:2010.14571",
"region:us"
] | null | The Open Super-large Crawled ALMAnaCH coRpus is a huge multilingual corpus obtained by language classification and filtering of the Common Crawl corpus using the goclassy architecture.\ | @inproceedings{ortiz-suarez-etal-2020-monolingual,
title = "A Monolingual Approach to Contextualized Word Embeddings for Mid-Resource Languages",
author = "Ortiz Su{\'a}rez, Pedro Javier and
Romary, Laurent and
Sagot, Benoit",
booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics",
month = jul,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.acl-main.156",
pages = "1703--1714",
abstract = "We use the multilingual OSCAR corpus, extracted from Common Crawl via language classification, filtering and cleaning, to train monolingual contextualized word embeddings (ELMo) for five mid-resource languages. We then compare the performance of OSCAR-based and Wikipedia-based ELMo embeddings for these languages on the part-of-speech tagging and parsing tasks. We show that, despite the noise in the Common-Crawl-based OSCAR data, embeddings trained on OSCAR perform much better than monolingual embeddings trained on Wikipedia. They actually equal or improve the current state of the art in tagging and parsing for all five languages. In particular, they also improve over multilingual Wikipedia-based contextual embeddings (multilingual BERT), which almost always constitutes the previous state of the art, thereby showing that the benefit of a larger, more diverse corpus surpasses the cross-lingual benefit of multilingual embedding architectures.",
}
@inproceedings{OrtizSuarezSagotRomary2019,
author = {Pedro Javier {Ortiz Su{\'a}rez} and Benoit Sagot and Laurent Romary},
title = {Asynchronous pipelines for processing huge corpora on medium to low resource infrastructures},
series = {Proceedings of the Workshop on Challenges in the Management of Large Corpora (CMLC-7) 2019. Cardiff, 22nd July 2019},
editor = {Piotr Bański and Adrien Barbaresi and Hanno Biber and Evelyn Breiteneder and Simon Clematide and Marc Kupietz and Harald L{\"u}ngen and Caroline Iliadi},
publisher = {Leibniz-Institut f{\"u}r Deutsche Sprache},
address = {Mannheim},
doi = {10.14618/ids-pub-9021},
url = {http://nbn-resolving.de/urn:nbn:de:bsz:mh39-90215},
pages = {9 -- 16},
year = {2019},
abstract = {Common Crawl is a considerably large, heterogeneous multilingual corpus comprised of crawled documents from the internet, surpassing 20TB of data and distributed as a set of more than 50 thousand plain text files where each contains many documents written in a wide variety of languages. Even though each document has a metadata block associated to it, this data lacks any information about the language in which each document is written, making it extremely difficult to use Common Crawl for monolingual applications. We propose a general, highly parallel, multithreaded pipeline to clean and classify Common Crawl by language; we specifically design it so that it runs efficiently on medium to low resource infrastructures where I/O speeds are the main constraint. We develop the pipeline so that it can be easily reapplied to any kind of heterogeneous corpus and so that it can be parameterised to a wide range of infrastructures. We also distribute a 6.3TB version of Common Crawl, filtered, classified by language, shuffled at line level in order to avoid copyright issues, and ready to be used for NLP applications.},
language = {en}
} | null | 122 | 6,620 | ---
pretty_name: OSCAR
annotations_creators:
- no-annotation
language_creators:
- found
language:
- af
- als
- am
- an
- ar
- arz
- as
- ast
- av
- az
- azb
- ba
- bar
- bcl
- be
- bg
- bh
- bn
- bo
- bpy
- br
- bs
- bxr
- ca
- cbk
- ce
- ceb
- ckb
- cs
- cv
- cy
- da
- de
- diq
- dsb
- dv
- el
- eml
- en
- eo
- es
- et
- eu
- fa
- fi
- fr
- frr
- fy
- ga
- gd
- gl
- gn
- gom
- gu
- he
- hi
- hr
- hsb
- ht
- hu
- hy
- ia
- id
- ie
- ilo
- io
- is
- it
- ja
- jbo
- jv
- ka
- kk
- km
- kn
- ko
- krc
- ku
- kv
- kw
- ky
- la
- lb
- lez
- li
- lmo
- lo
- lrc
- lt
- lv
- mai
- mg
- mhr
- min
- mk
- ml
- mn
- mr
- mrj
- ms
- mt
- mwl
- my
- myv
- mzn
- nah
- nap
- nds
- ne
- new
- nl
- nn
- 'no'
- oc
- or
- os
- pa
- pam
- pl
- pms
- pnb
- ps
- pt
- qu
- rm
- ro
- ru
- sa
- sah
- scn
- sd
- sh
- si
- sk
- sl
- so
- sq
- sr
- su
- sv
- sw
- ta
- te
- tg
- th
- tk
- tl
- tr
- tt
- tyv
- ug
- uk
- ur
- uz
- vec
- vi
- vo
- wa
- war
- wuu
- xal
- xmf
- yi
- yo
- yue
- zh
license:
- cc0-1.0
multilinguality:
- multilingual
size_categories:
- 100K<n<1M
- 100M<n<1B
- 10K<n<100K
- 10M<n<100M
- 1K<n<10K
- 1M<n<10M
- n<1K
source_datasets:
- original
task_categories:
- text-generation
- fill-mask
task_ids:
- language-modeling
- masked-language-modeling
paperswithcode_id: oscar
dataset_info:
- config_name: unshuffled_deduplicated_af
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 171320914
num_examples: 130640
download_size: 65989254
dataset_size: 171320914
- config_name: unshuffled_deduplicated_als
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2915912
num_examples: 4518
download_size: 1263294
dataset_size: 2915912
- config_name: unshuffled_deduplicated_arz
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 34893248
num_examples: 79928
download_size: 10027493
dataset_size: 34893248
- config_name: unshuffled_deduplicated_an
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 842246
num_examples: 2025
download_size: 133373
dataset_size: 842246
- config_name: unshuffled_deduplicated_ast
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2150022
num_examples: 5343
download_size: 856177
dataset_size: 2150022
- config_name: unshuffled_deduplicated_ba
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 93623739
num_examples: 27050
download_size: 25983491
dataset_size: 93623739
- config_name: unshuffled_deduplicated_am
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 215618603
num_examples: 43102
download_size: 61347279
dataset_size: 215618603
- config_name: unshuffled_deduplicated_as
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 73989818
num_examples: 9212
download_size: 15513004
dataset_size: 73989818
- config_name: unshuffled_deduplicated_azb
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 20001183
num_examples: 9985
download_size: 5191704
dataset_size: 20001183
- config_name: unshuffled_deduplicated_be
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1077152244
num_examples: 307405
download_size: 306700943
dataset_size: 1077152244
- config_name: unshuffled_deduplicated_bo
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 144506264
num_examples: 15762
download_size: 22365048
dataset_size: 144506264
- config_name: unshuffled_deduplicated_bxr
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 11325
num_examples: 36
download_size: 3666
dataset_size: 11325
- config_name: unshuffled_deduplicated_ceb
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 24439249
num_examples: 26145
download_size: 7124786
dataset_size: 24439249
- config_name: unshuffled_deduplicated_az
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1526935070
num_examples: 626796
download_size: 521744076
dataset_size: 1526935070
- config_name: unshuffled_deduplicated_bcl
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 900
num_examples: 1
download_size: 594
dataset_size: 900
- config_name: unshuffled_deduplicated_cy
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 140412555
num_examples: 98225
download_size: 53629697
dataset_size: 140412555
- config_name: unshuffled_deduplicated_dsb
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 7589
num_examples: 37
download_size: 3640
dataset_size: 7589
- config_name: unshuffled_deduplicated_bn
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 6233041155
num_examples: 1114481
download_size: 1257218381
dataset_size: 6233041155
- config_name: unshuffled_deduplicated_bs
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 125977
num_examples: 702
download_size: 38669
dataset_size: 125977
- config_name: unshuffled_deduplicated_ce
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 7021674
num_examples: 2984
download_size: 1862792
dataset_size: 7021674
- config_name: unshuffled_deduplicated_cv
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 27359554
num_examples: 10130
download_size: 7461982
dataset_size: 27359554
- config_name: unshuffled_deduplicated_diq
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 161
num_examples: 1
download_size: 331
dataset_size: 161
- config_name: unshuffled_deduplicated_eml
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 24657
num_examples: 80
download_size: 10055
dataset_size: 24657
- config_name: unshuffled_deduplicated_et
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2434152666
num_examples: 1172041
download_size: 966785545
dataset_size: 2434152666
- config_name: unshuffled_deduplicated_bg
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 14420684170
num_examples: 3398679
download_size: 3848659853
dataset_size: 14420684170
- config_name: unshuffled_deduplicated_bpy
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1725535
num_examples: 1770
download_size: 191472
dataset_size: 1725535
- config_name: unshuffled_deduplicated_ca
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 4544123629
num_examples: 2458067
download_size: 1734548117
dataset_size: 4544123629
- config_name: unshuffled_deduplicated_ckb
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 237229156
num_examples: 68210
download_size: 60319928
dataset_size: 237229156
- config_name: unshuffled_deduplicated_ar
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 33468271639
num_examples: 9006977
download_size: 9667185012
dataset_size: 33468271639
- config_name: unshuffled_deduplicated_av
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 334755
num_examples: 360
download_size: 75341
dataset_size: 334755
- config_name: unshuffled_deduplicated_bar
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 551
num_examples: 4
download_size: 354
dataset_size: 551
- config_name: unshuffled_deduplicated_bh
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 35216
num_examples: 82
download_size: 6003
dataset_size: 35216
- config_name: unshuffled_deduplicated_br
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 16712284
num_examples: 14724
download_size: 6468062
dataset_size: 16712284
- config_name: unshuffled_deduplicated_cbk
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 535
num_examples: 1
download_size: 247
dataset_size: 535
- config_name: unshuffled_deduplicated_da
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 10204168604
num_examples: 4771098
download_size: 3816376656
dataset_size: 10204168604
- config_name: unshuffled_deduplicated_dv
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 82122241
num_examples: 17024
download_size: 16836170
dataset_size: 82122241
- config_name: unshuffled_deduplicated_eo
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 239597935
num_examples: 84752
download_size: 92858714
dataset_size: 239597935
- config_name: unshuffled_deduplicated_fa
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 39986583410
num_examples: 8203495
download_size: 10459318520
dataset_size: 39986583410
- config_name: unshuffled_deduplicated_fy
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 26562554
num_examples: 20661
download_size: 10270434
dataset_size: 26562554
- config_name: unshuffled_deduplicated_gn
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 24545
num_examples: 68
download_size: 9566
dataset_size: 24545
- config_name: unshuffled_deduplicated_cs
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 25590158564
num_examples: 12308039
download_size: 10494256383
dataset_size: 25590158564
- config_name: unshuffled_deduplicated_hi
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 9550345517
num_examples: 1909387
download_size: 2007441283
dataset_size: 9550345517
- config_name: unshuffled_deduplicated_hu
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 19027456462
num_examples: 6582908
download_size: 7368098962
dataset_size: 19027456462
- config_name: unshuffled_deduplicated_ie
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1688
num_examples: 11
download_size: 649
dataset_size: 1688
- config_name: unshuffled_deduplicated_fr
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 147774253219
num_examples: 59448891
download_size: 55462770729
dataset_size: 147774253219
- config_name: unshuffled_deduplicated_gd
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1339050
num_examples: 3883
download_size: 420601
dataset_size: 1339050
- config_name: unshuffled_deduplicated_gu
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 758319353
num_examples: 169834
download_size: 162974870
dataset_size: 758319353
- config_name: unshuffled_deduplicated_hsb
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1821734
num_examples: 3084
download_size: 728158
dataset_size: 1821734
- config_name: unshuffled_deduplicated_ia
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 373710
num_examples: 529
download_size: 52722
dataset_size: 373710
- config_name: unshuffled_deduplicated_io
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 139493
num_examples: 617
download_size: 42813
dataset_size: 139493
- config_name: unshuffled_deduplicated_jbo
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 700428
num_examples: 617
download_size: 203506
dataset_size: 700428
- config_name: unshuffled_deduplicated_km
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 609886370
num_examples: 108346
download_size: 114480044
dataset_size: 609886370
- config_name: unshuffled_deduplicated_ku
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 62855449
num_examples: 29054
download_size: 23343869
dataset_size: 62855449
- config_name: unshuffled_deduplicated_la
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 8867995
num_examples: 18808
download_size: 3421499
dataset_size: 8867995
- config_name: unshuffled_deduplicated_lmo
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 458386
num_examples: 1374
download_size: 106048
dataset_size: 458386
- config_name: unshuffled_deduplicated_lv
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1895693807
num_examples: 843195
download_size: 710448932
dataset_size: 1895693807
- config_name: unshuffled_deduplicated_min
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 318749
num_examples: 166
download_size: 10233
dataset_size: 318749
- config_name: unshuffled_deduplicated_mr
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1487944837
num_examples: 212556
download_size: 299680349
dataset_size: 1487944837
- config_name: unshuffled_deduplicated_mwl
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1121
num_examples: 7
download_size: 797
dataset_size: 1121
- config_name: unshuffled_deduplicated_nah
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 11540
num_examples: 58
download_size: 2868
dataset_size: 11540
- config_name: unshuffled_deduplicated_new
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 4226557
num_examples: 2126
download_size: 830767
dataset_size: 4226557
- config_name: unshuffled_deduplicated_oc
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 3938772
num_examples: 6485
download_size: 1338194
dataset_size: 3938772
- config_name: unshuffled_deduplicated_pam
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 319
num_examples: 1
download_size: 366
dataset_size: 319
- config_name: unshuffled_deduplicated_ps
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 254360032
num_examples: 67921
download_size: 71823163
dataset_size: 254360032
- config_name: unshuffled_deduplicated_it
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 73843292670
num_examples: 28522082
download_size: 27931571784
dataset_size: 73843292670
- config_name: unshuffled_deduplicated_ka
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1982841952
num_examples: 372158
download_size: 377220437
dataset_size: 1982841952
- config_name: unshuffled_deduplicated_ro
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 11601264185
num_examples: 5044757
download_size: 4478423935
dataset_size: 11601264185
- config_name: unshuffled_deduplicated_scn
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2990
num_examples: 17
download_size: 1620
dataset_size: 2990
- config_name: unshuffled_deduplicated_ko
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 11956006533
num_examples: 3675420
download_size: 4462788278
dataset_size: 11956006533
- config_name: unshuffled_deduplicated_kw
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 14971
num_examples: 68
download_size: 6195
dataset_size: 14971
- config_name: unshuffled_deduplicated_lez
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 3075326
num_examples: 1381
download_size: 763936
dataset_size: 3075326
- config_name: unshuffled_deduplicated_lrc
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 65291
num_examples: 72
download_size: 16272
dataset_size: 65291
- config_name: unshuffled_deduplicated_mg
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 13516085
num_examples: 13343
download_size: 4303472
dataset_size: 13516085
- config_name: unshuffled_deduplicated_ml
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2685637627
num_examples: 453904
download_size: 496801596
dataset_size: 2685637627
- config_name: unshuffled_deduplicated_ms
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 45064684
num_examples: 183443
download_size: 16391407
dataset_size: 45064684
- config_name: unshuffled_deduplicated_myv
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1224
num_examples: 5
download_size: 705
dataset_size: 1224
- config_name: unshuffled_deduplicated_nds
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 13360483
num_examples: 8714
download_size: 5271194
dataset_size: 13360483
- config_name: unshuffled_deduplicated_nn
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 57286159
num_examples: 109118
download_size: 23583774
dataset_size: 57286159
- config_name: unshuffled_deduplicated_os
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 10962689
num_examples: 2559
download_size: 2829131
dataset_size: 10962689
- config_name: unshuffled_deduplicated_pms
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1996853
num_examples: 2859
download_size: 716837
dataset_size: 1996853
- config_name: unshuffled_deduplicated_qu
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 72587
num_examples: 411
download_size: 17501
dataset_size: 72587
- config_name: unshuffled_deduplicated_sa
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 38236039
num_examples: 7121
download_size: 7268337
dataset_size: 38236039
- config_name: unshuffled_deduplicated_sk
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 4768416160
num_examples: 2820821
download_size: 1960409934
dataset_size: 4768416160
- config_name: unshuffled_deduplicated_sh
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 6184582
num_examples: 17610
download_size: 1445894
dataset_size: 6184582
- config_name: unshuffled_deduplicated_so
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 16269
num_examples: 42
download_size: 2109
dataset_size: 16269
- config_name: unshuffled_deduplicated_sr
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2358255234
num_examples: 645747
download_size: 665025000
dataset_size: 2358255234
- config_name: unshuffled_deduplicated_ta
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 5477003981
num_examples: 833101
download_size: 971118176
dataset_size: 5477003981
- config_name: unshuffled_deduplicated_tk
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 7092199
num_examples: 4694
download_size: 2219582
dataset_size: 7092199
- config_name: unshuffled_deduplicated_tyv
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 8319
num_examples: 24
download_size: 2976
dataset_size: 8319
- config_name: unshuffled_deduplicated_uz
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 11834927
num_examples: 15074
download_size: 4300299
dataset_size: 11834927
- config_name: unshuffled_deduplicated_wa
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 214337
num_examples: 677
download_size: 79130
dataset_size: 214337
- config_name: unshuffled_deduplicated_xmf
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 4617445
num_examples: 2418
download_size: 943151
dataset_size: 4617445
- config_name: unshuffled_deduplicated_sv
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 26239415574
num_examples: 11014487
download_size: 10185393483
dataset_size: 26239415574
- config_name: unshuffled_deduplicated_tg
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 261233997
num_examples: 56259
download_size: 62908723
dataset_size: 261233997
- config_name: unshuffled_deduplicated_de
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 155723559907
num_examples: 62398034
download_size: 60797849113
dataset_size: 155723559907
- config_name: unshuffled_deduplicated_tr
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 28375018927
num_examples: 11596446
download_size: 10390754678
dataset_size: 28375018927
- config_name: unshuffled_deduplicated_el
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 28689398676
num_examples: 6521169
download_size: 7907952068
dataset_size: 28689398676
- config_name: unshuffled_deduplicated_uk
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 29791312367
num_examples: 7782375
download_size: 8037737457
dataset_size: 29791312367
- config_name: unshuffled_deduplicated_vi
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 33528331774
num_examples: 9897709
download_size: 10711506712
dataset_size: 33528331774
- config_name: unshuffled_deduplicated_wuu
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 33253
num_examples: 64
download_size: 7273
dataset_size: 33253
- config_name: unshuffled_deduplicated_yo
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 27169
num_examples: 49
download_size: 8925
dataset_size: 27169
- config_name: unshuffled_original_als
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 5297910
num_examples: 7324
download_size: 1489734
dataset_size: 5297910
- config_name: unshuffled_original_arz
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 70132423
num_examples: 158113
download_size: 15891255
dataset_size: 70132423
- config_name: unshuffled_original_az
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2964781192
num_examples: 912330
download_size: 927763846
dataset_size: 2964781192
- config_name: unshuffled_original_bcl
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 901
num_examples: 1
download_size: 581
dataset_size: 901
- config_name: unshuffled_original_bn
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 10771945233
num_examples: 1675515
download_size: 2139944099
dataset_size: 10771945233
- config_name: unshuffled_original_bs
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 482740
num_examples: 2143
download_size: 56419
dataset_size: 482740
- config_name: unshuffled_original_ce
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 8735740
num_examples: 4042
download_size: 2089184
dataset_size: 8735740
- config_name: unshuffled_original_cv
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 41047029
num_examples: 20281
download_size: 9400068
dataset_size: 41047029
- config_name: unshuffled_original_diq
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 162
num_examples: 1
download_size: 318
dataset_size: 162
- config_name: unshuffled_original_eml
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 26099
num_examples: 84
download_size: 10071
dataset_size: 26099
- config_name: unshuffled_original_et
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 5174800705
num_examples: 2093621
download_size: 1881328631
dataset_size: 5174800705
- config_name: unshuffled_deduplicated_zh
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 267614324325
num_examples: 41708901
download_size: 99982781539
dataset_size: 267614324325
- config_name: unshuffled_original_an
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1329433
num_examples: 2449
download_size: 148184
dataset_size: 1329433
- config_name: unshuffled_original_ast
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2539238
num_examples: 6999
download_size: 920730
dataset_size: 2539238
- config_name: unshuffled_original_ba
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 133704014
num_examples: 42551
download_size: 33215002
dataset_size: 133704014
- config_name: unshuffled_original_bg
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 33753811450
num_examples: 5869686
download_size: 8336964541
dataset_size: 33753811450
- config_name: unshuffled_original_bpy
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 4347467
num_examples: 6046
download_size: 336974
dataset_size: 4347467
- config_name: unshuffled_original_ca
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 8623251470
num_examples: 4390754
download_size: 3101954304
dataset_size: 8623251470
- config_name: unshuffled_original_ckb
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 510965919
num_examples: 103639
download_size: 111884006
dataset_size: 510965919
- config_name: unshuffled_deduplicated_es
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 160418075023
num_examples: 56326016
download_size: 60464970319
dataset_size: 160418075023
- config_name: unshuffled_original_da
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 16756455589
num_examples: 7664010
download_size: 6000579388
dataset_size: 16756455589
- config_name: unshuffled_original_dv
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 131628992
num_examples: 21018
download_size: 24914404
dataset_size: 131628992
- config_name: unshuffled_original_eo
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 314188336
num_examples: 121168
download_size: 117076019
dataset_size: 314188336
- config_name: unshuffled_deduplicated_fi
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 13945067515
num_examples: 5326443
download_size: 5380047103
dataset_size: 13945067515
- config_name: unshuffled_deduplicated_ga
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 63370688
num_examples: 46493
download_size: 22218633
dataset_size: 63370688
- config_name: unshuffled_deduplicated_gom
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1863089
num_examples: 484
download_size: 377051
dataset_size: 1863089
- config_name: unshuffled_deduplicated_hr
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 118047678
num_examples: 321484
download_size: 46731365
dataset_size: 118047678
- config_name: unshuffled_deduplicated_hy
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1559114836
num_examples: 396093
download_size: 393620208
dataset_size: 1559114836
- config_name: unshuffled_deduplicated_ilo
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 667896
num_examples: 1578
download_size: 230065
dataset_size: 667896
- config_name: unshuffled_original_fa
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 84209448803
num_examples: 13704702
download_size: 20956409096
dataset_size: 84209448803
- config_name: unshuffled_original_fy
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 36238452
num_examples: 33053
download_size: 12409774
dataset_size: 36238452
- config_name: unshuffled_original_gn
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 37427
num_examples: 106
download_size: 9761
dataset_size: 37427
- config_name: unshuffled_original_hi
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 17929286362
num_examples: 3264660
download_size: 3656636848
dataset_size: 17929286362
- config_name: unshuffled_original_hu
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 43074893842
num_examples: 11197780
download_size: 15693847091
dataset_size: 43074893842
- config_name: unshuffled_original_ie
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 25355
num_examples: 101
download_size: 783
dataset_size: 25355
- config_name: unshuffled_deduplicated_ja
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 113315056833
num_examples: 39496439
download_size: 40801218295
dataset_size: 113315056833
- config_name: unshuffled_deduplicated_kk
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1583064520
num_examples: 338073
download_size: 389111715
dataset_size: 1583064520
- config_name: unshuffled_deduplicated_krc
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2412731
num_examples: 1377
download_size: 615982
dataset_size: 2412731
- config_name: unshuffled_deduplicated_ky
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 407576051
num_examples: 86561
download_size: 106219565
dataset_size: 407576051
- config_name: unshuffled_deduplicated_li
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 28176
num_examples: 118
download_size: 11724
dataset_size: 28176
- config_name: unshuffled_deduplicated_lt
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 4185372402
num_examples: 1737411
download_size: 1653025558
dataset_size: 4185372402
- config_name: unshuffled_deduplicated_mhr
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 6247177
num_examples: 2515
download_size: 1622076
dataset_size: 6247177
- config_name: unshuffled_deduplicated_mn
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 880883961
num_examples: 197878
download_size: 219516471
dataset_size: 880883961
- config_name: unshuffled_deduplicated_mt
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 17539926
num_examples: 16383
download_size: 5898934
dataset_size: 17539926
- config_name: unshuffled_deduplicated_mzn
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 626534
num_examples: 917
download_size: 157541
dataset_size: 626534
- config_name: unshuffled_deduplicated_ne
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1239170286
num_examples: 219334
download_size: 240627361
dataset_size: 1239170286
- config_name: unshuffled_deduplicated_no
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 5077919278
num_examples: 3229940
download_size: 1960828800
dataset_size: 5077919278
- config_name: unshuffled_deduplicated_pa
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 482461302
num_examples: 87235
download_size: 102390579
dataset_size: 482461302
- config_name: unshuffled_deduplicated_pnb
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 9416915
num_examples: 3463
download_size: 2579976
dataset_size: 9416915
- config_name: unshuffled_deduplicated_rm
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 6932
num_examples: 34
download_size: 2679
dataset_size: 6932
- config_name: unshuffled_deduplicated_sah
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 27293316
num_examples: 8555
download_size: 7020207
dataset_size: 27293316
- config_name: unshuffled_deduplicated_si
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 841460012
num_examples: 120684
download_size: 175610997
dataset_size: 841460012
- config_name: unshuffled_deduplicated_sq
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1208425681
num_examples: 461598
download_size: 445358539
dataset_size: 1208425681
- config_name: unshuffled_deduplicated_sw
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 8747758
num_examples: 24803
download_size: 2946034
dataset_size: 8747758
- config_name: unshuffled_deduplicated_th
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 17082022564
num_examples: 3749826
download_size: 3536468931
dataset_size: 17082022564
- config_name: unshuffled_deduplicated_tt
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 320641922
num_examples: 82738
download_size: 85893621
dataset_size: 320641922
- config_name: unshuffled_deduplicated_ur
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1819253063
num_examples: 428674
download_size: 483593818
dataset_size: 1819253063
- config_name: unshuffled_deduplicated_vo
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2098461
num_examples: 3317
download_size: 301687
dataset_size: 2098461
- config_name: unshuffled_deduplicated_xal
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 114574
num_examples: 36
download_size: 31863
dataset_size: 114574
- config_name: unshuffled_deduplicated_yue
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2267
num_examples: 7
download_size: 646
dataset_size: 2267
- config_name: unshuffled_original_am
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 378060369
num_examples: 83663
download_size: 102789518
dataset_size: 378060369
- config_name: unshuffled_original_as
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 117733678
num_examples: 14985
download_size: 21437245
dataset_size: 117733678
- config_name: unshuffled_original_azb
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 28469069
num_examples: 15446
download_size: 6641415
dataset_size: 28469069
- config_name: unshuffled_original_be
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1877972506
num_examples: 586031
download_size: 498295673
dataset_size: 1877972506
- config_name: unshuffled_original_bo
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 195400209
num_examples: 26795
download_size: 28940995
dataset_size: 195400209
- config_name: unshuffled_original_bxr
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 13376
num_examples: 42
download_size: 3688
dataset_size: 13376
- config_name: unshuffled_original_ceb
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 40964537
num_examples: 56248
download_size: 11070392
dataset_size: 40964537
- config_name: unshuffled_original_cy
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 224933804
num_examples: 157698
download_size: 81736037
dataset_size: 224933804
- config_name: unshuffled_original_dsb
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 13761
num_examples: 65
download_size: 3753
dataset_size: 13761
- config_name: unshuffled_original_fr
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 303190338653
num_examples: 96742378
download_size: 105324330228
dataset_size: 303190338653
- config_name: unshuffled_original_gd
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2022000
num_examples: 5799
download_size: 525253
dataset_size: 2022000
- config_name: unshuffled_original_gu
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1094814909
num_examples: 240691
download_size: 232021129
dataset_size: 1094814909
- config_name: unshuffled_original_hsb
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 4482886
num_examples: 7959
download_size: 1389826
dataset_size: 4482886
- config_name: unshuffled_original_ia
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 689455
num_examples: 1040
download_size: 83325
dataset_size: 689455
- config_name: unshuffled_original_io
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 158808
num_examples: 694
download_size: 44548
dataset_size: 158808
- config_name: unshuffled_original_jbo
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 763027
num_examples: 832
download_size: 212962
dataset_size: 763027
- config_name: unshuffled_original_km
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1102616385
num_examples: 159363
download_size: 193286621
dataset_size: 1102616385
- config_name: unshuffled_original_ku
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 99062676
num_examples: 46535
download_size: 33376537
dataset_size: 99062676
- config_name: unshuffled_original_la
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 27801400
num_examples: 94588
download_size: 5458131
dataset_size: 27801400
- config_name: unshuffled_original_lmo
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 470001
num_examples: 1401
download_size: 109759
dataset_size: 470001
- config_name: unshuffled_original_lv
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 4266812625
num_examples: 1593820
download_size: 1486675302
dataset_size: 4266812625
- config_name: unshuffled_original_min
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 624991
num_examples: 220
download_size: 12379
dataset_size: 624991
- config_name: unshuffled_original_mr
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2816455519
num_examples: 326804
download_size: 525303459
dataset_size: 2816455519
- config_name: unshuffled_original_mwl
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1273
num_examples: 8
download_size: 789
dataset_size: 1273
- config_name: unshuffled_original_nah
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 12070
num_examples: 61
download_size: 2857
dataset_size: 12070
- config_name: unshuffled_original_new
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 5766053
num_examples: 4696
download_size: 1031042
dataset_size: 5766053
- config_name: unshuffled_original_oc
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 6127539
num_examples: 10709
download_size: 1574956
dataset_size: 6127539
- config_name: unshuffled_original_pam
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 800
num_examples: 3
download_size: 364
dataset_size: 800
- config_name: unshuffled_original_ps
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 379515973
num_examples: 98216
download_size: 103659691
dataset_size: 379515973
- config_name: unshuffled_original_ro
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 26869251055
num_examples: 9387265
download_size: 9534521905
dataset_size: 26869251055
- config_name: unshuffled_original_scn
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 3573
num_examples: 21
download_size: 1614
dataset_size: 3573
- config_name: unshuffled_original_sk
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 9808179461
num_examples: 5492194
download_size: 3708313186
dataset_size: 9808179461
- config_name: unshuffled_original_sr
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 4131922671
num_examples: 1013619
download_size: 1081129678
dataset_size: 4131922671
- config_name: unshuffled_original_ta
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 9934879441
num_examples: 1263280
download_size: 1737252172
dataset_size: 9934879441
- config_name: unshuffled_original_tk
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 10662991
num_examples: 6456
download_size: 2956150
dataset_size: 10662991
- config_name: unshuffled_original_tyv
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 12219
num_examples: 34
download_size: 3034
dataset_size: 12219
- config_name: unshuffled_original_uz
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 21464779
num_examples: 27537
download_size: 5775644
dataset_size: 21464779
- config_name: unshuffled_original_wa
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 291400
num_examples: 1001
download_size: 89942
dataset_size: 291400
- config_name: unshuffled_original_xmf
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 6120123
num_examples: 3783
download_size: 1048265
dataset_size: 6120123
- config_name: unshuffled_original_it
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 147378116499
num_examples: 46981781
download_size: 52157691650
dataset_size: 147378116499
- config_name: unshuffled_original_ka
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 3768832240
num_examples: 563916
download_size: 680732710
dataset_size: 3768832240
- config_name: unshuffled_original_ko
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 25292102197
num_examples: 7345075
download_size: 8807937093
dataset_size: 25292102197
- config_name: unshuffled_original_kw
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 47016
num_examples: 203
download_size: 6715
dataset_size: 47016
- config_name: unshuffled_original_lez
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 3378104
num_examples: 1485
download_size: 825648
dataset_size: 3378104
- config_name: unshuffled_original_lrc
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 78347
num_examples: 88
download_size: 16573
dataset_size: 78347
- config_name: unshuffled_original_mg
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 21789998
num_examples: 17957
download_size: 6213316
dataset_size: 21789998
- config_name: unshuffled_original_ml
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 5244279375
num_examples: 603937
download_size: 938681749
dataset_size: 5244279375
- config_name: unshuffled_original_ms
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 122326270
num_examples: 534016
download_size: 28458804
dataset_size: 122326270
- config_name: unshuffled_original_myv
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1436
num_examples: 6
download_size: 691
dataset_size: 1436
- config_name: unshuffled_original_nds
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 18238189
num_examples: 18174
download_size: 6744705
dataset_size: 18238189
- config_name: unshuffled_original_nn
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 90838777
num_examples: 185884
download_size: 32863375
dataset_size: 90838777
- config_name: unshuffled_original_os
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 12893477
num_examples: 5213
download_size: 3096133
dataset_size: 12893477
- config_name: unshuffled_original_pms
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2154710
num_examples: 3225
download_size: 756400
dataset_size: 2154710
- config_name: unshuffled_original_qu
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 85032
num_examples: 452
download_size: 17931
dataset_size: 85032
- config_name: unshuffled_original_sa
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 97055224
num_examples: 14291
download_size: 17517475
dataset_size: 97055224
- config_name: unshuffled_original_sh
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 25841505
num_examples: 36700
download_size: 3457359
dataset_size: 25841505
- config_name: unshuffled_original_so
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 63785
num_examples: 156
download_size: 2478
dataset_size: 63785
- config_name: unshuffled_original_sv
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 47000933560
num_examples: 17395625
download_size: 17182697021
dataset_size: 47000933560
- config_name: unshuffled_original_tg
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 397436494
num_examples: 89002
download_size: 90972727
dataset_size: 397436494
- config_name: unshuffled_original_tr
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 63581153419
num_examples: 18535253
download_size: 21961561999
dataset_size: 63581153419
- config_name: unshuffled_original_uk
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 56439494556
num_examples: 12973467
download_size: 14419203733
dataset_size: 56439494556
- config_name: unshuffled_original_vi
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 72226388484
num_examples: 14898250
download_size: 21503594095
dataset_size: 72226388484
- config_name: unshuffled_original_wuu
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 114041
num_examples: 214
download_size: 8780
dataset_size: 114041
- config_name: unshuffled_original_yo
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 58546
num_examples: 214
download_size: 9550
dataset_size: 58546
- config_name: unshuffled_original_zh
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 545607539477
num_examples: 60137667
download_size: 206003993405
dataset_size: 545607539477
- config_name: unshuffled_deduplicated_en
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1297616499791
num_examples: 304230423
download_size: 496496144465
dataset_size: 1297616499791
- config_name: unshuffled_deduplicated_eu
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 360674267
num_examples: 256513
download_size: 134683484
dataset_size: 360674267
- config_name: unshuffled_deduplicated_frr
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 4500
num_examples: 7
download_size: 540
dataset_size: 4500
- config_name: unshuffled_deduplicated_gl
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 404922022
num_examples: 284320
download_size: 155851883
dataset_size: 404922022
- config_name: unshuffled_deduplicated_he
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 10451408409
num_examples: 2375030
download_size: 3043383695
dataset_size: 10451408409
- config_name: unshuffled_deduplicated_ht
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 3439
num_examples: 9
download_size: 594
dataset_size: 3439
- config_name: unshuffled_deduplicated_id
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 16964948727
num_examples: 9948521
download_size: 5995510660
dataset_size: 16964948727
- config_name: unshuffled_deduplicated_is
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 891047926
num_examples: 389515
download_size: 332871764
dataset_size: 891047926
- config_name: unshuffled_deduplicated_jv
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 609713
num_examples: 1163
download_size: 208165
dataset_size: 609713
- config_name: unshuffled_deduplicated_kn
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1080985653
num_examples: 251064
download_size: 215526836
dataset_size: 1080985653
- config_name: unshuffled_deduplicated_kv
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1200609
num_examples: 924
download_size: 327479
dataset_size: 1200609
- config_name: unshuffled_deduplicated_lb
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 21242773
num_examples: 21735
download_size: 8300328
dataset_size: 21242773
- config_name: unshuffled_deduplicated_lo
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 119015146
num_examples: 32652
download_size: 23634237
dataset_size: 119015146
- config_name: unshuffled_deduplicated_mai
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 10721
num_examples: 25
download_size: 2267
dataset_size: 10721
- config_name: unshuffled_deduplicated_mk
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1186605123
num_examples: 299457
download_size: 303118518
dataset_size: 1186605123
- config_name: unshuffled_deduplicated_mrj
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1096428
num_examples: 669
download_size: 289048
dataset_size: 1096428
- config_name: unshuffled_deduplicated_my
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1112006614
num_examples: 136639
download_size: 207136614
dataset_size: 1112006614
- config_name: unshuffled_deduplicated_nap
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 13782
num_examples: 55
download_size: 4965
dataset_size: 13782
- config_name: unshuffled_deduplicated_nl
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 41726089054
num_examples: 20812149
download_size: 15734167112
dataset_size: 41726089054
- config_name: unshuffled_deduplicated_or
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 197401878
num_examples: 44230
download_size: 38726721
dataset_size: 197401878
- config_name: unshuffled_deduplicated_pl
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 50387595763
num_examples: 20682611
download_size: 20189161328
dataset_size: 50387595763
- config_name: unshuffled_deduplicated_pt
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 68162434231
num_examples: 26920397
download_size: 25997795946
dataset_size: 68162434231
- config_name: unshuffled_deduplicated_ru
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 611031071327
num_examples: 115954598
download_size: 166677136024
dataset_size: 611031071327
- config_name: unshuffled_deduplicated_sd
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 275327037
num_examples: 33925
download_size: 74169753
dataset_size: 275327037
- config_name: unshuffled_deduplicated_sl
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1311219223
num_examples: 886223
download_size: 523218283
dataset_size: 1311219223
- config_name: unshuffled_deduplicated_su
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 149921
num_examples: 511
download_size: 53164
dataset_size: 149921
- config_name: unshuffled_deduplicated_te
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1694004428
num_examples: 312644
download_size: 342429224
dataset_size: 1694004428
- config_name: unshuffled_deduplicated_tl
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 429427446
num_examples: 294132
download_size: 151342433
dataset_size: 429427446
- config_name: unshuffled_deduplicated_ug
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 86344782
num_examples: 15503
download_size: 20527752
dataset_size: 86344782
- config_name: unshuffled_deduplicated_vec
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 17303
num_examples: 64
download_size: 7647
dataset_size: 17303
- config_name: unshuffled_deduplicated_war
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2338532
num_examples: 9161
download_size: 546586
dataset_size: 2338532
- config_name: unshuffled_deduplicated_yi
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 87935052
num_examples: 32919
download_size: 22197718
dataset_size: 87935052
- config_name: unshuffled_original_af
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 254076274
num_examples: 201117
download_size: 85795254
dataset_size: 254076274
- config_name: unshuffled_original_ar
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 87935768938
num_examples: 16365602
download_size: 22232546836
dataset_size: 87935768938
- config_name: unshuffled_original_av
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 423603
num_examples: 456
download_size: 84767
dataset_size: 423603
- config_name: unshuffled_original_bar
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 555
num_examples: 4
download_size: 341
dataset_size: 555
- config_name: unshuffled_original_bh
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 116514
num_examples: 336
download_size: 7615
dataset_size: 116514
- config_name: unshuffled_original_br
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 30203875
num_examples: 37085
download_size: 9178158
dataset_size: 30203875
- config_name: unshuffled_original_cbk
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 536
num_examples: 1
download_size: 234
dataset_size: 536
- config_name: unshuffled_original_cs
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 57080142860
num_examples: 21001388
download_size: 21716697253
dataset_size: 57080142860
- config_name: unshuffled_original_de
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 331224484023
num_examples: 104913504
download_size: 119506267566
dataset_size: 331224484023
- config_name: unshuffled_original_el
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 66273231642
num_examples: 10425596
download_size: 17309601342
dataset_size: 66273231642
- config_name: unshuffled_original_es
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 298492270636
num_examples: 88199221
download_size: 106039137656
dataset_size: 298492270636
- config_name: unshuffled_original_fi
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 28571419204
num_examples: 8557453
download_size: 9970837279
dataset_size: 28571419204
- config_name: unshuffled_original_ga
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 92369035
num_examples: 83223
download_size: 29262282
dataset_size: 92369035
- config_name: unshuffled_original_gom
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2257169
num_examples: 640
download_size: 442950
dataset_size: 2257169
- config_name: unshuffled_original_hr
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 243829069
num_examples: 582219
download_size: 79417804
dataset_size: 243829069
- config_name: unshuffled_original_hy
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 3939672772
num_examples: 659430
download_size: 897364024
dataset_size: 3939672772
- config_name: unshuffled_original_ilo
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 925809
num_examples: 2638
download_size: 267451
dataset_size: 925809
- config_name: unshuffled_original_ja
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 232216718556
num_examples: 62721527
download_size: 79564645083
dataset_size: 232216718556
- config_name: unshuffled_original_kk
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2833778199
num_examples: 524591
download_size: 615067761
dataset_size: 2833778199
- config_name: unshuffled_original_krc
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2688672
num_examples: 1581
download_size: 656496
dataset_size: 2688672
- config_name: unshuffled_original_ky
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 630794622
num_examples: 146993
download_size: 152636608
dataset_size: 630794622
- config_name: unshuffled_original_li
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 31312
num_examples: 137
download_size: 11793
dataset_size: 31312
- config_name: unshuffled_original_lt
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 9445278312
num_examples: 2977757
download_size: 3439789726
dataset_size: 9445278312
- config_name: unshuffled_original_mhr
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 7553453
num_examples: 3212
download_size: 1834912
dataset_size: 7553453
- config_name: unshuffled_original_mn
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2332897881
num_examples: 395605
download_size: 472357548
dataset_size: 2332897881
- config_name: unshuffled_original_mt
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 24470330
num_examples: 26598
download_size: 7533204
dataset_size: 24470330
- config_name: unshuffled_original_mzn
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 720229
num_examples: 1055
download_size: 177817
dataset_size: 720229
- config_name: unshuffled_original_ne
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1866852959
num_examples: 299938
download_size: 355291639
dataset_size: 1866852959
- config_name: unshuffled_original_no
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 8652054976
num_examples: 5546211
download_size: 3106155643
dataset_size: 8652054976
- config_name: unshuffled_original_pa
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 801167879
num_examples: 127467
download_size: 164207256
dataset_size: 801167879
- config_name: unshuffled_original_pnb
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 12039418
num_examples: 4599
download_size: 3215579
dataset_size: 12039418
- config_name: unshuffled_original_rm
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 8027
num_examples: 41
download_size: 2691
dataset_size: 8027
- config_name: unshuffled_original_sah
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 43817239
num_examples: 22301
download_size: 9079982
dataset_size: 43817239
- config_name: unshuffled_original_si
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1469374795
num_examples: 203082
download_size: 310935021
dataset_size: 1469374795
- config_name: unshuffled_original_sq
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2440834375
num_examples: 672077
download_size: 861831806
dataset_size: 2440834375
- config_name: unshuffled_original_sw
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 14073775
num_examples: 41986
download_size: 3712739
dataset_size: 14073775
- config_name: unshuffled_original_th
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 38289228753
num_examples: 6064129
download_size: 7377469078
dataset_size: 38289228753
- config_name: unshuffled_original_tt
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 703412782
num_examples: 135923
download_size: 151056507
dataset_size: 703412782
- config_name: unshuffled_original_ur
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2802270961
num_examples: 638596
download_size: 712607161
dataset_size: 2802270961
- config_name: unshuffled_original_vo
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2118909
num_examples: 3366
download_size: 307184
dataset_size: 2118909
- config_name: unshuffled_original_xal
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 116043
num_examples: 39
download_size: 32117
dataset_size: 116043
- config_name: unshuffled_original_yue
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 3899
num_examples: 11
download_size: 647
dataset_size: 3899
- config_name: unshuffled_original_en
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2525437912097
num_examples: 455994980
download_size: 903830686146
dataset_size: 2525437912097
- config_name: unshuffled_original_eu
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 894836188
num_examples: 506883
download_size: 248190119
dataset_size: 894836188
- config_name: unshuffled_original_frr
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 4507
num_examples: 7
download_size: 527
dataset_size: 4507
- config_name: unshuffled_original_gl
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 656477422
num_examples: 544388
download_size: 235384299
dataset_size: 656477422
- config_name: unshuffled_original_he
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 21113706929
num_examples: 3808397
download_size: 5660026441
dataset_size: 21113706929
- config_name: unshuffled_original_ht
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 4083
num_examples: 13
download_size: 590
dataset_size: 4083
- config_name: unshuffled_original_id
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 32317679452
num_examples: 16236463
download_size: 10596988488
dataset_size: 32317679452
- config_name: unshuffled_original_is
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1524936467
num_examples: 625673
download_size: 533034495
dataset_size: 1524936467
- config_name: unshuffled_original_jv
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 691812
num_examples: 1445
download_size: 219246
dataset_size: 691812
- config_name: unshuffled_original_kn
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1763625096
num_examples: 350363
download_size: 342155433
dataset_size: 1763625096
- config_name: unshuffled_original_kv
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2379758
num_examples: 1549
download_size: 400725
dataset_size: 2379758
- config_name: unshuffled_original_lb
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 30595156
num_examples: 34807
download_size: 10725552
dataset_size: 30595156
- config_name: unshuffled_original_lo
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 182361509
num_examples: 52910
download_size: 33916738
dataset_size: 182361509
- config_name: unshuffled_original_mai
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 325990
num_examples: 123
download_size: 5563
dataset_size: 325990
- config_name: unshuffled_original_mk
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2202480390
num_examples: 437871
download_size: 508239918
dataset_size: 2202480390
- config_name: unshuffled_original_mrj
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1165977
num_examples: 757
download_size: 303447
dataset_size: 1165977
- config_name: unshuffled_original_my
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2021872493
num_examples: 232329
download_size: 369850157
dataset_size: 2021872493
- config_name: unshuffled_original_nap
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 17839
num_examples: 73
download_size: 5023
dataset_size: 17839
- config_name: unshuffled_original_nl
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 83230965323
num_examples: 34682142
download_size: 29352811750
dataset_size: 83230965323
- config_name: unshuffled_original_or
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 260151226
num_examples: 59463
download_size: 49834443
dataset_size: 260151226
- config_name: unshuffled_original_pl
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 117121370605
num_examples: 35440972
download_size: 42884898947
dataset_size: 117121370605
- config_name: unshuffled_original_pt
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 132635490139
num_examples: 42114520
download_size: 47257949300
dataset_size: 132635490139
- config_name: unshuffled_original_ru
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 1241627166551
num_examples: 161836003
download_size: 319755378587
dataset_size: 1241627166551
- config_name: unshuffled_original_sd
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 364256869
num_examples: 44280
download_size: 90621520
dataset_size: 364256869
- config_name: unshuffled_original_sl
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2675665926
num_examples: 1746604
download_size: 956197026
dataset_size: 2675665926
- config_name: unshuffled_original_su
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 225627
num_examples: 805
download_size: 59643
dataset_size: 225627
- config_name: unshuffled_original_te
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2611548765
num_examples: 475703
download_size: 522470115
dataset_size: 2611548765
- config_name: unshuffled_original_tl
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 606295665
num_examples: 458206
download_size: 204895159
dataset_size: 606295665
- config_name: unshuffled_original_ug
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 127419368
num_examples: 22255
download_size: 27923925
dataset_size: 127419368
- config_name: unshuffled_original_vec
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 19182
num_examples: 73
download_size: 7672
dataset_size: 19182
- config_name: unshuffled_original_war
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 2682430
num_examples: 9760
download_size: 644576
dataset_size: 2682430
- config_name: unshuffled_original_yi
features:
- name: id
dtype: int64
- name: text
dtype: string
splits:
- name: train
num_bytes: 147601654
num_examples: 59364
download_size: 33337157
dataset_size: 147601654
config_names:
- unshuffled_deduplicated_af
- unshuffled_deduplicated_als
- unshuffled_deduplicated_am
- unshuffled_deduplicated_an
- unshuffled_deduplicated_ar
- unshuffled_deduplicated_arz
- unshuffled_deduplicated_as
- unshuffled_deduplicated_ast
- unshuffled_deduplicated_av
- unshuffled_deduplicated_az
- unshuffled_deduplicated_azb
- unshuffled_deduplicated_ba
- unshuffled_deduplicated_bar
- unshuffled_deduplicated_bcl
- unshuffled_deduplicated_be
- unshuffled_deduplicated_bg
- unshuffled_deduplicated_bh
- unshuffled_deduplicated_bn
- unshuffled_deduplicated_bo
- unshuffled_deduplicated_bpy
- unshuffled_deduplicated_br
- unshuffled_deduplicated_bs
- unshuffled_deduplicated_bxr
- unshuffled_deduplicated_ca
- unshuffled_deduplicated_cbk
- unshuffled_deduplicated_ce
- unshuffled_deduplicated_ceb
- unshuffled_deduplicated_ckb
- unshuffled_deduplicated_cs
- unshuffled_deduplicated_cv
- unshuffled_deduplicated_cy
- unshuffled_deduplicated_da
- unshuffled_deduplicated_de
- unshuffled_deduplicated_diq
- unshuffled_deduplicated_dsb
- unshuffled_deduplicated_dv
- unshuffled_deduplicated_el
- unshuffled_deduplicated_eml
- unshuffled_deduplicated_en
- unshuffled_deduplicated_eo
- unshuffled_deduplicated_es
- unshuffled_deduplicated_et
- unshuffled_deduplicated_eu
- unshuffled_deduplicated_fa
- unshuffled_deduplicated_fi
- unshuffled_deduplicated_fr
- unshuffled_deduplicated_frr
- unshuffled_deduplicated_fy
- unshuffled_deduplicated_ga
- unshuffled_deduplicated_gd
- unshuffled_deduplicated_gl
- unshuffled_deduplicated_gn
- unshuffled_deduplicated_gom
- unshuffled_deduplicated_gu
- unshuffled_deduplicated_he
- unshuffled_deduplicated_hi
- unshuffled_deduplicated_hr
- unshuffled_deduplicated_hsb
- unshuffled_deduplicated_ht
- unshuffled_deduplicated_hu
- unshuffled_deduplicated_hy
- unshuffled_deduplicated_ia
- unshuffled_deduplicated_id
- unshuffled_deduplicated_ie
- unshuffled_deduplicated_ilo
- unshuffled_deduplicated_io
- unshuffled_deduplicated_is
- unshuffled_deduplicated_it
- unshuffled_deduplicated_ja
- unshuffled_deduplicated_jbo
- unshuffled_deduplicated_jv
- unshuffled_deduplicated_ka
- unshuffled_deduplicated_kk
- unshuffled_deduplicated_km
- unshuffled_deduplicated_kn
- unshuffled_deduplicated_ko
- unshuffled_deduplicated_krc
- unshuffled_deduplicated_ku
- unshuffled_deduplicated_kv
- unshuffled_deduplicated_kw
- unshuffled_deduplicated_ky
- unshuffled_deduplicated_la
- unshuffled_deduplicated_lb
- unshuffled_deduplicated_lez
- unshuffled_deduplicated_li
- unshuffled_deduplicated_lmo
- unshuffled_deduplicated_lo
- unshuffled_deduplicated_lrc
- unshuffled_deduplicated_lt
- unshuffled_deduplicated_lv
- unshuffled_deduplicated_mai
- unshuffled_deduplicated_mg
- unshuffled_deduplicated_mhr
- unshuffled_deduplicated_min
- unshuffled_deduplicated_mk
- unshuffled_deduplicated_ml
- unshuffled_deduplicated_mn
- unshuffled_deduplicated_mr
- unshuffled_deduplicated_mrj
- unshuffled_deduplicated_ms
- unshuffled_deduplicated_mt
- unshuffled_deduplicated_mwl
- unshuffled_deduplicated_my
- unshuffled_deduplicated_myv
- unshuffled_deduplicated_mzn
- unshuffled_deduplicated_nah
- unshuffled_deduplicated_nap
- unshuffled_deduplicated_nds
- unshuffled_deduplicated_ne
- unshuffled_deduplicated_new
- unshuffled_deduplicated_nl
- unshuffled_deduplicated_nn
- unshuffled_deduplicated_no
- unshuffled_deduplicated_oc
- unshuffled_deduplicated_or
- unshuffled_deduplicated_os
- unshuffled_deduplicated_pa
- unshuffled_deduplicated_pam
- unshuffled_deduplicated_pl
- unshuffled_deduplicated_pms
- unshuffled_deduplicated_pnb
- unshuffled_deduplicated_ps
- unshuffled_deduplicated_pt
- unshuffled_deduplicated_qu
- unshuffled_deduplicated_rm
- unshuffled_deduplicated_ro
- unshuffled_deduplicated_ru
- unshuffled_deduplicated_sa
- unshuffled_deduplicated_sah
- unshuffled_deduplicated_scn
- unshuffled_deduplicated_sd
- unshuffled_deduplicated_sh
- unshuffled_deduplicated_si
- unshuffled_deduplicated_sk
- unshuffled_deduplicated_sl
- unshuffled_deduplicated_so
- unshuffled_deduplicated_sq
- unshuffled_deduplicated_sr
- unshuffled_deduplicated_su
- unshuffled_deduplicated_sv
- unshuffled_deduplicated_sw
- unshuffled_deduplicated_ta
- unshuffled_deduplicated_te
- unshuffled_deduplicated_tg
- unshuffled_deduplicated_th
- unshuffled_deduplicated_tk
- unshuffled_deduplicated_tl
- unshuffled_deduplicated_tr
- unshuffled_deduplicated_tt
- unshuffled_deduplicated_tyv
- unshuffled_deduplicated_ug
- unshuffled_deduplicated_uk
- unshuffled_deduplicated_ur
- unshuffled_deduplicated_uz
- unshuffled_deduplicated_vec
- unshuffled_deduplicated_vi
- unshuffled_deduplicated_vo
- unshuffled_deduplicated_wa
- unshuffled_deduplicated_war
- unshuffled_deduplicated_wuu
- unshuffled_deduplicated_xal
- unshuffled_deduplicated_xmf
- unshuffled_deduplicated_yi
- unshuffled_deduplicated_yo
- unshuffled_deduplicated_yue
- unshuffled_deduplicated_zh
- unshuffled_original_af
- unshuffled_original_als
- unshuffled_original_am
- unshuffled_original_an
- unshuffled_original_ar
- unshuffled_original_arz
- unshuffled_original_as
- unshuffled_original_ast
- unshuffled_original_av
- unshuffled_original_az
- unshuffled_original_azb
- unshuffled_original_ba
- unshuffled_original_bar
- unshuffled_original_bcl
- unshuffled_original_be
- unshuffled_original_bg
- unshuffled_original_bh
- unshuffled_original_bn
- unshuffled_original_bo
- unshuffled_original_bpy
- unshuffled_original_br
- unshuffled_original_bs
- unshuffled_original_bxr
- unshuffled_original_ca
- unshuffled_original_cbk
- unshuffled_original_ce
- unshuffled_original_ceb
- unshuffled_original_ckb
- unshuffled_original_cs
- unshuffled_original_cv
- unshuffled_original_cy
- unshuffled_original_da
- unshuffled_original_de
- unshuffled_original_diq
- unshuffled_original_dsb
- unshuffled_original_dv
- unshuffled_original_el
- unshuffled_original_eml
- unshuffled_original_en
- unshuffled_original_eo
- unshuffled_original_es
- unshuffled_original_et
- unshuffled_original_eu
- unshuffled_original_fa
- unshuffled_original_fi
- unshuffled_original_fr
- unshuffled_original_frr
- unshuffled_original_fy
- unshuffled_original_ga
- unshuffled_original_gd
- unshuffled_original_gl
- unshuffled_original_gn
- unshuffled_original_gom
- unshuffled_original_gu
- unshuffled_original_he
- unshuffled_original_hi
- unshuffled_original_hr
- unshuffled_original_hsb
- unshuffled_original_ht
- unshuffled_original_hu
- unshuffled_original_hy
- unshuffled_original_ia
- unshuffled_original_id
- unshuffled_original_ie
- unshuffled_original_ilo
- unshuffled_original_io
- unshuffled_original_is
- unshuffled_original_it
- unshuffled_original_ja
- unshuffled_original_jbo
- unshuffled_original_jv
- unshuffled_original_ka
- unshuffled_original_kk
- unshuffled_original_km
- unshuffled_original_kn
- unshuffled_original_ko
- unshuffled_original_krc
- unshuffled_original_ku
- unshuffled_original_kv
- unshuffled_original_kw
- unshuffled_original_ky
- unshuffled_original_la
- unshuffled_original_lb
- unshuffled_original_lez
- unshuffled_original_li
- unshuffled_original_lmo
- unshuffled_original_lo
- unshuffled_original_lrc
- unshuffled_original_lt
- unshuffled_original_lv
- unshuffled_original_mai
- unshuffled_original_mg
- unshuffled_original_mhr
- unshuffled_original_min
- unshuffled_original_mk
- unshuffled_original_ml
- unshuffled_original_mn
- unshuffled_original_mr
- unshuffled_original_mrj
- unshuffled_original_ms
- unshuffled_original_mt
- unshuffled_original_mwl
- unshuffled_original_my
- unshuffled_original_myv
- unshuffled_original_mzn
- unshuffled_original_nah
- unshuffled_original_nap
- unshuffled_original_nds
- unshuffled_original_ne
- unshuffled_original_new
- unshuffled_original_nl
- unshuffled_original_nn
- unshuffled_original_no
- unshuffled_original_oc
- unshuffled_original_or
- unshuffled_original_os
- unshuffled_original_pa
- unshuffled_original_pam
- unshuffled_original_pl
- unshuffled_original_pms
- unshuffled_original_pnb
- unshuffled_original_ps
- unshuffled_original_pt
- unshuffled_original_qu
- unshuffled_original_rm
- unshuffled_original_ro
- unshuffled_original_ru
- unshuffled_original_sa
- unshuffled_original_sah
- unshuffled_original_scn
- unshuffled_original_sd
- unshuffled_original_sh
- unshuffled_original_si
- unshuffled_original_sk
- unshuffled_original_sl
- unshuffled_original_so
- unshuffled_original_sq
- unshuffled_original_sr
- unshuffled_original_su
- unshuffled_original_sv
- unshuffled_original_sw
- unshuffled_original_ta
- unshuffled_original_te
- unshuffled_original_tg
- unshuffled_original_th
- unshuffled_original_tk
- unshuffled_original_tl
- unshuffled_original_tr
- unshuffled_original_tt
- unshuffled_original_tyv
- unshuffled_original_ug
- unshuffled_original_uk
- unshuffled_original_ur
- unshuffled_original_uz
- unshuffled_original_vec
- unshuffled_original_vi
- unshuffled_original_vo
- unshuffled_original_wa
- unshuffled_original_war
- unshuffled_original_wuu
- unshuffled_original_xal
- unshuffled_original_xmf
- unshuffled_original_yi
- unshuffled_original_yo
- unshuffled_original_yue
- unshuffled_original_zh
---
# Dataset Card for "oscar"
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://oscar-corpus.com](https://oscar-corpus.com)
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Dataset Summary
OSCAR or **O**pen **S**uper-large **C**rawled [**A**LMAnaCH](https://team.inria.fr/almanach/) co**R**pus is a huge multilingual corpus obtained by language classification and filtering of the [Common Crawl](https://commoncrawl.org/) corpus using the [goclassy](https://github.com/pjox/goclassy) architecture. Data is distributed by language in both original and deduplicated form.
The version here is the original OSCAR 2019 release: https://oscar-project.org/post/oscar-2019/
For more recent versions, visit the [oscar-corpus](https://huggingface.co/oscar-corpus) organization on the Hub:
- OSCAR 22.01 (released in January 2022): [oscar-corpus/OSCAR-2201](https://huggingface.co/datasets/oscar-corpus/OSCAR-2201)
- OSCAR 21.09 (released in September 2021): [oscar-corpus/OSCAR-2109](https://huggingface.co/datasets/oscar-corpus/OSCAR-2109)
### Supported Tasks and Leaderboards
OSCAR is mainly inteded to pretrain language models and word represantations.
### Languages
All the data is distributed by language, both the original and the deduplicated versions of the data are available. 166 different languages are available. The table in subsection [Data Splits Sample Size](#data-splits-sample-size) provides the language code for each subcorpus as well as the number of words (space separated tokens), lines and sizes for both the original and the deduplicated versions of OSCAR.
## Dataset Structure
We show detailed information for all the configurations of the dataset.
### Data Instances
<details>
<summary>Click to expand the Data/size information for each language (deduplicated)</summary>
#### unshuffled_deduplicated_af
- **Size of downloaded dataset files:** 65.99 MB
- **Size of the generated dataset:** 172.30 MB
- **Total amount of disk used:** 238.29 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "aanlyn markte as gevolg van ons voortgesette 'n begrip opsie handel sakeplan pdf terwyl ons steeds die gereelde ons binêre opsies handel"
}
```
#### unshuffled_deduplicated_als
- **Size of downloaded dataset files:** 1.26 MB
- **Size of the generated dataset:** 2.96 MB
- **Total amount of disk used:** 4.22 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"De Nazionalpark hät e Flächi vo 170,3 km² und isch dodemit s grösti Naturschutzgebiet vo de Schwiz. Er ligt uf em Gebiet vo de ..."
}
```
#### unshuffled_deduplicated_am
- **Size of downloaded dataset files:** 61.35 MB
- **Size of the generated dataset:** 216.15 MB
- **Total amount of disk used:** 277.50 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"አየር መንገዱ ከአዲስ አበባ ወደ ሮም ጣሊያን በማምራት ላይ በነበረበት ጊዜ ረዳት አብራሪው የጉዞውን አቅጣጫ በመቀየር ጄኔቭ አውሮፓላን ማረፊያ በማሳረፍ እጁን ለፖሊስ ሰጥቷል።\\nየኢትዮጵያ መንግስት የ..."
}
```
#### unshuffled_deduplicated_an
- **Size of downloaded dataset files:** 0.14 MB
- **Size of the generated dataset:** 0.85 MB
- **Total amount of disk used:** 0.99 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"واااااااأسفاه الأمم تفتخر ب 0 أمي ووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووو..."
}
```
#### unshuffled_deduplicated_ar
- **Size of downloaded dataset files:** 9.67 GB
- **Size of the generated dataset:** 33.57 GB
- **Total amount of disk used:** 43.23 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"مرحبا بك عزيز الزائر نتمنى لك أوقاتاً سعيدة معنا وأن نزداد شرفا بخدمتك ولا تنسى التسجيل معنا لتستفيد بكل جديد\\nأهلا وسهلا بك زا..."
}
```
#### unshuffled_deduplicated_arz
- **Size of downloaded dataset files:** 10.02 MB
- **Size of the generated dataset:** 35.91 MB
- **Total amount of disk used:** 45.94 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"بنى عجل : قبيلة من عجل بن لجيم بن صعب بن على بن بكر بن وائل انتقل اغلبهم الى البصرة فى العراق و اصفهان و خراسان فى ايران و اذرب..."
}
```
#### unshuffled_deduplicated_as
- **Size of downloaded dataset files:** 15.51 MB
- **Size of the generated dataset:** 74.07 MB
- **Total amount of disk used:** 89.58 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"আমি, এই সংগঠনৰ সদস্য সকলে একেলগ হৈ অসমকে ধৰি ভাৰতৰ উত্তৰ পূৰ্বাঞ্চলৰ অমূল্য কলা-সাংস্কৃতিক সম্পদৰাজি বৃহত্তৰ অষ্ট্ৰেলিয়াৰ সন্মু..."
}
```
#### unshuffled_deduplicated_ast
- **Size of downloaded dataset files:** 0.86 MB
- **Size of the generated dataset:** 2.17 MB
- **Total amount of disk used:** 3.03 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"The Killers llanzaron el so álbum debú, Hot Fuss, en xunu de 2004 nel Reinu Xuníu, al traviés de la discográfica Lizard King, y..."
}
```
#### unshuffled_deduplicated_av
- **Size of downloaded dataset files:** 0.07 MB
- **Size of the generated dataset:** 0.34 MB
- **Total amount of disk used:** 0.41 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Жинда малъараб ва божизе бегьулеб рагІудаса кьуризе бегьуларо гьев. Гьес насихІат гьабизе кколелъул бацІцІадаб диналъул рахъалъ..."
}
```
#### unshuffled_deduplicated_az
- **Size of downloaded dataset files:** 521.74 MB
- **Size of the generated dataset:** 1.53 GB
- **Total amount of disk used:** 2.05 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"AZTV-Artıq 7 ildir ki, Abşeron rayonu dotasiya almadan bütün xərclərini yerli daxilolmalar hesabına maliyyələşdirir.\\nDünən, 10..."
}
```
#### unshuffled_deduplicated_azb
- **Size of downloaded dataset files:** 5.19 MB
- **Size of the generated dataset:** 20.08 MB
- **Total amount of disk used:** 25.27 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"لعلی ١٣-جو عصرده یاشاییب یاراتمیش گؤرکملی آذربایجان شاعرلریندندیر. ١٢٢٤-جی ایلده تبریزده آنادان اولموشدور، گنج یاشلاریندا تیجار..."
}
```
#### unshuffled_deduplicated_ba
- **Size of downloaded dataset files:** 25.98 MB
- **Size of the generated dataset:** 93.84 MB
- **Total amount of disk used:** 119.82 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Күҙәтеү ҡуласаһы моделен хәҙер Мифтахетдин Аҡмулла исемендәге Башҡорт дәүләт педагогия университетында ла эшләргә мөмкин\\t\\nКүҙ..."
}
```
#### unshuffled_deduplicated_bar
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": " vo"
}
```
#### unshuffled_deduplicated_bcl
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"& ÿ ó / í 0 - ø û ù ö ú ð ï ú \\u0014 ù þ ô ö í ÷ ò \\u0014 ÷ í ù û ö í \\u0001 û ñ ç þ \\u0001 ð \\u0007 þ ò ñ ñ ò ô \\u0017 û ö ô ÷..."
}
```
#### unshuffled_deduplicated_be
- **Size of downloaded dataset files:** 306.70 MB
- **Size of the generated dataset:** 1.08 GB
- **Total amount of disk used:** 1.39 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Брэсцкія ўлады не дазволілі прафсаюзу РЭП правесці пікетаванне ў парку Воінаў-інтэрнацыяналістаў 30 мая 2018 года.\\nСітуацыю пр..."
}
```
#### unshuffled_deduplicated_bg
- **Size of downloaded dataset files:** 3.85 GB
- **Size of the generated dataset:** 14.45 GB
- **Total amount of disk used:** 18.30 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ЖАЛБОПОДАТЕЛЯТ директор на Дирекция „ Обжалване и данъчно-осигурителна практика“- Бургас, редовно призован, се представлява от ..."
}
```
#### unshuffled_deduplicated_bh
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.04 MB
- **Total amount of disk used:** 0.04 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"सुकमा जिला भारत के छत्तीसगढ़ राज्य में एगो जिला बाटे। एकर मुख्यालय सुकमा शहर बाटे। एकर कुल रकबा 5636 वर्ग कि॰मी॰ बाटे।\"..."
}
```
#### unshuffled_deduplicated_bn
- **Size of downloaded dataset files:** 1.26 GB
- **Size of the generated dataset:** 6.24 GB
- **Total amount of disk used:** 7.50 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ভড়ং সর্বস্ব বাংলা আর্ট অ্যান্ড কালচারের হিসাব গুলিয়ে দেওয়ার ম্যাজিকের নাম ব্রাত্য রাইসু November 23, 2017\\nTagged with ডায়োজিনি..."
}
```
#### unshuffled_deduplicated_bo
- **Size of downloaded dataset files:** 22.37 MB
- **Size of the generated dataset:** 144.65 MB
- **Total amount of disk used:** 167.02 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"བོད་མི་འདི་དག་ནི་རང་རྒྱུད་སྒོ་རུ་ཕུད་དེ་གཞན་རྒྱུད་པང་དུ་ཉར་ནས་གསོ་སྐྱོང་བྱེད་དགོས་ཟེར་བ་དང་གཅིག་མཚུངས་རེད།\\nཚན་རིག་ནི་དང་ཐོག་རང..."
}
```
#### unshuffled_deduplicated_bpy
- **Size of downloaded dataset files:** 0.19 MB
- **Size of the generated dataset:** 1.78 MB
- **Total amount of disk used:** 1.97 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"পৌরসভা এহার আয়তন (লয়াহান) ২,৭৩০,.৬৩ বর্গ কিলোমিটার। পৌরসভা এহার মাপাহানর অক্ষাংশ বারো দ্রাঘিমাংশ ইলতাই 18.63° S 48.18° W ।[১]..."
}
```
#### unshuffled_deduplicated_br
- **Size of downloaded dataset files:** 6.47 MB
- **Size of the generated dataset:** 17.00 MB
- **Total amount of disk used:** 23.47 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Ar mank Magalhães(Daveoù a vank) a zo ur spesad evned, Spheniscus magellanicus an anv skiantel anezhañ.\\nGallout a reer implijo..."
}
```
#### unshuffled_deduplicated_bs
- **Size of downloaded dataset files:** 0.04 MB
- **Size of the generated dataset:** 0.15 MB
- **Total amount of disk used:** 0.18 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ž šř é ú šř šř ě šř ž é č ě ž ů ě ď éé ýš ě ě Ž č š ý ě ď é ýš ě ď ě éé ýš ě č ž ě š ý ď ě ýš é ú č ž č š ý ď ý ž é éě ď é č ýš..."
}
```
#### unshuffled_deduplicated_bxr
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.01 MB
- **Total amount of disk used:** 0.01 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"2002 оной хабар буряад хэлэ бэшэгэй һалбари Үндэһэтэнэй хүмүүнлиг ухаанай дээдэ һургуули болгогдожо өөршэлэгдөө.\\nХарин мүнөө б..."
}
```
#### unshuffled_deduplicated_ca
- **Size of downloaded dataset files:** 1.73 GB
- **Size of the generated dataset:** 4.57 GB
- **Total amount of disk used:** 6.30 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Daniel Vendrell, conegut com Vandrell, ha sigut un dels il•lustradors contemporanis més influents, representant a la nova onada..."
}
```
#### unshuffled_deduplicated_cbk
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano..."
}
```
#### unshuffled_deduplicated_ce
- **Size of downloaded dataset files:** 1.87 MB
- **Size of the generated dataset:** 7.04 MB
- **Total amount of disk used:** 8.90 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Шаьш анархисташ ду бохучу жигархойн дIахьедарехь дуьйцу, оьрсийн ницкъаллийн структурийн а, федералан каналан а Iалашонаш \\\"мар..."
}
```
#### unshuffled_deduplicated_ceb
- **Size of downloaded dataset files:** 7.12 MB
- **Size of the generated dataset:** 24.83 MB
- **Total amount of disk used:** 31.95 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Si Isko walay pupamilok nga nagtan-aw sa unahan, natugaw. “Naunsa ka gud diha Isko nga layo man kaayo ang imong panan-aw?” ni I..."
}
```
#### unshuffled_deduplicated_ckb
- **Size of downloaded dataset files:** 60.32 MB
- **Size of the generated dataset:** 237.72 MB
- **Total amount of disk used:** 298.05 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"رسی رۆژ - ساڵێک دوای بومەلەرزەی کرماشان میوانی بەرنامە : کاک سیاوەش حەیاتی چالاکی مەدەنی -قەسری شیرین\\nپارچە موزیک 30 / 10 / 20..."
}
```
#### unshuffled_deduplicated_cs
- **Size of downloaded dataset files:** 10.49 GB
- **Size of the generated dataset:** 25.71 GB
- **Total amount of disk used:** 36.20 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Akce anarchistů proti připravovanému novému služební řádu a nízkým mzdám 1903 – Historie českého anarchismu (1880 – 1939)\\nRost..."
}
```
#### unshuffled_deduplicated_cv
- **Size of downloaded dataset files:** 7.47 MB
- **Size of the generated dataset:** 27.49 MB
- **Total amount of disk used:** 34.95 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Шыранӑ чухне ӑнсӑртран латин кирилл саспаллисем вырӑнне латин саспаллисене ҫырсан, сайт эсир ҫырнине юсама тӑрӑшӗ.\\nКу сайтра ч..."
}
```
#### unshuffled_deduplicated_cy
- **Size of downloaded dataset files:** 53.63 MB
- **Size of the generated dataset:** 141.22 MB
- **Total amount of disk used:** 194.86 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Mae capeli Cymreig yr Andes ym Mhatagonia wedi cyhoeddi na fydd gwasanaethau yno weddill y mis, oherwydd yr eira trwm sydd wedi..."
}
```
#### unshuffled_deduplicated_da
- **Size of downloaded dataset files:** 3.82 GB
- **Size of the generated dataset:** 10.24 GB
- **Total amount of disk used:** 14.06 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Den 2.-5. februar 2016 løb det tredje kursus i uddannelsen af 4kommunesamarbejdets Local Impact Coaches, af stablen i Gentofte ..."
}
```
#### unshuffled_deduplicated_de
- **Size of downloaded dataset files:** 60.80 GB
- **Size of the generated dataset:** 156.30 GB
- **Total amount of disk used:** 217.10 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Auf dieser Seite gibt es mind. ein YouTube Video. Cookies für diese Website wurden abgelehnt. Dadurch können keine YouTube Vide..."
}
```
#### unshuffled_deduplicated_diq
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "Zıwanê Slawki, zıwano merdumanê Slawano. Zıwanê Slawki yew lızgeyê Zıwananê Hind u Ewropao. Keyeyê Zıwananê Slawki beno hirê letey:"
}
```
#### unshuffled_deduplicated_dsb
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.01 MB
- **Total amount of disk used:** 0.01 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Pśiklaskaju južo pśed pśedstajenim... 1500 źiśi njamóžo wěcej docakaś, měsćańska hala w Chóśebuzu - wupśedana."
}
```
#### unshuffled_deduplicated_dv
- **Size of downloaded dataset files:** 16.84 MB
- **Size of the generated dataset:** 82.19 MB
- **Total amount of disk used:** 99.03 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ބ. އަތޮޅުގައި ހުޅުވަން ތައްޔާރުވަމުން އަންނަ ވައްކަރު ރިސޯޓުގައި ވަޒީފާ އަދާކުރަން ޝައުގުވެރިވާ ފަރާތްތަކަށް ކުރިމަތިލުމުގެ ފުރ..."
}
```
#### unshuffled_deduplicated_el
- **Size of downloaded dataset files:** 7.91 GB
- **Size of the generated dataset:** 28.74 GB
- **Total amount of disk used:** 36.65 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Νεκρός εντοπίστηκε μέσα στο σπίτι του στην οδό Ηρώδου Αττικού στον αριθμό 7 ο επικεφαλής του προξενικού τμήματος της Ρωσικής πρ..."
}
```
#### unshuffled_deduplicated_eml
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.02 MB
- **Total amount of disk used:** 0.03 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"A séguit dal prucès ad rubutiśasiòṅ di abitànt dal pòpul ad Mikenes, Angoras 'l è finî dènt'r a 'n robot cun la tèsta dna rana ..."
}
```
#### unshuffled_deduplicated_en
- **Size of downloaded dataset files:** 496.50 GB
- **Size of the generated dataset:** 1299.75 GB
- **Total amount of disk used:** 1796.24 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Mtendere Village was inspired by the vision of Chief Napoleon Dzombe, which he shared with John Blanchard during his first visi..."
}
```
#### unshuffled_deduplicated_eo
- **Size of downloaded dataset files:** 92.86 MB
- **Size of the generated dataset:** 240.12 MB
- **Total amount of disk used:** 332.99 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Ĉu ... preĝi | mediti | ricevi instigojn || kanti | muziki || informiĝi | legi | studi || prepari Diservon\\nTemas pri kolekto d..."
}
```
#### unshuffled_deduplicated_es
- **Size of downloaded dataset files:** 60.46 GB
- **Size of the generated dataset:** 160.86 GB
- **Total amount of disk used:** 221.32 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Como se librará de la celulitis en el gimnasio La piel superflua en las manos después del adelgazamiento, Los bailes fáciles pa..."
}
```
#### unshuffled_deduplicated_et
- **Size of downloaded dataset files:** 966.79 MB
- **Size of the generated dataset:** 2.45 GB
- **Total amount of disk used:** 3.41 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"MTÜ AB Video järgib oma tegevuses kodanikuühenduste eetilise tegevuse üldtunnustatud põhimõtteid, mis on lühidalt kokkuvõetud 7..."
}
```
#### unshuffled_deduplicated_eu
- **Size of downloaded dataset files:** 134.68 MB
- **Size of the generated dataset:** 363.93 MB
- **Total amount of disk used:** 498.61 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "Gure jarduerek eraikuntzarekin, elkarbizitzarekin, hirigintzarekin eta ekologiarekin dute harremana, baita ideia eta konponbideak irudikatu eta garatzearekin ere, eraikuntza sektorea hobetuz, pertsonen erosotasuna eta bizi-kalitatea hobetzeko."
}
```
#### unshuffled_deduplicated_fa
- **Size of downloaded dataset files:** 10.46 GB
- **Size of the generated dataset:** 40.06 GB
- **Total amount of disk used:** 50.52 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"قـــــــــــــــــرار بود با هم کنـــــــــــــار بیایم نه اینکه از کنــــــــــــار هم رد بشیم...!!!\\nاگر روزی دلت لبریز غم بو..."
}
```
#### unshuffled_deduplicated_fi
- **Size of downloaded dataset files:** 5.38 GB
- **Size of the generated dataset:** 13.99 GB
- **Total amount of disk used:** 19.37 GB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Kiitos Deelle kaikesta - 1,5 viikkoa kulunut, kun Dee ei ole enää ollut omani. Reilu viikko sitten sunnuntaina vein Deen uuteen kotiinsa. Itselläni on ollut niin ristiriitaiset t..."
}
```
#### unshuffled_deduplicated_fr
- **Size of downloaded dataset files:** 55.46 GB
- **Size of the generated dataset:** 148.28 GB
- **Total amount of disk used:** 203.75 GB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "Média de débat d'idées, de culture et de littérature. Récits, décryptages, analyses, portraits et critiques autour de la vie des idées. Magazine engagé, ouvert aux autres et au monde.. Bring up to date in french"
}
```
#### unshuffled_deduplicated_frr
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Hiragana’ Practice’Sheet’1’(A -O)’ ’ Name:’________ __________________________’Section:’_______________ _’ ’ ’ ’ ’ ’ ’ ’ ’ ’ ’ ..."
}
```
#### unshuffled_deduplicated_fy
- **Size of downloaded dataset files:** 10.27 MB
- **Size of the generated dataset:** 26.73 MB
- **Total amount of disk used:** 37.00 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Nim in sêfte ride op Holmsjön, yn ien fan 'e lytse marren yn de omkriten, of nim se op avontueren lykas nonresidential. lâns Indalsälven wetter. Holm Sportklubb hawwe kano 's te huur, yn gearwurking mei de Baltyske Power konferinsje."
}
```
#### unshuffled_deduplicated_ga
- **Size of downloaded dataset files:** 22.22 MB
- **Size of the generated dataset:** 63.86 MB
- **Total amount of disk used:** 86.08 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Is fóram é seo chun plé a dhéanamh ar an leabhar atá roghnaithe do mhí na Samhna 2013 amháin. Ní féidir ach le baill chláraithe..."
}
```
#### unshuffled_deduplicated_gd
- **Size of downloaded dataset files:** 0.42 MB
- **Size of the generated dataset:** 1.36 MB
- **Total amount of disk used:** 1.78 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "Zhou Yujun, a 'phàrtaidh Rùnaire Comataidh Sgìre Yanfeng ann Hengyang bhaile agus a Sgìre pàrtaidh agus an riaghaltas a' bhuidheann-riochdachaidh a 'tighinn a chèilidh air ar companaidh air Apr. 14, 2017."
}
```
#### unshuffled_deduplicated_gl
- **Size of downloaded dataset files:** 155.85 MB
- **Size of the generated dataset:** 408.34 MB
- **Total amount of disk used:** 564.19 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"O persoal de Inditex da provincia de Pontevedra segue a reclamar iguais condicións laborais no conxunto do país - CIG: Confeder..."
}
```
#### unshuffled_deduplicated_gn
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.02 MB
- **Total amount of disk used:** 0.03 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"º ÑÆÚÓ À Ã Ð É Æ ¾ ÄÂ Î À ¼ Æ É ÄÛ = Ü Ý\\\"Þ ßà á â ã ä å æçè ã é ê â å àë ì æê íî é á ë ï í çì àð í Ü à ñ ê é ò ä ì\"..."
}
```
#### unshuffled_deduplicated_gom
- **Size of downloaded dataset files:** 0.38 MB
- **Size of the generated dataset:** 1.87 MB
- **Total amount of disk used:** 2.24 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"दुष्ट शीळ हें कौरवांचें । रामें सविस्तर देखूनि साचें । बोलिले वचनें जें दुर्वाचे । करी तयांचें अनुस्मरण ॥२२०॥\"..."
}
```
#### unshuffled_deduplicated_gu
- **Size of downloaded dataset files:** 162.97 MB
- **Size of the generated dataset:** 759.34 MB
- **Total amount of disk used:** 922.32 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"અધિક માસ ચાલે છે. સમગ્ર ભારતમાં અને તેમાંય ખાસ કરીને પવિત્ર કે ધાર્મિક કહેવાય છે તેવા સ્થાનક પર કથાનો દોર ચાલે છે. ઉનાળાની કાળઝ..."
}
```
#### unshuffled_deduplicated_he
- **Size of downloaded dataset files:** 3.04 GB
- **Size of the generated dataset:** 10.47 GB
- **Total amount of disk used:** 13.51 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"זקוקים לרשתות נגד יתושים? מחפשים רשת מתאימה לחלון צר וקטן? רשתות נגד יתושים אקורדיון של חברת קליר-מש הן הפתרון.\\nרשתות לחלונות ..."
}
```
#### unshuffled_deduplicated_hi
- **Size of downloaded dataset files:** 2.01 GB
- **Size of the generated dataset:** 9.57 GB
- **Total amount of disk used:** 11.58 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"'आइटम गर्ल' बनकर हिट हुई थीं राखी सावंत, आज करीना-कटरीना तक फॉलो कर रही हैं ट्रेंड नक्सलियों का दम निकालेगा बाइक ग्रेनेड लॉन्च..."
}
```
#### unshuffled_deduplicated_hr
- **Size of downloaded dataset files:** 46.74 MB
- **Size of the generated dataset:** 121.50 MB
- **Total amount of disk used:** 168.23 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"U raspravi je sudjelovao i HSS-ov saborski zastupnik rekavši kako poljoprivrednici ne osjete mjere o kojima ministar govori jer..."
}
```
#### unshuffled_deduplicated_hsb
- **Size of downloaded dataset files:** 0.72 MB
- **Size of the generated dataset:** 1.89 MB
- **Total amount of disk used:** 2.61 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Budyšin (SN/BŠe). Elektronikarjo mějachu lětsa cyle hinaši zazběh do swojeho wukubłanja. Wokrjesne rjemjeslnistwo bě mjenujcy w..."
}
```
#### unshuffled_deduplicated_ht
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan..."
}
```
#### unshuffled_deduplicated_hu
- **Size of downloaded dataset files:** 7.37 GB
- **Size of the generated dataset:** 19.09 GB
- **Total amount of disk used:** 26.46 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"monster - Amatőr, házi szex videók és kezdő csjaok pornó filmjei. - Free amateur, home made sex videos and online porn movies. ..."
}
```
#### unshuffled_deduplicated_hy
- **Size of downloaded dataset files:** 393.62 MB
- **Size of the generated dataset:** 1.56 GB
- **Total amount of disk used:** 1.96 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Արցախի Հանրապետության հռչակման 26-րդ տարեդարձի կապակցությամբ Շուշիի Արվեստի կենտրոնում կազմակերպվել է մոսկվաբնակ նկարիչներ՝ հայ..."
}
```
#### unshuffled_deduplicated_ia
- **Size of downloaded dataset files:** 0.05 MB
- **Size of the generated dataset:** 0.38 MB
- **Total amount of disk used:** 0.43 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha h..."
}
```
#### unshuffled_deduplicated_id
- **Size of downloaded dataset files:** 6.00 GB
- **Size of the generated dataset:** 17.05 GB
- **Total amount of disk used:** 23.05 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Perihal dari itu, kalau kunci hal yang demikian hilang, pemilik wajib melapor ke bengkel sah untuk dibuatkan kunci baru dengan ..."
}
```
#### unshuffled_deduplicated_ie
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "Plastic Yo Yo Metal Yo Yos Wooden Yo Yo Keychain Yo Yo Translucent Yo Yo Light Up Yo Yo Globe Yo Yo Stress Reliever Yo Yo Jellyfish Yo Yo Sports Ball Yo Yo Sound Yo Yo Miniature Yo Yo Promotional Yo Yo Novelty Yo Yo Video Game Yo Yo ECO Recycled Yo Yo"
}
```
#### unshuffled_deduplicated_ilo
- **Size of downloaded dataset files:** 0.23 MB
- **Size of the generated dataset:** 0.68 MB
- **Total amount of disk used:** 0.91 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Segun ken ni Ping-ay, ti yellow corn ti maysa kadagiti nadakamat a liberalized agricultural commodity iti daytoy a free trade k..."
}
```
#### unshuffled_deduplicated_io
- **Size of downloaded dataset files:** 0.04 MB
- **Size of the generated dataset:** 0.14 MB
- **Total amount of disk used:** 0.19 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Chekia esas parlamentala republiko. La chefo di stato esas la prezidanto. Til 2013 lu elektesis dal parlamento. Pos ta yaro, ol..."
}
```
#### unshuffled_deduplicated_is
- **Size of downloaded dataset files:** 332.87 MB
- **Size of the generated dataset:** 894.28 MB
- **Total amount of disk used:** 1.23 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Eyjar.net - upplýsinga- og fréttamiðill um Vestmannaeyjar - Fréttir - Nái núverandi stefna stjórnvalda fram að ganga mun það va..."
}
```
#### unshuffled_deduplicated_it
- **Size of downloaded dataset files:** 27.93 GB
- **Size of the generated dataset:** 74.09 GB
- **Total amount of disk used:** 102.03 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Jaundice - causes, treatment & pathology massaggio a osteochondrosis dellindizio di una controindicazione\\nTrattamento su un co..."
}
```
#### unshuffled_deduplicated_ja
- **Size of downloaded dataset files:** 40.80 GB
- **Size of the generated dataset:** 113.63 GB
- **Total amount of disk used:** 154.44 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"神社などへ一緒に同行して、様々な角度のショットで家族写真やお子様の写真を撮影致します!お好みに合わせて様々な写真を取ることができますので、その場でカメラマンへのリクエストも可能です!お子様の晴れ姿を、緊張していない自然な笑顔で残しませんか?\\n※七五三の..."
}
```
#### unshuffled_deduplicated_jbo
- **Size of downloaded dataset files:** 0.20 MB
- **Size of the generated dataset:** 0.70 MB
- **Total amount of disk used:** 0.91 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "ni'o 23 la cimast. cu 23moi djedi fi'o masti la cimast. noi ke'a cu cimoi masti .i 22 la cimast. cu purlamdei .ije 24 la cimast. cu bavlamdei"
}
```
#### unshuffled_deduplicated_jv
- **Size of downloaded dataset files:** 0.21 MB
- **Size of the generated dataset:** 0.62 MB
- **Total amount of disk used:** 0.82 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"José Mourinho (diwaca: [ʒuˈzɛ moˈɾiɲu]; lair ing Setubal, Portugal, 26 Januari 1963; umur 55 taun) iku salah siji pelatih bal k..."
}
```
#### unshuffled_deduplicated_ka
- **Size of downloaded dataset files:** 377.23 MB
- **Size of the generated dataset:** 1.99 GB
- **Total amount of disk used:** 2.36 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"წამიყვანე შენთან ერთად (ქართულად) / Возьми меня с собой (картулад) / (რუსული სერიალები ქართულად) (რუსების პორნო ონლაინში) (ruse..."
}
```
#### unshuffled_deduplicated_kk
- **Size of downloaded dataset files:** 389.12 MB
- **Size of the generated dataset:** 1.59 GB
- **Total amount of disk used:** 1.97 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Түлкібас ауданында «Латын негізді әліпби мен емле ережесі туралы насихат» жобасының тобы семинар өткізді\\nЕлорданың «Қазақстан»..."
}
```
#### unshuffled_deduplicated_km
- **Size of downloaded dataset files:** 114.48 MB
- **Size of the generated dataset:** 610.61 MB
- **Total amount of disk used:** 725.09 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ខ្សឹបដាក់ត្រចៀក៖ លោក សួស សុផានិត នាយផ្នែករដ្ឋបាលព្រៃឈើ ស្រុកភ្នំក្រវាញ់ ដែលទើបឡើងកាន់តំណែងថ្មី បើកដៃឲ្យឈ្នួញ ប្រព្រឹត្តបទល្មើស ..."
}
```
#### unshuffled_deduplicated_kn
- **Size of downloaded dataset files:** 215.52 MB
- **Size of the generated dataset:** 1.08 GB
- **Total amount of disk used:** 1.30 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ರಾಷ್ಟ್ರಪತಿ ಪ್ರಣಬ್ ಮುಖರ್ಜಿಯಿಂದ ಪದ್ಮ ಪ್ರಶಸ್ತಿ ಪ್ರದಾನ | President Pranab Mukherjee Confers Padma Awards | Photo Gallery on Kannada..."
}
```
#### unshuffled_deduplicated_ko
- **Size of downloaded dataset files:** 4.46 GB
- **Size of the generated dataset:** 12.00 GB
- **Total amount of disk used:** 16.47 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"CIA 프로젝트에서는 데이터베이스로 들어오는 요청을 중간에 수집(Sniffing)하고 수집한 데이터를 분석(Parsing)하여 그로 인한 결과를 판단하여 알릴 수 있는 시스템(Push Service)이 필요하다. 그리고 연구를 ..."
}
```
#### unshuffled_deduplicated_krc
- **Size of downloaded dataset files:** 0.62 MB
- **Size of the generated dataset:** 2.41 MB
- **Total amount of disk used:** 3.03 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Шамханланы, Бийлени къаршысына ябушуп, Батыр уланларыбызны къоллары булан «ортакъ ожакъ» къургъанбыз. Шо иш уллу зараллы иш бол..."
}
```
#### unshuffled_deduplicated_ku
- **Size of downloaded dataset files:** 23.34 MB
- **Size of the generated dataset:** 63.09 MB
- **Total amount of disk used:** 86.43 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Me di 114 bernameyên xwe yên berê da perçeyên ji berhemên zanyarî yên kurdzanên mezin bi wergera kurdî da ...\\nMe di 114 bernam..."
}
```
#### unshuffled_deduplicated_kv
- **Size of downloaded dataset files:** 0.33 MB
- **Size of the generated dataset:** 1.21 MB
- **Total amount of disk used:** 1.54 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Коми кытшыслӧн ыджытжык тор вӧр увтын куйлӧ, сійӧн и фаунасӧ татӧн аркмӧтӧны вӧрын олісь подаэз. Ассямаӧн лоӧ сія, мый кытшас с..."
}
```
#### unshuffled_deduplicated_kw
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.02 MB
- **Total amount of disk used:** 0.02 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼Pray without ceasing🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏..."
}
```
#### unshuffled_deduplicated_ky
- **Size of downloaded dataset files:** 106.22 MB
- **Size of the generated dataset:** 408.40 MB
- **Total amount of disk used:** 514.61 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Turmush: Бишкек шаардык кеңешинин кезексиз отурумунда мэрге ишенбөөчүлүк көрсөтүү маселеси каралат, - депутат Т.Сагынов\\nБишкек..."
}
```
#### unshuffled_deduplicated_la
- **Size of downloaded dataset files:** 3.42 MB
- **Size of the generated dataset:** 9.79 MB
- **Total amount of disk used:** 13.22 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Hæ sunt generationes Noë: Noë vir justus atque perfectus fuit in generationibus suis; cum Deo ambulavit.\\nEcce ego adducam aqua..."
}
```
#### unshuffled_deduplicated_lb
- **Size of downloaded dataset files:** 8.30 MB
- **Size of the generated dataset:** 21.42 MB
- **Total amount of disk used:** 29.72 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Während dem Gaardefestival \\\"Ambiance Jardins\\\" vum 15. bis de 17. Mee huet den SNJ nees zesumme mam Groupe Animateur en Inform..."
}
```
#### unshuffled_deduplicated_lez
- **Size of downloaded dataset files:** 0.77 MB
- **Size of the generated dataset:** 3.08 MB
- **Total amount of disk used:** 3.84 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Ахцегь хуьр, виридалай ч1ехи лезги хуьрерикая я. Ам Урусатдин виридалай къиблепатавай хуьрерикай я. Ин хуьр...\"..."
}
```
#### unshuffled_deduplicated_li
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.03 MB
- **Total amount of disk used:** 0.04 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"'t Good Goedenraad aan de Ezerbaek besjteit oet 'n kesjtièl mèt gesjlote haof en 'n park van 26 hectare. Hie in sjtoon väól beu..."
}
```
#### unshuffled_deduplicated_lmo
- **Size of downloaded dataset files:** 0.10 MB
- **Size of the generated dataset:** 0.46 MB
- **Total amount of disk used:** 0.57 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Serét (en tortonés: Sregh; en piemontés: Srèj) l'è 'n cümü italià, de la regiù del Piemónt, en Pruvìncia de Alessandria. El g'h..."
}
```
#### unshuffled_deduplicated_lo
- **Size of downloaded dataset files:** 23.63 MB
- **Size of the generated dataset:** 119.29 MB
- **Total amount of disk used:** 142.92 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"ຜູ້ພິພາກສາ ປະຈຳເຂດ ສຫລ ທ່ານນຶ່ງ ຕັດສິນວ່າ ໂຄງການເກັບກຳຂໍ້ມູນ ທາງໂທລະສັບ ຂອງອົງການ ຄວາມໝັ້ນຄົງແຫ່ງຊາດ ແມ່ນຖືກຕ້ອງ ຕາມກົດໝາຍ.\\nກະ..."
}
```
#### unshuffled_deduplicated_lrc
- **Size of downloaded dataset files:** 0.02 MB
- **Size of the generated dataset:** 0.06 MB
- **Total amount of disk used:** 0.08 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"آرلینگتون یئ گئل د شأریا ڤولاتچە ڤیرجینیا و یئ گئل د شأریا ڤولات ڤولاتچە یا یأکاگئرئتە ئمریکاە. ئی شأر دویومی کألوٙن شأر د راسا..."
}
```
#### unshuffled_deduplicated_lt
- **Size of downloaded dataset files:** 1.65 GB
- **Size of the generated dataset:** 4.20 GB
- **Total amount of disk used:** 5.86 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Čir vir vir pavasaris! Čia čia čia… dalinamės labai simpatiška video pamokėle, kurią pristato ab888art galerija.\\nBe galo papra..."
}
```
#### unshuffled_deduplicated_lv
- **Size of downloaded dataset files:** 710.45 MB
- **Size of the generated dataset:** 1.91 GB
- **Total amount of disk used:** 2.62 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Dekoratīvi sliekšņi MITSUBISHI OUTLANDER 2007, izgatavoti no ovālas formas, pulētas nerūsējošā tērauda caurules...\\ndažādas tūn..."
}
```
#### unshuffled_deduplicated_mai
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.01 MB
- **Total amount of disk used:** 0.01 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"१ · २ · ३ · ४ · ५ · ६ · ७ · ८ · ९ · १० · ११ · १२ · १३ · १४ · १५ · १६ · १७ · १८ · १९ · २० · २१ · २२ · २३ · २४ · २५ · २६ · २७ · २..."
}
```
#### unshuffled_deduplicated_mg
- **Size of downloaded dataset files:** 4.30 MB
- **Size of the generated dataset:** 13.59 MB
- **Total amount of disk used:** 17.89 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Nanamboatra taratasy apetaka sy soso-kevitra ho an'ny olona te-hanatevin-daharana ity fihetsiketsehana ity i Anocrena.\\nNosorat..."
}
```
#### unshuffled_deduplicated_mhr
- **Size of downloaded dataset files:** 1.63 MB
- **Size of the generated dataset:** 6.26 MB
- **Total amount of disk used:** 7.89 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Акрет жап годым Уганда кундемым Пигмей племена- влак айлен шогеныт. мемнан эран 1 курым гыч Банту племена влакат тиде кундемышк..."
}
```
#### unshuffled_deduplicated_min
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.31 MB
- **Total amount of disk used:** 0.33 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\" ..."
}
```
#### unshuffled_deduplicated_mk
- **Size of downloaded dataset files:** 303.12 MB
- **Size of the generated dataset:** 1.19 GB
- **Total amount of disk used:** 1.49 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"„Филм плус“ е насловен првиот филмски месечник во Македонија, чиј прв број ќе биде промовиран вечер во „Менада“. Новото македон..."
}
```
#### unshuffled_deduplicated_ml
- **Size of downloaded dataset files:** 496.80 MB
- **Size of the generated dataset:** 2.69 GB
- **Total amount of disk used:** 3.18 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"സ്ത്രീ പ്രവേശനം സര്ക്കാര് പൂര്ണമായും അംഗീകരിക്കുന്നുവെന്നും ശബരിമലയുടെ സുരക്ഷയില് ഇടപെടുമെന്നും സര്ക്കാര് ഹൈക്കോടതിയില്\\..."
}
```
#### unshuffled_deduplicated_mn
- **Size of downloaded dataset files:** 219.52 MB
- **Size of the generated dataset:** 883.46 MB
- **Total amount of disk used:** 1.10 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"МУБИС-ын багш мэргэжлийн хөрвөх сургалтыг төгссөн багшид багшлах эрх олгох тухай ~ БМДИ-ийн захирлын тушаал - Багшийн мэргэжил ..."
}
```
#### unshuffled_deduplicated_mr
- **Size of downloaded dataset files:** 299.68 MB
- **Size of the generated dataset:** 1.49 GB
- **Total amount of disk used:** 1.79 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Home / motivational marathi story / उद्योजकता (Entrepreneurship) / यांना हे जमलय, तर आपल्याला का नाही जमणार ?\\nयापैकी कोणाचीही ..."
}
```
#### unshuffled_deduplicated_mrj
- **Size of downloaded dataset files:** 0.29 MB
- **Size of the generated dataset:** 1.10 MB
- **Total amount of disk used:** 1.38 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Лӹпӹвлӓ (латинлӓ Lepidoptera ; алыкмарла лыве-влак) — капшангывлӓ йыхыш пырышы сӱмӓн нӹл шылдыран капшангывлӓ. Цилӓжӹ 180000 тӹ..."
}
```
#### unshuffled_deduplicated_ms
- **Size of downloaded dataset files:** 16.39 MB
- **Size of the generated dataset:** 49.45 MB
- **Total amount of disk used:** 65.85 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Sanad pertama daripada Zuhair bin Harb daripada ‘Affan daripada Hammad daripada Thabit daripada Anas.\\nSanad kedua daripada ‘Ab..."
}
```
#### unshuffled_deduplicated_mt
- **Size of downloaded dataset files:** 5.90 MB
- **Size of the generated dataset:** 17.68 MB
- **Total amount of disk used:** 23.58 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "tibgħat il-kawża lura lill-Qorti Ġenerali għall-annullament jew għat-tnaqqis tal-penalità imposta mill-Kummissjoni bid-deċiżjoni inizjali kif emendata bid-deċiżjoni ta’ rettifika;"
}
```
#### unshuffled_deduplicated_mwl
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Deciplina social i outónoma que angloba atebidades de ouserbaçon, de análeze, de çcriçon, cumparaçon, de sistematizaçon i de sp..."
}
```
#### unshuffled_deduplicated_my
- **Size of downloaded dataset files:** 207.14 MB
- **Size of the generated dataset:** 1.11 GB
- **Total amount of disk used:** 1.32 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ျမ၀တီ - ရန္ကုန္တိုင္းေဒသႀကီး ေျမာက္ဥကၠလာပႏွင္႕ ဗဟန္းၿမိဳ႔နယ္ မေကြးတိုင္း ေဒသႀကီး ပခုကၠဴၿမိဳ႔နယ္တို႔၌ ျမန္မာ႕တပ္မေတာ္အား ေထာက္ခံ..."
}
```
#### unshuffled_deduplicated_myv
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"2018 иень умарьковонь 6-це чистэ сась паро куля! Россиянь культурань Министерствась макссь невтемань конёв (прокатной удостовер..."
}
```
#### unshuffled_deduplicated_mzn
- **Size of downloaded dataset files:** 0.16 MB
- **Size of the generated dataset:** 0.63 MB
- **Total amount of disk used:** 0.79 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"قرآن یا قوران اسلام ِآسمونی کتاب هسته. مسلمونون گانّّه قرآن ره خدا، وحی جه برسنییه، «محمد معجزه» هسته و ثقلین حدیث دله ونه خَو..."
}
```
#### unshuffled_deduplicated_nah
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.01 MB
- **Total amount of disk used:** 0.01 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "In mācuīlpōhualxihuitl VI (inic chicuacē) in mācuīlpōhualli xiuhitl cāhuitl īhuīcpa 501 xihuitl oc 600 xihuitl."
}
```
#### unshuffled_deduplicated_nap
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.01 MB
- **Total amount of disk used:** 0.02 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ò AUDIT í Ç è î ÿ å å 30 ò ÿ ÿ é, õ ñ ì ÿ, ê ã- ò à ì. å â å í ç â à à é ñ è å é ó ó ë. å å å û è å î é è à. à è à AUDIT 1-7 â ..."
}
```
#### unshuffled_deduplicated_nds
- **Size of downloaded dataset files:** 5.27 MB
- **Size of the generated dataset:** 13.48 MB
- **Total amount of disk used:** 18.76 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Dor kann sik vun nu af an de hele plattdüütsche Welt – vun Niebüll bit New York, vun Helgoland bit Honolulu – drapen. Allens, w..."
}
```
#### unshuffled_deduplicated_ne
- **Size of downloaded dataset files:** 240.63 MB
- **Size of the generated dataset:** 1.24 GB
- **Total amount of disk used:** 1.48 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"बर्दिबास नगरपालिकाको तेस्रो नगर परिषदबाट पारित आ.व.२०७३।७४ को संशोधित र २०७४।७५ को प्रस्तावित नीति, कार्यक्रम तथा बजेट\\nअार्थिक..."
}
```
#### unshuffled_deduplicated_new
- **Size of downloaded dataset files:** 0.83 MB
- **Size of the generated dataset:** 4.26 MB
- **Total amount of disk used:** 5.09 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"थ्व शहरयागु अक्षांश ३४.७००१६४ उत्तर व देशान्तर ८६.३७६४६९ पश्चिम खः (34.700164° N 86.376469° W)। थ्व थासे ७२२६७३२ वर्ग मिटर (२.७..."
}
```
#### unshuffled_deduplicated_nl
- **Size of downloaded dataset files:** 15.73 GB
- **Size of the generated dataset:** 41.91 GB
- **Total amount of disk used:** 57.65 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Op vrijdag 31 augustus wordt het nieuwe studiejaar van de masteropleiding architectuur geopend met een dagexcursie naar Venlo.\\..."
}
```
#### unshuffled_deduplicated_nn
- **Size of downloaded dataset files:** 23.58 MB
- **Size of the generated dataset:** 58.32 MB
- **Total amount of disk used:** 81.90 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "Planomtale krav til innhald Bakgrunn: Spørsmål frå fleire kommunar om kva ein planomtale/planbeskrivelse bør innehalde Fylkeskommunen og fylkesmannen har i ein del saker reist motsegn på formelt grunnlag"
}
```
#### unshuffled_deduplicated_no
- **Size of downloaded dataset files:** 1.96 GB
- **Size of the generated dataset:** 5.11 GB
- **Total amount of disk used:** 7.07 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Ytterligere aktører i primærhelsetjenesten og andre NHS-virksomheter ble infisert, inkludert legekontor.Læreren vår er så attra..."
}
```
#### unshuffled_deduplicated_oc
- **Size of downloaded dataset files:** 1.34 MB
- **Size of the generated dataset:** 4.00 MB
- **Total amount of disk used:** 5.34 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": ".рф (rf, còdi punycode: .xn--p1ai)[1] es lo nom de domeni en rus per Russia. Foguèt activat lo 12 de mai de 2010. Lo còdi latin es .ru."
}
```
#### unshuffled_deduplicated_or
- **Size of downloaded dataset files:** 38.72 MB
- **Size of the generated dataset:** 197.63 MB
- **Total amount of disk used:** 236.36 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ଭୁବନେଶ୍ୱର, ୨୭/୧– (ଓଡ଼ିଆ ପୁଅ) ସିପିଆଇ ଜାତୀୟ ପରିଷଦର ଆହ୍ୱାନକ୍ରମେ ଗତକାଲି ଜାନୁୟାରୀ ୨୬ ସାଧାରଣତନ୍ତ୍ର ଦିବସକୁ ଦେଶ ବ୍ୟାପୀ ସମ୍ବିଧାନ ସୁରକ୍ଷା ..."
}
```
#### unshuffled_deduplicated_os
- **Size of downloaded dataset files:** 2.83 MB
- **Size of the generated dataset:** 11.00 MB
- **Total amount of disk used:** 13.83 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"1. Лæппу æмæ чызг казрæдзийы зæрдæмæ куы фæцæуынц æмæ, куы сфæнд кæнынц сæ цард баиу кæнын, уæд лæппу бар ракуры чызгæй, цæмæй ..."
}
```
#### unshuffled_deduplicated_pa
- **Size of downloaded dataset files:** 102.39 MB
- **Size of the generated dataset:** 483.04 MB
- **Total amount of disk used:** 585.42 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ਰਜਿ: ਨੰ: PB/JL-138/2018-20 ਜਿਲਦ 63, ਬਾਨੀ ਸੰਪਾਦਕ (ਸਵ:) ਡਾ: ਸਾਧੂ ਸਿੰਘ ਹਮਦਰਦ ਫ਼ੋਨ : 0181-2455961-62-63, 5032400, ਫੈਕਸ : 2455960, 2..."
}
```
#### unshuffled_deduplicated_pam
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Áku pu i Anak ning Aláya at ngeni ipákit kó kékayu ngan nûng makanánu lang susúlat détinang kulit a mágkas. Lauan ya ing tarátu..."
}
```
#### unshuffled_deduplicated_pl
- **Size of downloaded dataset files:** 20.19 GB
- **Size of the generated dataset:** 50.59 GB
- **Total amount of disk used:** 70.78 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"System informatyczny - Załącznik nr 1 do zarządzenia Wójta Gminy Podegrodzie Nr 530/2013 z dnia 27 maja 2013 r\\nSystem informat..."
}
```
#### unshuffled_deduplicated_pms
- **Size of downloaded dataset files:** 0.71 MB
- **Size of the generated dataset:** 2.00 MB
- **Total amount of disk used:** 2.72 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Louvigné-du-Désert a l'é na comun-a fransèisa ant la region aministrativa dla Brëtagna, ant ël dipartiment d'Ille-et-Vilaine. A..."
}
```
#### unshuffled_deduplicated_pnb
- **Size of downloaded dataset files:** 2.58 MB
- **Size of the generated dataset:** 9.44 MB
- **Total amount of disk used:** 12.02 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"ایہ فائل Wikimedia Commons توں اے تے دوجیاں ویونتاں تے وی ورتی جاےکدی اے۔ گل بات اس دے فائل گل بات صفہ تے تھلے دتی گئی۔\"..."
}
```
#### unshuffled_deduplicated_ps
- **Size of downloaded dataset files:** 71.83 MB
- **Size of the generated dataset:** 254.79 MB
- **Total amount of disk used:** 326.61 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Many people usually use the time period ‘business to business (B2B) advertising,’ however most of them do not know precisely wh..."
}
```
#### unshuffled_deduplicated_pt
- **Size of downloaded dataset files:** 26.00 GB
- **Size of the generated dataset:** 68.37 GB
- **Total amount of disk used:** 94.37 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Você pode estar lendo este texto no sofá, levantar pra pegar uma breja na geladeira, dar uma cagada e sentar novamente, sem int..."
}
```
#### unshuffled_deduplicated_qu
- **Size of downloaded dataset files:** 0.02 MB
- **Size of the generated dataset:** 0.07 MB
- **Total amount of disk used:** 0.09 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Warayu wichay (kastilla simipi: Ascensión de Guarayos) nisqaqa Buliwya mama llaqtapi, Santa Krus suyupi, huk llaqtam, Warayu pruwinsyap uma llaqtanmi."
}
```
#### unshuffled_deduplicated_rm
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.01 MB
- **Total amount of disk used:** 0.01 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"practicists agrars / practicistas agraras AFP pon far ina furmaziun da basa scursanida per cuntanscher in attestat federal da q..."
}
```
#### unshuffled_deduplicated_ro
- **Size of downloaded dataset files:** 4.48 GB
- **Size of the generated dataset:** 11.66 GB
- **Total amount of disk used:** 16.14 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"“În viață, oportunitatea nu este totul. Cine atrage Lumina, cineva bun în umbră. Timpul ne creează.” maestru\\nLyn.Evans: Ce mar..."
}
```
#### unshuffled_deduplicated_ru
- **Size of downloaded dataset files:** 166.68 GB
- **Size of the generated dataset:** 611.70 GB
- **Total amount of disk used:** 778.38 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Доступ к данному профилю для публичного просмотра закрыт администрацией сайта - профиль находится на модерации.\\nРазработчикам ..."
}
```
#### unshuffled_deduplicated_sa
- **Size of downloaded dataset files:** 7.27 MB
- **Size of the generated dataset:** 38.33 MB
- **Total amount of disk used:** 45.60 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"अनिरुद्धनगरे क्रीडिता रामलीला सम्प्रति समाप्ता अस्ति । तस्य कानिचन् चित्राणि पूर्वमेव प्रकाशितानि सन्ति । द्वौ चलचित्रौ अपि ..."
}
```
#### unshuffled_deduplicated_sah
- **Size of downloaded dataset files:** 7.01 MB
- **Size of the generated dataset:** 27.46 MB
- **Total amount of disk used:** 34.49 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████..."
}
```
#### unshuffled_deduplicated_scn
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "La gilusìa è nu sintimentu dulurusu ca nasci d'un disideriu di pussessu sclusivu ntê cunfrunti dâ pirsuna amata e dû timuri, dû suspettu o dâ cirtizza dâ sò nfidiltati."
}
```
#### unshuffled_deduplicated_sd
- **Size of downloaded dataset files:** 74.17 MB
- **Size of the generated dataset:** 275.48 MB
- **Total amount of disk used:** 349.66 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"هر ڪو ڄاڻي ٿو ته جڏهن توهان هڪ وڏي خريد ڪرڻ چاهيون ٿا, توهان پڄي ضروري حڪم ۾ ان جي ڪم ڪرڻ جي هٿ ۾ لاڳاپو ڪيو آهي. جي شيء آهي ته..."
}
```
#### unshuffled_deduplicated_sh
- **Size of downloaded dataset files:** 1.45 MB
- **Size of the generated dataset:** 6.44 MB
- **Total amount of disk used:** 7.87 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Opština Gornja Radgona se nalazi u sjeveroistočnoj Sloveniji i graniči s susjednom Austriji duž rijeke Mure. Sa tridesetim nase..."
}
```
#### unshuffled_deduplicated_si
- **Size of downloaded dataset files:** 175.62 MB
- **Size of the generated dataset:** 842.57 MB
- **Total amount of disk used:** 1.02 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"ලාංකීය සිතිවිලි සිංහල බ්ලොග් කියවනය කොත්තු සින්ඩිය ලංකා Blogger හත්මාළුව ලංකා බ්ලොග් කියවනය මාතලන්ගේ සින්ඩිය මොබයිල්lk\\nඅවකාශය ..."
}
```
#### unshuffled_deduplicated_sk
- **Size of downloaded dataset files:** 1.96 GB
- **Size of the generated dataset:** 4.80 GB
- **Total amount of disk used:** 6.76 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Aktivity | Agentúra podporovaného zamestnávania | vzdelávanie pre klientov, vzdelávanie pre odborníkov, kurzy\\nŠpecializované k..."
}
```
#### unshuffled_deduplicated_sl
- **Size of downloaded dataset files:** 523.22 MB
- **Size of the generated dataset:** 1.32 GB
- **Total amount of disk used:** 1.85 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Če Creatures, ki je želel, da pridejo na čas, predvsem je povedlo – razlikuje od ljubosumja začel grizenja kolen (ali zadnjica)..."
}
```
#### unshuffled_deduplicated_so
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.02 MB
- **Total amount of disk used:** 0.02 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"тттттттттттттттттттттттттттттттт тттттттттттттттттттттттттттттттт тттттттттттттттттттттттттттттттт ттттттттттттттттуууууууууууу..."
}
```
#### unshuffled_deduplicated_sq
- **Size of downloaded dataset files:** 445.36 MB
- **Size of the generated dataset:** 1.21 GB
- **Total amount of disk used:** 1.66 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Çfarë do të më pëlqente tek një femër ose çfarë do të më shndërronte në një shpërthim drite? – Albert Vataj\\nTë gjithëve një zo..."
}
```
#### unshuffled_deduplicated_sr
- **Size of downloaded dataset files:** 665.03 MB
- **Size of the generated dataset:** 2.36 GB
- **Total amount of disk used:** 3.03 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Корисни савети за сваки дан. На сајту су разне категорије, као што су љепота, мода, кување и поправка властитим рукама.\\nШколск..."
}
```
#### unshuffled_deduplicated_su
- **Size of downloaded dataset files:** 0.05 MB
- **Size of the generated dataset:** 0.16 MB
- **Total amount of disk used:** 0.21 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Kartu krédit nyaéta \"duit plastik\" anu dikaluarkeun ku bank pikeun alat pambayaran di tempat-tempat nu tangtu samisal jiga di hotél, réstoran, tempat rékréasi jeung sajabana.[1]"
}
```
#### unshuffled_deduplicated_sv
- **Size of downloaded dataset files:** 10.19 GB
- **Size of the generated dataset:** 26.33 GB
- **Total amount of disk used:** 36.51 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"1783 är ett viktigt årtal i den nya tidens historia. Det året slöts en fred i Paris och därmed blev de 13 brittiska kolonierna ..."
}
```
#### unshuffled_deduplicated_sw
- **Size of downloaded dataset files:** 2.95 MB
- **Size of the generated dataset:** 8.98 MB
- **Total amount of disk used:** 11.92 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Miripuko hiyo inakuja mwanzoni mwa Wiki Takatifu kuelekea Pasaka na ikiwa ni wiki chache tu kabla ya Papa Francis kuanza ziara yake katika nchi hiyo yenye idadi kubwa kabisa ya watu katika ulimwengu wa nchi za Kiarabu."
}
```
#### unshuffled_deduplicated_ta
- **Size of downloaded dataset files:** 971.12 MB
- **Size of the generated dataset:** 5.48 GB
- **Total amount of disk used:** 6.45 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"பொழுது சாய்ந்து வெகு நேரமாகிவிட்டது. கூலி வேலைக்குப் போயிருந்த 'சித்தாள் ' பெண்கள் எல்லோரும் வீடு திரும்பி விட்டார்கள். இன்னும்..."
}
```
#### unshuffled_deduplicated_te
- **Size of downloaded dataset files:** 342.43 MB
- **Size of the generated dataset:** 1.70 GB
- **Total amount of disk used:** 2.04 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"హర్యానాలో టోల్ దగ్గర సిబ్బంది.. స్థానిక ప్రజలు కొట్టుకున్నారు. కర్నాల్ అనే గ్రామానికి సమీపంలో టోల్ గేట్ ఉంది. అయితే సాధారణంగా స..."
}
```
#### unshuffled_deduplicated_tg
- **Size of downloaded dataset files:** 62.90 MB
- **Size of the generated dataset:** 261.68 MB
- **Total amount of disk used:** 324.60 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Ҳумайро гуфтааст, мухолифи низом аст, низоме, ки дар Тоҷикистон вуҷуд дорад. Ба ин маънӣ, худро мухолифи давлату ҳукумати Тоҷик..."
}
```
#### unshuffled_deduplicated_th
- **Size of downloaded dataset files:** 3.54 GB
- **Size of the generated dataset:** 17.11 GB
- **Total amount of disk used:** 20.65 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ฟันที่แลดูขาวสะอาดไม่มีเศษอาหารติดอยู่ เหงือกสีชมพู ไม่เจ็บ หรือมีเลือดออกเวลาแปรงฟันหรือขัดฟัน ไม่มีปัญหาเรื่องกลิ่นปาก ทำให้ก..."
}
```
#### unshuffled_deduplicated_tk
- **Size of downloaded dataset files:** 2.22 MB
- **Size of the generated dataset:** 7.12 MB
- **Total amount of disk used:** 9.34 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Türkmenistanyň Prezidenti agyr atletika boýunça dünýä çempionatyna taýýarlyk işleriniň barşy bilen tanyşdy\\nHalallykdan kemal t..."
}
```
#### unshuffled_deduplicated_tl
- **Size of downloaded dataset files:** 151.34 MB
- **Size of the generated dataset:** 431.69 MB
- **Total amount of disk used:** 583.04 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"“Gusto ko manawagan sa mga Unit Head ng Chanel 2 Salve. Kasi napapansin ko iyon mga alaga ko ang taping halos once a week lang,..."
}
```
#### unshuffled_deduplicated_tr
- **Size of downloaded dataset files:** 10.39 GB
- **Size of the generated dataset:** 28.47 GB
- **Total amount of disk used:** 38.86 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Son yıllarda görülen ay tutulmalarına göre daha etkili olacağı söylenen Kanlı veya Kırmızı Ay Tutulmasına saatler kaldı. Bu akş..."
}
```
#### unshuffled_deduplicated_tt
- **Size of downloaded dataset files:** 85.89 MB
- **Size of the generated dataset:** 321.37 MB
- **Total amount of disk used:** 407.26 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"\\\"Иремнең вафатына 40 көн узгач, Алмаз да безнең өйгә кереп үлде\\\". Арчада 35 яшьлек ир өстенә кондызлар ега башлаган агач төшк..."
}
```
#### unshuffled_deduplicated_tyv
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.01 MB
- **Total amount of disk used:** 0.01 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Экии, хүндүлуг аалчылар болгаш тыва дылдың деткикчилери! Тыва дылдың болгаш чогаалдың ховар бир башкызынга, Менги Ооржакка, ажы..."
}
```
#### unshuffled_deduplicated_ug
- **Size of downloaded dataset files:** 20.53 MB
- **Size of the generated dataset:** 86.44 MB
- **Total amount of disk used:** 106.97 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"زاڭ-ءتۇزىم | عىلىم-تەحنيكا | ءتىل-ادەبيەت | تۇرمىس | دەنە تاربيە | ساياحات-ورتا | سۋرەتتى حابار | سىر سۇحبات | ارناۋلى تاقىرىپ ..."
}
```
#### unshuffled_deduplicated_uk
- **Size of downloaded dataset files:** 8.04 GB
- **Size of the generated dataset:** 29.86 GB
- **Total amount of disk used:** 37.90 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Про надання роз'яснення (щодо форми письмового зобов'язання громадян про зворотне ввезення/вивезення товарів), Державна митна с..."
}
```
#### unshuffled_deduplicated_ur
- **Size of downloaded dataset files:** 483.59 MB
- **Size of the generated dataset:** 1.82 GB
- **Total amount of disk used:** 2.31 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"آئیے اہم اسلامی کتب کو یونیکوڈ میں انٹرنیٹ پر پیش کرنے کے لئے مل جل کر آن لائن ٹائپنگ کریں۔ محدث ٹائپنگ پراجیکٹ کے ذریعے آپ روز..."
}
```
#### unshuffled_deduplicated_uz
- **Size of downloaded dataset files:** 4.30 MB
- **Size of the generated dataset:** 12.00 MB
- **Total amount of disk used:** 16.29 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Qurama tog'lari tizmasining Toshkentdan 154 km uzoqlikdagi Toshkent-Ush yo'li yeqasidaxushmanzara tabiat qo'ynida joylashgan maydoni 30 ga.\nBolalarni sog'lomlashtirish oromgohi Bo'stonliq tumani Oqtosh muntaqasining soy-salqin gushasida joylashgan."
}
```
#### unshuffled_deduplicated_vec
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.02 MB
- **Total amount of disk used:** 0.02 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Par ogni pónto, ła derivada ła xe ła pendensa de ła reta tangente a ła curva de ła funsion f. Ła reta de cołor róso l'è senpre ..."
}
```
#### unshuffled_deduplicated_vi
- **Size of downloaded dataset files:** 10.71 GB
- **Size of the generated dataset:** 33.60 GB
- **Total amount of disk used:** 44.31 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Canh chua cá bông lau không chỉ là món ăn giải nhiệt, thanh mát ngày hè mà còn là món siêu bổ dưỡng, rất tốt cho người gầy ốm. ..."
}
```
#### unshuffled_deduplicated_vo
- **Size of downloaded dataset files:** 0.30 MB
- **Size of the generated dataset:** 2.10 MB
- **Total amount of disk used:** 2.40 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Sarniguet binon zif in ziläk: Hautes-Pyrénées, in topäd: Midi-Pyrénées, in Fransän. Sarniguet topon videtü 43°19’ 7’’ N e lunetü 0°5’ 19’’ L."
}
```
#### unshuffled_deduplicated_wa
- **Size of downloaded dataset files:** 0.08 MB
- **Size of the generated dataset:** 0.22 MB
- **Total amount of disk used:** 0.29 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Cisse pådje ci n' est co k' on djermon, dj' ô bén k' el pådje est djusse sibåtcheye, eyet co trop tene; et s' divreut ele ecråxhî ene miete."
}
```
#### unshuffled_deduplicated_war
- **Size of downloaded dataset files:** 0.55 MB
- **Size of the generated dataset:** 2.36 MB
- **Total amount of disk used:** 2.90 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "An Honce amo in usa ka baryo ngan munisipalidad ha distrito han Rožňava ha rehiyon han Košice ha nasod han Slovakia.\nAn Rumegies amo in usa ka komyun ha departamento han Nord ngan ha rehiyon han Nord-Pas-de-Calais ha nasod han Fransya."
}
```
#### unshuffled_deduplicated_wuu
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.03 MB
- **Total amount of disk used:** 0.04 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"伊春元旦天气 伊春腊八天气 伊春春节天气 伊春情人节天气 伊春元宵节天气 伊春愚人节天气 伊春清明节天气 伊春劳动节天气 伊春母亲节天气 伊春端午节天气 伊春七夕节天气 伊春教师节天气 伊春中秋节天气 伊春国庆节天气 伊春重阳节天气 伊春万圣节天气 伊春..."
}
```
#### unshuffled_deduplicated_xal
- **Size of downloaded dataset files:** 0.03 MB
- **Size of the generated dataset:** 0.12 MB
- **Total amount of disk used:** 0.15 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Арнгудин Орн гисн Европд бәәдг һазр. 2007 җилин тooһaр эн орн нутгт 3,600,523 әмтн бәәдг билә. Арнгудин Орнин хотл балһсна нерн..."
}
```
#### unshuffled_deduplicated_xmf
- **Size of downloaded dataset files:** 0.94 MB
- **Size of the generated dataset:** 4.63 MB
- **Total amount of disk used:** 5.58 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"მოჩამილი ტექსტი წჷმორინელი რე Creative Commons Attribution-ShareAlike ლიცენზიათ; შილებე გეძინელი პირობეფიშ არსებუა. კილიშკილიშა..."
}
```
#### unshuffled_deduplicated_yi
- **Size of downloaded dataset files:** 22.20 MB
- **Size of the generated dataset:** 88.29 MB
- **Total amount of disk used:** 110.49 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ממשותדיק - חבֿרה, איך אַרבעט איצט אױף אַ זשורנאַל. טאָמער איר האָט עפּעס צוצוגעבן זאָלט איר שיקן מיר אַן אָנזאָג. ס'װעט הײסן \\\"..."
}
```
#### unshuffled_deduplicated_yo
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.03 MB
- **Total amount of disk used:** 0.04 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Copyright © 2018 BBC. BBC kò mọ̀ nípa àwọn ohun tí ó wà ní àwọn ojú òpó tí ó wà ní ìta. Ọwọ́ tí a fi mú ìbáṣepọ̀ ti ìta.\"..."
}
```
#### unshuffled_deduplicated_yue
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"我 灌 我 灌 我 灌 灌 灌 我 灌 我 灌 我 灌 灌 灌 我 灌 我 灌 我 灌 灌 灌 我 灌 我 灌 我 灌 灌 灌 我 灌 我 灌 我 灌 灌 灌 我 灌 我 灌 我 灌 灌 灌 你還不爆 我累了 投降輸一半可以嗎\"..."
}
```
#### unshuffled_deduplicated_zh
- **Size of downloaded dataset files:** 99.98 GB
- **Size of the generated dataset:** 267.88 GB
- **Total amount of disk used:** 367.86 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"中国铝灰网 中国有色金属矿产网 中国黄莲网 中国水轮发电机网 中国抽油泵网 中国数控雕刻机网 中国不锈钢抛光网 中国磨具加工网 中国压铸铝网 中国耐水腻子网 中国手机摄像头网 中国粗粮网 中国车门锁网 中国钛粉网 中国轮圈网\\n天天中奖彩票图 天天中彩票..."
}
```
</details>
<details>
<summary>Click to expand the Data/size information for each language (original)</summary>
#### unshuffled_original_af
- **Size of downloaded dataset files:** 85.79 MB
- **Size of the generated dataset:** 254.08 MB
- **Total amount of disk used:** 339.87 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "aanlyn markte as gevolg van ons voortgesette 'n begrip opsie handel sakeplan pdf terwyl ons steeds die gereelde ons binêre opsies handel"
}
```
#### unshuffled_original_als
- **Size of downloaded dataset files:** 1.49 MB
- **Size of the generated dataset:** 5.30 MB
- **Total amount of disk used:** 6.78 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"De Nazionalpark hät e Flächi vo 170,3 km² und isch dodemit s grösti Naturschutzgebiet vo de Schwiz. Er ligt uf em Gebiet vo de ..."
}
```
#### unshuffled_original_am
- **Size of downloaded dataset files:** 102.79 MB
- **Size of the generated dataset:** 378.06 MB
- **Total amount of disk used:** 480.85 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"አየር መንገዱ ከአዲስ አበባ ወደ ሮም ጣሊያን በማምራት ላይ በነበረበት ጊዜ ረዳት አብራሪው የጉዞውን አቅጣጫ በመቀየር ጄኔቭ አውሮፓላን ማረፊያ በማሳረፍ እጁን ለፖሊስ ሰጥቷል።\\nየኢትዮጵያ መንግስት የ..."
}
```
#### unshuffled_original_an
- **Size of downloaded dataset files:** 0.15 MB
- **Size of the generated dataset:** 1.33 MB
- **Total amount of disk used:** 1.48 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"واااااااأسفاه الأمم تفتخر ب 0 أمي ووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووووو..."
}
```
#### unshuffled_original_ar
- **Size of downloaded dataset files:** 22.23 GB
- **Size of the generated dataset:** 87.94 GB
- **Total amount of disk used:** 110.17 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"مرحبا بك عزيز الزائر نتمنى لك أوقاتاً سعيدة معنا وأن نزداد شرفا بخدمتك ولا تنسى التسجيل معنا لتستفيد بكل جديد\\nأهلا وسهلا بك زا..."
}
```
#### unshuffled_original_arz
- **Size of downloaded dataset files:** 15.90 MB
- **Size of the generated dataset:** 70.13 MB
- **Total amount of disk used:** 86.03 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"بنى عجل : قبيلة من عجل بن لجيم بن صعب بن على بن بكر بن وائل انتقل اغلبهم الى البصرة فى العراق و اصفهان و خراسان فى ايران و اذرب..."
}
```
#### unshuffled_original_as
- **Size of downloaded dataset files:** 21.43 MB
- **Size of the generated dataset:** 117.73 MB
- **Total amount of disk used:** 139.17 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"আমি, এই সংগঠনৰ সদস্য সকলে একেলগ হৈ অসমকে ধৰি ভাৰতৰ উত্তৰ পূৰ্বাঞ্চলৰ অমূল্য কলা-সাংস্কৃতিক সম্পদৰাজি বৃহত্তৰ অষ্ট্ৰেলিয়াৰ সন্মু..."
}
```
#### unshuffled_original_ast
- **Size of downloaded dataset files:** 0.92 MB
- **Size of the generated dataset:** 2.54 MB
- **Total amount of disk used:** 3.46 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"The Killers llanzaron el so álbum debú, Hot Fuss, en xunu de 2004 nel Reinu Xuníu, al traviés de la discográfica Lizard King, y..."
}
```
#### unshuffled_original_av
- **Size of downloaded dataset files:** 0.08 MB
- **Size of the generated dataset:** 0.42 MB
- **Total amount of disk used:** 0.50 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Жинда малъараб ва божизе бегьулеб рагІудаса кьуризе бегьуларо гьев. Гьес насихІат гьабизе кколелъул бацІцІадаб диналъул рахъалъ..."
}
```
#### unshuffled_original_az
- **Size of downloaded dataset files:** 927.76 MB
- **Size of the generated dataset:** 2.96 GB
- **Total amount of disk used:** 3.89 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"AZTV-Artıq 7 ildir ki, Abşeron rayonu dotasiya almadan bütün xərclərini yerli daxilolmalar hesabına maliyyələşdirir.\\nDünən, 10..."
}
```
#### unshuffled_original_azb
- **Size of downloaded dataset files:** 6.64 MB
- **Size of the generated dataset:** 28.47 MB
- **Total amount of disk used:** 35.11 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"لعلی ١٣-جو عصرده یاشاییب یاراتمیش گؤرکملی آذربایجان شاعرلریندندیر. ١٢٢٤-جی ایلده تبریزده آنادان اولموشدور، گنج یاشلاریندا تیجار..."
}
```
#### unshuffled_original_ba
- **Size of downloaded dataset files:** 33.22 MB
- **Size of the generated dataset:** 133.70 MB
- **Total amount of disk used:** 166.92 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Күҙәтеү ҡуласаһы моделен хәҙер Мифтахетдин Аҡмулла исемендәге Башҡорт дәүләт педагогия университетында ла эшләргә мөмкин\\t\\nКүҙ..."
}
```
#### unshuffled_original_bar
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": " vo"
}
```
#### unshuffled_original_bcl
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"& ÿ ó / í 0 - ø û ù ö ú ð ï ú \\u0014 ù þ ô ö í ÷ ò \\u0014 ÷ í ù û ö í \\u0001 û ñ ç þ \\u0001 ð \\u0007 þ ò ñ ñ ò ô \\u0017 û ö ô ÷..."
}
```
#### unshuffled_original_be
- **Size of downloaded dataset files:** 498.29 MB
- **Size of the generated dataset:** 1.88 GB
- **Total amount of disk used:** 2.38 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Брэсцкія ўлады не дазволілі прафсаюзу РЭП правесці пікетаванне ў парку Воінаў-інтэрнацыяналістаў 30 мая 2018 года.\\nСітуацыю пр..."
}
```
#### unshuffled_original_bg
- **Size of downloaded dataset files:** 8.34 GB
- **Size of the generated dataset:** 33.75 GB
- **Total amount of disk used:** 42.09 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ЖАЛБОПОДАТЕЛЯТ директор на Дирекция „ Обжалване и данъчно-осигурителна практика“- Бургас, редовно призован, се представлява от ..."
}
```
#### unshuffled_original_bh
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.12 MB
- **Total amount of disk used:** 0.13 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"सुकमा जिला भारत के छत्तीसगढ़ राज्य में एगो जिला बाटे। एकर मुख्यालय सुकमा शहर बाटे। एकर कुल रकबा 5636 वर्ग कि॰मी॰ बाटे।\"..."
}
```
#### unshuffled_original_bn
- **Size of downloaded dataset files:** 2.14 GB
- **Size of the generated dataset:** 10.77 GB
- **Total amount of disk used:** 12.91 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ভড়ং সর্বস্ব বাংলা আর্ট অ্যান্ড কালচারের হিসাব গুলিয়ে দেওয়ার ম্যাজিকের নাম ব্রাত্য রাইসু November 23, 2017\\nভড়ং সর্বস্ব বাংলা আর..."
}
```
#### unshuffled_original_bo
- **Size of downloaded dataset files:** 28.94 MB
- **Size of the generated dataset:** 195.40 MB
- **Total amount of disk used:** 224.34 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"བོད་མི་འདི་དག་ནི་རང་རྒྱུད་སྒོ་རུ་ཕུད་དེ་གཞན་རྒྱུད་པང་དུ་ཉར་ནས་གསོ་སྐྱོང་བྱེད་དགོས་ཟེར་བ་དང་གཅིག་མཚུངས་རེད།\\nཚན་རིག་ནི་དང་ཐོག་རང..."
}
```
#### unshuffled_original_bpy
- **Size of downloaded dataset files:** 0.34 MB
- **Size of the generated dataset:** 4.35 MB
- **Total amount of disk used:** 4.69 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"পৌরসভা এহার আয়তন (লয়াহান) ২,৭৩০,.৬৩ বর্গ কিলোমিটার। পৌরসভা এহার মাপাহানর অক্ষাংশ বারো দ্রাঘিমাংশ ইলতাই 18.63° S 48.18° W ।[১]..."
}
```
#### unshuffled_original_br
- **Size of downloaded dataset files:** 9.18 MB
- **Size of the generated dataset:** 30.20 MB
- **Total amount of disk used:** 39.38 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Ar mank Magalhães(Daveoù a vank) a zo ur spesad evned, Spheniscus magellanicus an anv skiantel anezhañ.\\nGallout a reer implijo..."
}
```
#### unshuffled_original_bs
- **Size of downloaded dataset files:** 0.05 MB
- **Size of the generated dataset:** 0.48 MB
- **Total amount of disk used:** 0.53 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ž šř é ú šř šř ě šř ž é č ě ž ů ě ď éé ýš ě ě Ž č š ý ě ď é ýš ě ď ě éé ýš ě č ž ě š ý ď ě ýš é ú č ž č š ý ď ý ž é éě ď é č ýš..."
}
```
#### unshuffled_original_bxr
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.01 MB
- **Total amount of disk used:** 0.02 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"2002 оной хабар буряад хэлэ бэшэгэй һалбари Үндэһэтэнэй хүмүүнлиг ухаанай дээдэ һургуули болгогдожо өөршэлэгдөө.\\nХарин мүнөө б..."
}
```
#### unshuffled_original_ca
- **Size of downloaded dataset files:** 3.10 GB
- **Size of the generated dataset:** 8.62 GB
- **Total amount of disk used:** 11.73 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Daniel Vendrell, conegut com Vandrell, ha sigut un dels il•lustradors contemporanis més influents, representant a la nova onada..."
}
```
#### unshuffled_original_cbk
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano yo gano..."
}
```
#### unshuffled_original_ce
- **Size of downloaded dataset files:** 2.09 MB
- **Size of the generated dataset:** 8.73 MB
- **Total amount of disk used:** 10.82 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Шаьш анархисташ ду бохучу жигархойн дIахьедарехь дуьйцу, оьрсийн ницкъаллийн структурийн а, федералан каналан а Iалашонаш \\\"мар..."
}
```
#### unshuffled_original_ceb
- **Size of downloaded dataset files:** 11.07 MB
- **Size of the generated dataset:** 40.97 MB
- **Total amount of disk used:** 52.03 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Si Isko walay pupamilok nga nagtan-aw sa unahan, natugaw. “Naunsa ka gud diha Isko nga layo man kaayo ang imong panan-aw?” ni I..."
}
```
#### unshuffled_original_ckb
- **Size of downloaded dataset files:** 111.88 MB
- **Size of the generated dataset:** 510.97 MB
- **Total amount of disk used:** 622.85 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"رسی رۆژ - ساڵێک دوای بومەلەرزەی کرماشان میوانی بەرنامە : کاک سیاوەش حەیاتی چالاکی مەدەنی -قەسری شیرین\\nپارچە موزیک 30 / 10 / 20..."
}
```
#### unshuffled_original_cs
- **Size of downloaded dataset files:** 21.72 GB
- **Size of the generated dataset:** 57.08 GB
- **Total amount of disk used:** 78.80 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Akce anarchistů proti připravovanému novému služební řádu a nízkým mzdám 1903 – Historie českého anarchismu (1880 – 1939)\\nRost..."
}
```
#### unshuffled_original_cv
- **Size of downloaded dataset files:** 9.40 MB
- **Size of the generated dataset:** 41.05 MB
- **Total amount of disk used:** 50.45 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Шыранӑ чухне ӑнсӑртран латин кирилл саспаллисем вырӑнне латин саспаллисене ҫырсан, сайт эсир ҫырнине юсама тӑрӑшӗ.\\nКу сайтра ч..."
}
```
#### unshuffled_original_cy
- **Size of downloaded dataset files:** 81.74 MB
- **Size of the generated dataset:** 224.93 MB
- **Total amount of disk used:** 306.67 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Mae capeli Cymreig yr Andes ym Mhatagonia wedi cyhoeddi na fydd gwasanaethau yno weddill y mis, oherwydd yr eira trwm sydd wedi..."
}
```
#### unshuffled_original_da
- **Size of downloaded dataset files:** 6.00 GB
- **Size of the generated dataset:** 16.76 GB
- **Total amount of disk used:** 22.76 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Den 2.-5. februar 2016 løb det tredje kursus i uddannelsen af 4kommunesamarbejdets Local Impact Coaches, af stablen i Gentofte ..."
}
```
#### unshuffled_original_de
- **Size of downloaded dataset files:** 119.51 GB
- **Size of the generated dataset:** 331.22 GB
- **Total amount of disk used:** 450.73 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Auf dieser Seite gibt es mind. ein YouTube Video. Cookies für diese Website wurden abgelehnt. Dadurch können keine YouTube Vide..."
}
```
#### unshuffled_original_diq
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "Zıwanê Slawki, zıwano merdumanê Slawano. Zıwanê Slawki yew lızgeyê Zıwananê Hind u Ewropao. Keyeyê Zıwananê Slawki beno hirê letey:"
}
```
#### unshuffled_original_dsb
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.01 MB
- **Total amount of disk used:** 0.02 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Pśiklaskaju južo pśed pśedstajenim... 1500 źiśi njamóžo wěcej docakaś, měsćańska hala w Chóśebuzu - wupśedana."
}
```
#### unshuffled_original_dv
- **Size of downloaded dataset files:** 24.91 MB
- **Size of the generated dataset:** 131.63 MB
- **Total amount of disk used:** 156.54 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ބ. އަތޮޅުގައި ހުޅުވަން ތައްޔާރުވަމުން އަންނަ ވައްކަރު ރިސޯޓުގައި ވަޒީފާ އަދާކުރަން ޝައުގުވެރިވާ ފަރާތްތަކަށް ކުރިމަތިލުމުގެ ފުރ..."
}
```
#### unshuffled_original_el
- **Size of downloaded dataset files:** 17.31 GB
- **Size of the generated dataset:** 66.27 GB
- **Total amount of disk used:** 83.58 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Νεκρός εντοπίστηκε μέσα στο σπίτι του στην οδό Ηρώδου Αττικού στον αριθμό 7 ο επικεφαλής του προξενικού τμήματος της Ρωσικής πρ..."
}
```
#### unshuffled_original_eml
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.02 MB
- **Total amount of disk used:** 0.03 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"A séguit dal prucès ad rubutiśasiòṅ di abitànt dal pòpul ad Mikenes, Angoras 'l è finî dènt'r a 'n robot cun la tèsta dna rana ..."
}
```
#### unshuffled_original_en
- **Size of downloaded dataset files:** 903.83 GB
- **Size of the generated dataset:** 2525.44 GB
- **Total amount of disk used:** 3429.27 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Mtendere Village was inspired by the vision of Chief Napoleon Dzombe, which he shared with John Blanchard during his first visi..."
}
```
#### unshuffled_original_eo
- **Size of downloaded dataset files:** 117.07 MB
- **Size of the generated dataset:** 314.18 MB
- **Total amount of disk used:** 431.27 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Ĉu ... preĝi | mediti | ricevi instigojn || kanti | muziki || informiĝi | legi | studi || prepari Diservon\\nTemas pri kolekto d..."
}
```
#### unshuffled_original_es
- **Size of downloaded dataset files:** 106.04 GB
- **Size of the generated dataset:** 298.49 GB
- **Total amount of disk used:** 404.53 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Como se librará de la celulitis en el gimnasio La piel superflua en las manos después del adelgazamiento, Los bailes fáciles pa..."
}
```
#### unshuffled_original_et
- **Size of downloaded dataset files:** 1.88 GB
- **Size of the generated dataset:** 5.17 GB
- **Total amount of disk used:** 7.06 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"MTÜ AB Video järgib oma tegevuses kodanikuühenduste eetilise tegevuse üldtunnustatud põhimõtteid, mis on lühidalt kokkuvõetud 7..."
}
```
#### unshuffled_original_eu
- **Size of downloaded dataset files:** 248.19 MB
- **Size of the generated dataset:** 894.83 MB
- **Total amount of disk used:** 1.14 GB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "Gure jarduerek eraikuntzarekin, elkarbizitzarekin, hirigintzarekin eta ekologiarekin dute harremana, baita ideia eta konponbideak irudikatu eta garatzearekin ere, eraikuntza sektorea hobetuz, pertsonen erosotasuna eta bizi-kalitatea hobetzeko."
}
```
#### unshuffled_original_fa
- **Size of downloaded dataset files:** 20.96 GB
- **Size of the generated dataset:** 84.21 GB
- **Total amount of disk used:** 105.17 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"قـــــــــــــــــرار بود با هم کنـــــــــــــار بیایم نه اینکه از کنــــــــــــار هم رد بشیم...!!!\\nاگر روزی دلت لبریز غم بو..."
}
```
#### unshuffled_original_fi
- **Size of downloaded dataset files:** 9.97 GB
- **Size of the generated dataset:** 28.57 GB
- **Total amount of disk used:** 38.54 GB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Kiitos Deelle kaikesta - 1,5 viikkoa kulunut, kun Dee ei ole enää ollut omani. Reilu viikko sitten sunnuntaina vein Deen uuteen kotiinsa. Itselläni on ollut niin ristiriitaiset t..."
}
```
#### unshuffled_original_fr
- **Size of downloaded dataset files:** 105.32 GB
- **Size of the generated dataset:** 303.19 GB
- **Total amount of disk used:** 408.51 GB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "Média de débat d'idées, de culture et de littérature. Récits, décryptages, analyses, portraits et critiques autour de la vie des idées. Magazine engagé, ouvert aux autres et au monde.. Bring up to date in french"
}
```
#### unshuffled_original_frr
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Hiragana’ Practice’Sheet’1’(A -O)’ ’ Name:’________ __________________________’Section:’_______________ _’ ’ ’ ’ ’ ’ ’ ’ ’ ’ ’ ..."
}
```
#### unshuffled_original_fy
- **Size of downloaded dataset files:** 12.40 MB
- **Size of the generated dataset:** 36.24 MB
- **Total amount of disk used:** 48.64 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Nim in sêfte ride op Holmsjön, yn ien fan 'e lytse marren yn de omkriten, of nim se op avontueren lykas nonresidential. lâns Indalsälven wetter. Holm Sportklubb hawwe kano 's te huur, yn gearwurking mei de Baltyske Power konferinsje."
}
```
#### unshuffled_original_ga
- **Size of downloaded dataset files:** 29.27 MB
- **Size of the generated dataset:** 92.37 MB
- **Total amount of disk used:** 121.63 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Is fóram é seo chun plé a dhéanamh ar an leabhar atá roghnaithe do mhí na Samhna 2013 amháin. Ní féidir ach le baill chláraithe..."
}
```
#### unshuffled_original_gd
- **Size of downloaded dataset files:** 0.52 MB
- **Size of the generated dataset:** 2.02 MB
- **Total amount of disk used:** 2.55 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "Zhou Yujun, a 'phàrtaidh Rùnaire Comataidh Sgìre Yanfeng ann Hengyang bhaile agus a Sgìre pàrtaidh agus an riaghaltas a' bhuidheann-riochdachaidh a 'tighinn a chèilidh air ar companaidh air Apr. 14, 2017."
}
```
#### unshuffled_original_gl
- **Size of downloaded dataset files:** 235.38 MB
- **Size of the generated dataset:** 656.48 MB
- **Total amount of disk used:** 891.87 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"O persoal de Inditex da provincia de Pontevedra segue a reclamar iguais condicións laborais no conxunto do país - CIG: Confeder..."
}
```
#### unshuffled_original_gn
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.04 MB
- **Total amount of disk used:** 0.05 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"º ÑÆÚÓ À Ã Ð É Æ ¾ ÄÂ Î À ¼ Æ É ÄÛ = Ü Ý\\\"Þ ßà á â ã ä å æçè ã é ê â å àë ì æê íî é á ë ï í çì àð í Ü à ñ ê é ò ä ì\"..."
}
```
#### unshuffled_original_gom
- **Size of downloaded dataset files:** 0.44 MB
- **Size of the generated dataset:** 2.25 MB
- **Total amount of disk used:** 2.71 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"दुष्ट शीळ हें कौरवांचें । रामें सविस्तर देखूनि साचें । बोलिले वचनें जें दुर्वाचे । करी तयांचें अनुस्मरण ॥२२०॥\"..."
}
```
#### unshuffled_original_gu
- **Size of downloaded dataset files:** 232.02 MB
- **Size of the generated dataset:** 1.09 GB
- **Total amount of disk used:** 1.33 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"અધિક માસ ચાલે છે. સમગ્ર ભારતમાં અને તેમાંય ખાસ કરીને પવિત્ર કે ધાર્મિક કહેવાય છે તેવા સ્થાનક પર કથાનો દોર ચાલે છે. ઉનાળાની કાળઝ..."
}
```
#### unshuffled_original_he
- **Size of downloaded dataset files:** 5.66 GB
- **Size of the generated dataset:** 21.11 GB
- **Total amount of disk used:** 26.77 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"זקוקים לרשתות נגד יתושים? מחפשים רשת מתאימה לחלון צר וקטן? רשתות נגד יתושים אקורדיון של חברת קליר-מש הן הפתרון.\\nרשתות לחלונות ..."
}
```
#### unshuffled_original_hi
- **Size of downloaded dataset files:** 3.66 GB
- **Size of the generated dataset:** 17.93 GB
- **Total amount of disk used:** 21.59 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"'आइटम गर्ल' बनकर हिट हुई थीं राखी सावंत, आज करीना-कटरीना तक फॉलो कर रही हैं ट्रेंड नक्सलियों का दम निकालेगा बाइक ग्रेनेड लॉन्च..."
}
```
#### unshuffled_original_hr
- **Size of downloaded dataset files:** 79.42 MB
- **Size of the generated dataset:** 243.83 MB
- **Total amount of disk used:** 323.24 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"U raspravi je sudjelovao i HSS-ov saborski zastupnik rekavši kako poljoprivrednici ne osjete mjere o kojima ministar govori jer..."
}
```
#### unshuffled_original_hsb
- **Size of downloaded dataset files:** 1.39 MB
- **Size of the generated dataset:** 4.49 MB
- **Total amount of disk used:** 5.87 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Budyšin (SN/BŠe). Elektronikarjo mějachu lětsa cyle hinaši zazběh do swojeho wukubłanja. Wokrjesne rjemjeslnistwo bě mjenujcy w..."
}
```
#### unshuffled_original_ht
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan..."
}
```
#### unshuffled_original_hu
- **Size of downloaded dataset files:** 15.69 GB
- **Size of the generated dataset:** 43.07 GB
- **Total amount of disk used:** 58.77 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"monster - Amatőr, házi szex videók és kezdő csjaok pornó filmjei. - Free amateur, home made sex videos and online porn movies. ..."
}
```
#### unshuffled_original_hy
- **Size of downloaded dataset files:** 897.36 MB
- **Size of the generated dataset:** 3.94 GB
- **Total amount of disk used:** 4.84 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Արցախի Հանրապետության հռչակման 26-րդ տարեդարձի կապակցությամբ Շուշիի Արվեստի կենտրոնում կազմակերպվել է մոսկվաբնակ նկարիչներ՝ հայ..."
}
```
#### unshuffled_original_ia
- **Size of downloaded dataset files:** 0.08 MB
- **Size of the generated dataset:** 0.69 MB
- **Total amount of disk used:** 0.78 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha ha h..."
}
```
#### unshuffled_original_id
- **Size of downloaded dataset files:** 10.60 GB
- **Size of the generated dataset:** 32.32 GB
- **Total amount of disk used:** 42.91 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Perihal dari itu, kalau kunci hal yang demikian hilang, pemilik wajib melapor ke bengkel sah untuk dibuatkan kunci baru dengan ..."
}
```
#### unshuffled_original_ie
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.02 MB
- **Total amount of disk used:** 0.02 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "Plastic Yo Yo Metal Yo Yos Wooden Yo Yo Keychain Yo Yo Translucent Yo Yo Light Up Yo Yo Globe Yo Yo Stress Reliever Yo Yo Jellyfish Yo Yo Sports Ball Yo Yo Sound Yo Yo Miniature Yo Yo Promotional Yo Yo Novelty Yo Yo Video Game Yo Yo ECO Recycled Yo Yo"
}
```
#### unshuffled_original_ilo
- **Size of downloaded dataset files:** 0.27 MB
- **Size of the generated dataset:** 0.92 MB
- **Total amount of disk used:** 1.20 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Segun ken ni Ping-ay, ti yellow corn ti maysa kadagiti nadakamat a liberalized agricultural commodity iti daytoy a free trade k..."
}
```
#### unshuffled_original_io
- **Size of downloaded dataset files:** 0.04 MB
- **Size of the generated dataset:** 0.16 MB
- **Total amount of disk used:** 0.20 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Chekia esas parlamentala republiko. La chefo di stato esas la prezidanto. Til 2013 lu elektesis dal parlamento. Pos ta yaro, ol..."
}
```
#### unshuffled_original_is
- **Size of downloaded dataset files:** 533.03 MB
- **Size of the generated dataset:** 1.52 GB
- **Total amount of disk used:** 2.06 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Eyjar.net - upplýsinga- og fréttamiðill um Vestmannaeyjar - Fréttir - Nái núverandi stefna stjórnvalda fram að ganga mun það va..."
}
```
#### unshuffled_original_it
- **Size of downloaded dataset files:** 52.16 GB
- **Size of the generated dataset:** 147.38 GB
- **Total amount of disk used:** 199.54 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Jaundice - causes, treatment & pathology massaggio a osteochondrosis dellindizio di una controindicazione\\nTrattamento su un co..."
}
```
#### unshuffled_original_ja
- **Size of downloaded dataset files:** 79.56 GB
- **Size of the generated dataset:** 232.22 GB
- **Total amount of disk used:** 311.78 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"神社などへ一緒に同行して、様々な角度のショットで家族写真やお子様の写真を撮影致します!お好みに合わせて様々な写真を取ることができますので、その場でカメラマンへのリクエストも可能です!お子様の晴れ姿を、緊張していない自然な笑顔で残しませんか?\\n※七五三の..."
}
```
#### unshuffled_original_jbo
- **Size of downloaded dataset files:** 0.21 MB
- **Size of the generated dataset:** 0.77 MB
- **Total amount of disk used:** 0.98 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "ni'o 23 la cimast. cu 23moi djedi fi'o masti la cimast. noi ke'a cu cimoi masti .i 22 la cimast. cu purlamdei .ije 24 la cimast. cu bavlamdei"
}
```
#### unshuffled_original_jv
- **Size of downloaded dataset files:** 0.22 MB
- **Size of the generated dataset:** 0.69 MB
- **Total amount of disk used:** 0.91 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"José Mourinho (diwaca: [ʒuˈzɛ moˈɾiɲu]; lair ing Setubal, Portugal, 26 Januari 1963; umur 55 taun) iku salah siji pelatih bal k..."
}
```
#### unshuffled_original_ka
- **Size of downloaded dataset files:** 680.74 MB
- **Size of the generated dataset:** 3.77 GB
- **Total amount of disk used:** 4.45 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"წამიყვანე შენთან ერთად (ქართულად) / Возьми меня с собой (картулад) / (რუსული სერიალები ქართულად) (რუსების პორნო ონლაინში) (ruse..."
}
```
#### unshuffled_original_kk
- **Size of downloaded dataset files:** 615.06 MB
- **Size of the generated dataset:** 2.83 GB
- **Total amount of disk used:** 3.45 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Түлкібас ауданында «Латын негізді әліпби мен емле ережесі туралы насихат» жобасының тобы семинар өткізді\\nЕлорданың «Қазақстан»..."
}
```
#### unshuffled_original_km
- **Size of downloaded dataset files:** 193.28 MB
- **Size of the generated dataset:** 1.10 GB
- **Total amount of disk used:** 1.30 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ខ្សឹបដាក់ត្រចៀក៖ លោក សួស សុផានិត នាយផ្នែករដ្ឋបាលព្រៃឈើ ស្រុកភ្នំក្រវាញ់ ដែលទើបឡើងកាន់តំណែងថ្មី បើកដៃឲ្យឈ្នួញ ប្រព្រឹត្តបទល្មើស ..."
}
```
#### unshuffled_original_kn
- **Size of downloaded dataset files:** 342.15 MB
- **Size of the generated dataset:** 1.76 GB
- **Total amount of disk used:** 2.11 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ರಾಷ್ಟ್ರಪತಿ ಪ್ರಣಬ್ ಮುಖರ್ಜಿಯಿಂದ ಪದ್ಮ ಪ್ರಶಸ್ತಿ ಪ್ರದಾನ | President Pranab Mukherjee Confers Padma Awards | Photo Gallery on Kannada..."
}
```
#### unshuffled_original_ko
- **Size of downloaded dataset files:** 8.81 GB
- **Size of the generated dataset:** 25.29 GB
- **Total amount of disk used:** 34.10 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"CIA 프로젝트에서는 데이터베이스로 들어오는 요청을 중간에 수집(Sniffing)하고 수집한 데이터를 분석(Parsing)하여 그로 인한 결과를 판단하여 알릴 수 있는 시스템(Push Service)이 필요하다. 그리고 연구를 ..."
}
```
#### unshuffled_original_krc
- **Size of downloaded dataset files:** 0.66 MB
- **Size of the generated dataset:** 2.68 MB
- **Total amount of disk used:** 3.34 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Шамханланы, Бийлени къаршысына ябушуп, Батыр уланларыбызны къоллары булан «ортакъ ожакъ» къургъанбыз. Шо иш уллу зараллы иш бол..."
}
```
#### unshuffled_original_ku
- **Size of downloaded dataset files:** 33.38 MB
- **Size of the generated dataset:** 99.06 MB
- **Total amount of disk used:** 132.44 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Me di 114 bernameyên xwe yên berê da perçeyên ji berhemên zanyarî yên kurdzanên mezin bi wergera kurdî da ...\\nMe di 114 bernam..."
}
```
#### unshuffled_original_kv
- **Size of downloaded dataset files:** 0.40 MB
- **Size of the generated dataset:** 2.38 MB
- **Total amount of disk used:** 2.78 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Коми кытшыслӧн ыджытжык тор вӧр увтын куйлӧ, сійӧн и фаунасӧ татӧн аркмӧтӧны вӧрын олісь подаэз. Ассямаӧн лоӧ сія, мый кытшас с..."
}
```
#### unshuffled_original_kw
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.04 MB
- **Total amount of disk used:** 0.05 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼Pray without ceasing🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏🏼🙏..."
}
```
#### unshuffled_original_ky
- **Size of downloaded dataset files:** 152.64 MB
- **Size of the generated dataset:** 630.79 MB
- **Total amount of disk used:** 783.43 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Turmush: Бишкек шаардык кеңешинин кезексиз отурумунда мэрге ишенбөөчүлүк көрсөтүү маселеси каралат, - депутат Т.Сагынов\\nБишкек..."
}
```
#### unshuffled_original_la
- **Size of downloaded dataset files:** 5.46 MB
- **Size of the generated dataset:** 27.80 MB
- **Total amount of disk used:** 33.26 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Hæ sunt generationes Noë: Noë vir justus atque perfectus fuit in generationibus suis; cum Deo ambulavit.\\nEcce ego adducam aqua..."
}
```
#### unshuffled_original_lb
- **Size of downloaded dataset files:** 10.73 MB
- **Size of the generated dataset:** 30.60 MB
- **Total amount of disk used:** 41.32 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Während dem Gaardefestival \\\"Ambiance Jardins\\\" vum 15. bis de 17. Mee huet den SNJ nees zesumme mam Groupe Animateur en Inform..."
}
```
#### unshuffled_original_lez
- **Size of downloaded dataset files:** 0.83 MB
- **Size of the generated dataset:** 3.38 MB
- **Total amount of disk used:** 4.20 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Ахцегь хуьр, виридалай ч1ехи лезги хуьрерикая я. Ам Урусатдин виридалай къиблепатавай хуьрерикай я. Ин хуьр...\"..."
}
```
#### unshuffled_original_li
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.03 MB
- **Total amount of disk used:** 0.04 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"'t Good Goedenraad aan de Ezerbaek besjteit oet 'n kesjtièl mèt gesjlote haof en 'n park van 26 hectare. Hie in sjtoon väól beu..."
}
```
#### unshuffled_original_lmo
- **Size of downloaded dataset files:** 0.10 MB
- **Size of the generated dataset:** 0.47 MB
- **Total amount of disk used:** 0.58 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Serét (en tortonés: Sregh; en piemontés: Srèj) l'è 'n cümü italià, de la regiù del Piemónt, en Pruvìncia de Alessandria. El g'h..."
}
```
#### unshuffled_original_lo
- **Size of downloaded dataset files:** 33.92 MB
- **Size of the generated dataset:** 182.36 MB
- **Total amount of disk used:** 216.28 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"ຜູ້ພິພາກສາ ປະຈຳເຂດ ສຫລ ທ່ານນຶ່ງ ຕັດສິນວ່າ ໂຄງການເກັບກຳຂໍ້ມູນ ທາງໂທລະສັບ ຂອງອົງການ ຄວາມໝັ້ນຄົງແຫ່ງຊາດ ແມ່ນຖືກຕ້ອງ ຕາມກົດໝາຍ.\\nກະ..."
}
```
#### unshuffled_original_lrc
- **Size of downloaded dataset files:** 0.02 MB
- **Size of the generated dataset:** 0.07 MB
- **Total amount of disk used:** 0.09 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"آرلینگتون یئ گئل د شأریا ڤولاتچە ڤیرجینیا و یئ گئل د شأریا ڤولات ڤولاتچە یا یأکاگئرئتە ئمریکاە. ئی شأر دویومی کألوٙن شأر د راسا..."
}
```
#### unshuffled_original_lt
- **Size of downloaded dataset files:** 3.44 GB
- **Size of the generated dataset:** 9.45 GB
- **Total amount of disk used:** 12.89 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Čir vir vir pavasaris! Čia čia čia… dalinamės labai simpatiška video pamokėle, kurią pristato ab888art galerija.\\nBe galo papra..."
}
```
#### unshuffled_original_lv
- **Size of downloaded dataset files:** 1.49 GB
- **Size of the generated dataset:** 4.27 GB
- **Total amount of disk used:** 5.75 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Dekoratīvi sliekšņi MITSUBISHI OUTLANDER 2007, izgatavoti no ovālas formas, pulētas nerūsējošā tērauda caurules...\\ndažādas tūn..."
}
```
#### unshuffled_original_mai
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.33 MB
- **Total amount of disk used:** 0.34 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"१ · २ · ३ · ४ · ५ · ६ · ७ · ८ · ९ · १० · ११ · १२ · १३ · १४ · १५ · १६ · १७ · १८ · १९ · २० · २१ · २२ · २३ · २४ · २५ · २६ · २७ · २..."
}
```
#### unshuffled_original_mg
- **Size of downloaded dataset files:** 6.22 MB
- **Size of the generated dataset:** 21.79 MB
- **Total amount of disk used:** 28.01 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Nanamboatra taratasy apetaka sy soso-kevitra ho an'ny olona te-hanatevin-daharana ity fihetsiketsehana ity i Anocrena.\\nNosorat..."
}
```
#### unshuffled_original_mhr
- **Size of downloaded dataset files:** 1.84 MB
- **Size of the generated dataset:** 7.55 MB
- **Total amount of disk used:** 9.38 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Акрет жап годым Уганда кундемым Пигмей племена- влак айлен шогеныт. мемнан эран 1 курым гыч Банту племена влакат тиде кундемышк..."
}
```
#### unshuffled_original_min
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.63 MB
- **Total amount of disk used:** 0.64 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\" ..."
}
```
#### unshuffled_original_mk
- **Size of downloaded dataset files:** 508.24 MB
- **Size of the generated dataset:** 2.20 GB
- **Total amount of disk used:** 2.71 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"„Филм плус“ е насловен првиот филмски месечник во Македонија, чиј прв број ќе биде промовиран вечер во „Менада“. Новото македон..."
}
```
#### unshuffled_original_ml
- **Size of downloaded dataset files:** 938.69 MB
- **Size of the generated dataset:** 5.24 GB
- **Total amount of disk used:** 6.18 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"സ്ത്രീ പ്രവേശനം സര്ക്കാര് പൂര്ണമായും അംഗീകരിക്കുന്നുവെന്നും ശബരിമലയുടെ സുരക്ഷയില് ഇടപെടുമെന്നും സര്ക്കാര് ഹൈക്കോടതിയില്\\..."
}
```
#### unshuffled_original_mn
- **Size of downloaded dataset files:** 472.36 MB
- **Size of the generated dataset:** 2.33 GB
- **Total amount of disk used:** 2.81 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Монгол улс, Улаанбаатар хот - 14191 Энхтайваны өргөн чөлөө - 10, Багш хөгжлийн ордон, Багшийн мэргэжил дээшлүүлэх институт\\nБаг..."
}
```
#### unshuffled_original_mr
- **Size of downloaded dataset files:** 525.31 MB
- **Size of the generated dataset:** 2.82 GB
- **Total amount of disk used:** 3.34 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Home / motivational marathi story / उद्योजकता (Entrepreneurship) / यांना हे जमलय, तर आपल्याला का नाही जमणार ?\\nयापैकी कोणाचीही ..."
}
```
#### unshuffled_original_mrj
- **Size of downloaded dataset files:** 0.30 MB
- **Size of the generated dataset:** 1.16 MB
- **Total amount of disk used:** 1.47 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Лӹпӹвлӓ (латинлӓ Lepidoptera ; алыкмарла лыве-влак) — капшангывлӓ йыхыш пырышы сӱмӓн нӹл шылдыран капшангывлӓ. Цилӓжӹ 180000 тӹ..."
}
```
#### unshuffled_original_ms
- **Size of downloaded dataset files:** 28.46 MB
- **Size of the generated dataset:** 122.33 MB
- **Total amount of disk used:** 150.79 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Sanad pertama daripada Zuhair bin Harb daripada ‘Affan daripada Hammad daripada Thabit daripada Anas.\\nSanad kedua daripada ‘Ab..."
}
```
#### unshuffled_original_mt
- **Size of downloaded dataset files:** 7.53 MB
- **Size of the generated dataset:** 24.47 MB
- **Total amount of disk used:** 32.00 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "tibgħat il-kawża lura lill-Qorti Ġenerali għall-annullament jew għat-tnaqqis tal-penalità imposta mill-Kummissjoni bid-deċiżjoni inizjali kif emendata bid-deċiżjoni ta’ rettifika;"
}
```
#### unshuffled_original_mwl
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Deciplina social i outónoma que angloba atebidades de ouserbaçon, de análeze, de çcriçon, cumparaçon, de sistematizaçon i de sp..."
}
```
#### unshuffled_original_my
- **Size of downloaded dataset files:** 369.85 MB
- **Size of the generated dataset:** 2.02 GB
- **Total amount of disk used:** 2.39 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ျမ၀တီ - ရန္ကုန္တိုင္းေဒသႀကီး ေျမာက္ဥကၠလာပႏွင္႕ ဗဟန္းၿမိဳ႔နယ္ မေကြးတိုင္း ေဒသႀကီး ပခုကၠဴၿမိဳ႔နယ္တို႔၌ ျမန္မာ႕တပ္မေတာ္အား ေထာက္ခံ..."
}
```
#### unshuffled_original_myv
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"2018 иень умарьковонь 6-це чистэ сась паро куля! Россиянь культурань Министерствась макссь невтемань конёв (прокатной удостовер..."
}
```
#### unshuffled_original_mzn
- **Size of downloaded dataset files:** 0.18 MB
- **Size of the generated dataset:** 0.72 MB
- **Total amount of disk used:** 0.90 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"قرآن یا قوران اسلام ِآسمونی کتاب هسته. مسلمونون گانّّه قرآن ره خدا، وحی جه برسنییه، «محمد معجزه» هسته و ثقلین حدیث دله ونه خَو..."
}
```
#### unshuffled_original_nah
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.01 MB
- **Total amount of disk used:** 0.01 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "In mācuīlpōhualxihuitl VI (inic chicuacē) in mācuīlpōhualli xiuhitl cāhuitl īhuīcpa 501 xihuitl oc 600 xihuitl."
}
```
#### unshuffled_original_nap
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.02 MB
- **Total amount of disk used:** 0.02 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ò AUDIT í Ç è î ÿ å å 30 ò ÿ ÿ é, õ ñ ì ÿ, ê ã- ò à ì. å â å í ç â à à é ñ è å é ó ó ë. å å å û è å î é è à. à è à AUDIT 1-7 â ..."
}
```
#### unshuffled_original_nds
- **Size of downloaded dataset files:** 6.74 MB
- **Size of the generated dataset:** 18.23 MB
- **Total amount of disk used:** 24.99 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Dor kann sik vun nu af an de hele plattdüütsche Welt – vun Niebüll bit New York, vun Helgoland bit Honolulu – drapen. Allens, w..."
}
```
#### unshuffled_original_ne
- **Size of downloaded dataset files:** 355.29 MB
- **Size of the generated dataset:** 1.87 GB
- **Total amount of disk used:** 2.22 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"बर्दिबास नगरपालिकाको तेस्रो नगर परिषदबाट पारित आ.व.२०७३।७४ को संशोधित र २०७४।७५ को प्रस्तावित नीति, कार्यक्रम तथा बजेट\\nअार्थिक..."
}
```
#### unshuffled_original_new
- **Size of downloaded dataset files:** 1.03 MB
- **Size of the generated dataset:** 5.77 MB
- **Total amount of disk used:** 6.79 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"थ्व शहरयागु अक्षांश ३४.७००१६४ उत्तर व देशान्तर ८६.३७६४६९ पश्चिम खः (34.700164° N 86.376469° W)। थ्व थासे ७२२६७३२ वर्ग मिटर (२.७..."
}
```
#### unshuffled_original_nl
- **Size of downloaded dataset files:** 29.35 GB
- **Size of the generated dataset:** 83.23 GB
- **Total amount of disk used:** 112.58 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Op vrijdag 31 augustus wordt het nieuwe studiejaar van de masteropleiding architectuur geopend met een dagexcursie naar Venlo.\\..."
}
```
#### unshuffled_original_nn
- **Size of downloaded dataset files:** 32.86 MB
- **Size of the generated dataset:** 90.84 MB
- **Total amount of disk used:** 123.70 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "Planomtale krav til innhald Bakgrunn: Spørsmål frå fleire kommunar om kva ein planomtale/planbeskrivelse bør innehalde Fylkeskommunen og fylkesmannen har i ein del saker reist motsegn på formelt grunnlag"
}
```
#### unshuffled_original_no
- **Size of downloaded dataset files:** 3.11 GB
- **Size of the generated dataset:** 8.65 GB
- **Total amount of disk used:** 11.76 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Ytterligere aktører i primærhelsetjenesten og andre NHS-virksomheter ble infisert, inkludert legekontor.Læreren vår er så attra..."
}
```
#### unshuffled_original_oc
- **Size of downloaded dataset files:** 1.57 MB
- **Size of the generated dataset:** 6.12 MB
- **Total amount of disk used:** 7.71 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": ".рф (rf, còdi punycode: .xn--p1ai)[1] es lo nom de domeni en rus per Russia. Foguèt activat lo 12 de mai de 2010. Lo còdi latin es .ru."
}
```
#### unshuffled_original_or
- **Size of downloaded dataset files:** 49.84 MB
- **Size of the generated dataset:** 260.15 MB
- **Total amount of disk used:** 309.99 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ଭୁବନେଶ୍ୱର, ୨୭/୧– (ଓଡ଼ିଆ ପୁଅ) ସିପିଆଇ ଜାତୀୟ ପରିଷଦର ଆହ୍ୱାନକ୍ରମେ ଗତକାଲି ଜାନୁୟାରୀ ୨୬ ସାଧାରଣତନ୍ତ୍ର ଦିବସକୁ ଦେଶ ବ୍ୟାପୀ ସମ୍ବିଧାନ ସୁରକ୍ଷା ..."
}
```
#### unshuffled_original_os
- **Size of downloaded dataset files:** 3.09 MB
- **Size of the generated dataset:** 12.90 MB
- **Total amount of disk used:** 15.99 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"1. Лæппу æмæ чызг казрæдзийы зæрдæмæ куы фæцæуынц æмæ, куы сфæнд кæнынц сæ цард баиу кæнын, уæд лæппу бар ракуры чызгæй, цæмæй ..."
}
```
#### unshuffled_original_pa
- **Size of downloaded dataset files:** 164.21 MB
- **Size of the generated dataset:** 801.16 MB
- **Total amount of disk used:** 965.37 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ਰਜਿ: ਨੰ: PB/JL-138/2018-20 ਜਿਲਦ 63, ਬਾਨੀ ਸੰਪਾਦਕ (ਸਵ:) ਡਾ: ਸਾਧੂ ਸਿੰਘ ਹਮਦਰਦ ਫ਼ੋਨ : 0181-2455961-62-63, 5032400, ਫੈਕਸ : 2455960, 2..."
}
```
#### unshuffled_original_pam
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Áku pu i Anak ning Aláya at ngeni ipákit kó kékayu ngan nûng makanánu lang susúlat détinang kulit a mágkas. Lauan ya ing tarátu..."
}
```
#### unshuffled_original_pl
- **Size of downloaded dataset files:** 42.88 GB
- **Size of the generated dataset:** 117.12 GB
- **Total amount of disk used:** 160.01 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"System informatyczny - Załącznik nr 1 do zarządzenia Wójta Gminy Podegrodzie Nr 530/2013 z dnia 27 maja 2013 r\\nSystem informat..."
}
```
#### unshuffled_original_pms
- **Size of downloaded dataset files:** 0.75 MB
- **Size of the generated dataset:** 2.15 MB
- **Total amount of disk used:** 2.92 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Louvigné-du-Désert a l'é na comun-a fransèisa ant la region aministrativa dla Brëtagna, ant ël dipartiment d'Ille-et-Vilaine. A..."
}
```
#### unshuffled_original_pnb
- **Size of downloaded dataset files:** 3.22 MB
- **Size of the generated dataset:** 12.04 MB
- **Total amount of disk used:** 15.26 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"ایہ فائل Wikimedia Commons توں اے تے دوجیاں ویونتاں تے وی ورتی جاےکدی اے۔ گل بات اس دے فائل گل بات صفہ تے تھلے دتی گئی۔\"..."
}
```
#### unshuffled_original_ps
- **Size of downloaded dataset files:** 103.66 MB
- **Size of the generated dataset:** 379.51 MB
- **Total amount of disk used:** 483.17 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Many people usually use the time period ‘business to business (B2B) advertising,’ however most of them do not know precisely wh..."
}
```
#### unshuffled_original_pt
- **Size of downloaded dataset files:** 47.26 GB
- **Size of the generated dataset:** 132.64 GB
- **Total amount of disk used:** 179.89 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Você pode estar lendo este texto no sofá, levantar pra pegar uma breja na geladeira, dar uma cagada e sentar novamente, sem int..."
}
```
#### unshuffled_original_qu
- **Size of downloaded dataset files:** 0.02 MB
- **Size of the generated dataset:** 0.08 MB
- **Total amount of disk used:** 0.10 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Warayu wichay (kastilla simipi: Ascensión de Guarayos) nisqaqa Buliwya mama llaqtapi, Santa Krus suyupi, huk llaqtam, Warayu pruwinsyap uma llaqtanmi."
}
```
#### unshuffled_original_rm
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.01 MB
- **Total amount of disk used:** 0.01 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"practicists agrars / practicistas agraras AFP pon far ina furmaziun da basa scursanida per cuntanscher in attestat federal da q..."
}
```
#### unshuffled_original_ro
- **Size of downloaded dataset files:** 9.53 GB
- **Size of the generated dataset:** 26.87 GB
- **Total amount of disk used:** 36.40 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"“În viață, oportunitatea nu este totul. Cine atrage Lumina, cineva bun în umbră. Timpul ne creează.” maestru\\nLyn.Evans: Ce mar..."
}
```
#### unshuffled_original_ru
- **Size of downloaded dataset files:** 319.76 GB
- **Size of the generated dataset:** 1241.63 GB
- **Total amount of disk used:** 1561.38 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Доступ к данному профилю для публичного просмотра закрыт администрацией сайта - профиль находится на модерации.\\nРазработчикам ..."
}
```
#### unshuffled_original_sa
- **Size of downloaded dataset files:** 17.52 MB
- **Size of the generated dataset:** 97.06 MB
- **Total amount of disk used:** 114.58 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"अनिरुद्धनगरे क्रीडिता रामलीला सम्प्रति समाप्ता अस्ति । तस्य कानिचन् चित्राणि पूर्वमेव प्रकाशितानि सन्ति । द्वौ चलचित्रौ अपि ..."
}
```
#### unshuffled_original_sah
- **Size of downloaded dataset files:** 9.08 MB
- **Size of the generated dataset:** 43.82 MB
- **Total amount of disk used:** 52.90 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████..."
}
```
#### unshuffled_original_scn
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
{
"id": 0,
"text": "La gilusìa è nu sintimentu dulurusu ca nasci d'un disideriu di pussessu sclusivu ntê cunfrunti dâ pirsuna amata e dû timuri, dû suspettu o dâ cirtizza dâ sò nfidiltati."
}
```
#### unshuffled_original_sd
- **Size of downloaded dataset files:** 90.62 MB
- **Size of the generated dataset:** 364.25 MB
- **Total amount of disk used:** 454.88 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"هر ڪو ڄاڻي ٿو ته جڏهن توهان هڪ وڏي خريد ڪرڻ چاهيون ٿا, توهان پڄي ضروري حڪم ۾ ان جي ڪم ڪرڻ جي هٿ ۾ لاڳاپو ڪيو آهي. جي شيء آهي ته..."
}
```
#### unshuffled_original_sh
- **Size of downloaded dataset files:** 3.46 MB
- **Size of the generated dataset:** 25.84 MB
- **Total amount of disk used:** 29.30 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Opština Gornja Radgona se nalazi u sjeveroistočnoj Sloveniji i graniči s susjednom Austriji duž rijeke Mure. Sa tridesetim nase..."
}
```
#### unshuffled_original_si
- **Size of downloaded dataset files:** 310.93 MB
- **Size of the generated dataset:** 1.47 GB
- **Total amount of disk used:** 1.78 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"ලාංකීය සිතිවිලි සිංහල බ්ලොග් කියවනය කොත්තු සින්ඩිය ලංකා Blogger හත්මාළුව ලංකා බ්ලොග් කියවනය මාතලන්ගේ සින්ඩිය මොබයිල්lk\\nඅවකාශය ..."
}
```
#### unshuffled_original_sk
- **Size of downloaded dataset files:** 3.71 GB
- **Size of the generated dataset:** 9.81 GB
- **Total amount of disk used:** 13.52 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Aktivity | Agentúra podporovaného zamestnávania | vzdelávanie pre klientov, vzdelávanie pre odborníkov, kurzy\\nŠpecializované k..."
}
```
#### unshuffled_original_sl
- **Size of downloaded dataset files:** 956.20 MB
- **Size of the generated dataset:** 2.68 GB
- **Total amount of disk used:** 3.63 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Če Creatures, ki je želel, da pridejo na čas, predvsem je povedlo – razlikuje od ljubosumja začel grizenja kolen (ali zadnjica)..."
}
```
#### unshuffled_original_so
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.06 MB
- **Total amount of disk used:** 0.06 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"тттттттттттттттттттттттттттттттт тттттттттттттттттттттттттттттттт тттттттттттттттттттттттттттттттт ттттттттттттттттуууууууууууу..."
}
```
#### unshuffled_original_sq
- **Size of downloaded dataset files:** 861.84 MB
- **Size of the generated dataset:** 2.44 GB
- **Total amount of disk used:** 3.30 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Çfarë do të më pëlqente tek një femër ose çfarë do të më shndërronte në një shpërthim drite? – Albert Vataj\\nTë gjithëve një zo..."
}
```
#### unshuffled_original_sr
- **Size of downloaded dataset files:** 1.08 GB
- **Size of the generated dataset:** 4.13 GB
- **Total amount of disk used:** 5.21 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Корисни савети за сваки дан. На сајту су разне категорије, као што су љепота, мода, кување и поправка властитим рукама.\\nШколск..."
}
```
#### unshuffled_original_su
- **Size of downloaded dataset files:** 0.06 MB
- **Size of the generated dataset:** 0.23 MB
- **Total amount of disk used:** 0.28 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Kartu krédit nyaéta \"duit plastik\" anu dikaluarkeun ku bank pikeun alat pambayaran di tempat-tempat nu tangtu samisal jiga di hotél, réstoran, tempat rékréasi jeung sajabana.[1]"
}
```
#### unshuffled_original_sv
- **Size of downloaded dataset files:** 17.18 GB
- **Size of the generated dataset:** 47.00 GB
- **Total amount of disk used:** 64.18 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"1783 är ett viktigt årtal i den nya tidens historia. Det året slöts en fred i Paris och därmed blev de 13 brittiska kolonierna ..."
}
```
#### unshuffled_original_sw
- **Size of downloaded dataset files:** 3.71 MB
- **Size of the generated dataset:** 14.07 MB
- **Total amount of disk used:** 17.78 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Miripuko hiyo inakuja mwanzoni mwa Wiki Takatifu kuelekea Pasaka na ikiwa ni wiki chache tu kabla ya Papa Francis kuanza ziara yake katika nchi hiyo yenye idadi kubwa kabisa ya watu katika ulimwengu wa nchi za Kiarabu."
}
```
#### unshuffled_original_ta
- **Size of downloaded dataset files:** 1.74 GB
- **Size of the generated dataset:** 9.93 GB
- **Total amount of disk used:** 11.67 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"பொழுது சாய்ந்து வெகு நேரமாகிவிட்டது. கூலி வேலைக்குப் போயிருந்த 'சித்தாள் ' பெண்கள் எல்லோரும் வீடு திரும்பி விட்டார்கள். இன்னும்..."
}
```
#### unshuffled_original_te
- **Size of downloaded dataset files:** 522.47 MB
- **Size of the generated dataset:** 2.61 GB
- **Total amount of disk used:** 3.13 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"హర్యానాలో టోల్ దగ్గర సిబ్బంది.. స్థానిక ప్రజలు కొట్టుకున్నారు. కర్నాల్ అనే గ్రామానికి సమీపంలో టోల్ గేట్ ఉంది. అయితే సాధారణంగా స..."
}
```
#### unshuffled_original_tg
- **Size of downloaded dataset files:** 90.97 MB
- **Size of the generated dataset:** 397.43 MB
- **Total amount of disk used:** 488.41 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Ҳумайро гуфтааст, мухолифи низом аст, низоме, ки дар Тоҷикистон вуҷуд дорад. Ба ин маънӣ, худро мухолифи давлату ҳукумати Тоҷик..."
}
```
#### unshuffled_original_th
- **Size of downloaded dataset files:** 7.38 GB
- **Size of the generated dataset:** 38.29 GB
- **Total amount of disk used:** 45.67 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ฟันที่แลดูขาวสะอาดไม่มีเศษอาหารติดอยู่ เหงือกสีชมพู ไม่เจ็บ หรือมีเลือดออกเวลาแปรงฟันหรือขัดฟัน ไม่มีปัญหาเรื่องกลิ่นปาก ทำให้ก..."
}
```
#### unshuffled_original_tk
- **Size of downloaded dataset files:** 2.96 MB
- **Size of the generated dataset:** 10.66 MB
- **Total amount of disk used:** 13.62 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"Türkmenistanyň Prezidenti agyr atletika boýunça dünýä çempionatyna taýýarlyk işleriniň barşy bilen tanyşdy\\nHalallykdan kemal t..."
}
```
#### unshuffled_original_tl
- **Size of downloaded dataset files:** 204.89 MB
- **Size of the generated dataset:** 606.30 MB
- **Total amount of disk used:** 811.19 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"“Gusto ko manawagan sa mga Unit Head ng Chanel 2 Salve. Kasi napapansin ko iyon mga alaga ko ang taping halos once a week lang,..."
}
```
#### unshuffled_original_tr
- **Size of downloaded dataset files:** 21.96 GB
- **Size of the generated dataset:** 63.58 GB
- **Total amount of disk used:** 85.54 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Son yıllarda görülen ay tutulmalarına göre daha etkili olacağı söylenen Kanlı veya Kırmızı Ay Tutulmasına saatler kaldı. Bu akş..."
}
```
#### unshuffled_original_tt
- **Size of downloaded dataset files:** 151.06 MB
- **Size of the generated dataset:** 703.42 MB
- **Total amount of disk used:** 854.47 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"\\\"Иремнең вафатына 40 көн узгач, Алмаз да безнең өйгә кереп үлде\\\". Арчада 35 яшьлек ир өстенә кондызлар ега башлаган агач төшк..."
}
```
#### unshuffled_original_tyv
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.01 MB
- **Total amount of disk used:** 0.01 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Экии, хүндүлуг аалчылар болгаш тыва дылдың деткикчилери! Тыва дылдың болгаш чогаалдың ховар бир башкызынга, Менги Ооржакка, ажы..."
}
```
#### unshuffled_original_ug
- **Size of downloaded dataset files:** 27.92 MB
- **Size of the generated dataset:** 127.42 MB
- **Total amount of disk used:** 155.35 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"زاڭ-ءتۇزىم | عىلىم-تەحنيكا | ءتىل-ادەبيەت | تۇرمىس | دەنە تاربيە | ساياحات-ورتا | سۋرەتتى حابار | سىر سۇحبات | ارناۋلى تاقىرىپ ..."
}
```
#### unshuffled_original_uk
- **Size of downloaded dataset files:** 14.42 GB
- **Size of the generated dataset:** 56.44 GB
- **Total amount of disk used:** 70.86 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Про надання роз'яснення (щодо форми письмового зобов'язання громадян про зворотне ввезення/вивезення товарів), Державна митна с..."
}
```
#### unshuffled_original_ur
- **Size of downloaded dataset files:** 712.61 MB
- **Size of the generated dataset:** 2.80 GB
- **Total amount of disk used:** 3.51 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"آئیے اہم اسلامی کتب کو یونیکوڈ میں انٹرنیٹ پر پیش کرنے کے لئے مل جل کر آن لائن ٹائپنگ کریں۔ محدث ٹائپنگ پراجیکٹ کے ذریعے آپ روز..."
}
```
#### unshuffled_original_uz
- **Size of downloaded dataset files:** 5.78 MB
- **Size of the generated dataset:** 21.46 MB
- **Total amount of disk used:** 27.24 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Qurama tog'lari tizmasining Toshkentdan 154 km uzoqlikdagi Toshkent-Ush yo'li yeqasidaxushmanzara tabiat qo'ynida joylashgan maydoni 30 ga.\nBolalarni sog'lomlashtirish oromgohi Bo'stonliq tumani Oqtosh muntaqasining soy-salqin gushasida joylashgan."
}
```
#### unshuffled_original_vec
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.02 MB
- **Total amount of disk used:** 0.03 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Par ogni pónto, ła derivada ła xe ła pendensa de ła reta tangente a ła curva de ła funsion f. Ła reta de cołor róso l'è senpre ..."
}
```
#### unshuffled_original_vi
- **Size of downloaded dataset files:** 21.50 GB
- **Size of the generated dataset:** 72.23 GB
- **Total amount of disk used:** 93.73 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Canh chua cá bông lau không chỉ là món ăn giải nhiệt, thanh mát ngày hè mà còn là món siêu bổ dưỡng, rất tốt cho người gầy ốm. ..."
}
```
#### unshuffled_original_vo
- **Size of downloaded dataset files:** 0.30 MB
- **Size of the generated dataset:** 2.12 MB
- **Total amount of disk used:** 2.42 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Sarniguet binon zif in ziläk: Hautes-Pyrénées, in topäd: Midi-Pyrénées, in Fransän. Sarniguet topon videtü 43°19’ 7’’ N e lunetü 0°5’ 19’’ L."
}
```
#### unshuffled_original_wa
- **Size of downloaded dataset files:** 0.09 MB
- **Size of the generated dataset:** 0.29 MB
- **Total amount of disk used:** 0.38 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "Cisse pådje ci n' est co k' on djermon, dj' ô bén k' el pådje est djusse sibåtcheye, eyet co trop tene; et s' divreut ele ecråxhî ene miete."
}
```
#### unshuffled_original_war
- **Size of downloaded dataset files:** 0.64 MB
- **Size of the generated dataset:** 2.68 MB
- **Total amount of disk used:** 3.32 MB
An example of 'train' looks as follows.
```
{
"id": 1,
"text": "An Honce amo in usa ka baryo ngan munisipalidad ha distrito han Rožňava ha rehiyon han Košice ha nasod han Slovakia.\nAn Rumegies amo in usa ka komyun ha departamento han Nord ngan ha rehiyon han Nord-Pas-de-Calais ha nasod han Fransya."
}
```
#### unshuffled_original_wuu
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.12 MB
- **Total amount of disk used:** 0.13 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"伊春元旦天气 伊春腊八天气 伊春春节天气 伊春情人节天气 伊春元宵节天气 伊春愚人节天气 伊春清明节天气 伊春劳动节天气 伊春母亲节天气 伊春端午节天气 伊春七夕节天气 伊春教师节天气 伊春中秋节天气 伊春国庆节天气 伊春重阳节天气 伊春万圣节天气 伊春..."
}
```
#### unshuffled_original_xal
- **Size of downloaded dataset files:** 0.03 MB
- **Size of the generated dataset:** 0.12 MB
- **Total amount of disk used:** 0.15 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Арнгудин Орн гисн Европд бәәдг һазр. 2007 җилин тooһaр эн орн нутгт 3,600,523 әмтн бәәдг билә. Арнгудин Орнин хотл балһсна нерн..."
}
```
#### unshuffled_original_xmf
- **Size of downloaded dataset files:** 1.05 MB
- **Size of the generated dataset:** 6.12 MB
- **Total amount of disk used:** 7.17 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"მოჩამილი ტექსტი წჷმორინელი რე Creative Commons Attribution-ShareAlike ლიცენზიათ; შილებე გეძინელი პირობეფიშ არსებუა. კილიშკილიშა..."
}
```
#### unshuffled_original_yi
- **Size of downloaded dataset files:** 33.33 MB
- **Size of the generated dataset:** 147.60 MB
- **Total amount of disk used:** 180.94 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"ממשותדיק - חבֿרה, איך אַרבעט איצט אױף אַ זשורנאַל. טאָמער איר האָט עפּעס צוצוגעבן זאָלט איר שיקן מיר אַן אָנזאָג. ס'װעט הײסן \\\"..."
}
```
#### unshuffled_original_yo
- **Size of downloaded dataset files:** 0.01 MB
- **Size of the generated dataset:** 0.06 MB
- **Total amount of disk used:** 0.06 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 0,
"text": "\"Copyright © 2018 BBC. BBC kò mọ̀ nípa àwọn ohun tí ó wà ní àwọn ojú òpó tí ó wà ní ìta. Ọwọ́ tí a fi mú ìbáṣepọ̀ ti ìta.\"..."
}
```
#### unshuffled_original_yue
- **Size of downloaded dataset files:** 0.00 MB
- **Size of the generated dataset:** 0.00 MB
- **Total amount of disk used:** 0.00 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"我 灌 我 灌 我 灌 灌 灌 我 灌 我 灌 我 灌 灌 灌 我 灌 我 灌 我 灌 灌 灌 我 灌 我 灌 我 灌 灌 灌 我 灌 我 灌 我 灌 灌 灌 我 灌 我 灌 我 灌 灌 灌 你還不爆 我累了 投降輸一半可以嗎\"..."
}
```
#### unshuffled_original_zh
- **Size of downloaded dataset files:** 206.00 GB
- **Size of the generated dataset:** 545.61 GB
- **Total amount of disk used:** 751.61 GB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"id": 1,
"text": "\"中国铝灰网 中国有色金属矿产网 中国黄莲网 中国水轮发电机网 中国抽油泵网 中国数控雕刻机网 中国不锈钢抛光网 中国磨具加工网 中国压铸铝网 中国耐水腻子网 中国手机摄像头网 中国粗粮网 中国车门锁网 中国钛粉网 中国轮圈网\\n天天中奖彩票图 天天中彩票..."
}
```
</details>
### Data Fields
The data fields are the same among all configs.
- `id`: a `int64` feature.
- `text`: a `string` feature.
### Data Splits
<details>
<summary>Click to expand the number of samples per configuration</summary>
| Language | Language code | Name original | Train original | Words original | Size original | Name deduplicated | Train deduplicated | Words deduplicated | Size deduplicated |
| ----------------- | ------------- | ----------------------- | -------------- | --------------- | ------------- | --------------------------- | ------------------ | ------------------ | ----------------- |
| Afrikaans | af | unshuffled_original_af | 201117 | 43,482,801 | 241M | unshuffled_deduplicated_af | 130640 | 29,533,437 | 163M |
| Albanian | sq | unshuffled_original_sq | 672077 | 374,196,110 | 2.3G | unshuffled_deduplicated_sq | 461598 | 186,856,699 | 1.2G |
| Alemannic | als | unshuffled_original_als | 7324 | 841,750 | 5.0M | unshuffled_deduplicated_als | 4518 | 459,001 | 2.8M |
| Amharic | am | unshuffled_original_am | 83663 | 28,301,601 | 360M | unshuffled_deduplicated_am | 43102 | 16,086,628 | 206M |
| Arabic | ar | unshuffled_original_ar | 16365602 | 8,117,162,828 | 82G | unshuffled_deduplicated_ar | 9006977 | 3,171,221,354 | 32G |
| Aragonese | an | unshuffled_original_an | 2449 | 52,896 | 1.3M | unshuffled_deduplicated_an | 2025 | 45,669 | 801K |
| Armenian | hy | unshuffled_original_hy | 659430 | 273,919,388 | 3.7G | unshuffled_deduplicated_hy | 396093 | 110,196,043 | 1.5G |
| Assamese | as | unshuffled_original_as | 14985 | 6,956,663 | 113M | unshuffled_deduplicated_as | 9212 | 4,366,570 | 71M |
| Asturian | ast | unshuffled_original_ast | 6999 | 381,005 | 2.4M | unshuffled_deduplicated_ast | 5343 | 325,237 | 2.0M |
| Avaric | av | unshuffled_original_av | 456 | 24,720 | 409K | unshuffled_deduplicated_av | 360 | 19,478 | 324K |
| Azerbaijani | az | unshuffled_original_az | 912330 | 322,641,710 | 2.8G | unshuffled_deduplicated_az | 626796 | 167,742,296 | 1.5G |
| Bashkir | ba | unshuffled_original_ba | 42551 | 9,796,764 | 128M | unshuffled_deduplicated_ba | 27050 | 6,922,589 | 90M |
| Basque | eu | unshuffled_original_eu | 506883 | 120,456,652 | 848M | unshuffled_deduplicated_eu | 256513 | 45,359,710 | 342M |
| Bavarian | bar | unshuffled_original_bar | 4 | 399 | 503 | unshuffled_deduplicated_bar | 4 | 399 | 503 |
| Belarusian | be | unshuffled_original_be | 586031 | 144,579,630 | 1.8G | unshuffled_deduplicated_be | 307405 | 83,499,037 | 1.1G |
| Bengali | bn | unshuffled_original_bn | 1675515 | 623,575,733 | 11G | unshuffled_deduplicated_bn | 1114481 | 363,766,143 | 5.8G |
| Bihari | bh | unshuffled_original_bh | 336 | 8,848 | 110K | unshuffled_deduplicated_bh | 82 | 2,875 | 34K |
| Bishnupriya | bpy | unshuffled_original_bpy | 6046 | 198,286 | 4.1M | unshuffled_deduplicated_bpy | 1770 | 96,940 | 1.7M |
| Bosnian | bs | unshuffled_original_bs | 2143 | 106,448 | 447K | unshuffled_deduplicated_bs | 702 | 20,485 | 116K |
| Breton | br | unshuffled_original_br | 37085 | 5,013,241 | 29M | unshuffled_deduplicated_br | 14724 | 2,890,384 | 16M |
| Bulgarian | bg | unshuffled_original_bg | 5869686 | 2,947,648,106 | 32G | unshuffled_deduplicated_bg | 3398679 | 1,268,114,977 | 14G |
| Burmese | my | unshuffled_original_my | 232329 | 56,111,184 | 1.9G | unshuffled_deduplicated_my | 136639 | 30,102,173 | 1.1G |
| Catalan | ca | unshuffled_original_ca | 4390754 | 1,360,212,450 | 8.0G | unshuffled_deduplicated_ca | 2458067 | 729,333,440 | 4.3G |
| Cebuano | ceb | unshuffled_original_ceb | 56248 | 6,603,567 | 39M | unshuffled_deduplicated_ceb | 26145 | 3,675,024 | 24M |
| Central Bikol | bcl | unshuffled_original_bcl | 1 | 312 | 885 | unshuffled_deduplicated_bcl | 1 | 312 | 885 |
| Central Khmer | km | unshuffled_original_km | 159363 | 20,690,610 | 1.1G | unshuffled_deduplicated_km | 108346 | 10,082,245 | 581M |
| Central Kurdish | ckb | unshuffled_original_ckb | 103639 | 48,478,334 | 487M | unshuffled_deduplicated_ckb | 68210 | 18,726,721 | 226M |
| Chavacano | cbk | unshuffled_original_cbk | 1 | 130 | 520 | unshuffled_deduplicated_cbk | 1 | 130 | 520 |
| Chechen | ce | unshuffled_original_ce | 4042 | 711,051 | 8.3M | unshuffled_deduplicated_ce | 2984 | 568,146 | 6.7M |
| Chinese | zh | unshuffled_original_zh | 60137667 | 14,986,424,850 | 508G | unshuffled_deduplicated_zh | 41708901 | 6,350,215,113 | 249G |
| Chuvash | cv | unshuffled_original_cv | 20281 | 3,041,614 | 39M | unshuffled_deduplicated_cv | 10130 | 2,054,810 | 26M |
| Cornish | kw | unshuffled_original_kw | 203 | 8,329 | 44K | unshuffled_deduplicated_kw | 68 | 2,704 | 14K |
| Croatian | hr | unshuffled_original_hr | 582219 | 34,232,765 | 226M | unshuffled_deduplicated_hr | 321484 | 16,727,640 | 110M |
| Czech | cs | unshuffled_original_cs | 21001388 | 7,715,977,441 | 53G | unshuffled_deduplicated_cs | 12308039 | 3,540,997,509 | 24G |
| Danish | da | unshuffled_original_da | 7664010 | 2,637,463,889 | 16G | unshuffled_deduplicated_da | 4771098 | 1,620,091,317 | 9.5G |
| Dhivehi | dv | unshuffled_original_dv | 21018 | 7,559,472 | 126M | unshuffled_deduplicated_dv | 17024 | 4,726,660 | 79M |
| Dimli | diq | unshuffled_original_diq | 1 | 19 | 146 | unshuffled_deduplicated_diq | 1 | 19 | 146 |
| Dutch | nl | unshuffled_original_nl | 34682142 | 13,020,136,373 | 78G | unshuffled_deduplicated_nl | 20812149 | 6,598,786,137 | 39G |
| Eastern Mari | mhr | unshuffled_original_mhr | 3212 | 565,992 | 7.2M | unshuffled_deduplicated_mhr | 2515 | 469,297 | 6.0M |
| Egyptian Arabic | arz | unshuffled_original_arz | 158113 | 7,305,151 | 66M | unshuffled_deduplicated_arz | 79928 | 3,659,419 | 33M |
| Emilian-Romagnol | eml | unshuffled_original_eml | 84 | 6,376 | 25K | unshuffled_deduplicated_eml | 80 | 6,121 | 24K |
| English | en | unshuffled_original_en | 455994980 | 418,187,793,408 | 2.3T | unshuffled_deduplicated_en | 304230423 | 215,841,256,971 | 1.2T |
| Erzya | myv | unshuffled_original_myv | 6 | 90 | 1.4K | unshuffled_deduplicated_myv | 5 | 78 | 1.2K |
| Esperanto | eo | unshuffled_original_eo | 121171 | 48,486,161 | 299M | unshuffled_deduplicated_eo | 84752 | 37,324,446 | 228M |
| Estonian | et | unshuffled_original_et | 2093621 | 643,163,730 | 4.8G | unshuffled_deduplicated_et | 1172041 | 309,931,463 | 2.3G |
| Finnish | fi | unshuffled_original_fi | 8557453 | 3,196,666,419 | 27G | unshuffled_deduplicated_fi | 5326443 | 1,597,855,468 | 13G |
| French | fr | unshuffled_original_fr | 96742378 | 46,896,036,417 | 282G | unshuffled_deduplicated_fr | 59448891 | 23,206,776,649 | 138G |
| Galician | gl | unshuffled_original_gl | 544388 | 102,011,291 | 620M | unshuffled_deduplicated_gl | 284320 | 63,600,602 | 384M |
| Georgian | ka | unshuffled_original_ka | 563916 | 171,950,621 | 3.6G | unshuffled_deduplicated_ka | 372158 | 91,569,739 | 1.9G |
| German | de | unshuffled_original_de | 104913504 | 44,878,908,446 | 308G | unshuffled_deduplicated_de | 62398034 | 21,529,164,172 | 145G |
| Goan Konkani | gom | unshuffled_original_gom | 640 | 124,277 | 2.2M | unshuffled_deduplicated_gom | 484 | 102,306 | 1.8M |
| Guarani | gn | unshuffled_original_gn | 106 | 7,382 | 36K | unshuffled_deduplicated_gn | 68 | 4,680 | 24K |
| Gujarati | gu | unshuffled_original_gu | 240691 | 72,045,701 | 1.1G | unshuffled_deduplicated_gu | 169834 | 50,023,432 | 722M |
| Haitian | ht | unshuffled_original_ht | 13 | 1,014 | 3.9K | unshuffled_deduplicated_ht | 9 | 832 | 3.3K |
| Hebrew | he | unshuffled_original_he | 3808397 | 2,067,753,528 | 20G | unshuffled_deduplicated_he | 2375030 | 1,032,018,056 | 9.8G |
| Hindi | hi | unshuffled_original_hi | 3264660 | 1,372,234,782 | 17G | unshuffled_deduplicated_hi | 1909387 | 745,774,934 | 8.9G |
| Hungarian | hu | unshuffled_original_hu | 11197780 | 5,163,936,345 | 40G | unshuffled_deduplicated_hu | 6582908 | 2,339,127,555 | 18G |
| Icelandic | is | unshuffled_original_is | 625673 | 219,900,094 | 1.5G | unshuffled_deduplicated_is | 389515 | 129,818,331 | 846M |
| Ido | io | unshuffled_original_io | 694 | 25,702 | 147K | unshuffled_deduplicated_io | 617 | 22,773 | 130K |
| Iloko | ilo | unshuffled_original_ilo | 2638 | 142,942 | 874K | unshuffled_deduplicated_ilo | 1578 | 105,564 | 636K |
| Indonesian | id | unshuffled_original_id | 16236463 | 4,574,692,265 | 30G | unshuffled_deduplicated_id | 9948521 | 2,394,957,629 | 16G |
| Interlingua | ia | unshuffled_original_ia | 1040 | 180,231 | 662K | unshuffled_deduplicated_ia | 529 | 100,019 | 360K |
| Interlingue | ie | unshuffled_original_ie | 101 | 5,352 | 24K | unshuffled_deduplicated_ie | 11 | 602 | 1.6K |
| Irish | ga | unshuffled_original_ga | 83223 | 14,483,593 | 88M | unshuffled_deduplicated_ga | 46493 | 10,017,303 | 60M |
| Italian | it | unshuffled_original_it | 46981781 | 22,248,707,341 | 137G | unshuffled_deduplicated_it | 28522082 | 11,250,012,896 | 69G |
| Japanese | ja | unshuffled_original_ja | 62721527 | 4,962,979,182 | 216G | unshuffled_deduplicated_ja | 39496439 | 1,123,067,063 | 106G |
| Javanese | jv | unshuffled_original_jv | 1445 | 104,896 | 659K | unshuffled_deduplicated_jv | 1163 | 86,654 | 583K |
| Kalmyk | xal | unshuffled_original_xal | 39 | 10,277 | 113K | unshuffled_deduplicated_xal | 36 | 10,155 | 112K |
| Kannada | kn | unshuffled_original_kn | 350363 | 81,186,863 | 1.7G | unshuffled_deduplicated_kn | 251064 | 49,343,462 | 1.1G |
| Karachay-Balkar | krc | unshuffled_original_krc | 1581 | 185,436 | 2.6M | unshuffled_deduplicated_krc | 1377 | 166,496 | 2.3M |
| Kazakh | kk | unshuffled_original_kk | 524591 | 191,126,469 | 2.7G | unshuffled_deduplicated_kk | 338073 | 108,388,743 | 1.5G |
| Kirghiz | ky | unshuffled_original_ky | 146993 | 44,194,823 | 600M | unshuffled_deduplicated_ky | 86561 | 28,982,620 | 388M |
| Komi | kv | unshuffled_original_kv | 1549 | 201,404 | 2.3M | unshuffled_deduplicated_kv | 924 | 95,243 | 1.2M |
| Korean | ko | unshuffled_original_ko | 7345075 | 2,368,765,142 | 24G | unshuffled_deduplicated_ko | 3675420 | 1,120,375,149 | 12G |
| Kurdish | ku | unshuffled_original_ku | 46535 | 15,561,003 | 94M | unshuffled_deduplicated_ku | 29054 | 9,946,440 | 60M |
| Lao | lo | unshuffled_original_lo | 52910 | 4,133,311 | 174M | unshuffled_deduplicated_lo | 32652 | 2,583,342 | 114M |
| Latin | la | unshuffled_original_la | 94588 | 4,122,201 | 26M | unshuffled_deduplicated_la | 18808 | 1,328,038 | 8.3M |
| Latvian | lv | unshuffled_original_lv | 1593820 | 520,761,977 | 4.0G | unshuffled_deduplicated_lv | 843195 | 236,428,905 | 1.8G |
| Lezghian | lez | unshuffled_original_lez | 1485 | 247,646 | 3.3M | unshuffled_deduplicated_lez | 1381 | 224,871 | 3.0M |
| Limburgan | li | unshuffled_original_li | 137 | 4,730 | 29K | unshuffled_deduplicated_li | 118 | 4,283 | 27K |
| Lithuanian | lt | unshuffled_original_lt | 2977757 | 1,159,661,742 | 8.8G | unshuffled_deduplicated_lt | 1737411 | 516,183,525 | 3.9G |
| Lojban | jbo | unshuffled_original_jbo | 832 | 154,330 | 736K | unshuffled_deduplicated_jbo | 617 | 141,973 | 678K |
| Lombard | lmo | unshuffled_original_lmo | 1401 | 75,229 | 443K | unshuffled_deduplicated_lmo | 1374 | 73,665 | 433K |
| Low German | nds | unshuffled_original_nds | 18174 | 2,906,347 | 18M | unshuffled_deduplicated_nds | 8714 | 2,146,417 | 13M |
| Lower Sorbian | dsb | unshuffled_original_dsb | 65 | 1,787 | 13K | unshuffled_deduplicated_dsb | 37 | 966 | 7.1K |
| Luxembourgish | lb | unshuffled_original_lb | 34807 | 4,403,577 | 29M | unshuffled_deduplicated_lb | 21735 | 3,087,650 | 21M |
| Macedonian | mk | unshuffled_original_mk | 437871 | 189,289,873 | 2.1G | unshuffled_deduplicated_mk | 299457 | 102,849,595 | 1.2G |
| Maithili | mai | unshuffled_original_mai | 123 | 69,161 | 317K | unshuffled_deduplicated_mai | 25 | 874 | 11K |
| Malagasy | mg | unshuffled_original_mg | 17957 | 3,068,360 | 21M | unshuffled_deduplicated_mg | 13343 | 1,872,044 | 13M |
| Malay | ms | unshuffled_original_ms | 534016 | 16,696,882 | 111M | unshuffled_deduplicated_ms | 183443 | 6,045,753 | 42M |
| Malayalam | ml | unshuffled_original_ml | 603937 | 189,534,472 | 4.9G | unshuffled_deduplicated_ml | 453904 | 95,892,551 | 2.5G |
| Maltese | mt | unshuffled_original_mt | 26598 | 2,995,654 | 24M | unshuffled_deduplicated_mt | 16383 | 2,163,358 | 17M |
| Marathi | mr | unshuffled_original_mr | 326804 | 162,609,404 | 2.7G | unshuffled_deduplicated_mr | 212556 | 82,130,803 | 1.4G |
| Mazanderani | mzn | unshuffled_original_mzn | 1055 | 73,870 | 691K | unshuffled_deduplicated_mzn | 917 | 64,481 | 602K |
| Minangkabau | min | unshuffled_original_min | 220 | 5,682 | 608K | unshuffled_deduplicated_min | 166 | 4,825 | 310K |
| Mingrelian | xmf | unshuffled_original_xmf | 3783 | 299,098 | 5.8M | unshuffled_deduplicated_xmf | 2418 | 228,629 | 4.4M |
| Mirandese | mwl | unshuffled_original_mwl | 8 | 171 | 1.2K | unshuffled_deduplicated_mwl | 7 | 152 | 1.1K |
| Modern Greek | el | unshuffled_original_el | 10425596 | 5,479,180,137 | 62G | unshuffled_deduplicated_el | 6521169 | 2,412,419,435 | 27G |
| Mongolian | mn | unshuffled_original_mn | 395605 | 181,307,167 | 2.2G | unshuffled_deduplicated_mn | 197878 | 68,362,013 | 838M |
| Nahuatl languages | nah | unshuffled_original_nah | 61 | 1,234 | 12K | unshuffled_deduplicated_nah | 58 | 1,193 | 11K |
| Neapolitan | nap | unshuffled_original_nap | 73 | 5,282 | 17K | unshuffled_deduplicated_nap | 55 | 4,147 | 13K |
| Nepali | ne | unshuffled_original_ne | 299938 | 107,448,208 | 1.8G | unshuffled_deduplicated_ne | 219334 | 71,628,317 | 1.2G |
| Newari | new | unshuffled_original_new | 4696 | 564,697 | 5.5M | unshuffled_deduplicated_new | 2126 | 288,995 | 4.1M |
| Northern Frisian | frr | unshuffled_original_frr | 7 | 1,516 | 4.4K | unshuffled_deduplicated_frr | 7 | 1,516 | 4.4K |
| Northern Luri | lrc | unshuffled_original_lrc | 88 | 8,022 | 76K | unshuffled_deduplicated_lrc | 72 | 6,740 | 63K |
| Norwegian | no | unshuffled_original_no | 5546211 | 1,344,326,388 | 8.0G | unshuffled_deduplicated_no | 3229940 | 804,894,377 | 4.7G |
| Norwegian Nynorsk | nn | unshuffled_original_nn | 185884 | 14,764,980 | 85M | unshuffled_deduplicated_nn | 109118 | 9,435,139 | 54M |
| Occitan | oc | unshuffled_original_oc | 10709 | 750,301 | 5.8M | unshuffled_deduplicated_oc | 6485 | 512,678 | 3.7M |
| Oriya | or | unshuffled_original_or | 59463 | 14,938,567 | 248M | unshuffled_deduplicated_or | 44230 | 11,321,740 | 188M |
| Ossetian | os | unshuffled_original_os | 5213 | 1,031,268 | 13M | unshuffled_deduplicated_os | 2559 | 878,765 | 11M |
| Pampanga | pam | unshuffled_original_pam | 3 | 130 | 760 | unshuffled_deduplicated_pam | 1 | 52 | 304 |
| Panjabi | pa | unshuffled_original_pa | 127467 | 61,847,806 | 763M | unshuffled_deduplicated_pa | 87235 | 37,555,835 | 460M |
| Persian | fa | unshuffled_original_fa | 13704702 | 9,096,554,121 | 79G | unshuffled_deduplicated_fa | 8203495 | 4,363,505,319 | 38G |
| Piemontese | pms | unshuffled_original_pms | 3225 | 362,013 | 2.1M | unshuffled_deduplicated_pms | 2859 | 337,246 | 1.9M |
| Polish | pl | unshuffled_original_pl | 35440972 | 15,277,255,137 | 109G | unshuffled_deduplicated_pl | 20682611 | 6,708,709,674 | 47G |
| Portuguese | pt | unshuffled_original_pt | 42114520 | 20,641,903,898 | 124G | unshuffled_deduplicated_pt | 26920397 | 10,751,156,918 | 64G |
| Pushto | ps | unshuffled_original_ps | 98216 | 46,559,441 | 361M | unshuffled_deduplicated_ps | 67921 | 31,347,348 | 242M |
| Quechua | qu | unshuffled_original_qu | 452 | 10,186 | 78K | unshuffled_deduplicated_qu | 411 | 8,691 | 67K |
| Romanian | ro | unshuffled_original_ro | 9387265 | 3,984,317,058 | 25G | unshuffled_deduplicated_ro | 5044757 | 1,741,794,069 | 11G |
| Romansh | rm | unshuffled_original_rm | 41 | 1,093 | 7.4K | unshuffled_deduplicated_rm | 34 | 960 | 6.5K |
| Russia Buriat | bxr | unshuffled_original_bxr | 42 | 963 | 13K | unshuffled_deduplicated_bxr | 36 | 809 | 11K |
| Russian | ru | unshuffled_original_ru | 161836003 | 92,522,407,837 | 1.2T | unshuffled_deduplicated_ru | 115954598 | 46,692,691,520 | 568G |
| Sanskrit | sa | unshuffled_original_sa | 14291 | 4,331,569 | 93M | unshuffled_deduplicated_sa | 7121 | 1,713,930 | 37M |
| Scottish Gaelic | gd | unshuffled_original_gd | 5799 | 310,689 | 1.9M | unshuffled_deduplicated_gd | 3883 | 207,110 | 1.3M |
| Serbian | sr | unshuffled_original_sr | 1013619 | 364,395,411 | 3.9G | unshuffled_deduplicated_sr | 645747 | 207,561,168 | 2.2G |
| Serbo-Croatian | sh | unshuffled_original_sh | 36700 | 5,292,184 | 25M | unshuffled_deduplicated_sh | 17610 | 1,040,573 | 5.8M |
| Sicilian | scn | unshuffled_original_scn | 21 | 554 | 3.3K | unshuffled_deduplicated_scn | 17 | 468 | 2.8K |
| Sindhi | sd | unshuffled_original_sd | 44280 | 43,530,158 | 347M | unshuffled_deduplicated_sd | 33925 | 33,028,015 | 263M |
| Sinhala | si | unshuffled_original_si | 203082 | 93,053,465 | 1.4G | unshuffled_deduplicated_si | 120684 | 50,864,857 | 802M |
| Slovak | sk | unshuffled_original_sk | 5492194 | 1,322,247,763 | 9.1G | unshuffled_deduplicated_sk | 2820821 | 656,346,179 | 4.5G |
| Slovenian | sl | unshuffled_original_sl | 1746604 | 387,399,700 | 2.5G | unshuffled_deduplicated_sl | 886223 | 193,926,684 | 1.3G |
| Somali | so | unshuffled_original_so | 156 | 1,202 | 61K | unshuffled_deduplicated_so | 42 | 472 | 16K |
| South Azerbaijani | azb | unshuffled_original_azb | 15446 | 2,175,054 | 27M | unshuffled_deduplicated_azb | 9985 | 1,528,709 | 19M |
| Spanish | es | unshuffled_original_es | 88199221 | 47,545,122,279 | 278G | unshuffled_deduplicated_es | 56326016 | 25,928,290,729 | 149G |
| Sundanese | su | unshuffled_original_su | 805 | 30,321 | 211K | unshuffled_deduplicated_su | 511 | 20,278 | 141K |
| Swahili | sw | unshuffled_original_sw | 41986 | 2,211,927 | 13M | unshuffled_deduplicated_sw | 24803 | 1,376,963 | 8.1M |
| Swedish | sv | unshuffled_original_sv | 17395625 | 7,155,994,312 | 44G | unshuffled_deduplicated_sv | 11014487 | 4,106,120,608 | 25G |
| Tagalog | tl | unshuffled_original_tl | 458206 | 98,949,299 | 573M | unshuffled_deduplicated_tl | 294132 | 70,121,601 | 407M |
| Tajik | tg | unshuffled_original_tg | 89002 | 31,758,142 | 379M | unshuffled_deduplicated_tg | 56259 | 21,029,893 | 249M |
| Tamil | ta | unshuffled_original_ta | 1263280 | 420,537,132 | 9.3G | unshuffled_deduplicated_ta | 833101 | 226,013,330 | 5.1G |
| Tatar | tt | unshuffled_original_tt | 135923 | 51,034,893 | 670M | unshuffled_deduplicated_tt | 82738 | 23,825,695 | 305M |
| Telugu | te | unshuffled_original_te | 475703 | 123,711,517 | 2.5G | unshuffled_deduplicated_te | 312644 | 79,094,167 | 1.6G |
| Thai | th | unshuffled_original_th | 6064129 | 951,743,087 | 36G | unshuffled_deduplicated_th | 3749826 | 368,965,202 | 16G |
| Tibetan | bo | unshuffled_original_bo | 26795 | 1,483,589 | 187M | unshuffled_deduplicated_bo | 15762 | 936,556 | 138M |
| Turkish | tr | unshuffled_original_tr | 18535253 | 7,577,388,700 | 60G | unshuffled_deduplicated_tr | 11596446 | 3,365,734,289 | 27G |
| Turkmen | tk | unshuffled_original_tk | 6456 | 1,113,869 | 11M | unshuffled_deduplicated_tk | 4694 | 752,326 | 6.8M |
| Tuvinian | tyv | unshuffled_original_tyv | 34 | 759 | 12K | unshuffled_deduplicated_tyv | 24 | 540 | 7.9K |
| Uighur | ug | unshuffled_original_ug | 22255 | 8,657,141 | 122M | unshuffled_deduplicated_ug | 15503 | 5,852,225 | 83M |
| Ukrainian | uk | unshuffled_original_uk | 12973467 | 4,204,381,276 | 53G | unshuffled_deduplicated_uk | 7782375 | 2,252,380,351 | 28G |
| Upper Sorbian | hsb | unshuffled_original_hsb | 7959 | 545,351 | 4.2M | unshuffled_deduplicated_hsb | 3084 | 236,867 | 1.8M |
| Urdu | ur | unshuffled_original_ur | 638596 | 331,817,982 | 2.7G | unshuffled_deduplicated_ur | 428674 | 218,030,228 | 1.7G |
| Uzbek | uz | unshuffled_original_uz | 27537 | 2,450,256 | 21M | unshuffled_deduplicated_uz | 15074 | 1,381,644 | 12M |
| Venetian | vec | unshuffled_original_vec | 73 | 3,492 | 18K | unshuffled_deduplicated_vec | 64 | 3,199 | 17K |
| Vietnamese | vi | unshuffled_original_vi | 14898250 | 12,036,845,359 | 68G | unshuffled_deduplicated_vi | 9897709 | 5,577,159,843 | 32G |
| Volapük | vo | unshuffled_original_vo | 3366 | 321,121 | 2.0M | unshuffled_deduplicated_vo | 3317 | 318,568 | 2.0M |
| Walloon | wa | unshuffled_original_wa | 1001 | 50,720 | 273K | unshuffled_deduplicated_wa | 677 | 37,543 | 203K |
| Waray | war | unshuffled_original_war | 9760 | 397,315 | 2.5M | unshuffled_deduplicated_war | 9161 | 336,311 | 2.2M |
| Welsh | cy | unshuffled_original_cy | 157698 | 37,422,441 | 213M | unshuffled_deduplicated_cy | 98225 | 23,574,673 | 133M |
| Western Frisian | fy | unshuffled_original_fy | 33053 | 5,691,077 | 35M | unshuffled_deduplicated_fy | 20661 | 4,223,816 | 26M |
| Western Mari | mrj | unshuffled_original_mrj | 757 | 93,338 | 1.2M | unshuffled_deduplicated_mrj | 669 | 87,780 | 1.1M |
| Western Panjabi | pnb | unshuffled_original_pnb | 4599 | 1,426,986 | 12M | unshuffled_deduplicated_pnb | 3463 | 1,111,112 | 9.0M |
| Wu Chinese | wuu | unshuffled_original_wuu | 214 | 11,189 | 109K | unshuffled_deduplicated_wuu | 64 | 4,333 | 32K |
| Yakut | sah | unshuffled_original_sah | 22301 | 2,547,623 | 42M | unshuffled_deduplicated_sah | 8555 | 1,789,174 | 26M |
| Yiddish | yi | unshuffled_original_yi | 59364 | 13,834,320 | 141M | unshuffled_deduplicated_yi | 32919 | 8,212,970 | 84M |
| Yoruba | yo | unshuffled_original_yo | 214 | 8,906 | 55K | unshuffled_deduplicated_yo | 49 | 3,518 | 27K |
| Yue Chinese | yue | unshuffled_original_yue | 11 | 186 | 3.7K | unshuffled_deduplicated_yue | 7 | 128 | 2.2K |
</details>
## Dataset Creation
### Curation Rationale
OSCAR was constructed new pipeline derived from the [fastText's one](https://github.com/facebookresearch/fastText), called [_goclassy_](https://github.com/pjox/goclassy). Goclassy reuses the [fastText linear classifier](https://fasttext.cc) and the pre-trained fastText model for language recognition, but it completely rewrites and parallelises their pipeline in an asynchronous manner.
The order of operations is more or less the same as in the fastText pre-processing pipeline but instead of clustering multiple operations into a single blocking process, a worker is launched for each operation but bounding the number of possible parallel operations at a given time by the number of available threads instead of the number of CPUs. Goclassy is implemented in the [Go programming language](https://golang.org/) so it lets the [Go runtime](https://golang.org/src/runtime/mprof.go) handle the scheduling of the processes. Thus the goclassy's pipeline one does not have to wait for a whole WET file to download, decompress and classify in order to start downloading and processing the next one, a new file will start downloading and processing as soon as the scheduler is able to allocate a new process.
Filtering and cleaning processes at line level are done before feeding each line to the classifier. Lines shorter than 100 UTF-8 characters and lines containing invalid UTF-8 characters are discarted and are not classified. After all files are proccesed the deduplicated versions are constructed and everything is then splitted in shards and compressed.
### Source Data
#### Initial Data Collection and Normalization
[Common Crawl](https://commoncrawl.org/) is a non-profit foundation which produces and maintains an open repository of web crawled data that is both accessible and analysable. Common Crawl's complete web archive consists of petabytes of data collected over 8 years of web crawling. The repository contains raw web page HTML data (WARC files), metdata extracts (WAT files) and plain text extracts (WET files). The organisation's crawlers has always respected [nofollow](http://microformats.org/wiki/rel-nofollow) and [robots.txt](https://www.robotstxt.org/) policies.
Each monthly Common Crawl snapshot is in itself a massive multilingual corpus, where every single file contains data coming from multiple web pages written in a large variety of languages and covering all possible types of topics.
To construct OSCAR the WET files of Common Crawl were used. These contain the extracted plain texts from the websites mostly converted to UTF-8, as well as headers containing the metatada of each crawled document. Each WET file comes compressed in gzip format and is stored on Amazon Web Services. In the case of OSCAR, the **November 2018** snapshot was used. It surpasses 20TB of uncompressed data and contains more than 50 thousand plain text files where each file consists of the plain text from multiple websites along its metadata header.
#### Who are the source language producers?
The data comes from multiple web pages in a large variety of languages.
### Annotations
The dataset does not contain any additional annotations.
#### Annotation process
N/A
#### Who are the annotators?
N/A
### Personal and Sensitive Information
Being constructed from Common Crawl, Personal and sensitive information might be present. This **must** be considered before training deep learning models with OSCAR, specially in the case of text-generation models.
## Considerations for Using the Data
### Social Impact of Dataset
OSCAR is intended to bring more data to a wide variety of lanuages, the aim of the corpus is to make large amounts of data available to lower resource languages in order to facilitate the pre-training of state-of-the-art language modeling architectures.
### Discussion of Biases
OSCAR is not properly filtered yet and this can be reflected on the models trained with it. Care is advised specially concerning biases of the resulting models.
### Other Known Limitations
The [fastText linear classifier](https://fasttext.cc) is limed both in performance and the variety of languages it can recognize, so the quality of some OSCAR sub-corpora might be lower than expected, specially for the lowest-resource langiuages. Some audits have already been done by [third parties](https://arxiv.org/abs/2010.14571).
## Additional Information
### Dataset Curators
The corpus was put together by [Pedro J. Ortiz](https://pjortiz.eu/), [Benoît Sagot](http://pauillac.inria.fr/~sagot/), and [Laurent Romary](https://cv.archives-ouvertes.fr/laurentromary), during work done at [Inria](https://www.inria.fr/en), particularly at the [ALMAnaCH team](https://team.inria.fr/almanach/).
### Licensing Information
These data are released under this licensing scheme
We do not own any of the text from which these data has been extracted.
We license the actual packaging of these data under the Creative Commons CC0 license ("no rights reserved") http://creativecommons.org/publicdomain/zero/1.0/
To the extent possible under law, Inria has waived all copyright and related or neighboring rights to OSCAR
This work is published from: France.
Should you consider that our data contains material that is owned by you and should therefore not be reproduced here, please:
* Clearly identify yourself, with detailed contact data such as an address, telephone number or email address at which you can be contacted.
* Clearly identify the copyrighted work claimed to be infringed.
* Clearly identify the material that is claimed to be infringing and information reasonably sufficient to allow us to locate the material.
We will comply to legitimate requests by removing the affected sources from the next release of the corpus.
### Citation Information
```
@inproceedings{ortiz-suarez-etal-2020-monolingual,
title = "A Monolingual Approach to Contextualized Word Embeddings for Mid-Resource Languages",
author = "Ortiz Su{'a}rez, Pedro Javier and
Romary, Laurent and
Sagot, Benoit",
booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics",
month = jul,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.acl-main.156",
pages = "1703--1714",
abstract = "We use the multilingual OSCAR corpus, extracted from Common Crawl via language classification, filtering and cleaning, to train monolingual contextualized word embeddings (ELMo) for five mid-resource languages. We then compare the performance of OSCAR-based and Wikipedia-based ELMo embeddings for these languages on the part-of-speech tagging and parsing tasks. We show that, despite the noise in the Common-Crawl-based OSCAR data, embeddings trained on OSCAR perform much better than monolingual embeddings trained on Wikipedia. They actually equal or improve the current state of the art in tagging and parsing for all five languages. In particular, they also improve over multilingual Wikipedia-based contextual embeddings (multilingual BERT), which almost always constitutes the previous state of the art, thereby showing that the benefit of a larger, more diverse corpus surpasses the cross-lingual benefit of multilingual embedding architectures.",
}
@inproceedings{OrtizSuarezSagotRomary2019,
author = {Pedro Javier {Ortiz Su{'a}rez} and Benoit Sagot and Laurent Romary},
title = {Asynchronous pipelines for processing huge corpora on medium to low resource infrastructures},
series = {Proceedings of the Workshop on Challenges in the Management of Large Corpora (CMLC-7) 2019. Cardiff, 22nd July 2019},
editor = {Piotr Bański and Adrien Barbaresi and Hanno Biber and Evelyn Breiteneder and Simon Clematide and Marc Kupietz and Harald L{"u}ngen and Caroline Iliadi},
publisher = {Leibniz-Institut f{"u}r Deutsche Sprache},
address = {Mannheim},
doi = {10.14618/ids-pub-9021},
url = {http://nbn-resolving.de/urn:nbn:de:bsz:mh39-90215},
pages = {9 -- 16},
year = {2019},
abstract = {Common Crawl is a considerably large, heterogeneous multilingual corpus comprised of crawled documents from the internet, surpassing 20TB of data and distributed as a set of more than 50 thousand plain text files where each contains many documents written in a wide variety of languages. Even though each document has a metadata block associated to it, this data lacks any information about the language in which each document is written, making it extremely difficult to use Common Crawl for monolingual applications. We propose a general, highly parallel, multithreaded pipeline to clean and classify Common Crawl by language; we specifically design it so that it runs efficiently on medium to low resource infrastructures where I/O speeds are the main constraint. We develop the pipeline so that it can be easily reapplied to any kind of heterogeneous corpus and so that it can be parameterised to a wide range of infrastructures. We also distribute a 6.3TB version of Common Crawl, filtered, classified by language, shuffled at line level in order to avoid copyright issues, and ready to be used for NLP applications.},
language = {en}
}
```
### Contributions
Thanks to [@pjox](https://github.com/pjox) and [@lhoestq](https://github.com/lhoestq) for adding this dataset. |
llm-lens/descriptors-text-davinci-003 | 2023-06-29T02:39:27.000Z | [
"region:us"
] | llm-lens | null | null | null | 0 | 6,614 | ---
dataset_info:
features:
- name: vocab
dtype: string
- name: descriptions
sequence: string
- name: prompt_descriptions
sequence: string
splits:
- name: birdsnap
num_bytes: 322488
num_examples: 500
- name: caltech101
num_bytes: 56880
num_examples: 102
- name: cifar100
num_bytes: 54081
num_examples: 100
- name: coco
num_bytes: 45186
num_examples: 80
- name: dtd
num_bytes: 25204
num_examples: 47
- name: fgvc
num_bytes: 74126
num_examples: 100
- name: food101
num_bytes: 58525
num_examples: 101
- name: full
num_bytes: 3000578
num_examples: 4951
- name: imagenet21k
num_bytes: 604897
num_examples: 998
- name: lvis
num_bytes: 679195
num_examples: 1198
- name: oxfordflowers
num_bytes: 58560
num_examples: 102
- name: oxfordpets
num_bytes: 22322
num_examples: 37
- name: stanfordcars
num_bytes: 157786
num_examples: 196
- name: sun397
num_bytes: 243017
num_examples: 362
- name: visualgenome
num_bytes: 1092697
num_examples: 1913
download_size: 3257945
dataset_size: 6495542
---
# Dataset Card for "descriptors-text-davinci-003"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
open-llm-leaderboard/details_GOAT-AI__GOAT-7B-Community | 2023-09-22T17:15:05.000Z | [
"region:us"
] | open-llm-leaderboard | null | null | null | 0 | 6,469 | ---
pretty_name: Evaluation run of GOAT-AI/GOAT-7B-Community
dataset_summary: "Dataset automatically created during the evaluation run of model\
\ [GOAT-AI/GOAT-7B-Community](https://huggingface.co/GOAT-AI/GOAT-7B-Community)\
\ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\
\nThe dataset is composed of 64 configuration, each one coresponding to one of the\
\ evaluated task.\n\nThe dataset has been created from 2 run(s). Each run can be\
\ found as a specific split in each configuration, the split being named using the\
\ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\
\nAn additional configuration \"results\" store all the aggregated results of the\
\ run (and is used to compute and display the agregated metrics on the [Open LLM\
\ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\
\nTo load the details from a run, you can for instance do the following:\n```python\n\
from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_GOAT-AI__GOAT-7B-Community\"\
,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\
These are the [latest results from run 2023-09-22T17:14:52.967997](https://huggingface.co/datasets/open-llm-leaderboard/details_GOAT-AI__GOAT-7B-Community/blob/main/results_2023-09-22T17-14-52.967997.json)(note\
\ that their might be results for other tasks in the repos if successive evals didn't\
\ cover the same tasks. You find each in the results and the \"latest\" split for\
\ each eval):\n\n```python\n{\n \"all\": {\n \"em\": 0.005243288590604027,\n\
\ \"em_stderr\": 0.0007396052260778031,\n \"f1\": 0.06909395973154382,\n\
\ \"f1_stderr\": 0.0015832414439852427,\n \"acc\": 0.3838492484021702,\n\
\ \"acc_stderr\": 0.009135888573374731\n },\n \"harness|drop|3\": {\n\
\ \"em\": 0.005243288590604027,\n \"em_stderr\": 0.0007396052260778031,\n\
\ \"f1\": 0.06909395973154382,\n \"f1_stderr\": 0.0015832414439852427\n\
\ },\n \"harness|gsm8k|5\": {\n \"acc\": 0.04473085670962851,\n \
\ \"acc_stderr\": 0.005693886131407048\n },\n \"harness|winogrande|5\"\
: {\n \"acc\": 0.7229676400947119,\n \"acc_stderr\": 0.012577891015342414\n\
\ }\n}\n```"
repo_url: https://huggingface.co/GOAT-AI/GOAT-7B-Community
leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
point_of_contact: clementine@hf.co
configs:
- config_name: harness_arc_challenge_25
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|arc:challenge|25_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|arc:challenge|25_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_drop_3
data_files:
- split: 2023_09_22T17_14_52.967997
path:
- '**/details_harness|drop|3_2023-09-22T17-14-52.967997.parquet'
- split: latest
path:
- '**/details_harness|drop|3_2023-09-22T17-14-52.967997.parquet'
- config_name: harness_gsm8k_5
data_files:
- split: 2023_09_22T17_14_52.967997
path:
- '**/details_harness|gsm8k|5_2023-09-22T17-14-52.967997.parquet'
- split: latest
path:
- '**/details_harness|gsm8k|5_2023-09-22T17-14-52.967997.parquet'
- config_name: harness_hellaswag_10
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hellaswag|10_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hellaswag|10_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-anatomy|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-astronomy|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-business_ethics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-college_biology|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-college_medicine|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-college_physics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-computer_security|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-econometrics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-formal_logic|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-global_facts|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-human_aging|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-international_law|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-machine_learning|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-management|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-marketing|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-nutrition|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-philosophy|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-prehistory|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-professional_law|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-public_relations|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-security_studies|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-sociology|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-virology|5_2023-07-25T12:51:32.230763.parquet'
- '**/details_harness|hendrycksTest-world_religions|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_abstract_algebra_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-abstract_algebra|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_anatomy_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-anatomy|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_astronomy_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-astronomy|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_business_ethics_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-business_ethics|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_clinical_knowledge_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_college_biology_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_biology|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_college_chemistry_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_chemistry|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_college_computer_science_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_computer_science|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_college_mathematics_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_mathematics|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_college_medicine_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_medicine|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_college_physics_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-college_physics|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_computer_security_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-computer_security|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_conceptual_physics_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-conceptual_physics|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_econometrics_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-econometrics|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_electrical_engineering_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-electrical_engineering|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_elementary_mathematics_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_formal_logic_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-formal_logic|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_global_facts_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-global_facts|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_high_school_biology_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_biology|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_high_school_chemistry_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_high_school_computer_science_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_high_school_european_history_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_european_history|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_high_school_geography_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_geography|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_high_school_government_and_politics_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_high_school_macroeconomics_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_high_school_mathematics_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_high_school_microeconomics_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_high_school_physics_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_physics|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_high_school_psychology_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_psychology|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_high_school_statistics_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_statistics|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_high_school_us_history_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_us_history|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_high_school_world_history_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-high_school_world_history|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_human_aging_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_aging|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_human_sexuality_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-human_sexuality|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_international_law_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-international_law|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_jurisprudence_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-jurisprudence|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_logical_fallacies_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-logical_fallacies|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_machine_learning_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-machine_learning|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_management_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-management|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-management|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_marketing_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-marketing|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_medical_genetics_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-medical_genetics|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_miscellaneous_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-miscellaneous|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_moral_disputes_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_disputes|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_moral_scenarios_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-moral_scenarios|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_nutrition_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-nutrition|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_philosophy_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-philosophy|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_prehistory_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-prehistory|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_professional_accounting_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_accounting|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_professional_law_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_law|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_professional_medicine_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_medicine|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_professional_psychology_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-professional_psychology|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_public_relations_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-public_relations|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_security_studies_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-security_studies|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_sociology_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-sociology|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_us_foreign_policy_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_virology_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-virology|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-virology|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_hendrycksTest_world_religions_5
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|hendrycksTest-world_religions|5_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_truthfulqa_mc_0
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- '**/details_harness|truthfulqa:mc|0_2023-07-25T12:51:32.230763.parquet'
- split: latest
path:
- '**/details_harness|truthfulqa:mc|0_2023-07-25T12:51:32.230763.parquet'
- config_name: harness_winogrande_5
data_files:
- split: 2023_09_22T17_14_52.967997
path:
- '**/details_harness|winogrande|5_2023-09-22T17-14-52.967997.parquet'
- split: latest
path:
- '**/details_harness|winogrande|5_2023-09-22T17-14-52.967997.parquet'
- config_name: results
data_files:
- split: 2023_07_25T12_51_32.230763
path:
- results_2023-07-25T12:51:32.230763.parquet
- split: 2023_09_22T17_14_52.967997
path:
- results_2023-09-22T17-14-52.967997.parquet
- split: latest
path:
- results_2023-09-22T17-14-52.967997.parquet
---
# Dataset Card for Evaluation run of GOAT-AI/GOAT-7B-Community
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/GOAT-AI/GOAT-7B-Community
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** clementine@hf.co
### Dataset Summary
Dataset automatically created during the evaluation run of model [GOAT-AI/GOAT-7B-Community](https://huggingface.co/GOAT-AI/GOAT-7B-Community) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 64 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 2 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_GOAT-AI__GOAT-7B-Community",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-09-22T17:14:52.967997](https://huggingface.co/datasets/open-llm-leaderboard/details_GOAT-AI__GOAT-7B-Community/blob/main/results_2023-09-22T17-14-52.967997.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"em": 0.005243288590604027,
"em_stderr": 0.0007396052260778031,
"f1": 0.06909395973154382,
"f1_stderr": 0.0015832414439852427,
"acc": 0.3838492484021702,
"acc_stderr": 0.009135888573374731
},
"harness|drop|3": {
"em": 0.005243288590604027,
"em_stderr": 0.0007396052260778031,
"f1": 0.06909395973154382,
"f1_stderr": 0.0015832414439852427
},
"harness|gsm8k|5": {
"acc": 0.04473085670962851,
"acc_stderr": 0.005693886131407048
},
"harness|winogrande|5": {
"acc": 0.7229676400947119,
"acc_stderr": 0.012577891015342414
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] |
swag | 2023-01-25T14:45:08.000Z | [
"task_categories:text-classification",
"task_ids:natural-language-inference",
"annotations_creators:crowdsourced",
"annotations_creators:machine-generated",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:100K<n<1M",
"source_datasets:original",
"language:en",
"license:unknown",
"arxiv:1808.05326",
"region:us"
] | null | Given a partial description like "she opened the hood of the car,"
humans can reason about the situation and anticipate what might come
next ("then, she examined the engine"). SWAG (Situations With Adversarial Generations)
is a large-scale dataset for this task of grounded commonsense
inference, unifying natural language inference and physically grounded reasoning.
The dataset consists of 113k multiple choice questions about grounded situations
(73k training, 20k validation, 20k test).
Each question is a video caption from LSMDC or ActivityNet Captions,
with four answer choices about what might happen next in the scene.
The correct answer is the (real) video caption for the next event in the video;
the three incorrect answers are adversarially generated and human verified,
so as to fool machines but not humans. SWAG aims to be a benchmark for
evaluating grounded commonsense NLI and for learning representations.
The full data contain more information,
but the regular configuration will be more interesting for modeling
(note that the regular data are shuffled). The test set for leaderboard submission
is under the regular configuration. | @inproceedings{zellers2018swagaf,
title={SWAG: A Large-Scale Adversarial Dataset for Grounded Commonsense Inference},
author={Zellers, Rowan and Bisk, Yonatan and Schwartz, Roy and Choi, Yejin},
booktitle = "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (EMNLP)",
year={2018}
} | null | 11 | 6,407 | ---
annotations_creators:
- crowdsourced
- machine-generated
language_creators:
- found
language:
- en
license:
- unknown
multilinguality:
- monolingual
size_categories:
- 100K<n<1M
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- natural-language-inference
paperswithcode_id: swag
pretty_name: Situations With Adversarial Generations
dataset_info:
- config_name: regular
features:
- name: video-id
dtype: string
- name: fold-ind
dtype: string
- name: startphrase
dtype: string
- name: sent1
dtype: string
- name: sent2
dtype: string
- name: gold-source
dtype: string
- name: ending0
dtype: string
- name: ending1
dtype: string
- name: ending2
dtype: string
- name: ending3
dtype: string
- name: label
dtype:
class_label:
names:
'0': '0'
'1': '1'
'2': '2'
'3': '3'
splits:
- name: train
num_bytes: 30274672
num_examples: 73546
- name: validation
num_bytes: 8451771
num_examples: 20006
- name: test
num_bytes: 8417644
num_examples: 20005
download_size: 43954806
dataset_size: 47144087
- config_name: full
features:
- name: video-id
dtype: string
- name: fold-ind
dtype: string
- name: startphrase
dtype: string
- name: gold-ending
dtype: string
- name: distractor-0
dtype: string
- name: distractor-1
dtype: string
- name: distractor-2
dtype: string
- name: distractor-3
dtype: string
- name: gold-source
dtype: string
- name: gold-type
dtype: string
- name: distractor-0-type
dtype: string
- name: distractor-1-type
dtype: string
- name: distractor-2-type
dtype: string
- name: distractor-3-type
dtype: string
- name: sent1
dtype: string
- name: sent2
dtype: string
splits:
- name: train
num_bytes: 34941649
num_examples: 73546
- name: validation
num_bytes: 9832603
num_examples: 20006
download_size: 40537624
dataset_size: 44774252
---
# Dataset Card for Situations With Adversarial Generations
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [SWAG AF](https://rowanzellers.com/swag/)
- **Repository:** [Github repository](https://github.com/rowanz/swagaf/tree/master/data)
- **Paper:** [SWAG: A Large-Scale Adversarial Dataset for Grounded Commonsense Inference](https://arxiv.org/abs/1808.05326)
- **Leaderboard:** [SWAG Leaderboard](https://leaderboard.allenai.org/swag)
- **Point of Contact:** [Rowan Zellers](https://rowanzellers.com/#contact)
### Dataset Summary
Given a partial description like "she opened the hood of the car,"
humans can reason about the situation and anticipate what might come
next ("then, she examined the engine"). SWAG (Situations With Adversarial Generations)
is a large-scale dataset for this task of grounded commonsense
inference, unifying natural language inference and physically grounded reasoning.
The dataset consists of 113k multiple choice questions about grounded situations
(73k training, 20k validation, 20k test).
Each question is a video caption from LSMDC or ActivityNet Captions,
with four answer choices about what might happen next in the scene.
The correct answer is the (real) video caption for the next event in the video;
the three incorrect answers are adversarially generated and human verified,
so as to fool machines but not humans. SWAG aims to be a benchmark for
evaluating grounded commonsense NLI and for learning representations.
### Supported Tasks and Leaderboards
The dataset introduces the task of grounded commonsense inference, unifying natural language inference and commonsense reasoning.
### Languages
The text in the dataset is in English. The associated BCP-47 code is `en`.
## Dataset Structure
### Data Instances
The `regular` configuration should be used for modeling. An example looks like this:
```
{
"video-id": "anetv_dm5WXFiQZUQ",
"fold-ind": "18419",
"startphrase", "He rides the motorcycle down the hall and into the elevator. He",
"sent1": "He rides the motorcycle down the hall and into the elevator."
"sent2": "He",
"gold-source": "gold",
"ending0": "looks at a mirror in the mirror as he watches someone walk through a door.",
"ending1": "stops, listening to a cup of coffee with the seated woman, who's standing.",
"ending2": "exits the building and rides the motorcycle into a casino where he performs several tricks as people watch.",
"ending3": "pulls the bag out of his pocket and hands it to someone's grandma.",
"label": 2,
}
```
Note that the test are reseved for blind submission on the leaderboard.
The full train and validation sets provide more information regarding the collection process.
### Data Fields
- `video-id`: identification
- `fold-ind`: identification
- `startphrase`: the context to be filled
- `sent1`: the first sentence
- `sent2`: the start of the second sentence (to be filled)
- `gold-source`: generated or comes from the found completion
- `ending0`: first proposition
- `ending1`: second proposition
- `ending2`: third proposition
- `ending3`: fourth proposition
- `label`: the correct proposition
More info concerning the fields can be found [on the original repo](https://github.com/rowanz/swagaf/tree/master/data).
### Data Splits
The dataset consists of 113k multiple choice questions about grounded situations: 73k for training, 20k for validation, and 20k for (blind) test.
## Dataset Creation
### Curation Rationale
The authors seek dataset diversity while minimizing annotation artifacts, conditional stylistic patterns such as length and word-preference biases. To avoid introducing easily “gamed” patterns, they introduce Adversarial Filtering (AF), a generally- applicable treatment involving the iterative refinement of a set of assignments to increase the entropy under a chosen model family. The dataset is then human verified by paid crowdsourcers.
### Source Data
This section describes the source data (e.g. news text and headlines, social media posts, translated sentences,...)
#### Initial Data Collection and Normalization
The dataset is derived from pairs of consecutive video captions from [ActivityNet Captions](https://cs.stanford.edu/people/ranjaykrishna/densevid/) and the [Large Scale Movie Description Challenge](https://sites.google.com/site/describingmovies/). The two datasets are slightly different in nature and allow us to achieve broader coverage: ActivityNet contains 20k YouTube clips containing one of 203 activity types (such as doing gymnastics or playing guitar); LSMDC consists of 128k movie captions (audio descriptions and scripts).
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
Annotations are first machine generated and then adversarially filtered. Finally, the remaining examples are human-verified by paid crowdsourcers.
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
Unknown
### Citation Information
```
@inproceedings{zellers2018swagaf,
title={SWAG: A Large-Scale Adversarial Dataset for Grounded Commonsense Inference},
author={Zellers, Rowan and Bisk, Yonatan and Schwartz, Roy and Choi, Yejin},
booktitle = "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (EMNLP)",
year={2018}
}
```
### Contributions
Thanks to [@VictorSanh](https://github.com/VictorSanh) for adding this dataset. |
cosmos_qa | 2023-04-05T10:02:42.000Z | [
"task_categories:multiple-choice",
"task_ids:multiple-choice-qa",
"annotations_creators:crowdsourced",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:cc-by-4.0",
"arxiv:1909.00277",
"region:us"
] | null | Cosmos QA is a large-scale dataset of 35.6K problems that require commonsense-based reading comprehension, formulated as multiple-choice questions. It focuses on reading between the lines over a diverse collection of people's everyday narratives, asking questions concerning on the likely causes or effects of events that require reasoning beyond the exact text spans in the context | @inproceedings{huang-etal-2019-cosmos,
title = "Cosmos {QA}: Machine Reading Comprehension with Contextual Commonsense Reasoning",
author = "Huang, Lifu and
Le Bras, Ronan and
Bhagavatula, Chandra and
Choi, Yejin",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)",
month = nov,
year = "2019",
address = "Hong Kong, China",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/D19-1243",
doi = "10.18653/v1/D19-1243",
pages = "2391--2401",
} | null | 9 | 6,357 | ---
annotations_creators:
- crowdsourced
language:
- en
language_creators:
- found
license:
- cc-by-4.0
multilinguality:
- monolingual
pretty_name: CosmosQA
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- multiple-choice
task_ids:
- multiple-choice-qa
paperswithcode_id: cosmosqa
dataset_info:
features:
- name: id
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answer0
dtype: string
- name: answer1
dtype: string
- name: answer2
dtype: string
- name: answer3
dtype: string
- name: label
dtype: int32
splits:
- name: train
num_bytes: 17159918
num_examples: 25262
- name: test
num_bytes: 5121479
num_examples: 6963
- name: validation
num_bytes: 2186987
num_examples: 2985
download_size: 24399475
dataset_size: 24468384
---
# Dataset Card for "cosmos_qa"
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://wilburone.github.io/cosmos/](https://wilburone.github.io/cosmos/)
- **Repository:** https://github.com/wilburOne/cosmosqa/
- **Paper:** [Cosmos QA: Machine Reading Comprehension with Contextual Commonsense Reasoning](https://arxiv.org/abs/1909.00277)
- **Point of Contact:** [Lifu Huang](mailto:warrior.fu@gmail.com)
- **Size of downloaded dataset files:** 24.40 MB
- **Size of the generated dataset:** 24.51 MB
- **Total amount of disk used:** 48.91 MB
### Dataset Summary
Cosmos QA is a large-scale dataset of 35.6K problems that require commonsense-based reading comprehension, formulated as multiple-choice questions. It focuses on reading between the lines over a diverse collection of people's everyday narratives, asking questions concerning on the likely causes or effects of events that require reasoning beyond the exact text spans in the context
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Dataset Structure
### Data Instances
#### default
- **Size of downloaded dataset files:** 24.40 MB
- **Size of the generated dataset:** 24.51 MB
- **Total amount of disk used:** 48.91 MB
An example of 'validation' looks as follows.
```
This example was too long and was cropped:
{
"answer0": "If he gets married in the church he wo nt have to get a divorce .",
"answer1": "He wants to get married to a different person .",
"answer2": "He wants to know if he does nt like this girl can he divorce her ?",
"answer3": "None of the above choices .",
"context": "\"Do i need to go for a legal divorce ? I wanted to marry a woman but she is not in the same religion , so i am not concern of th...",
"id": "3BFF0DJK8XA7YNK4QYIGCOG1A95STE##3180JW2OT5AF02OISBX66RFOCTG5J7##A2LTOS0AZ3B28A##Blog_56156##q1_a1##378G7J1SJNCDAAIN46FM2P7T6KZEW2",
"label": 1,
"question": "Why is this person asking about divorce ?"
}
```
### Data Fields
The data fields are the same among all splits.
#### default
- `id`: a `string` feature.
- `context`: a `string` feature.
- `question`: a `string` feature.
- `answer0`: a `string` feature.
- `answer1`: a `string` feature.
- `answer2`: a `string` feature.
- `answer3`: a `string` feature.
- `label`: a `int32` feature.
### Data Splits
| name |train|validation|test|
|-------|----:|---------:|---:|
|default|25262| 2985|6963|
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
As reported via email by Yejin Choi, the dataset is licensed under [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) license.
### Citation Information
```
@inproceedings{huang-etal-2019-cosmos,
title = "Cosmos {QA}: Machine Reading Comprehension with Contextual Commonsense Reasoning",
author = "Huang, Lifu and
Le Bras, Ronan and
Bhagavatula, Chandra and
Choi, Yejin",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)",
month = nov,
year = "2019",
address = "Hong Kong, China",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/D19-1243",
doi = "10.18653/v1/D19-1243",
pages = "2391--2401",
}
```
### Contributions
Thanks to [@patrickvonplaten](https://github.com/patrickvonplaten), [@lewtun](https://github.com/lewtun), [@albertvillanova](https://github.com/albertvillanova), [@thomwolf](https://github.com/thomwolf) for adding this dataset. |
tau/zero_scrolls | 2023-06-30T17:21:02.000Z | [
"task_categories:question-answering",
"task_categories:summarization",
"task_categories:text-generation",
"task_ids:multiple-choice-qa",
"language:en",
"query-based-summarization",
"long-texts",
"arxiv:2104.02112",
"arxiv:2104.07091",
"arxiv:2104.05938",
"arxiv:2205.11465",
"arxiv:2105.03011",
"arxiv:1712.07040",
"arxiv:2112.08608",
"arxiv:2108.00573",
"arxiv:2305.14196",
"region:us"
] | tau | ZeroSCROLLS: Zero-Shot CompaRison Over Long Language Sequences.
A zero shot benchmark for long text reasoning.
https://zero.scrolls-benchmark.com/ | @misc{shaham2023zeroscrolls,
title={ZeroSCROLLS: A Zero-Shot Benchmark for Long Text Understanding},
author={Uri Shaham and Maor Ivgi and Avia Efrat and Jonathan Berant and Omer Levy},
year={2023},
eprint={2305.14196},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
Note that each ZeroSCROLLS task has its own citation. Please see the source to
get the correct citation for each one. | null | 4 | 6,326 | ---
language:
- en
task_categories:
- question-answering
- summarization
- text-generation
task_ids:
- multiple-choice-qa
tags:
- query-based-summarization
- long-texts
---
## Dataset Description
- **Homepage:** [ZeroSCROLLS](https://www.zero.scrolls-benchmark.com/)
- **Leaderboard:** [Leaderboard](https://www.zero.scrolls-benchmark.com/leaderboard)
- **Point of Contact:** [scrolls-benchmark-contact@googlegroups.com](scrolls-benchmark-contact@googlegroups.com)
# Dataset Card for ZeroSCROLLS
## Overview
ZeroSCROLLS is a zero-shot benchmark for natural language understanding over long texts.
The validation sets contain only ~20 examples per task and are meant for eyeballing alone.
## Leaderboard
The ZeroSCROLLS benchmark leaderboard can be found [here](https://www.zero.scrolls-benchmark.com/leaderboard).
## Tasks
ZeroSCROLLS contains the following tasks:
#### GovReport ([Huang et al., 2021](https://arxiv.org/pdf/2104.02112.pdf))
GovReport is a summarization dataset of reports addressing various national policy issues published by the
Congressional Research Service and the U.S. Government Accountability Office, where each document is paired with a hand-written executive summary.
The reports and their summaries are longer than their equivalents in other popular long-document summarization datasets;
for example, GovReport's documents are approximately 1.5 and 2.5 times longer than the documents in Arxiv and PubMed, respectively.
#### SummScreenFD ([Chen et al., 2022](https://arxiv.org/pdf/2104.07091.pdf))
SummScreenFD is a summarization dataset in the domain of TV shows (e.g. Friends, Game of Thrones).
Given a transcript of a specific episode, the goal is to produce the episode's recap.
The original dataset is divided into two complementary subsets, based on the source of its community contributed transcripts.
For SCROLLS, we use the ForeverDreaming (FD) subset, as it incorporates 88 different shows,
making it a more diverse alternative to the TV MegaSite (TMS) subset, which has only 10 shows.
Community-authored recaps for the ForeverDreaming transcripts were collected from English Wikipedia and TVMaze.
#### QMSum ([Zhong et al., 2021](https://arxiv.org/pdf/2104.05938.pdf))
QMSum is a query-based summarization dataset, consisting of 232 meetings transcripts from multiple domains.
The corpus covers academic group meetings at the International Computer Science Institute and their summaries, industrial product meetings for designing a remote control,
and committee meetings of the Welsh and Canadian Parliaments, dealing with a variety of public policy issues.
Annotators were tasked with writing queries about the broad contents of the meetings, as well as specific questions about certain topics or decisions,
while ensuring that the relevant text for answering each query spans at least 200 words or 10 turns.
#### SQuALITY ([Wang et al., 2022](https://arxiv.org/pdf/2205.11465.pdf))
SQuALITY (Wang et al., 2022) is a question-focused summarization dataset, where given a story from Project Gutenberg,
the task is to produce a summary of the story or aspects of it based on a guiding question.
The questions and summaries are original and crowdsourced; experienced writers were guided to design questions that require reading significant parts of the story to answer correctly.
#### Qasper ([Dasigi et al., 2021](https://arxiv.org/pdf/2105.03011.pdf))
Qasper is a question answering dataset over NLP papers filtered from the Semantic Scholar Open Research Corpus (S2ORC).
Questions were written by NLP practitioners after reading only the title and abstract of the papers,
while another set of NLP practitioners annotated the answers given the entire document.
Qasper contains abstractive, extractive, and yes/no questions, as well as unanswerable ones.
#### NarrativeQA ([Kočiský et al., 2018](https://arxiv.org/pdf/1712.07040.pdf))
NarrativeQA (Kočiský et al., 2021) is an established question answering dataset over entire books from Project Gutenberg and movie scripts from different websites.
Annotators were given summaries of the books and scripts obtained from Wikipedia, and asked to generate question-answer pairs,
resulting in about 30 questions and answers for each of the 1,567 books and scripts.
They were encouraged to use their own words rather then copying, and avoid asking yes/no questions or ones about the cast.
Each question was then answered by an additional annotator, providing each question with two reference answers (unless both answers are identical).
#### QuALITY ([Pang et al., 2022](https://arxiv.org/pdf/2112.08608.pdf))
QuALITY is a multiple-choice question answering dataset over articles and stories sourced from Project Gutenberg,
the Open American National Corpus, and more.
Experienced writers wrote questions and distractors, and were incentivized to write answerable, unambiguous questions such that in order to correctly answer them,
human annotators must read large portions of the given document.
Reference answers were then calculated using the majority vote between of the annotators and writer's answers.
To measure the difficulty of their questions, Pang et al. conducted a speed validation process,
where another set of annotators were asked to answer questions given only a short period of time to skim through the document.
As a result, 50% of the questions in QuALITY are labeled as hard, i.e. the majority of the annotators in the speed validation setting chose the wrong answer.
#### MuSiQue ([Trivedi et al., 2022](https://arxiv.org/pdf/2108.00573.pdf))
MuSiQue is a multi-hop question answering dataset, where the inputs are 20 Wikipedia paragraphs and a question that requires multiple hops between different paragraphs.
In the original dataset, each question also has an unanswerable twin question, where the correct answer is not present in the paragraphs.
#### SpaceDigest (New)
SpaceDigest is a new sentiment aggregation task. Given 50 hotel reviews (without their ratings) from the Space dataset (Angelidis et al., 2021), the task is to determine the percentage of positive reviews.
#### BookSumSort (New)
BookSumSort is a new task based on the BookSum dataset (Kry ́sci ́nski et al., 2022), which contains summaries of chapters (or parts) of novels, plays, and long poems from various sources.
Given a shuffled list of chapter summaries, the task is to reorder them according to the original order of summaries in BookSum.
## Data Fields
Most datasets in the benchmark are in the same input-output format
- `input`: a `string` feature. The input document.
- `output`: this feature is always None, as ZeroSCROLLS contains only test sets.
- `id`: a `string` feature. Unique per input.
- `pid`: a `string` feature, identical to 'id`. Facilitates evaluating tasks with multiple refrences per input.
- `document_start_index`: an `int32` feature. Character index that enables easy parsing of the context document.
- `document_end_index`: an `int32` feature. Character index that enables easy parsing of the context document.
- `query_start_index`: an `int32` feature. Character index that enables easy parsing of the query, if exists.
- `query_end_index`: an `int32` feature. Character index that enables easy parsing of the query, if exists.
- `truncation_seperator`: a `string` feature. The string used to append to a trimmed context document, mentioning the context was trimmed.
Datasets containing multiple documents inside the `input` feature are MuSiQue, SpaceDigest, and BookSumSort. They also have the following feature:
- `inner_docs_start_indices`: a sequence of `int32` feature. Character indexes that enables easy parsing of the the inner documents, e.g. Reviews, of Summaries.
## Citation
If you use the ZeroSCROLLS data, **please make sure to cite all of the original dataset papers.** [[bibtex](https://zero-scrolls-tau.s3.us-east-2.amazonaws.com/zero_scrolls_datasets.bib)]
```
@misc{shaham2023zeroscrolls,
title={ZeroSCROLLS: A Zero-Shot Benchmark for Long Text Understanding},
author={Uri Shaham and Maor Ivgi and Avia Efrat and Jonathan Berant and Omer Levy},
year={2023},
eprint={2305.14196},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
``` |
paws | 2023-06-01T14:59:56.000Z | [
"task_categories:text-classification",
"task_ids:semantic-similarity-classification",
"task_ids:semantic-similarity-scoring",
"task_ids:text-scoring",
"task_ids:multi-input-text-classification",
"annotations_creators:expert-generated",
"annotations_creators:machine-generated",
"language_creators:machine-generated",
"multilinguality:monolingual",
"size_categories:100K<n<1M",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:other",
"paraphrase-identification",
"arxiv:1904.01130",
"region:us"
] | null | PAWS: Paraphrase Adversaries from Word Scrambling
This dataset contains 108,463 human-labeled and 656k noisily labeled pairs that feature
the importance of modeling structure, context, and word order information for the problem
of paraphrase identification. The dataset has two subsets, one based on Wikipedia and the
other one based on the Quora Question Pairs (QQP) dataset.
For further details, see the accompanying paper: PAWS: Paraphrase Adversaries from Word Scrambling
(https://arxiv.org/abs/1904.01130)
PAWS-QQP is not available due to license of QQP. It must be reconstructed by downloading the original
data and then running our scripts to produce the data and attach the labels.
NOTE: There might be some missing or wrong labels in the dataset and we have replaced them with -1. | @InProceedings{paws2019naacl,
title = {{PAWS: Paraphrase Adversaries from Word Scrambling}},
author = {Zhang, Yuan and Baldridge, Jason and He, Luheng},
booktitle = {Proc. of NAACL},
year = {2019}
} | null | 16 | 6,295 | ---
annotations_creators:
- expert-generated
- machine-generated
language_creators:
- machine-generated
language:
- en
license:
- other
multilinguality:
- monolingual
size_categories:
- 100K<n<1M
- 10K<n<100K
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- semantic-similarity-classification
- semantic-similarity-scoring
- text-scoring
- multi-input-text-classification
paperswithcode_id: paws
pretty_name: 'PAWS: Paraphrase Adversaries from Word Scrambling'
tags:
- paraphrase-identification
dataset_info:
- config_name: labeled_final
features:
- name: id
dtype: int32
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': '0'
'1': '1'
splits:
- name: train
num_bytes: 12239978
num_examples: 49401
- name: test
num_bytes: 1987802
num_examples: 8000
- name: validation
num_bytes: 1975870
num_examples: 8000
download_size: 4687157
dataset_size: 16203650
- config_name: labeled_swap
features:
- name: id
dtype: int32
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': '0'
'1': '1'
splits:
- name: train
num_bytes: 7963651
num_examples: 30397
download_size: 2257283
dataset_size: 7963651
- config_name: unlabeled_final
features:
- name: id
dtype: int32
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': '0'
'1': '1'
splits:
- name: train
num_bytes: 157806996
num_examples: 645652
- name: validation
num_bytes: 2442173
num_examples: 10000
download_size: 47393331
dataset_size: 160249169
config_names:
- labeled_final
- labeled_swap
- unlabeled_final
---
# Dataset Card for PAWS: Paraphrase Adversaries from Word Scrambling
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [PAWS](https://github.com/google-research-datasets/paws)
- **Repository:** [PAWS](https://github.com/google-research-datasets/paws)
- **Paper:** [PAWS: Paraphrase Adversaries from Word Scrambling](https://arxiv.org/abs/1904.01130)
- **Point of Contact:** [Yuan Zhang](zhangyua@google.com)
### Dataset Summary
PAWS: Paraphrase Adversaries from Word Scrambling
This dataset contains 108,463 human-labeled and 656k noisily labeled pairs that feature the importance of modeling structure, context, and word order information for the problem of paraphrase identification. The dataset has two subsets, one based on Wikipedia and the other one based on the Quora Question Pairs (QQP) dataset.
For further details, see the accompanying paper: PAWS: Paraphrase Adversaries from Word Scrambling (https://arxiv.org/abs/1904.01130)
PAWS-QQP is not available due to license of QQP. It must be reconstructed by downloading the original data and then running our scripts to produce the data and attach the labels.
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
The text in the dataset is in English.
## Dataset Structure
### Data Instances
Below are two examples from the dataset:
| | Sentence 1 | Sentence 2 | Label |
| :-- | :---------------------------- | :---------------------------- | :---- |
| (1) | Although interchangeable, the body pieces on the 2 cars are not similar. | Although similar, the body parts are not interchangeable on the 2 cars. | 0 |
| (2) | Katz was born in Sweden in 1947 and moved to New York City at the age of 1. | Katz was born in 1947 in Sweden and moved to New York at the age of one. | 1 |
The first pair has different semantic meaning while the second pair is a paraphrase. State-of-the-art models trained on existing datasets have dismal performance on PAWS (<40% accuracy); however, including PAWS training data for these models improves their accuracy to 85% while maintaining performance on existing datasets such as the [Quora Question Pairs](https://data.quora.com/First-Quora-Dataset-Release-Question-Pairs).
### Data Fields
This corpus contains pairs generated from Wikipedia pages, and can be downloaded
here:
* **PAWS-Wiki Labeled (Final)**: containing pairs that are generated from both word swapping and back translation methods. All pairs have human judgements on both paraphrasing and fluency and they are split into Train/Dev/Test sections.
* **PAWS-Wiki Labeled (Swap-only)**: containing pairs that have no back translation counterparts and therefore they are not included in the first set. Nevertheless, they are high-quality pairs with human judgements on both paraphrasing and fluency, and they can be included as an auxiliary training set.
* **PAWS-Wiki Unlabeled (Final)**: Pairs in this set have noisy labels without human judgments and can also be used as an auxiliary training set. They are generated from both word swapping and back translation methods.
All files are in the tsv format with four columns:
Column Name | Data
:------------ | :--------------------------
id | A unique id for each pair
sentence1 | The first sentence
sentence2 | The second sentence
(noisy_)label | (Noisy) label for each pair
Each label has two possible values: `0` indicates the pair has different meaning, while `1` indicates the pair is a paraphrase.
### Data Splits
The number of examples and the proportion of paraphrase (Yes%) pairs are shown
below:
Data | Train | Dev | Test | Yes%
:------------------ | ------: | -----: | ----: | ----:
Labeled (Final) | 49,401 | 8,000 | 8,000 | 44.2%
Labeled (Swap-only) | 30,397 | -- | -- | 9.6%
Unlabeled (Final) | 645,652 | 10,000 | -- | 50.0%
## Dataset Creation
### Curation Rationale
Existing paraphrase identification datasets lack sentence pairs that have high lexical overlap without being paraphrases. Models trained on such data fail to distinguish pairs like *flights from New York to Florida* and *flights from Florida to New York*.
### Source Data
#### Initial Data Collection and Normalization
Their automatic generation method is based on two ideas. The first swaps words to generate a sentence pair with the same BOW, controlled by a language model. The second uses back translation to generate paraphrases with high BOW overlap but different word order. These two strategies generate high-quality, diverse PAWS pairs, balanced evenly between paraphrases and non-paraphrases.
#### Who are the source language producers?
Mentioned above.
### Annotations
#### Annotation process
Sentence pairs are presented to five annotators, each of which gives a binary judgment as to whether they are paraphrases or not. They chose binary judgments to make dataset have the same label schema as the QQP corpus. Overall, human agreement is high on both Quora (92.0%) and Wikipedia (94.7%) and each label only takes about 24 seconds. As such, answers are usually straight-forward to human raters.
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
List the people involved in collecting the dataset and their affiliation(s). If funding information is known, include it here.
### Licensing Information
The dataset may be freely used for any purpose, although acknowledgement of Google LLC ("Google") as the data source would be appreciated. The dataset is provided "AS IS" without any warranty, express or implied. Google disclaims all liability for any damages, direct or indirect, resulting from the use of the dataset.
### Citation Information
```
@InProceedings{paws2019naacl,
title = {{PAWS: Paraphrase Adversaries from Word Scrambling}},
author = {Zhang, Yuan and Baldridge, Jason and He, Luheng},
booktitle = {Proc. of NAACL},
year = {2019}
}
```
### Contributions
Thanks to [@bhavitvyamalik](https://github.com/bhavitvyamalik) for adding this dataset. |
nq_open | 2022-11-03T16:32:11.000Z | [
"task_categories:question-answering",
"task_ids:open-domain-qa",
"annotations_creators:expert-generated",
"language_creators:other",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:extended|natural_questions",
"language:en",
"license:cc-by-sa-3.0",
"region:us"
] | null | The NQ-Open task, introduced by Lee et.al. 2019,
is an open domain question answering benchmark that is derived from Natural Questions.
The goal is to predict an English answer string for an input English question.
All questions can be answered using the contents of English Wikipedia. | @article{doi:10.1162/tacl_a_00276,
author = {Kwiatkowski, Tom and Palomaki, Jennimaria and Redfield, Olivia and Collins, Michael and Parikh, Ankur and Alberti, Chris and Epstein, Danielle and Polosukhin, Illia and Devlin, Jacob and Lee, Kenton and Toutanova, Kristina and Jones, Llion and Kelcey, Matthew and Chang, Ming-Wei and Dai, Andrew M. and Uszkoreit, Jakob and Le, Quoc and Petrov, Slav},
title = {Natural Questions: A Benchmark for Question Answering Research},
journal = {Transactions of the Association for Computational Linguistics},
volume = {7},
number = {},
pages = {453-466},
year = {2019},
doi = {10.1162/tacl_a_00276},
URL = {
https://doi.org/10.1162/tacl_a_00276
},
eprint = {
https://doi.org/10.1162/tacl_a_00276
},
abstract = { We present the Natural Questions corpus, a question answering data set. Questions consist of real anonymized, aggregated queries issued to the Google search engine. An annotator is presented with a question along with a Wikipedia page from the top 5 search results, and annotates a long answer (typically a paragraph) and a short answer (one or more entities) if present on the page, or marks null if no long/short answer is present. The public release consists of 307,373 training examples with single annotations; 7,830 examples with 5-way annotations for development data; and a further 7,842 examples with 5-way annotated sequestered as test data. We present experiments validating quality of the data. We also describe analysis of 25-way annotations on 302 examples, giving insights into human variability on the annotation task. We introduce robust metrics for the purposes of evaluating question answering systems; demonstrate high human upper bounds on these metrics; and establish baseline results using competitive methods drawn from related literature. }
}
@inproceedings{lee-etal-2019-latent,
title = "Latent Retrieval for Weakly Supervised Open Domain Question Answering",
author = "Lee, Kenton and
Chang, Ming-Wei and
Toutanova, Kristina",
booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics",
month = jul,
year = "2019",
address = "Florence, Italy",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/P19-1612",
doi = "10.18653/v1/P19-1612",
pages = "6086--6096",
abstract = "Recent work on open domain question answering (QA) assumes strong supervision of the supporting evidence and/or assumes a blackbox information retrieval (IR) system to retrieve evidence candidates. We argue that both are suboptimal, since gold evidence is not always available, and QA is fundamentally different from IR. We show for the first time that it is possible to jointly learn the retriever and reader from question-answer string pairs and without any IR system. In this setting, evidence retrieval from all of Wikipedia is treated as a latent variable. Since this is impractical to learn from scratch, we pre-train the retriever with an Inverse Cloze Task. We evaluate on open versions of five QA datasets. On datasets where the questioner already knows the answer, a traditional IR system such as BM25 is sufficient. On datasets where a user is genuinely seeking an answer, we show that learned retrieval is crucial, outperforming BM25 by up to 19 points in exact match.",
} | null | 5 | 6,245 | ---
annotations_creators:
- expert-generated
language_creators:
- other
language:
- en
license:
- cc-by-sa-3.0
multilinguality:
- monolingual
pretty_name: NQ-Open
size_categories:
- 10K<n<100K
source_datasets:
- extended|natural_questions
task_categories:
- question-answering
task_ids:
- open-domain-qa
paperswithcode_id: null
dataset_info:
features:
- name: question
dtype: string
- name: answer
sequence: string
config_name: nq_open
splits:
- name: train
num_bytes: 6651344
num_examples: 87925
- name: validation
num_bytes: 313841
num_examples: 3610
download_size: 8913614
dataset_size: 6965185
---
# Dataset Card for nq_open
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** https://efficientqa.github.io/
- **Repository:** https://github.com/google-research-datasets/natural-questions/tree/master/nq_open
- **Paper:** https://www.aclweb.org/anthology/P19-1612.pdf
- **Leaderboard:** https://ai.google.com/research/NaturalQuestions/efficientqa
- **Point of Contact:** [Mailing List](efficientqa@googlegroups.com)
### Dataset Summary
The NQ-Open task, introduced by Lee et.al. 2019,
is an open domain question answering benchmark that is derived from Natural Questions.
The goal is to predict an English answer string for an input English question.
All questions can be answered using the contents of English Wikipedia.
### Supported Tasks and Leaderboards
Open Domain Question-Answering,
EfficientQA Leaderboard: https://ai.google.com/research/NaturalQuestions/efficientqa
### Languages
English (`en`)
## Dataset Structure
### Data Instances
```
{
"question": "names of the metropolitan municipalities in south africa",
"answer": [
"Mangaung Metropolitan Municipality",
"Nelson Mandela Bay Metropolitan Municipality",
"eThekwini Metropolitan Municipality",
"City of Tshwane Metropolitan Municipality",
"City of Johannesburg Metropolitan Municipality",
"Buffalo City Metropolitan Municipality",
"City of Ekurhuleni Metropolitan Municipality"
]
}
```
### Data Fields
- `question` - Input open domain question.
- `answer` - List of possible answers to the question
### Data Splits
- Train : 87925
- validation : 1800
## Dataset Creation
### Curation Rationale
[Needs More Information]
### Source Data
#### Initial Data Collection and Normalization
Natural Questions contains question from aggregated queries to Google Search (Kwiatkowski et al., 2019). To gather an open version of this dataset, we only keep questions with short answers and discard the given evidence document. Answers with many tokens often resemble extractive snippets rather than canonical answers, so we discard answers with more than 5 tokens.
#### Who are the source language producers?
[Needs More Information]
### Annotations
#### Annotation process
[Needs More Information]
#### Who are the annotators?
[Needs More Information]
### Personal and Sensitive Information
[Needs More Information]
## Considerations for Using the Data
### Social Impact of Dataset
[Needs More Information]
### Discussion of Biases
Evaluating on this diverse set of question-answer pairs is crucial, because all existing datasets have inherent biases that are problematic for open domain QA systems with learned retrieval.
In the Natural Questions dataset the question askers do not already know the answer. This accurately reflects a distribution of genuine information-seeking questions.
However, annotators must separately find correct answers, which requires assistance from automatic tools and can introduce a moderate bias towards results from the tool.
### Other Known Limitations
[Needs More Information]
## Additional Information
### Dataset Curators
[Needs More Information]
### Licensing Information
All of the Natural Questions data is released under the
[CC BY-SA 3.0](https://creativecommons.org/licenses/by-sa/3.0/) license.
### Citation Information
```
@article{doi:10.1162/tacl\_a\_00276,
author = {Kwiatkowski, Tom and Palomaki, Jennimaria and Redfield, Olivia and Collins, Michael and Parikh, Ankur and Alberti, Chris and Epstein, Danielle and Polosukhin, Illia and Devlin, Jacob and Lee, Kenton and Toutanova, Kristina and Jones, Llion and Kelcey, Matthew and Chang, Ming-Wei and Dai, Andrew M. and Uszkoreit, Jakob and Le, Quoc and Petrov, Slav},
title = {Natural Questions: A Benchmark for Question Answering Research},
journal = {Transactions of the Association for Computational Linguistics},
volume = {7},
number = {},
pages = {453-466},
year = {2019},
doi = {10.1162/tacl\_a\_00276},
URL = {
https://doi.org/10.1162/tacl_a_00276
},
eprint = {
https://doi.org/10.1162/tacl_a_00276
},
abstract = { We present the Natural Questions corpus, a question answering data set. Questions consist of real anonymized, aggregated queries issued to the Google search engine. An annotator is presented with a question along with a Wikipedia page from the top 5 search results, and annotates a long answer (typically a paragraph) and a short answer (one or more entities) if present on the page, or marks null if no long/short answer is present. The public release consists of 307,373 training examples with single annotations; 7,830 examples with 5-way annotations for development data; and a further 7,842 examples with 5-way annotated sequestered as test data. We present experiments validating quality of the data. We also describe analysis of 25-way annotations on 302 examples, giving insights into human variability on the annotation task. We introduce robust metrics for the purposes of evaluating question answering systems; demonstrate high human upper bounds on these metrics; and establish baseline results using competitive methods drawn from related literature. }
}
@inproceedings{lee-etal-2019-latent,
title = "Latent Retrieval for Weakly Supervised Open Domain Question Answering",
author = "Lee, Kenton and
Chang, Ming-Wei and
Toutanova, Kristina",
booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics",
month = jul,
year = "2019",
address = "Florence, Italy",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/P19-1612",
doi = "10.18653/v1/P19-1612",
pages = "6086--6096",
abstract = "Recent work on open domain question answering (QA) assumes strong supervision of the supporting evidence and/or assumes a blackbox information retrieval (IR) system to retrieve evidence candidates. We argue that both are suboptimal, since gold evidence is not always available, and QA is fundamentally different from IR. We show for the first time that it is possible to jointly learn the retriever and reader from question-answer string pairs and without any IR system. In this setting, evidence retrieval from all of Wikipedia is treated as a latent variable. Since this is impractical to learn from scratch, we pre-train the retriever with an Inverse Cloze Task. We evaluate on open versions of five QA datasets. On datasets where the questioner already knows the answer, a traditional IR system such as BM25 is sufficient. On datasets where a user is genuinely seeking an answer, we show that learned retrieval is crucial, outperforming BM25 by up to 19 points in exact match.",
}
```
### Contributions
Thanks to [@Nilanshrajput](https://github.com/Nilanshrajput) for adding this dataset. |
mit-han-lab/pile-val-backup | 2023-08-21T21:37:19.000Z | [
"region:us"
] | mit-han-lab | null | null | null | 2 | 6,091 | This is a backup for the pile val dataset downloaded from here: `https://the-eye.eu/public/AI/pile/val.jsonl.zst`
Please respect the original license of the dataset. |
web_nlg | 2023-06-01T14:59:54.000Z | [
"task_categories:tabular-to-text",
"task_ids:rdf-to-text",
"annotations_creators:found",
"language_creators:crowdsourced",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:extended|other-db_pedia",
"source_datasets:original",
"language:en",
"language:ru",
"license:cc-by-sa-3.0",
"license:cc-by-nc-sa-4.0",
"license:gfdl",
"region:us"
] | null | The WebNLG challenge consists in mapping data to text. The training data consists
of Data/Text pairs where the data is a set of triples extracted from DBpedia and the text is a verbalisation
of these triples. For instance, given the 3 DBpedia triples shown in (a), the aim is to generate a text such as (b).
a. (John_E_Blaha birthDate 1942_08_26) (John_E_Blaha birthPlace San_Antonio) (John_E_Blaha occupation Fighter_pilot)
b. John E Blaha, born in San Antonio on 1942-08-26, worked as a fighter pilot
As the example illustrates, the task involves specific NLG subtasks such as sentence segmentation
(how to chunk the input data into sentences), lexicalisation (of the DBpedia properties),
aggregation (how to avoid repetitions) and surface realisation
(how to build a syntactically correct and natural sounding text). | @inproceedings{web_nlg,
author = {Claire Gardent and
Anastasia Shimorina and
Shashi Narayan and
Laura Perez{-}Beltrachini},
editor = {Regina Barzilay and
Min{-}Yen Kan},
title = {Creating Training Corpora for {NLG} Micro-Planners},
booktitle = {Proceedings of the 55th Annual Meeting of the
Association for Computational Linguistics,
{ACL} 2017, Vancouver, Canada, July 30 - August 4,
Volume 1: Long Papers},
pages = {179--188},
publisher = {Association for Computational Linguistics},
year = {2017},
url = {https://doi.org/10.18653/v1/P17-1017},
doi = {10.18653/v1/P17-1017}
} | null | 10 | 6,039 | ---
annotations_creators:
- found
language_creators:
- crowdsourced
language:
- en
- ru
license:
- cc-by-sa-3.0
- cc-by-nc-sa-4.0
- gfdl
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- extended|other-db_pedia
- original
task_categories:
- tabular-to-text
task_ids:
- rdf-to-text
paperswithcode_id: webnlg
pretty_name: WebNLG
dataset_info:
- config_name: webnlg_challenge_2017
features:
- name: category
dtype: string
- name: size
dtype: int32
- name: eid
dtype: string
- name: original_triple_sets
sequence:
- name: otriple_set
sequence: string
- name: modified_triple_sets
sequence:
- name: mtriple_set
sequence: string
- name: shape
dtype: string
- name: shape_type
dtype: string
- name: lex
sequence:
- name: comment
dtype: string
- name: lid
dtype: string
- name: text
dtype: string
- name: lang
dtype: string
- name: test_category
dtype: string
- name: dbpedia_links
sequence: string
- name: links
sequence: string
splits:
- name: train
num_bytes: 5594812
num_examples: 6940
- name: dev
num_bytes: 706653
num_examples: 872
- name: test
num_bytes: 3122533
num_examples: 4615
download_size: 25499351
dataset_size: 9423998
- config_name: release_v1
features:
- name: category
dtype: string
- name: size
dtype: int32
- name: eid
dtype: string
- name: original_triple_sets
sequence:
- name: otriple_set
sequence: string
- name: modified_triple_sets
sequence:
- name: mtriple_set
sequence: string
- name: shape
dtype: string
- name: shape_type
dtype: string
- name: lex
sequence:
- name: comment
dtype: string
- name: lid
dtype: string
- name: text
dtype: string
- name: lang
dtype: string
- name: test_category
dtype: string
- name: dbpedia_links
sequence: string
- name: links
sequence: string
splits:
- name: full
num_bytes: 11684308
num_examples: 14237
download_size: 25499351
dataset_size: 11684308
- config_name: release_v2
features:
- name: category
dtype: string
- name: size
dtype: int32
- name: eid
dtype: string
- name: original_triple_sets
sequence:
- name: otriple_set
sequence: string
- name: modified_triple_sets
sequence:
- name: mtriple_set
sequence: string
- name: shape
dtype: string
- name: shape_type
dtype: string
- name: lex
sequence:
- name: comment
dtype: string
- name: lid
dtype: string
- name: text
dtype: string
- name: lang
dtype: string
- name: test_category
dtype: string
- name: dbpedia_links
sequence: string
- name: links
sequence: string
splits:
- name: train
num_bytes: 10830413
num_examples: 12876
- name: dev
num_bytes: 1360033
num_examples: 1619
- name: test
num_bytes: 1324934
num_examples: 1600
download_size: 25499351
dataset_size: 13515380
- config_name: release_v2_constrained
features:
- name: category
dtype: string
- name: size
dtype: int32
- name: eid
dtype: string
- name: original_triple_sets
sequence:
- name: otriple_set
sequence: string
- name: modified_triple_sets
sequence:
- name: mtriple_set
sequence: string
- name: shape
dtype: string
- name: shape_type
dtype: string
- name: lex
sequence:
- name: comment
dtype: string
- name: lid
dtype: string
- name: text
dtype: string
- name: lang
dtype: string
- name: test_category
dtype: string
- name: dbpedia_links
sequence: string
- name: links
sequence: string
splits:
- name: train
num_bytes: 10853434
num_examples: 12895
- name: dev
num_bytes: 1421590
num_examples: 1594
- name: test
num_bytes: 1243182
num_examples: 1606
download_size: 25499351
dataset_size: 13518206
- config_name: release_v2.1
features:
- name: category
dtype: string
- name: size
dtype: int32
- name: eid
dtype: string
- name: original_triple_sets
sequence:
- name: otriple_set
sequence: string
- name: modified_triple_sets
sequence:
- name: mtriple_set
sequence: string
- name: shape
dtype: string
- name: shape_type
dtype: string
- name: lex
sequence:
- name: comment
dtype: string
- name: lid
dtype: string
- name: text
dtype: string
- name: lang
dtype: string
- name: test_category
dtype: string
- name: dbpedia_links
sequence: string
- name: links
sequence: string
splits:
- name: train
num_bytes: 10848793
num_examples: 12876
- name: dev
num_bytes: 1362072
num_examples: 1619
- name: test
num_bytes: 1325860
num_examples: 1600
download_size: 25499351
dataset_size: 13536725
- config_name: release_v2.1_constrained
features:
- name: category
dtype: string
- name: size
dtype: int32
- name: eid
dtype: string
- name: original_triple_sets
sequence:
- name: otriple_set
sequence: string
- name: modified_triple_sets
sequence:
- name: mtriple_set
sequence: string
- name: shape
dtype: string
- name: shape_type
dtype: string
- name: lex
sequence:
- name: comment
dtype: string
- name: lid
dtype: string
- name: text
dtype: string
- name: lang
dtype: string
- name: test_category
dtype: string
- name: dbpedia_links
sequence: string
- name: links
sequence: string
splits:
- name: train
num_bytes: 11040016
num_examples: 12895
- name: dev
num_bytes: 1284044
num_examples: 1594
- name: test
num_bytes: 1212665
num_examples: 1606
download_size: 25499351
dataset_size: 13536725
- config_name: release_v3.0_en
features:
- name: category
dtype: string
- name: size
dtype: int32
- name: eid
dtype: string
- name: original_triple_sets
sequence:
- name: otriple_set
sequence: string
- name: modified_triple_sets
sequence:
- name: mtriple_set
sequence: string
- name: shape
dtype: string
- name: shape_type
dtype: string
- name: lex
sequence:
- name: comment
dtype: string
- name: lid
dtype: string
- name: text
dtype: string
- name: lang
dtype: string
- name: test_category
dtype: string
- name: dbpedia_links
sequence: string
- name: links
sequence: string
splits:
- name: train
num_bytes: 11084860
num_examples: 13211
- name: dev
num_bytes: 1394243
num_examples: 1667
- name: test
num_bytes: 4039282
num_examples: 5713
download_size: 25499351
dataset_size: 16518385
- config_name: release_v3.0_ru
features:
- name: category
dtype: string
- name: size
dtype: int32
- name: eid
dtype: string
- name: original_triple_sets
sequence:
- name: otriple_set
sequence: string
- name: modified_triple_sets
sequence:
- name: mtriple_set
sequence: string
- name: shape
dtype: string
- name: shape_type
dtype: string
- name: lex
sequence:
- name: comment
dtype: string
- name: lid
dtype: string
- name: text
dtype: string
- name: lang
dtype: string
- name: test_category
dtype: string
- name: dbpedia_links
sequence: string
- name: links
sequence: string
splits:
- name: train
num_bytes: 9550340
num_examples: 5573
- name: dev
num_bytes: 1314226
num_examples: 790
- name: test
num_bytes: 3656501
num_examples: 3410
download_size: 25499351
dataset_size: 14521067
config_names:
- release_v1
- release_v2
- release_v2.1
- release_v2.1_constrained
- release_v2_constrained
- release_v3.0_en
- release_v3.0_ru
- webnlg_challenge_2017
---
# Dataset Card for WebNLG
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [WebNLG challenge website](https://webnlg-challenge.loria.fr/)
- **Repository:** [WebNLG GitLab repository](https://gitlab.com/shimorina/webnlg-dataset/-/tree/master/)
- **Paper:** [Creating Training Corpora for NLG Micro-Planning](https://www.aclweb.org/anthology/P17-1017.pdf)
- **Leaderboard:** [WebNLG leaderboards](https://gerbil-nlg.dice-research.org/gerbil/webnlg2020results)
- **Point of Contact:** [anastasia.shimorina@loria.fr](anastasia.shimorina@loria.fr)
### Dataset Summary
The WebNLG challenge consists in mapping data to text. The training data consists
of Data/Text pairs where the data is a set of triples extracted from DBpedia and the text is a verbalisation
of these triples. For instance, given the 3 DBpedia triples shown in (a), the aim is to generate a text such as (b).
```
a. (John_E_Blaha birthDate 1942_08_26) (John_E_Blaha birthPlace San_Antonio) (John_E_Blaha occupation Fighter_pilot)
b. John E Blaha, born in San Antonio on 1942-08-26, worked as a fighter pilot
```
As the example illustrates, the task involves specific NLG subtasks such as sentence segmentation
(how to chunk the input data into sentences), lexicalisation (of the DBpedia properties),
aggregation (how to avoid repetitions) and surface realisation
(how to build a syntactically correct and natural sounding text).
### Supported Tasks and Leaderboards
The dataset supports a Structured to Text task which requires a model takes a set of RDF (Resource Description Format) triples from a database (DBpedia) of the form (subject, property, object) as input and write out a natural language sentence expressing the information contained in the triples. The dataset has supportd two challenges: the [WebNLG2017](https://www.aclweb.org/anthology/W17-3518/) and [WebNLG2020](https://gerbil-nlg.dice-research.org/gerbil/webnlg2020results) challenge. Results were ordered by their [METEOR](https://huggingface.co/metrics/meteor) to the reference, but the leaderboards report a range of other metrics including [BLEU](https://huggingface.co/metrics/bleu), [BERTscore](https://huggingface.co/metrics/bertscore), and [BLEURT](https://huggingface.co/metrics/bleurt). The v3 release (`release_v3.0_en`, `release_v3.0_ru`) for the WebNLG2020 challenge also supports a semantic `parsing` task.
### Languages
All releases contain English (`en`) data. The v3 release (`release_v3.0_ru`) also contains Russian (`ru`) examples.
## Dataset Structure
### Data Instances
A typical example contains the original RDF triples in the set, a modified version which presented to crowd workers, and a set of possible verbalizations for this set of triples:
```
{'2017_test_category': '',
'category': 'Politician',
'eid': 'Id10',
'lex': {'comment': ['good', 'good', 'good'],
'lid': ['Id1', 'Id2', 'Id3'],
'text': ['World War II had Chiang Kai-shek as a commander and United States Army soldier Abner W. Sibal.',
'Abner W. Sibal served in the United States Army during the Second World War and during that war Chiang Kai-shek was one of the commanders.',
'Abner W. Sibal, served in the United States Army and fought in World War II, one of the commanders of which, was Chiang Kai-shek.']},
'modified_triple_sets': {'mtriple_set': [['Abner_W._Sibal | battle | World_War_II',
'World_War_II | commander | Chiang_Kai-shek',
'Abner_W._Sibal | militaryBranch | United_States_Army']]},
'original_triple_sets': {'otriple_set': [['Abner_W._Sibal | battles | World_War_II', 'World_War_II | commander | Chiang_Kai-shek', 'Abner_W._Sibal | branch | United_States_Army'],
['Abner_W._Sibal | militaryBranch | United_States_Army',
'Abner_W._Sibal | battles | World_War_II',
'World_War_II | commander | Chiang_Kai-shek']]},
'shape': '(X (X) (X (X)))',
'shape_type': 'mixed',
'size': 3}
```
### Data Fields
The following fields can be found in the instances:
- `category`: the category of the DBpedia entities present in the RDF triples.
- `eid`: an example ID, only unique per split per category.
- `size`: number of RDF triples in the set.
- `shape`: (since v2) Each set of RDF-triples is a tree, which is characterised by its shape and shape type. `shape` is a string representation of the tree with nested parentheses where X is a node (see [Newick tree format](https://en.wikipedia.org/wiki/Newick_format))
- `shape_type`: (since v2) is a type of the tree shape, which can be: `chain` (the object of one triple is the subject of the other); `sibling` (triples with a shared subject); `mixed` (both chain and sibling types present).
- `test_category`: (for `webnlg_challenge_2017` and `v3`) tells whether the set of RDF triples was present in the training set or not. Several splits of the test set are available: with and without references, and for RDF-to-text generation / for semantic parsing.
- `lex`: the lexicalizations, with:
- `text`: the text to be predicted.
- `lid`: a lexicalization ID, unique per example.
- `comment`: the lexicalizations were rated by crowd workers are either `good` or `bad`
- `lang`: (for `release_v3.0_ru`) the language used because original English texts were kept in the Russian version.
Russian data has additional optional fields comparing to English:
- `dbpedialinks`: RDF triples extracted from DBpedia between English and Russian entities by means of the property `sameAs`.
- `links`: RDF triples created manually for some entities to serve as pointers to translators. There are two types of them:
* with `sameAs` (`Spaniards | sameAs | испанцы`)
* with `includes` (`Tomatoes, guanciale, cheese, olive oil | includes | гуанчиале`). Those were mostly created for string literals to translate some parts of them.
### Data Splits
For `v3.0` releases:
| English (v3.0) | Train | Dev | Test (data-to-text) |
|-----------------|--------|-------|-------|
| **triple sets** | 13,211 | 1,667 | 1,779 |
| **texts** | 35,426 | 4,464 | 5,150 |
|**properties** | 372 | 290 | 220 |
| Russian (v3.0) | Train | Dev | Test (data-to-text) |
|-----------------|--------|-------|---------------------|
| **triple sets** | 5,573 | 790 | 1,102 |
| **texts** | 14,239 | 2,026 | 2,780 |
|**properties** | 226 | 115 | 192 |
## Dataset Creation
### Curation Rationale
The WebNLG dataset was created to promote the development _(i)_ of RDF verbalisers and _(ii)_ of microplanners able to handle a wide range of linguistic constructions. The dataset aims at covering knowledge in different domains ("categories"). The same properties and entities can appear in several categories.
### Source Data
The data was compiled from raw DBpedia triples. [This paper](https://www.aclweb.org/anthology/C16-1141/) explains how the triples were selected.
#### Initial Data Collection and Normalization
Initial triples extracted from DBpedia were modified in several ways. See [official documentation](https://webnlg-challenge.loria.fr/docs/) for the most frequent changes that have been made. An original tripleset and a modified tripleset usually represent a one-to-one mapping. However, there are cases with many-to-one mappings when several original triplesets are mapped to one modified tripleset.
Entities that served as roots of RDF trees are listed in [this file](https://gitlab.com/shimorina/webnlg-dataset/-/blob/master/supplementary/entities_dict.json).
The English WebNLG 2020 dataset (v3.0) for training comprises data-text pairs for 16 distinct DBpedia categories:
- The 10 seen categories used in the 2017 version: Airport, Astronaut, Building, City, ComicsCharacter, Food, Monument, SportsTeam, University, and WrittenWork.
- The 5 unseen categories of 2017, which are now part of the seen data: Athlete, Artist, CelestialBody, MeanOfTransportation, Politician.
- 1 new category: Company.
The Russian dataset (v3.0) comprises data-text pairs for 9 distinct categories: Airport, Astronaut, Building, CelestialBody, ComicsCharacter, Food, Monument, SportsTeam, and University.
#### Who are the source language producers?
There are no source texts, all textual material was compiled during the annotation process.
### Annotations
#### Annotation process
Annotators were first asked to create sentences that verbalise single triples. In a second round, annotators were asked to combine single-triple sentences together into sentences that cover 2 triples. And so on until 7 triples. Quality checks were performed to ensure the quality of the annotations. See Section 3.3 in [the dataset paper](https://www.aclweb.org/anthology/P17-1017.pdf).
Russian data was translated from English with an MT system and then was post-edited by crowdworkers. See Section 2.2 of [this paper](https://webnlg-challenge.loria.fr/files/2020.webnlg-papers.7.pdf).
#### Who are the annotators?
All references were collected through crowdsourcing platforms (CrowdFlower/Figure 8 and Amazon Mechanical Turk). For Russian, post-editing was done using the Yandex.Toloka crowdsourcing platform.
### Personal and Sensitive Information
Neither the dataset as published or the annotation process involves the collection or sharing of any kind of personal / demographic information.
## Considerations for Using the Data
### Social Impact of Dataset
We do not foresee any negative social impact in particular from this dataset or task.
Positive outlooks: Being able to generate good quality text from RDF data would permit, e.g., making this data more accessible to lay users, enriching existing text with information drawn from knowledge bases such as DBpedia or describing, comparing and relating entities present in these knowledge bases.
### Discussion of Biases
This dataset is created using DBpedia RDF triples which naturally exhibit biases that have been found to exist in Wikipedia such as some forms of, e.g., gender bias.
The choice of [entities](https://gitlab.com/shimorina/webnlg-dataset/-/blob/master/supplementary/entities_dict.json), described by RDF trees, was not controlled. As such, they may contain gender biases; for instance, all the astronauts described by RDF triples are male. Hence, in texts, pronouns _he/him/his_ occur more often. Similarly, entities can be related to the Western culture more often than to other cultures.
### Other Known Limitations
The quality of the crowdsourced references is limited, in particular in terms of fluency/naturalness of the collected texts.
Russian data was machine-translated and then post-edited by crowdworkers, so some examples may still exhibit issues related to bad translations.
## Additional Information
### Dataset Curators
The principle curator of the dataset is Anastasia Shimorina (Université de Lorraine / LORIA, France). Throughout the WebNLG releases, several people contributed to their construction: Claire Gardent (CNRS / LORIA, France), Shashi Narayan (Google, UK), Laura Perez-Beltrachini (University of Edinburgh, UK), Elena Khasanova, and Thiago Castro Ferreira (Federal University of Minas Gerais, Brazil).
The dataset construction was funded by the French National Research Agency (ANR).
### Licensing Information
The dataset uses the `cc-by-nc-sa-4.0` license. The source DBpedia project uses the `cc-by-sa-3.0` and `gfdl-1.1` licenses.
### Citation Information
- If you use the WebNLG corpus, cite:
```
@inproceedings{web_nlg,
author = {Claire Gardent and
Anastasia Shimorina and
Shashi Narayan and
Laura Perez{-}Beltrachini},
editor = {Regina Barzilay and
Min{-}Yen Kan},
title = {Creating Training Corpora for {NLG} Micro-Planners},
booktitle = {Proceedings of the 55th Annual Meeting of the Association for Computational
Linguistics, {ACL} 2017, Vancouver, Canada, July 30 - August 4, Volume
1: Long Papers},
pages = {179--188},
publisher = {Association for Computational Linguistics},
year = {2017},
url = {https://doi.org/10.18653/v1/P17-1017},
doi = {10.18653/v1/P17-1017}
}
```
- If you use `release_v2_constrained` in particular, cite:
```
@InProceedings{shimorina2018handling,
author = "Shimorina, Anastasia
and Gardent, Claire",
title = "Handling Rare Items in Data-to-Text Generation",
booktitle = "Proceedings of the 11th International Conference on Natural Language Generation",
year = "2018",
publisher = "Association for Computational Linguistics",
pages = "360--370",
location = "Tilburg University, The Netherlands",
url = "http://aclweb.org/anthology/W18-6543"
}
```
### Contributions
Thanks to [@Shimorina](https://github.com/Shimorina), [@yjernite](https://github.com/yjernite) for adding this dataset. |
kunishou/databricks-dolly-15k-ja | 2023-09-10T13:47:12.000Z | [
"license:cc-by-sa-3.0",
"region:us"
] | kunishou | null | null | null | 52 | 6,002 | ---
license: cc-by-sa-3.0
---
This dataset was created by automatically translating "databricks-dolly-15k" into Japanese.
This dataset is licensed under CC-BY-SA-3.0
Last Update : 2023-05-11
databricks-dolly-15k-ja
https://github.com/kunishou/databricks-dolly-15k-ja
databricks-dolly-15k
https://github.com/databrickslabs/dolly/tree/master/data |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.