author
stringlengths
2
29
cardData
null
citation
stringlengths
0
9.58k
description
stringlengths
0
5.93k
disabled
bool
1 class
downloads
float64
1
1M
gated
bool
2 classes
id
stringlengths
2
108
lastModified
stringlengths
24
24
paperswithcode_id
stringlengths
2
45
private
bool
2 classes
sha
stringlengths
40
40
siblings
list
tags
list
readme_url
stringlengths
57
163
readme
stringlengths
0
977k
khalidalt
null
@book{book, author = {Misra, Rishabh and Grover, Jigyasa}, year = {2021}, month = {01}, pages = {}, title = {Sculpting Data for ML: The first act of Machine Learning}, isbn = {978-0-578-83125-1} } @dataset{dataset, author = {Misra, Rishabh}, year = {2018}, month = {06}, pages = {}, title = {News Category Dataset}, doi = {10.13140/RG.2.2.20331.18729} }
A dataset of approximately 200K news headlines from the year 2012 to 2018 collected from HuffPost.
false
112
false
khalidalt/HuffPost
2022-04-29T05:13:38.000Z
null
false
e45f5ed7b3c1309b54a4b62e9003b958a1e27b20
[]
[ "license:cc0-1.0" ]
https://huggingface.co/datasets/khalidalt/HuffPost/resolve/main/README.md
--- license: cc0-1.0 --- # Dataset Card for HuffPost ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:https://www.kaggle.com/datasets/rmisra/news-category-dataset/metadata** ### Dataset Summary A dataset of approximately 200K news headlines from the year 2012 to 2018 collected from HuffPost. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information license: cc0-1.0 ### Citation Information ``` @book{book, author = {Misra, Rishabh and Grover, Jigyasa}, year = {2021}, month = {01}, pages = {}, title = {Sculpting Data for ML: The first act of Machine Learning}, isbn = {978-0-578-83125-1} } @dataset{dataset, author = {Misra, Rishabh}, year = {2018}, month = {06}, pages = {}, title = {News Category Dataset}, doi = {10.13140/RG.2.2.20331.18729} } ``` ### Contributions Thanks to [@github-username](https://github.com/<github-username>) for adding this dataset.
blinoff
null
@article{blinov2013research, title={Research of lexical approach and machine learning methods for sentiment analysis}, author={Blinov, PD and Klekovkina, Maria and Kotelnikov, Eugeny and Pestov, Oleg}, journal={Computational Linguistics and Intellectual Technologies}, volume={2}, number={12}, pages={48--58}, year={2013} }
null
false
3
false
blinoff/kinopoisk
2022-10-23T16:51:58.000Z
null
false
726a7cb5d4eab90c9035bd55b7bde3018c3bd06b
[]
[ "language:ru", "multilinguality:monolingual", "size_categories:10K<n<100K", "task_categories:text-classification", "task_ids:sentiment-classification" ]
https://huggingface.co/datasets/blinoff/kinopoisk/resolve/main/README.md
--- language: - ru multilinguality: - monolingual pretty_name: Kinopoisk size_categories: - 10K<n<100K task_categories: - text-classification task_ids: - sentiment-classification --- ### Dataset Summary Kinopoisk movie reviews dataset (TOP250 & BOTTOM100 rank lists). In total it contains 36,591 reviews from July 2004 to November 2012. With following distribution along the 3-point sentiment scale: - Good: 27,264; - Bad: 4,751; - Neutral: 4,576. ### Data Fields Each sample contains the following fields: - **part**: rank list top250 or bottom100; - **movie_name**; - **review_id**; - **author**: review author; - **date**: date of a review; - **title**: review title; - **grade3**: sentiment score Good, Bad or Neutral; - **grade10**: sentiment score on a 10-point scale parsed from text; - **content**: review text. ### Python ```python3 import pandas as pd df = pd.read_json('kinopoisk.jsonl', lines=True) df.sample(5) ``` ### Citation ``` @article{blinov2013research, title={Research of lexical approach and machine learning methods for sentiment analysis}, author={Blinov, PD and Klekovkina, Maria and Kotelnikov, Eugeny and Pestov, Oleg}, journal={Computational Linguistics and Intellectual Technologies}, volume={2}, number={12}, pages={48--58}, year={2013} } ```
hady
null
null
null
false
2
false
hady/kurdiabadulhady
2022-04-26T10:31:42.000Z
null
false
485c67807d91e92466571c44279eacd217042b76
[]
[]
https://huggingface.co/datasets/hady/kurdiabadulhady/resolve/main/README.md
sxu
null
null
null
false
2
false
sxu/CANLI
2022-10-11T12:53:46.000Z
null
false
1ef455d5237f4d2c61719194e015a2d41e77073e
[]
[ "license:afl-3.0", "annotations_creators:expert-generated", "language:cn", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:1K<n<10K" ]
https://huggingface.co/datasets/sxu/CANLI/resolve/main/README.md
--- license: afl-3.0 annotations_creators: - expert-generated language: - cn language_creators: - expert-generated multilinguality: - monolingual size_categories: - 1K<n<10K --- # Dataset Card for CANLI ### Dataset Summary [CANLI: The Chinese Causative-Passive Homonymy Disambiguation: an Adversarial Dataset for NLI and a Probing Task](http://www.lrec-conf.org/proceedings/lrec2022/pdf/2022.lrec-1.460.pdf) The disambiguation of causative-passive homonymy (CPH) is potentially tricky for machines, as the causative and the passive are not distinguished by the sentences syntactic structure. By transforming CPH disambiguation to a challenging natural language inference (NLI) task, we present the first Chinese Adversarial NLI challenge set (CANLI). We show that the pretrained transformer model RoBERTa, fine-tuned on an existing large-scale Chinese NLI benchmark dataset, performs poorly on CANLI. We also employ Word Sense Disambiguation as a probing task to investigate to what extent the CPH feature is captured in the models internal representation. We find that the models performance on CANLI does not correspond to its internal representation of CPH, which is the crucial linguistic ability central to the CANLI dataset. ### Languages Chinese Mandarin # Citation Information @inproceedings{xu-markert-2022-chinese, title = "The {C}hinese Causative-Passive Homonymy Disambiguation: an adversarial Dataset for {NLI} and a Probing Task", author = "Xu, Shanshan and Markert, Katja", booktitle = "Proceedings of the Thirteenth Language Resources and Evaluation Conference", month = jun, year = "2022", address = "Marseille, France", publisher = "European Language Resources Association", url = "https://aclanthology.org/2022.lrec-1.460", pages = "4316--4323", }
ENM
null
null
null
false
1
false
ENM/dataset-prueba
2022-10-25T10:12:20.000Z
null
false
7455d89e3da5e569b49d6ae1005fd52e89eb5087
[]
[ "language:en" ]
https://huggingface.co/datasets/ENM/dataset-prueba/resolve/main/README.md
--- language: - en pretty_name: ScientificPapers --- # Dataset Card for "scientific_papers" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://github.com/armancohan/long-summarization](https://github.com/armancohan/long-summarization) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 8591.93 MB - **Size of the generated dataset:** 9622.19 MB - **Total amount of disk used:** 18214.12 MB ### Dataset Summary Scientific papers datasets contains two sets of long and structured documents. The datasets are obtained from ArXiv and PubMed OpenAccess repositories. Both "arxiv" and "pubmed" have two features: - article: the body of the document, pagragraphs seperated by "/n". - abstract: the abstract of the document, pagragraphs seperated by "/n". - section_names: titles of sections, seperated by "/n". ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure We show detailed information for up to 5 configurations of the dataset. ### Data Instances #### arxiv - **Size of downloaded dataset files:** 4295.97 MB - **Size of the generated dataset:** 7231.70 MB - **Total amount of disk used:** 11527.66 MB An example of 'train' looks as follows. ``` This example was too long and was cropped: { "abstract": "\" we have studied the leptonic decay @xmath0 , via the decay channel @xmath1 , using a sample of tagged @xmath2 decays collected...", "article": "\"the leptonic decays of a charged pseudoscalar meson @xmath7 are processes of the type @xmath8 , where @xmath9 , @xmath10 , or @...", "section_names": "[sec:introduction]introduction\n[sec:detector]data and the cleo- detector\n[sec:analysys]analysis method\n[sec:conclusion]summary" } ``` #### pubmed - **Size of downloaded dataset files:** 4295.97 MB - **Size of the generated dataset:** 2390.49 MB - **Total amount of disk used:** 6686.46 MB An example of 'validation' looks as follows. ``` This example was too long and was cropped: { "abstract": "\" background and aim : there is lack of substantial indian data on venous thromboembolism ( vte ) . \\n the aim of this study was...", "article": "\"approximately , one - third of patients with symptomatic vte manifests pe , whereas two - thirds manifest dvt alone .\\nboth dvt...", "section_names": "\"Introduction\\nSubjects and Methods\\nResults\\nDemographics and characteristics of venous thromboembolism patients\\nRisk factors ..." } ``` ### Data Fields The data fields are the same among all splits. #### arxiv - `article`: a `string` feature. - `abstract`: a `string` feature. - `section_names`: a `string` feature. #### pubmed - `article`: a `string` feature. - `abstract`: a `string` feature. - `section_names`: a `string` feature. ### Data Splits | name |train |validation|test| |------|-----:|---------:|---:| |arxiv |203037| 6436|6440| |pubmed|119924| 6633|6658| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @article{Cohan_2018, title={A Discourse-Aware Attention Model for Abstractive Summarization of Long Documents}, url={http://dx.doi.org/10.18653/v1/n18-2097}, DOI={10.18653/v1/n18-2097}, journal={Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers)}, publisher={Association for Computational Linguistics}, author={Cohan, Arman and Dernoncourt, Franck and Kim, Doo Soon and Bui, Trung and Kim, Seokhwan and Chang, Walter and Goharian, Nazli}, year={2018} } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@jplu](https://github.com/jplu), [@lewtun](https://github.com/lewtun), [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset.
SocialGrep
null
null
A comprehensive dataset of Reddit's NFT discussion.
false
2
false
SocialGrep/the-reddit-nft-dataset
2022-07-01T17:52:49.000Z
null
false
5fc63ea7788cd5b4edb6aeba801cdc7083cf07e9
[]
[ "annotations_creators:lexyr", "language_creators:crowdsourced", "language:en", "license:cc-by-4.0", "multilinguality:monolingual", "size_categories:1M<n<10M", "source_datasets:original" ]
https://huggingface.co/datasets/SocialGrep/the-reddit-nft-dataset/resolve/main/README.md
--- annotations_creators: - lexyr language_creators: - crowdsourced language: - en license: - cc-by-4.0 multilinguality: - monolingual size_categories: - 1M<n<10M source_datasets: - original paperswithcode_id: null --- # Dataset Card for the-reddit-nft-dataset ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Licensing Information](#licensing-information) ## Dataset Description - **Homepage:** [https://socialgrep.com/datasets](https://socialgrep.com/datasets/the-reddit-nft-dataset?utm_source=huggingface&utm_medium=link&utm_campaign=theredditnftdataset) - **Point of Contact:** [Website](https://socialgrep.com/contact?utm_source=huggingface&utm_medium=link&utm_campaign=theredditnftdataset) ### Dataset Summary A comprehensive dataset of Reddit's NFT discussion. ### Languages Mainly English. ## Dataset Structure ### Data Instances A data point is a post or a comment. Due to the separate nature of the two, those exist in two different files - even though many fields are shared. ### Data Fields - 'type': the type of the data point. Can be 'post' or 'comment'. - 'id': the base-36 Reddit ID of the data point. Unique when combined with type. - 'subreddit.id': the base-36 Reddit ID of the data point's host subreddit. Unique. - 'subreddit.name': the human-readable name of the data point's host subreddit. - 'subreddit.nsfw': a boolean marking the data point's host subreddit as NSFW or not. - 'created_utc': a UTC timestamp for the data point. - 'permalink': a reference link to the data point on Reddit. - 'score': score of the data point on Reddit. - 'domain': (Post only) the domain of the data point's link. - 'url': (Post only) the destination of the data point's link, if any. - 'selftext': (Post only) the self-text of the data point, if any. - 'title': (Post only) the title of the post data point. - 'body': (Comment only) the body of the comment data point. - 'sentiment': (Comment only) the result of an in-house sentiment analysis pipeline. Used for exploratory analysis. ## Additional Information ### Licensing Information CC-BY v4.0
aakanksha
null
@inproceedings{nivre-etal-2020-universal, title = "{U}niversal {D}ependencies v2: An Evergrowing Multilingual Treebank Collection", author = "Nivre, Joakim and de Marneffe, Marie-Catherine and Ginter, Filip and Haji{\v{c}}, Jan and Manning, Christopher D. and Pyysalo, Sampo and Schuster, Sebastian and Tyers, Francis and Zeman, Daniel", booktitle = "Proceedings of the 12th Language Resources and Evaluation Conference", month = may, year = "2020", address = "Marseille, France", publisher = "European Language Resources Association", url = "https://aclanthology.org/2020.lrec-1.497", pages = "4034--4043", abstract = "Universal Dependencies is an open community effort to create cross-linguistically consistent treebank annotation for many languages within a dependency-based lexicalist framework. The annotation consists in a linguistically motivated word segmentation; a morphological layer comprising lemmas, universal part-of-speech tags, and standardized morphological features; and a syntactic layer focusing on syntactic relations between predicates, arguments and modifiers. In this paper, we describe version 2 of the universal guidelines (UD v2), discuss the major changes from UD v1 to UD v2, and give an overview of the currently available treebanks for 90 languages.", language = "English", ISBN = "979-10-95546-34-4", }
Universal Dependencies is an open community effort to create cross-linguistically consistent treebank annotation for many languages within a dependency-based lexicalist framework. The annotation consists in a linguistically motivated word segmentation; a morphological layer comprising lemmas, universal part-of-speech tags, and standardized morphological features; and a syntactic layer focusing on syntactic relations between predicates, arguments and modifiers.
false
1
false
aakanksha/udpos
2022-04-27T19:21:57.000Z
null
false
739825f9dbb674e44f71019730d403f626aac4be
[]
[]
https://huggingface.co/datasets/aakanksha/udpos/resolve/main/README.md
POS tagging on the Universal Dependencies dataset
fut501
null
null
null
false
1
false
fut501/ds1
2022-05-10T01:23:18.000Z
null
false
a4ce90c2d3cd20978a678e6a108119716f235310
[]
[ "license:apache-2.0" ]
https://huggingface.co/datasets/fut501/ds1/resolve/main/README.md
--- license: apache-2.0 ---
zaraTahhhir
null
null
null
false
1
false
zaraTahhhir/urduprusdataset
2022-04-27T07:18:05.000Z
null
false
eb634bef6c528fb9df2acf63b56fdf82f0d41684
[]
[ "license:mit" ]
https://huggingface.co/datasets/zaraTahhhir/urduprusdataset/resolve/main/README.md
--- license: mit ---
Zaratahir123
null
null
null
false
1
false
Zaratahir123/urduprusdataset
2022-04-27T07:52:02.000Z
null
false
c340865212e37f0f37823ffb6cc4ed1c8a960c0e
[]
[ "license:mit" ]
https://huggingface.co/datasets/Zaratahir123/urduprusdataset/resolve/main/README.md
--- license: mit ---
junliang
null
null
null
false
1
false
junliang/symptom
2022-05-11T12:57:22.000Z
null
false
578d877dd50601749b406d53805a4bd332b63091
[]
[]
https://huggingface.co/datasets/junliang/symptom/resolve/main/README.md
annotations_creators: - found language_creators: - found languages: - zh licenses: - other-my-license multilinguality: - monolingual pretty_name: symptom size_categories: - 1K<n<10K source_datasets: - original task_categories: - token-classification task_ids: - word-sense-disambiguation
janck
null
null
null
false
1
false
janck/bigscience-lama
2022-10-21T08:16:23.000Z
lama
false
a3fc132b1a1b550f82e0801e9ded2ae475b659ea
[]
[ "annotations_creators:machine-generated", "language_creators:machine-generated", "language:en", "license:cc-by-4.0", "multilinguality:monolingual", "size_categories:1M<n<10M", "task_categories:text-retrieval", "task_categories:text-classification", "task_ids:fact-checking-retrieval", "task_ids:tex...
https://huggingface.co/datasets/janck/bigscience-lama/resolve/main/README.md
--- annotations_creators: - machine-generated language_creators: - machine-generated language: - en license: - cc-by-4.0 multilinguality: - monolingual size_categories: trex: - 1M<n<10M task_categories: - text-retrieval - text-classification task_ids: - fact-checking-retrieval - text-scoring paperswithcode_id: lama pretty_name: 'LAMA: LAnguage Model Analysis - BigScience version' tags: - probing --- # Dataset Card for LAMA: LAnguage Model Analysis - a dataset for probing and analyzing the factual and commonsense knowledge contained in pretrained language models. ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://github.com/facebookresearch/LAMA - **Repository:** https://github.com/facebookresearch/LAMA - **Paper:** @inproceedings{petroni2019language, title={Language Models as Knowledge Bases?}, author={F. Petroni, T. Rockt{\"{a}}schel, A. H. Miller, P. Lewis, A. Bakhtin, Y. Wu and S. Riedel}, booktitle={In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2019}, year={2019} } @inproceedings{petroni2020how, title={How Context Affects Language Models' Factual Predictions}, author={Fabio Petroni and Patrick Lewis and Aleksandra Piktus and Tim Rockt{\"a}schel and Yuxiang Wu and Alexander H. Miller and Sebastian Riedel}, booktitle={Automated Knowledge Base Construction}, year={2020}, url={https://openreview.net/forum?id=025X0zPfn} } ### Dataset Summary This dataset provides the data for LAMA. This dataset only contains TRex (subset of wikidata triples). The dataset includes some cleanup, and addition of a masked sentence and associated answers for the [MASK] token. The accuracy in predicting the [MASK] token shows how well the language model knows facts and common sense information. The [MASK] tokens are only for the "object" slots. This version also contains questions instead of templates that can be used to probe also non-masking models. See the paper for more details. For more information, also see: https://github.com/facebookresearch/LAMA ### Languages en ## Dataset Structure ### Data Instances The trex config has the following fields: `` {'uuid': 'a37257ae-4cbb-4309-a78a-623036c96797', 'sub_label': 'Pianos Become the Teeth', 'predicate_id': 'P740', 'obj_label': 'Baltimore', 'template': '[X] was founded in [Y] .', 'type': 'N-1', 'question': 'Where was [X] founded?'} 34039 `` ### Data Splits There are no data splits. ## Dataset Creation ### Curation Rationale This dataset was gathered and created to probe what language models understand. ### Source Data #### Initial Data Collection and Normalization See the reaserch paper and website for more detail. The dataset was created gathered from various other datasets with cleanups for probing. #### Who are the source language producers? The LAMA authors and the original authors of the various configs. ### Annotations #### Annotation process Human annotations under the original datasets (conceptnet), and various machine annotations. #### Who are the annotators? Human annotations and machine annotations. ### Personal and Sensitive Information Unkown, but likely names of famous people. ## Considerations for Using the Data ### Social Impact of Dataset The goal for the work is to probe the understanding of language models. ### Discussion of Biases Since the data is from human annotators, there is likely to be baises. [More Information Needed] ### Other Known Limitations The original documentation for the datafields are limited. ## Additional Information ### Dataset Curators The authors of LAMA at Facebook and the authors of the original datasets. ### Licensing Information The Creative Commons Attribution-Noncommercial 4.0 International License. see https://github.com/facebookresearch/LAMA/blob/master/LICENSE ### Citation Information @inproceedings{petroni2019language, title={Language Models as Knowledge Bases?}, author={F. Petroni, T. Rockt{\"{a}}schel, A. H. Miller, P. Lewis, A. Bakhtin, Y. Wu and S. Riedel}, booktitle={In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2019}, year={2019} } @inproceedings{petroni2020how, title={How Context Affects Language Models' Factual Predictions}, author={Fabio Petroni and Patrick Lewis and Aleksandra Piktus and Tim Rockt{\"a}schel and Yuxiang Wu and Alexander H. Miller and Sebastian Riedel}, booktitle={Automated Knowledge Base Construction}, year={2020}, url={https://openreview.net/forum?id=025X0zPfn} }
Zaratahir123
null
null
null
false
1
false
Zaratahir123/groupData
2022-04-28T16:33:38.000Z
null
false
60f3c3a4a3340dd7f3e8e7895e064c4790d38239
[]
[ "license:mit" ]
https://huggingface.co/datasets/Zaratahir123/groupData/resolve/main/README.md
--- license: mit ---
Zaratahir123
null
null
null
false
1
false
Zaratahir123/test
2022-04-27T10:45:17.000Z
null
false
b3413a70bbc5e925ec9a604338eb4ffab031b9a0
[]
[ "license:mit" ]
https://huggingface.co/datasets/Zaratahir123/test/resolve/main/README.md
--- license: mit ---
shreyasmani
null
null
null
false
1
false
shreyasmani/whrdata2021
2022-04-27T11:11:59.000Z
null
false
d3f7c9ee72dd91ddcb840a8585d248deffa0e5a0
[]
[ "license:other" ]
https://huggingface.co/datasets/shreyasmani/whrdata2021/resolve/main/README.md
--- license: other ---
PolyAI
null
@inproceedings{Casanueva2020, author = {I{\~{n}}igo Casanueva and Tadas Temcinas and Daniela Gerz and Matthew Henderson and Ivan Vulic}, title = {Efficient Intent Detection with Dual Sentence Encoders}, year = {2020}, month = {mar}, note = {Data available at https://github.com/PolyAI-LDN/task-specific-datasets}, url = {https://arxiv.org/abs/2003.04807}, booktitle = {Proceedings of the 2nd Workshop on NLP for ConvAI - ACL 2020} }
BANKING77 dataset provides a very fine-grained set of intents in a banking domain. It comprises 13,083 customer service queries labeled with 77 intents. It focuses on fine-grained single-domain intent detection.
false
20
false
PolyAI/banking77
2022-10-25T10:12:22.000Z
null
false
96a234bd25c04939c4a79213eb764ae90e4d0d81
[]
[ "arxiv:2003.04807", "annotations_creators:expert-generated", "extended:original", "language_creators:expert-generated", "language:en", "license:cc-by-4.0", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "task_categories:text-classification", "task_ids:in...
https://huggingface.co/datasets/PolyAI/banking77/resolve/main/README.md
--- annotations_creators: - expert-generated extended: - original language_creators: - expert-generated language: - en license: - cc-by-4.0 multilinguality: - monolingual size_categories: - 10K<n<100K source_datasets: - original task_categories: - text-classification task_ids: - intent-classification - multi-class-classification paperswithcode_id: null pretty_name: BANKING77 --- # Dataset Card for BANKING77 ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Github](https://github.com/PolyAI-LDN/task-specific-datasets) - **Repository:** [Github](https://github.com/PolyAI-LDN/task-specific-datasets) - **Paper:** [ArXiv](https://arxiv.org/abs/2003.04807) - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Dataset composed of online banking queries annotated with their corresponding intents. BANKING77 dataset provides a very fine-grained set of intents in a banking domain. It comprises 13,083 customer service queries labeled with 77 intents. It focuses on fine-grained single-domain intent detection. ### Supported Tasks and Leaderboards Intent classification, intent detection ### Languages English ## Dataset Structure ### Data Instances An example of 'train' looks as follows: ``` { 'label': 11, # integer label corresponding to "card_arrival" intent 'text': 'I am still waiting on my card?' } ``` ### Data Fields - `text`: a string feature. - `label`: One of classification labels (0-76) corresponding to unique intents. Intent names are mapped to `label` in the following way: | label | intent (category) | |---:|:-------------------------------------------------| | 0 | activate_my_card | | 1 | age_limit | | 2 | apple_pay_or_google_pay | | 3 | atm_support | | 4 | automatic_top_up | | 5 | balance_not_updated_after_bank_transfer | | 6 | balance_not_updated_after_cheque_or_cash_deposit | | 7 | beneficiary_not_allowed | | 8 | cancel_transfer | | 9 | card_about_to_expire | | 10 | card_acceptance | | 11 | card_arrival | | 12 | card_delivery_estimate | | 13 | card_linking | | 14 | card_not_working | | 15 | card_payment_fee_charged | | 16 | card_payment_not_recognised | | 17 | card_payment_wrong_exchange_rate | | 18 | card_swallowed | | 19 | cash_withdrawal_charge | | 20 | cash_withdrawal_not_recognised | | 21 | change_pin | | 22 | compromised_card | | 23 | contactless_not_working | | 24 | country_support | | 25 | declined_card_payment | | 26 | declined_cash_withdrawal | | 27 | declined_transfer | | 28 | direct_debit_payment_not_recognised | | 29 | disposable_card_limits | | 30 | edit_personal_details | | 31 | exchange_charge | | 32 | exchange_rate | | 33 | exchange_via_app | | 34 | extra_charge_on_statement | | 35 | failed_transfer | | 36 | fiat_currency_support | | 37 | get_disposable_virtual_card | | 38 | get_physical_card | | 39 | getting_spare_card | | 40 | getting_virtual_card | | 41 | lost_or_stolen_card | | 42 | lost_or_stolen_phone | | 43 | order_physical_card | | 44 | passcode_forgotten | | 45 | pending_card_payment | | 46 | pending_cash_withdrawal | | 47 | pending_top_up | | 48 | pending_transfer | | 49 | pin_blocked | | 50 | receiving_money | | 51 | Refund_not_showing_up | | 52 | request_refund | | 53 | reverted_card_payment? | | 54 | supported_cards_and_currencies | | 55 | terminate_account | | 56 | top_up_by_bank_transfer_charge | | 57 | top_up_by_card_charge | | 58 | top_up_by_cash_or_cheque | | 59 | top_up_failed | | 60 | top_up_limits | | 61 | top_up_reverted | | 62 | topping_up_by_card | | 63 | transaction_charged_twice | | 64 | transfer_fee_charged | | 65 | transfer_into_account | | 66 | transfer_not_received_by_recipient | | 67 | transfer_timing | | 68 | unable_to_verify_identity | | 69 | verify_my_identity | | 70 | verify_source_of_funds | | 71 | verify_top_up | | 72 | virtual_card_not_working | | 73 | visa_or_mastercard | | 74 | why_verify_identity | | 75 | wrong_amount_of_cash_received | | 76 | wrong_exchange_rate_for_cash_withdrawal | ### Data Splits | Dataset statistics | Train | Test | | --- | --- | --- | | Number of examples | 10 003 | 3 080 | | Average character length | 59.5 | 54.2 | | Number of intents | 77 | 77 | | Number of domains | 1 | 1 | ## Dataset Creation ### Curation Rationale Previous intent detection datasets such as Web Apps, Ask Ubuntu, the Chatbot Corpus or SNIPS are limited to small number of classes (<10), which oversimplifies the intent detection task and does not emulate the true environment of commercial systems. Although there exist large scale *multi-domain* datasets ([HWU64](https://github.com/xliuhw/NLU-Evaluation-Data) and [CLINC150](https://github.com/clinc/oos-eval)), the examples per each domain may not sufficiently capture the full complexity of each domain as encountered "in the wild". This dataset tries to fill the gap and provides a very fine-grained set of intents in a *single-domain* i.e. **banking**. Its focus on fine-grained single-domain intent detection makes it complementary to the other two multi-domain datasets. ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process The dataset does not contain any additional annotations. #### Who are the annotators? [N/A] ### Personal and Sensitive Information [N/A] ## Considerations for Using the Data ### Social Impact of Dataset The purpose of this dataset it to help develop better intent detection systems. Any comprehensive intent detection evaluation should involve both coarser-grained multi-domain datasets and a fine-grained single-domain dataset such as BANKING77. ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [PolyAI](https://github.com/PolyAI-LDN) ### Licensing Information Creative Commons Attribution 4.0 International ### Citation Information ``` @inproceedings{Casanueva2020, author = {I{\~{n}}igo Casanueva and Tadas Temcinas and Daniela Gerz and Matthew Henderson and Ivan Vulic}, title = {Efficient Intent Detection with Dual Sentence Encoders}, year = {2020}, month = {mar}, note = {Data available at https://github.com/PolyAI-LDN/task-specific-datasets}, url = {https://arxiv.org/abs/2003.04807}, booktitle = {Proceedings of the 2nd Workshop on NLP for ConvAI - ACL 2020} } ``` ### Contributions Thanks to [@dkajtoch](https://github.com/dkajtoch) for adding this dataset.
EAST
null
null
null
false
1
false
EAST/autotrain-data-Rule
2022-10-25T10:12:41.000Z
null
false
e504e76976d00dbe5d2ee3945bf6d42c65e2bd1d
[]
[ "language:zh", "task_categories:text-classification" ]
https://huggingface.co/datasets/EAST/autotrain-data-Rule/resolve/main/README.md
--- language: - zh task_categories: - text-classification --- # AutoTrain Dataset for project: Rule ## Dataset Descritpion This dataset has been automatically processed by AutoTrain for project Rule. ### Languages The BCP-47 code for the dataset's language is zh. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json [ { "text": "\u672c\u516c\u53f8\u4f1a\u5728\u60a8\u767b\u5f55\u53ca\u7248\u672c\u66f4\u65b0\u65f6\u4ee5\u63a8\u9001\u901a\u77e5\u3001\u5f39\u6846\u7684\u5f62\u5f0f\u5411\u60a8\u5c55\u793a\u53d8\u66f4\u540e\u7684\u9690\u79c1\u653f\u7b56", "target": 1 }, { "text": "\u6211\u4eec\u53ef\u80fd\u9002\u65f6\u4f1a\u5bf9\u672c\u9690\u79c1\u6743\u653f\u7b56\u8fdb\u884c\u8c03\u6574\u6216\u53d8\u66f4\uff0c\u672c\u9690\u79c1\u6743\u653f\u7b56\u7684\u4efb\u4f55\u66f4\u65b0\u5c06\u4ee5\u6807\u6ce8\u66f4\u65b0\u65f6\u95f4\u7684\u65b9\u5f0f\u516c\u5e03\u5728\u6211\u4eec\u7f51\u7ad9\u4e0a\uff0c\u9664\u6cd5\u5f8b\u6cd5\u89c4\u6216\u76d1\u7ba1\u89c4\u5b9a\u53e6\u6709\u5f3a\u5236\u6027\u89c4\u5b9a\u5916\uff0c\u7ecf\u8c03\u6574\u6216\u53d8\u66f4\u7684\u5185\u5bb9\u4e00\u7ecf\u901a\u77e5\u6216\u516c\u5e03\u540e\u76847\u65e5\u540e\u751f\u6548", "target": 0 } ] ``` ### Dataset Fields The dataset has the following fields (also called "features"): ```json { "text": "Value(dtype='string', id=None)", "target": "ClassLabel(num_classes=2, names=['0', '1'], id=None)" } ``` ### Dataset Splits This dataset is split into a train and validation split. The split sizes are as follow: | Split name | Num samples | | ------------ | ------------------- | | train | 70 | | valid | 19 |
osyvokon
null
null
null
false
1
false
osyvokon/pavlick-formality-scores
2022-10-25T10:12:43.000Z
null
false
90e53b922d4e6574623b352a4be4ae4ae6cdff61
[]
[ "annotations_creators:crowdsourced", "language_creators:found", "language:en-US", "license:cc-by-3.0", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "task_categories:text-classification", "task_ids:text-scoring" ]
https://huggingface.co/datasets/osyvokon/pavlick-formality-scores/resolve/main/README.md
--- annotations_creators: - crowdsourced language_creators: - found language: - en-US license: - cc-by-3.0 multilinguality: - monolingual pretty_name: 'Sentence-level formality annotations for news, blogs, email and QA forums. Published in "An Empirical Analysis of Formality in Online Communication" (Pavlick and Tetreault, 2016) ' size_categories: - 10K<n<100K source_datasets: - original task_categories: - text-classification task_ids: - text-scoring --- This dataset contains sentence-level formality annotations used in the 2016 TACL paper "An Empirical Analysis of Formality in Online Communication" (Pavlick and Tetreault, 2016). It includes sentences from four genres (news, blogs, email, and QA forums), all annotated by humans on Amazon Mechanical Turk. The news and blog data was collected by Shibamouli Lahiri, and we are redistributing it here for the convenience of other researchers. We collected the email and answers data ourselves, using a similar annotation setup to Shibamouli. In the original dataset, `answers` and `email` were tokenized. In this version, Oleksiy Syvokon detokenized them with `moses-detokenizer` and a bunch of additional regexps. If you use this data in your work, please cite BOTH of the below papers: ``` @article{PavlickAndTetreault-2016:TACL, author = {Ellie Pavlick and Joel Tetreault}, title = {An Empirical Analysis of Formality in Online Communication}, journal = {Transactions of the Association for Computational Linguistics}, year = {2016}, publisher = {Association for Computational Linguistics} } @article{Lahiri-2015:arXiv, title={{SQUINKY! A} Corpus of Sentence-level Formality, Informativeness, and Implicature}, author={Lahiri, Shibamouli}, journal={arXiv preprint arXiv:1506.02306}, year={2015} } ``` ## Contents The annotated data files and number of lines in each are as follows: * 4977 answers -- Annotated sentences from a random sample of posts from the Yahoo! Answers forums: https://answers.yahoo.com/ * 1821 blog -- Annotated sentences from the top 100 blogs listed on http://technorati.com/ on October 31, 2009. * 1701 email -- Annotated sentences from a random sample of emails from the Jeb Bush email archive: http://americanbridgepac.org/jeb-bushs-gubernatorial-email-archive/ * 2775 news -- Annotated sentences from the "breaking", "recent", and "local" news sections of the following 20 news sites: CNN, CBS News, ABC News, Reuters, BBC News Online, New York Times, Los Angeles Times, The Guardian (U.K.), Voice of America, Boston Globe, Chicago Tribune, San Francisco Chronicle, Times Online (U.K.), news.com.au, Xinhua, The Times of India, Seattle Post Intelligencer, Daily Mail, and Bloomberg L.P. ## Format Each record contains the following fields: 1. `avg_score`: the mean formality rating, which ranges from -3 to 3 where lower scores indicate less formal sentences 2. `sentence`
NLPC-UOM
null
null
null
false
1
false
NLPC-UOM/Writing-style-classification
2022-10-25T10:12:46.000Z
null
false
2d04a4fde26d92fbed2ae7604677736e6b901a4d
[]
[ "language_creators:crowdsourced", "language:si", "license:mit", "multilinguality:monolingual", "task_categories:text-classification" ]
https://huggingface.co/datasets/NLPC-UOM/Writing-style-classification/resolve/main/README.md
--- annotations_creators: [] language_creators: - crowdsourced language: - si license: - mit multilinguality: - monolingual pretty_name: sinhala-writing-style-classification size_categories: [] source_datasets: [] task_categories: - text-classification task_ids: [] --- This file contains news texts (sentences) belonging to different writing styles. The original dataset created by {*Upeksha, D., Wijayarathna, C., Siriwardena, M., Lasandun, L., Wimalasuriya, C., de Silva, N., and Dias, G. (2015). Implementing a corpus for Sinhala language. 01*}is processed and cleaned. If you use this dataset, please cite {*Dhananjaya et al. BERTifying Sinhala - A Comprehensive Analysis of Pre-trained Language Models for Sinhala Text Classification, 2022*} and the above mentioned paper.
bigscience-catalogue-data
null
""" # TODO: Add description of the dataset here # You can copy an official description _DESCRIPTION =
This is a preliminary version of the bias SHADES dataset for evaluating LMs for social biases.
false
65
true
bigscience-catalogue-data/bias-shades
2022-05-01T15:18:02.000Z
null
false
e220d18727665580a083db6162dd78f9e0f61438
[]
[ "arxiv:2010.00133", "license:cc-by-sa-4.0" ]
https://huggingface.co/datasets/bigscience-catalogue-data/bias-shades/resolve/main/README.md
mrm8488
null
null
null
false
211
false
mrm8488/ImageNet1K-val
2022-04-27T19:16:51.000Z
null
false
f004a913f9170e5ff39e63f6172ee3ae918197c9
[]
[]
https://huggingface.co/datasets/mrm8488/ImageNet1K-val/resolve/main/README.md
mapping: ``` n01440764 tench, Tinca tinca n01443537 goldfish, Carassius auratus n01484850 great white shark, white shark, man-eater, man-eating shark, Carcharodon carcharias n01491361 tiger shark, Galeocerdo cuvieri n01494475 hammerhead, hammerhead shark n01496331 electric ray, crampfish, numbfish, torpedo n01498041 stingray n01514668 cock n01514859 hen n01518878 ostrich, Struthio camelus n01530575 brambling, Fringilla montifringilla n01531178 goldfinch, Carduelis carduelis n01532829 house finch, linnet, Carpodacus mexicanus n01534433 junco, snowbird n01537544 indigo bunting, indigo finch, indigo bird, Passerina cyanea n01558993 robin, American robin, Turdus migratorius n01560419 bulbul n01580077 jay n01582220 magpie n01592084 chickadee n01601694 water ouzel, dipper n01608432 kite n01614925 bald eagle, American eagle, Haliaeetus leucocephalus n01616318 vulture n01622779 great grey owl, great gray owl, Strix nebulosa n01629819 European fire salamander, Salamandra salamandra n01630670 common newt, Triturus vulgaris n01631663 eft n01632458 spotted salamander, Ambystoma maculatum n01632777 axolotl, mud puppy, Ambystoma mexicanum n01641577 bullfrog, Rana catesbeiana n01644373 tree frog, tree-frog n01644900 tailed frog, bell toad, ribbed toad, tailed toad, Ascaphus trui n01664065 loggerhead, loggerhead turtle, Caretta caretta n01665541 leatherback turtle, leatherback, leathery turtle, Dermochelys coriacea n01667114 mud turtle n01667778 terrapin n01669191 box turtle, box tortoise n01675722 banded gecko n01677366 common iguana, iguana, Iguana iguana n01682714 American chameleon, anole, Anolis carolinensis n01685808 whiptail, whiptail lizard n01687978 agama n01688243 frilled lizard, Chlamydosaurus kingi n01689811 alligator lizard n01692333 Gila monster, Heloderma suspectum n01693334 green lizard, Lacerta viridis n01694178 African chameleon, Chamaeleo chamaeleon n01695060 Komodo dragon, Komodo lizard, dragon lizard, giant lizard, Varanus komodoensis n01697457 African crocodile, Nile crocodile, Crocodylus niloticus n01698640 American alligator, Alligator mississipiensis n01704323 triceratops n01728572 thunder snake, worm snake, Carphophis amoenus n01728920 ringneck snake, ring-necked snake, ring snake n01729322 hognose snake, puff adder, sand viper n01729977 green snake, grass snake n01734418 king snake, kingsnake n01735189 garter snake, grass snake n01737021 water snake n01739381 vine snake n01740131 night snake, Hypsiglena torquata n01742172 boa constrictor, Constrictor constrictor n01744401 rock python, rock snake, Python sebae n01748264 Indian cobra, Naja naja n01749939 green mamba n01751748 sea snake n01753488 horned viper, cerastes, sand viper, horned asp, Cerastes cornutus n01755581 diamondback, diamondback rattlesnake, Crotalus adamanteus n01756291 sidewinder, horned rattlesnake, Crotalus cerastes n01768244 trilobite n01770081 harvestman, daddy longlegs, Phalangium opilio n01770393 scorpion n01773157 black and gold garden spider, Argiope aurantia n01773549 barn spider, Araneus cavaticus n01773797 garden spider, Aranea diademata n01774384 black widow, Latrodectus mactans n01774750 tarantula n01775062 wolf spider, hunting spider n01776313 tick n01784675 centipede n01795545 black grouse n01796340 ptarmigan n01797886 ruffed grouse, partridge, Bonasa umbellus n01798484 prairie chicken, prairie grouse, prairie fowl n01806143 peacock n01806567 quail n01807496 partridge n01817953 African grey, African gray, Psittacus erithacus n01818515 macaw n01819313 sulphur-crested cockatoo, Kakatoe galerita, Cacatua galerita n01820546 lorikeet n01824575 coucal n01828970 bee eater n01829413 hornbill n01833805 hummingbird n01843065 jacamar n01843383 toucan n01847000 drake n01855032 red-breasted merganser, Mergus serrator n01855672 goose n01860187 black swan, Cygnus atratus n01871265 tusker n01872401 echidna, spiny anteater, anteater n01873310 platypus, duckbill, duckbilled platypus, duck-billed platypus, Ornithorhynchus anatinus n01877812 wallaby, brush kangaroo n01882714 koala, koala bear, kangaroo bear, native bear, Phascolarctos cinereus n01883070 wombat n01910747 jellyfish n01914609 sea anemone, anemone n01917289 brain coral n01924916 flatworm, platyhelminth n01930112 nematode, nematode worm, roundworm n01943899 conch n01944390 snail n01945685 slug n01950731 sea slug, nudibranch n01955084 chiton, coat-of-mail shell, sea cradle, polyplacophore n01968897 chambered nautilus, pearly nautilus, nautilus n01978287 Dungeness crab, Cancer magister n01978455 rock crab, Cancer irroratus n01980166 fiddler crab n01981276 king crab, Alaska crab, Alaskan king crab, Alaska king crab, Paralithodes camtschatica n01983481 American lobster, Northern lobster, Maine lobster, Homarus americanus n01984695 spiny lobster, langouste, rock lobster, crawfish, crayfish, sea crawfish n01985128 crayfish, crawfish, crawdad, crawdaddy n01986214 hermit crab n01990800 isopod n02002556 white stork, Ciconia ciconia n02002724 black stork, Ciconia nigra n02006656 spoonbill n02007558 flamingo n02009229 little blue heron, Egretta caerulea n02009912 American egret, great white heron, Egretta albus n02011460 bittern n02012849 crane n02013706 limpkin, Aramus pictus n02017213 European gallinule, Porphyrio porphyrio n02018207 American coot, marsh hen, mud hen, water hen, Fulica americana n02018795 bustard n02025239 ruddy turnstone, Arenaria interpres n02027492 red-backed sandpiper, dunlin, Erolia alpina n02028035 redshank, Tringa totanus n02033041 dowitcher n02037110 oystercatcher, oyster catcher n02051845 pelican n02056570 king penguin, Aptenodytes patagonica n02058221 albatross, mollymawk n02066245 grey whale, gray whale, devilfish, Eschrichtius gibbosus, Eschrichtius robustus n02071294 killer whale, killer, orca, grampus, sea wolf, Orcinus orca n02074367 dugong, Dugong dugon n02077923 sea lion n02085620 Chihuahua n02085782 Japanese spaniel n02085936 Maltese dog, Maltese terrier, Maltese n02086079 Pekinese, Pekingese, Peke n02086240 Shih-Tzu n02086646 Blenheim spaniel n02086910 papillon n02087046 toy terrier n02087394 Rhodesian ridgeback n02088094 Afghan hound, Afghan n02088238 basset, basset hound n02088364 beagle n02088466 bloodhound, sleuthhound n02088632 bluetick n02089078 black-and-tan coonhound n02089867 Walker hound, Walker foxhound n02089973 English foxhound n02090379 redbone n02090622 borzoi, Russian wolfhound n02090721 Irish wolfhound n02091032 Italian greyhound n02091134 whippet n02091244 Ibizan hound, Ibizan Podenco n02091467 Norwegian elkhound, elkhound n02091635 otterhound, otter hound n02091831 Saluki, gazelle hound n02092002 Scottish deerhound, deerhound n02092339 Weimaraner n02093256 Staffordshire bullterrier, Staffordshire bull terrier n02093428 American Staffordshire terrier, Staffordshire terrier, American pit bull terrier, pit bull terrier n02093647 Bedlington terrier n02093754 Border terrier n02093859 Kerry blue terrier n02093991 Irish terrier n02094114 Norfolk terrier n02094258 Norwich terrier n02094433 Yorkshire terrier n02095314 wire-haired fox terrier n02095570 Lakeland terrier n02095889 Sealyham terrier, Sealyham n02096051 Airedale, Airedale terrier n02096177 cairn, cairn terrier n02096294 Australian terrier n02096437 Dandie Dinmont, Dandie Dinmont terrier n02096585 Boston bull, Boston terrier n02097047 miniature schnauzer n02097130 giant schnauzer n02097209 standard schnauzer n02097298 Scotch terrier, Scottish terrier, Scottie n02097474 Tibetan terrier, chrysanthemum dog n02097658 silky terrier, Sydney silky n02098105 soft-coated wheaten terrier n02098286 West Highland white terrier n02098413 Lhasa, Lhasa apso n02099267 flat-coated retriever n02099429 curly-coated retriever n02099601 golden retriever n02099712 Labrador retriever n02099849 Chesapeake Bay retriever n02100236 German short-haired pointer n02100583 vizsla, Hungarian pointer n02100735 English setter n02100877 Irish setter, red setter n02101006 Gordon setter n02101388 Brittany spaniel n02101556 clumber, clumber spaniel n02102040 English springer, English springer spaniel n02102177 Welsh springer spaniel n02102318 cocker spaniel, English cocker spaniel, cocker n02102480 Sussex spaniel n02102973 Irish water spaniel n02104029 kuvasz n02104365 schipperke n02105056 groenendael n02105162 malinois n02105251 briard n02105412 kelpie n02105505 komondor n02105641 Old English sheepdog, bobtail n02105855 Shetland sheepdog, Shetland sheep dog, Shetland n02106030 collie n02106166 Border collie n02106382 Bouvier des Flandres, Bouviers des Flandres n02106550 Rottweiler n02106662 German shepherd, German shepherd dog, German police dog, alsatian n02107142 Doberman, Doberman pinscher n02107312 miniature pinscher n02107574 Greater Swiss Mountain dog n02107683 Bernese mountain dog n02107908 Appenzeller n02108000 EntleBucher n02108089 boxer n02108422 bull mastiff n02108551 Tibetan mastiff n02108915 French bulldog n02109047 Great Dane n02109525 Saint Bernard, St Bernard n02109961 Eskimo dog, husky n02110063 malamute, malemute, Alaskan malamute n02110185 Siberian husky n02110341 dalmatian, coach dog, carriage dog n02110627 affenpinscher, monkey pinscher, monkey dog n02110806 basenji n02110958 pug, pug-dog n02111129 Leonberg n02111277 Newfoundland, Newfoundland dog n02111500 Great Pyrenees n02111889 Samoyed, Samoyede n02112018 Pomeranian n02112137 chow, chow chow n02112350 keeshond n02112706 Brabancon griffon n02113023 Pembroke, Pembroke Welsh corgi n02113186 Cardigan, Cardigan Welsh corgi n02113624 toy poodle n02113712 miniature poodle n02113799 standard poodle n02113978 Mexican hairless n02114367 timber wolf, grey wolf, gray wolf, Canis lupus n02114548 white wolf, Arctic wolf, Canis lupus tundrarum n02114712 red wolf, maned wolf, Canis rufus, Canis niger n02114855 coyote, prairie wolf, brush wolf, Canis latrans n02115641 dingo, warrigal, warragal, Canis dingo n02115913 dhole, Cuon alpinus n02116738 African hunting dog, hyena dog, Cape hunting dog, Lycaon pictus n02117135 hyena, hyaena n02119022 red fox, Vulpes vulpes n02119789 kit fox, Vulpes macrotis n02120079 Arctic fox, white fox, Alopex lagopus n02120505 grey fox, gray fox, Urocyon cinereoargenteus n02123045 tabby, tabby cat n02123159 tiger cat n02123394 Persian cat n02123597 Siamese cat, Siamese n02124075 Egyptian cat n02125311 cougar, puma, catamount, mountain lion, painter, panther, Felis concolor n02127052 lynx, catamount n02128385 leopard, Panthera pardus n02128757 snow leopard, ounce, Panthera uncia n02128925 jaguar, panther, Panthera onca, Felis onca n02129165 lion, king of beasts, Panthera leo n02129604 tiger, Panthera tigris n02130308 cheetah, chetah, Acinonyx jubatus n02132136 brown bear, bruin, Ursus arctos n02133161 American black bear, black bear, Ursus americanus, Euarctos americanus n02134084 ice bear, polar bear, Ursus Maritimus, Thalarctos maritimus n02134418 sloth bear, Melursus ursinus, Ursus ursinus n02137549 mongoose n02138441 meerkat, mierkat n02165105 tiger beetle n02165456 ladybug, ladybeetle, lady beetle, ladybird, ladybird beetle n02167151 ground beetle, carabid beetle n02168699 long-horned beetle, longicorn, longicorn beetle n02169497 leaf beetle, chrysomelid n02172182 dung beetle n02174001 rhinoceros beetle n02177972 weevil n02190166 fly n02206856 bee n02219486 ant, emmet, pismire n02226429 grasshopper, hopper n02229544 cricket n02231487 walking stick, walkingstick, stick insect n02233338 cockroach, roach n02236044 mantis, mantid n02256656 cicada, cicala n02259212 leafhopper n02264363 lacewing, lacewing fly n02268443 dragonfly, darning needle, devil's darning needle, sewing needle, snake feeder, snake doctor, mosquito hawk, skeeter hawk n02268853 damselfly n02276258 admiral n02277742 ringlet, ringlet butterfly n02279972 monarch, monarch butterfly, milkweed butterfly, Danaus plexippus n02280649 cabbage butterfly n02281406 sulphur butterfly, sulfur butterfly n02281787 lycaenid, lycaenid butterfly n02317335 starfish, sea star n02319095 sea urchin n02321529 sea cucumber, holothurian n02325366 wood rabbit, cottontail, cottontail rabbit n02326432 hare n02328150 Angora, Angora rabbit n02342885 hamster n02346627 porcupine, hedgehog n02356798 fox squirrel, eastern fox squirrel, Sciurus niger n02361337 marmot n02363005 beaver n02364673 guinea pig, Cavia cobaya n02389026 sorrel n02391049 zebra n02395406 hog, pig, grunter, squealer, Sus scrofa n02396427 wild boar, boar, Sus scrofa n02397096 warthog n02398521 hippopotamus, hippo, river horse, Hippopotamus amphibius n02403003 ox n02408429 water buffalo, water ox, Asiatic buffalo, Bubalus bubalis n02410509 bison n02412080 ram, tup n02415577 bighorn, bighorn sheep, cimarron, Rocky Mountain bighorn, Rocky Mountain sheep, Ovis canadensis n02417914 ibex, Capra ibex n02422106 hartebeest n02422699 impala, Aepyceros melampus n02423022 gazelle n02437312 Arabian camel, dromedary, Camelus dromedarius n02437616 llama n02441942 weasel n02442845 mink n02443114 polecat, fitch, foulmart, foumart, Mustela putorius n02443484 black-footed ferret, ferret, Mustela nigripes n02444819 otter n02445715 skunk, polecat, wood pussy n02447366 badger n02454379 armadillo n02457408 three-toed sloth, ai, Bradypus tridactylus n02480495 orangutan, orang, orangutang, Pongo pygmaeus n02480855 gorilla, Gorilla gorilla n02481823 chimpanzee, chimp, Pan troglodytes n02483362 gibbon, Hylobates lar n02483708 siamang, Hylobates syndactylus, Symphalangus syndactylus n02484975 guenon, guenon monkey n02486261 patas, hussar monkey, Erythrocebus patas n02486410 baboon n02487347 macaque n02488291 langur n02488702 colobus, colobus monkey n02489166 proboscis monkey, Nasalis larvatus n02490219 marmoset n02492035 capuchin, ringtail, Cebus capucinus n02492660 howler monkey, howler n02493509 titi, titi monkey n02493793 spider monkey, Ateles geoffroyi n02494079 squirrel monkey, Saimiri sciureus n02497673 Madagascar cat, ring-tailed lemur, Lemur catta n02500267 indri, indris, Indri indri, Indri brevicaudatus n02504013 Indian elephant, Elephas maximus n02504458 African elephant, Loxodonta africana n02509815 lesser panda, red panda, panda, bear cat, cat bear, Ailurus fulgens n02510455 giant panda, panda, panda bear, coon bear, Ailuropoda melanoleuca n02514041 barracouta, snoek n02526121 eel n02536864 coho, cohoe, coho salmon, blue jack, silver salmon, Oncorhynchus kisutch n02606052 rock beauty, Holocanthus tricolor n02607072 anemone fish n02640242 sturgeon n02641379 gar, garfish, garpike, billfish, Lepisosteus osseus n02643566 lionfish n02655020 puffer, pufferfish, blowfish, globefish n02666196 abacus n02667093 abaya n02669723 academic gown, academic robe, judge's robe n02672831 accordion, piano accordion, squeeze box n02676566 acoustic guitar n02687172 aircraft carrier, carrier, flattop, attack aircraft carrier n02690373 airliner n02692877 airship, dirigible n02699494 altar n02701002 ambulance n02704792 amphibian, amphibious vehicle n02708093 analog clock n02727426 apiary, bee house n02730930 apron n02747177 ashcan, trash can, garbage can, wastebin, ash bin, ash-bin, ashbin, dustbin, trash barrel, trash bin n02749479 assault rifle, assault gun n02769748 backpack, back pack, knapsack, packsack, rucksack, haversack n02776631 bakery, bakeshop, bakehouse n02777292 balance beam, beam n02782093 balloon n02783161 ballpoint, ballpoint pen, ballpen, Biro n02786058 Band Aid n02787622 banjo n02788148 bannister, banister, balustrade, balusters, handrail n02790996 barbell n02791124 barber chair n02791270 barbershop n02793495 barn n02794156 barometer n02795169 barrel, cask n02797295 barrow, garden cart, lawn cart, wheelbarrow n02799071 baseball n02802426 basketball n02804414 bassinet n02804610 bassoon n02807133 bathing cap, swimming cap n02808304 bath towel n02808440 bathtub, bathing tub, bath, tub n02814533 beach wagon, station wagon, wagon, estate car, beach waggon, station waggon, waggon n02814860 beacon, lighthouse, beacon light, pharos n02815834 beaker n02817516 bearskin, busby, shako n02823428 beer bottle n02823750 beer glass n02825657 bell cote, bell cot n02834397 bib n02835271 bicycle-built-for-two, tandem bicycle, tandem n02837789 bikini, two-piece n02840245 binder, ring-binder n02841315 binoculars, field glasses, opera glasses n02843684 birdhouse n02859443 boathouse n02860847 bobsled, bobsleigh, bob n02865351 bolo tie, bolo, bola tie, bola n02869837 bonnet, poke bonnet n02870880 bookcase n02871525 bookshop, bookstore, bookstall n02877765 bottlecap n02879718 bow n02883205 bow tie, bow-tie, bowtie n02892201 brass, memorial tablet, plaque n02892767 brassiere, bra, bandeau n02894605 breakwater, groin, groyne, mole, bulwark, seawall, jetty n02895154 breastplate, aegis, egis n02906734 broom n02909870 bucket, pail n02910353 buckle n02916936 bulletproof vest n02917067 bullet train, bullet n02927161 butcher shop, meat market n02930766 cab, hack, taxi, taxicab n02939185 caldron, cauldron n02948072 candle, taper, wax light n02950826 cannon n02951358 canoe n02951585 can opener, tin opener n02963159 cardigan n02965783 car mirror n02966193 carousel, carrousel, merry-go-round, roundabout, whirligig n02966687 carpenter's kit, tool kit n02971356 carton n02974003 car wheel n02977058 cash machine, cash dispenser, automated teller machine, automatic teller machine, automated teller, automatic teller, ATM n02978881 cassette n02979186 cassette player n02980441 castle n02981792 catamaran n02988304 CD player n02992211 cello, violoncello n02992529 cellular telephone, cellular phone, cellphone, cell, mobile phone n02999410 chain n03000134 chainlink fence n03000247 chain mail, ring mail, mail, chain armor, chain armour, ring armor, ring armour n03000684 chain saw, chainsaw n03014705 chest n03016953 chiffonier, commode n03017168 chime, bell, gong n03018349 china cabinet, china closet n03026506 Christmas stocking n03028079 church, church building n03032252 cinema, movie theater, movie theatre, movie house, picture palace n03041632 cleaver, meat cleaver, chopper n03042490 cliff dwelling n03045698 cloak n03047690 clog, geta, patten, sabot n03062245 cocktail shaker n03063599 coffee mug n03063689 coffeepot n03065424 coil, spiral, volute, whorl, helix n03075370 combination lock n03085013 computer keyboard, keypad n03089624 confectionery, confectionary, candy store n03095699 container ship, containership, container vessel n03100240 convertible n03109150 corkscrew, bottle screw n03110669 cornet, horn, trumpet, trump n03124043 cowboy boot n03124170 cowboy hat, ten-gallon hat n03125729 cradle n03126707 crane n03127747 crash helmet n03127925 crate n03131574 crib, cot n03133878 Crock Pot n03134739 croquet ball n03141823 crutch n03146219 cuirass n03160309 dam, dike, dyke n03179701 desk n03180011 desktop computer n03187595 dial telephone, dial phone n03188531 diaper, nappy, napkin n03196217 digital clock n03197337 digital watch n03201208 dining table, board n03207743 dishrag, dishcloth n03207941 dishwasher, dish washer, dishwashing machine n03208938 disk brake, disc brake n03216828 dock, dockage, docking facility n03218198 dogsled, dog sled, dog sleigh n03220513 dome n03223299 doormat, welcome mat n03240683 drilling platform, offshore rig n03249569 drum, membranophone, tympan n03250847 drumstick n03255030 dumbbell n03259280 Dutch oven n03271574 electric fan, blower n03272010 electric guitar n03272562 electric locomotive n03290653 entertainment center n03291819 envelope n03297495 espresso maker n03314780 face powder n03325584 feather boa, boa n03337140 file, file cabinet, filing cabinet n03344393 fireboat n03345487 fire engine, fire truck n03347037 fire screen, fireguard n03355925 flagpole, flagstaff n03372029 flute, transverse flute n03376595 folding chair n03379051 football helmet n03384352 forklift n03388043 fountain n03388183 fountain pen n03388549 four-poster n03393912 freight car n03394916 French horn, horn n03400231 frying pan, frypan, skillet n03404251 fur coat n03417042 garbage truck, dustcart n03424325 gasmask, respirator, gas helmet n03425413 gas pump, gasoline pump, petrol pump, island dispenser n03443371 goblet n03444034 go-kart n03445777 golf ball n03445924 golfcart, golf cart n03447447 gondola n03447721 gong, tam-tam n03450230 gown n03452741 grand piano, grand n03457902 greenhouse, nursery, glasshouse n03459775 grille, radiator grille n03461385 grocery store, grocery, food market, market n03467068 guillotine n03476684 hair slide n03476991 hair spray n03478589 half track n03481172 hammer n03482405 hamper n03483316 hand blower, blow dryer, blow drier, hair dryer, hair drier n03485407 hand-held computer, hand-held microcomputer n03485794 handkerchief, hankie, hanky, hankey n03492542 hard disc, hard disk, fixed disk n03494278 harmonica, mouth organ, harp, mouth harp n03495258 harp n03496892 harvester, reaper n03498962 hatchet n03527444 holster n03529860 home theater, home theatre n03530642 honeycomb n03532672 hook, claw n03534580 hoopskirt, crinoline n03535780 horizontal bar, high bar n03538406 horse cart, horse-cart n03544143 hourglass n03584254 iPod n03584829 iron, smoothing iron n03590841 jack-o'-lantern n03594734 jean, blue jean, denim n03594945 jeep, landrover n03595614 jersey, T-shirt, tee shirt n03598930 jigsaw puzzle n03599486 jinrikisha, ricksha, rickshaw n03602883 joystick n03617480 kimono n03623198 knee pad n03627232 knot n03630383 lab coat, laboratory coat n03633091 ladle n03637318 lampshade, lamp shade n03642806 laptop, laptop computer n03649909 lawn mower, mower n03657121 lens cap, lens cover n03658185 letter opener, paper knife, paperknife n03661043 library n03662601 lifeboat n03666591 lighter, light, igniter, ignitor n03670208 limousine, limo n03673027 liner, ocean liner n03676483 lipstick, lip rouge n03680355 Loafer n03690938 lotion n03691459 loudspeaker, speaker, speaker unit, loudspeaker system, speaker system n03692522 loupe, jeweler's loupe n03697007 lumbermill, sawmill n03706229 magnetic compass n03709823 mailbag, postbag n03710193 mailbox, letter box n03710637 maillot n03710721 maillot, tank suit n03717622 manhole cover n03720891 maraca n03721384 marimba, xylophone n03724870 mask n03729826 matchstick n03733131 maypole n03733281 maze, labyrinth n03733805 measuring cup n03742115 medicine chest, medicine cabinet n03743016 megalith, megalithic structure n03759954 microphone, mike n03761084 microwave, microwave oven n03763968 military uniform n03764736 milk can n03769881 minibus n03770439 miniskirt, mini n03770679 minivan n03773504 missile n03775071 mitten n03775546 mixing bowl n03776460 mobile home, manufactured home n03777568 Model T n03777754 modem n03781244 monastery n03782006 monitor n03785016 moped n03786901 mortar n03787032 mortarboard n03788195 mosque n03788365 mosquito net n03791053 motor scooter, scooter n03792782 mountain bike, all-terrain bike, off-roader n03792972 mountain tent n03793489 mouse, computer mouse n03794056 mousetrap n03796401 moving van n03803284 muzzle n03804744 nail n03814639 neck brace n03814906 necklace n03825788 nipple n03832673 notebook, notebook computer n03837869 obelisk n03838899 oboe, hautboy, hautbois n03840681 ocarina, sweet potato n03841143 odometer, hodometer, mileometer, milometer n03843555 oil filter n03854065 organ, pipe organ n03857828 oscilloscope, scope, cathode-ray oscilloscope, CRO n03866082 overskirt n03868242 oxcart n03868863 oxygen mask n03871628 packet n03873416 paddle, boat paddle n03874293 paddlewheel, paddle wheel n03874599 padlock n03876231 paintbrush n03877472 pajama, pyjama, pj's, jammies n03877845 palace n03884397 panpipe, pandean pipe, syrinx n03887697 paper towel n03888257 parachute, chute n03888605 parallel bars, bars n03891251 park bench n03891332 parking meter n03895866 passenger car, coach, carriage n03899768 patio, terrace n03902125 pay-phone, pay-station n03903868 pedestal, plinth, footstall n03908618 pencil box, pencil case n03908714 pencil sharpener n03916031 perfume, essence n03920288 Petri dish n03924679 photocopier n03929660 pick, plectrum, plectron n03929855 pickelhaube n03930313 picket fence, paling n03930630 pickup, pickup truck n03933933 pier n03935335 piggy bank, penny bank n03937543 pill bottle n03938244 pillow n03942813 ping-pong ball n03944341 pinwheel n03947888 pirate, pirate ship n03950228 pitcher, ewer n03954731 plane, carpenter's plane, woodworking plane n03956157 planetarium n03958227 plastic bag n03961711 plate rack n03967562 plow, plough n03970156 plunger, plumber's helper n03976467 Polaroid camera, Polaroid Land camera n03976657 pole n03977966 police van, police wagon, paddy wagon, patrol wagon, wagon, black Maria n03980874 poncho n03982430 pool table, billiard table, snooker table n03983396 pop bottle, soda bottle n03991062 pot, flowerpot n03992509 potter's wheel n03995372 power drill n03998194 prayer rug, prayer mat n04004767 printer n04005630 prison, prison house n04008634 projectile, missile n04009552 projector n04019541 puck, hockey puck n04023962 punching bag, punch bag, punching ball, punchball n04026417 purse n04033901 quill, quill pen n04033995 quilt, comforter, comfort, puff n04037443 racer, race car, racing car n04039381 racket, racquet n04040759 radiator n04041544 radio, wireless n04044716 radio telescope, radio reflector n04049303 rain barrel n04065272 recreational vehicle, RV, R.V. n04067472 reel n04069434 reflex camera n04070727 refrigerator, icebox n04074963 remote control, remote n04081281 restaurant, eating house, eating place, eatery n04086273 revolver, six-gun, six-shooter n04090263 rifle n04099969 rocking chair, rocker n04111531 rotisserie n04116512 rubber eraser, rubber, pencil eraser n04118538 rugby ball n04118776 rule, ruler n04120489 running shoe n04125021 safe n04127249 safety pin n04131690 saltshaker, salt shaker n04133789 sandal n04136333 sarong n04141076 sax, saxophone n04141327 scabbard n04141975 scale, weighing machine n04146614 school bus n04147183 schooner n04149813 scoreboard n04152593 screen, CRT screen n04153751 screw n04154565 screwdriver n04162706 seat belt, seatbelt n04179913 sewing machine n04192698 shield, buckler n04200800 shoe shop, shoe-shop, shoe store n04201297 shoji n04204238 shopping basket n04204347 shopping cart n04208210 shovel n04209133 shower cap n04209239 shower curtain n04228054 ski n04229816 ski mask n04235860 sleeping bag n04238763 slide rule, slipstick n04239074 sliding door n04243546 slot, one-armed bandit n04251144 snorkel n04252077 snowmobile n04252225 snowplow, snowplough n04254120 soap dispenser n04254680 soccer ball n04254777 sock n04258138 solar dish, solar collector, solar furnace n04259630 sombrero n04263257 soup bowl n04264628 space bar n04265275 space heater n04266014 space shuttle n04270147 spatula n04273569 speedboat n04275548 spider web, spider's web n04277352 spindle n04285008 sports car, sport car n04286575 spotlight, spot n04296562 stage n04310018 steam locomotive n04311004 steel arch bridge n04311174 steel drum n04317175 stethoscope n04325704 stole n04326547 stone wall n04328186 stopwatch, stop watch n04330267 stove n04332243 strainer n04335435 streetcar, tram, tramcar, trolley, trolley car n04336792 stretcher n04344873 studio couch, day bed n04346328 stupa, tope n04347754 submarine, pigboat, sub, U-boat n04350905 suit, suit of clothes n04355338 sundial n04355933 sunglass n04356056 sunglasses, dark glasses, shades n04357314 sunscreen, sunblock, sun blocker n04366367 suspension bridge n04367480 swab, swob, mop n04370456 sweatshirt n04371430 swimming trunks, bathing trunks n04371774 swing n04372370 switch, electric switch, electrical switch n04376876 syringe n04380533 table lamp n04389033 tank, army tank, armored combat vehicle, armoured combat vehicle n04392985 tape player n04398044 teapot n04399382 teddy, teddy bear n04404412 television, television system n04409515 tennis ball n04417672 thatch, thatched roof n04418357 theater curtain, theatre curtain n04423845 thimble n04428191 thresher, thrasher, threshing machine n04429376 throne n04435653 tile roof n04442312 toaster n04443257 tobacco shop, tobacconist shop, tobacconist n04447861 toilet seat n04456115 torch n04458633 totem pole n04461696 tow truck, tow car, wrecker n04462240 toyshop n04465501 tractor n04467665 trailer truck, tractor trailer, trucking rig, rig, articulated lorry, semi n04476259 tray n04479046 trench coat n04482393 tricycle, trike, velocipede n04483307 trimaran n04485082 tripod n04486054 triumphal arch n04487081 trolleybus, trolley coach, trackless trolley n04487394 trombone n04493381 tub, vat n04501370 turnstile n04505470 typewriter keyboard n04507155 umbrella n04509417 unicycle, monocycle n04515003 upright, upright piano n04517823 vacuum, vacuum cleaner n04522168 vase n04523525 vault n04525038 velvet n04525305 vending machine n04532106 vestment n04532670 viaduct n04536866 violin, fiddle n04540053 volleyball n04542943 waffle iron n04548280 wall clock n04548362 wallet, billfold, notecase, pocketbook n04550184 wardrobe, closet, press n04552348 warplane, military plane n04553703 washbasin, handbasin, washbowl, lavabo, wash-hand basin n04554684 washer, automatic washer, washing machine n04557648 water bottle n04560804 water jug n04562935 water tower n04579145 whiskey jug n04579432 whistle n04584207 wig n04589890 window screen n04590129 window shade n04591157 Windsor tie n04591713 wine bottle n04592741 wing n04596742 wok n04597913 wooden spoon n04599235 wool, woolen, woollen n04604644 worm fence, snake fence, snake-rail fence, Virginia fence n04606251 wreck n04612504 yawl n04613696 yurt n06359193 web site, website, internet site, site n06596364 comic book n06785654 crossword puzzle, crossword n06794110 street sign n06874185 traffic light, traffic signal, stoplight n07248320 book jacket, dust cover, dust jacket, dust wrapper n07565083 menu n07579787 plate n07583066 guacamole n07584110 consomme n07590611 hot pot, hotpot n07613480 trifle n07614500 ice cream, icecream n07615774 ice lolly, lolly, lollipop, popsicle n07684084 French loaf n07693725 bagel, beigel n07695742 pretzel n07697313 cheeseburger n07697537 hotdog, hot dog, red hot n07711569 mashed potato n07714571 head cabbage n07714990 broccoli n07715103 cauliflower n07716358 zucchini, courgette n07716906 spaghetti squash n07717410 acorn squash n07717556 butternut squash n07718472 cucumber, cuke n07718747 artichoke, globe artichoke n07720875 bell pepper n07730033 cardoon n07734744 mushroom n07742313 Granny Smith n07745940 strawberry n07747607 orange n07749582 lemon n07753113 fig n07753275 pineapple, ananas n07753592 banana n07754684 jackfruit, jak, jack n07760859 custard apple n07768694 pomegranate n07802026 hay n07831146 carbonara n07836838 chocolate sauce, chocolate syrup n07860988 dough n07871810 meat loaf, meatloaf n07873807 pizza, pizza pie n07875152 potpie n07880968 burrito n07892512 red wine n07920052 espresso n07930864 cup n07932039 eggnog n09193705 alp n09229709 bubble n09246464 cliff, drop, drop-off n09256479 coral reef n09288635 geyser n09332890 lakeside, lakeshore n09399592 promontory, headland, head, foreland n09421951 sandbar, sand bar n09428293 seashore, coast, seacoast, sea-coast n09468604 valley, vale n09472597 volcano n09835506 ballplayer, baseball player n10148035 groom, bridegroom n10565667 scuba diver n11879895 rapeseed n11939491 daisy n12057211 yellow lady's slipper, yellow lady-slipper, Cypripedium calceolus, Cypripedium parviflorum n12144580 corn n12267677 acorn n12620546 hip, rose hip, rosehip n12768682 buckeye, horse chestnut, conker n12985857 coral fungus n12998815 agaric n13037406 gyromitra n13040303 stinkhorn, carrion fungus n13044778 earthstar n13052670 hen-of-the-woods, hen of the woods, Polyporus frondosus, Grifola frondosa n13054560 bolete n13133613 ear, spike, capitulum n15075141 toilet tissue, toilet paper, bathroom tissue ```
mrm8488
null
null
null
false
28
false
mrm8488/ImageNet1K-train
2022-04-28T11:06:11.000Z
null
false
006c4bc87abe217e728bbd7bfcd76f9f828c11e2
[]
[]
https://huggingface.co/datasets/mrm8488/ImageNet1K-train/resolve/main/README.md
mapping: ``` n01440764 tench, Tinca tinca n01443537 goldfish, Carassius auratus n01484850 great white shark, white shark, man-eater, man-eating shark, Carcharodon carcharias n01491361 tiger shark, Galeocerdo cuvieri n01494475 hammerhead, hammerhead shark n01496331 electric ray, crampfish, numbfish, torpedo n01498041 stingray n01514668 cock n01514859 hen n01518878 ostrich, Struthio camelus n01530575 brambling, Fringilla montifringilla n01531178 goldfinch, Carduelis carduelis n01532829 house finch, linnet, Carpodacus mexicanus n01534433 junco, snowbird n01537544 indigo bunting, indigo finch, indigo bird, Passerina cyanea n01558993 robin, American robin, Turdus migratorius n01560419 bulbul n01580077 jay n01582220 magpie n01592084 chickadee n01601694 water ouzel, dipper n01608432 kite n01614925 bald eagle, American eagle, Haliaeetus leucocephalus n01616318 vulture n01622779 great grey owl, great gray owl, Strix nebulosa n01629819 European fire salamander, Salamandra salamandra n01630670 common newt, Triturus vulgaris n01631663 eft n01632458 spotted salamander, Ambystoma maculatum n01632777 axolotl, mud puppy, Ambystoma mexicanum n01641577 bullfrog, Rana catesbeiana n01644373 tree frog, tree-frog n01644900 tailed frog, bell toad, ribbed toad, tailed toad, Ascaphus trui n01664065 loggerhead, loggerhead turtle, Caretta caretta n01665541 leatherback turtle, leatherback, leathery turtle, Dermochelys coriacea n01667114 mud turtle n01667778 terrapin n01669191 box turtle, box tortoise n01675722 banded gecko n01677366 common iguana, iguana, Iguana iguana n01682714 American chameleon, anole, Anolis carolinensis n01685808 whiptail, whiptail lizard n01687978 agama n01688243 frilled lizard, Chlamydosaurus kingi n01689811 alligator lizard n01692333 Gila monster, Heloderma suspectum n01693334 green lizard, Lacerta viridis n01694178 African chameleon, Chamaeleo chamaeleon n01695060 Komodo dragon, Komodo lizard, dragon lizard, giant lizard, Varanus komodoensis n01697457 African crocodile, Nile crocodile, Crocodylus niloticus n01698640 American alligator, Alligator mississipiensis n01704323 triceratops n01728572 thunder snake, worm snake, Carphophis amoenus n01728920 ringneck snake, ring-necked snake, ring snake n01729322 hognose snake, puff adder, sand viper n01729977 green snake, grass snake n01734418 king snake, kingsnake n01735189 garter snake, grass snake n01737021 water snake n01739381 vine snake n01740131 night snake, Hypsiglena torquata n01742172 boa constrictor, Constrictor constrictor n01744401 rock python, rock snake, Python sebae n01748264 Indian cobra, Naja naja n01749939 green mamba n01751748 sea snake n01753488 horned viper, cerastes, sand viper, horned asp, Cerastes cornutus n01755581 diamondback, diamondback rattlesnake, Crotalus adamanteus n01756291 sidewinder, horned rattlesnake, Crotalus cerastes n01768244 trilobite n01770081 harvestman, daddy longlegs, Phalangium opilio n01770393 scorpion n01773157 black and gold garden spider, Argiope aurantia n01773549 barn spider, Araneus cavaticus n01773797 garden spider, Aranea diademata n01774384 black widow, Latrodectus mactans n01774750 tarantula n01775062 wolf spider, hunting spider n01776313 tick n01784675 centipede n01795545 black grouse n01796340 ptarmigan n01797886 ruffed grouse, partridge, Bonasa umbellus n01798484 prairie chicken, prairie grouse, prairie fowl n01806143 peacock n01806567 quail n01807496 partridge n01817953 African grey, African gray, Psittacus erithacus n01818515 macaw n01819313 sulphur-crested cockatoo, Kakatoe galerita, Cacatua galerita n01820546 lorikeet n01824575 coucal n01828970 bee eater n01829413 hornbill n01833805 hummingbird n01843065 jacamar n01843383 toucan n01847000 drake n01855032 red-breasted merganser, Mergus serrator n01855672 goose n01860187 black swan, Cygnus atratus n01871265 tusker n01872401 echidna, spiny anteater, anteater n01873310 platypus, duckbill, duckbilled platypus, duck-billed platypus, Ornithorhynchus anatinus n01877812 wallaby, brush kangaroo n01882714 koala, koala bear, kangaroo bear, native bear, Phascolarctos cinereus n01883070 wombat n01910747 jellyfish n01914609 sea anemone, anemone n01917289 brain coral n01924916 flatworm, platyhelminth n01930112 nematode, nematode worm, roundworm n01943899 conch n01944390 snail n01945685 slug n01950731 sea slug, nudibranch n01955084 chiton, coat-of-mail shell, sea cradle, polyplacophore n01968897 chambered nautilus, pearly nautilus, nautilus n01978287 Dungeness crab, Cancer magister n01978455 rock crab, Cancer irroratus n01980166 fiddler crab n01981276 king crab, Alaska crab, Alaskan king crab, Alaska king crab, Paralithodes camtschatica n01983481 American lobster, Northern lobster, Maine lobster, Homarus americanus n01984695 spiny lobster, langouste, rock lobster, crawfish, crayfish, sea crawfish n01985128 crayfish, crawfish, crawdad, crawdaddy n01986214 hermit crab n01990800 isopod n02002556 white stork, Ciconia ciconia n02002724 black stork, Ciconia nigra n02006656 spoonbill n02007558 flamingo n02009229 little blue heron, Egretta caerulea n02009912 American egret, great white heron, Egretta albus n02011460 bittern n02012849 crane n02013706 limpkin, Aramus pictus n02017213 European gallinule, Porphyrio porphyrio n02018207 American coot, marsh hen, mud hen, water hen, Fulica americana n02018795 bustard n02025239 ruddy turnstone, Arenaria interpres n02027492 red-backed sandpiper, dunlin, Erolia alpina n02028035 redshank, Tringa totanus n02033041 dowitcher n02037110 oystercatcher, oyster catcher n02051845 pelican n02056570 king penguin, Aptenodytes patagonica n02058221 albatross, mollymawk n02066245 grey whale, gray whale, devilfish, Eschrichtius gibbosus, Eschrichtius robustus n02071294 killer whale, killer, orca, grampus, sea wolf, Orcinus orca n02074367 dugong, Dugong dugon n02077923 sea lion n02085620 Chihuahua n02085782 Japanese spaniel n02085936 Maltese dog, Maltese terrier, Maltese n02086079 Pekinese, Pekingese, Peke n02086240 Shih-Tzu n02086646 Blenheim spaniel n02086910 papillon n02087046 toy terrier n02087394 Rhodesian ridgeback n02088094 Afghan hound, Afghan n02088238 basset, basset hound n02088364 beagle n02088466 bloodhound, sleuthhound n02088632 bluetick n02089078 black-and-tan coonhound n02089867 Walker hound, Walker foxhound n02089973 English foxhound n02090379 redbone n02090622 borzoi, Russian wolfhound n02090721 Irish wolfhound n02091032 Italian greyhound n02091134 whippet n02091244 Ibizan hound, Ibizan Podenco n02091467 Norwegian elkhound, elkhound n02091635 otterhound, otter hound n02091831 Saluki, gazelle hound n02092002 Scottish deerhound, deerhound n02092339 Weimaraner n02093256 Staffordshire bullterrier, Staffordshire bull terrier n02093428 American Staffordshire terrier, Staffordshire terrier, American pit bull terrier, pit bull terrier n02093647 Bedlington terrier n02093754 Border terrier n02093859 Kerry blue terrier n02093991 Irish terrier n02094114 Norfolk terrier n02094258 Norwich terrier n02094433 Yorkshire terrier n02095314 wire-haired fox terrier n02095570 Lakeland terrier n02095889 Sealyham terrier, Sealyham n02096051 Airedale, Airedale terrier n02096177 cairn, cairn terrier n02096294 Australian terrier n02096437 Dandie Dinmont, Dandie Dinmont terrier n02096585 Boston bull, Boston terrier n02097047 miniature schnauzer n02097130 giant schnauzer n02097209 standard schnauzer n02097298 Scotch terrier, Scottish terrier, Scottie n02097474 Tibetan terrier, chrysanthemum dog n02097658 silky terrier, Sydney silky n02098105 soft-coated wheaten terrier n02098286 West Highland white terrier n02098413 Lhasa, Lhasa apso n02099267 flat-coated retriever n02099429 curly-coated retriever n02099601 golden retriever n02099712 Labrador retriever n02099849 Chesapeake Bay retriever n02100236 German short-haired pointer n02100583 vizsla, Hungarian pointer n02100735 English setter n02100877 Irish setter, red setter n02101006 Gordon setter n02101388 Brittany spaniel n02101556 clumber, clumber spaniel n02102040 English springer, English springer spaniel n02102177 Welsh springer spaniel n02102318 cocker spaniel, English cocker spaniel, cocker n02102480 Sussex spaniel n02102973 Irish water spaniel n02104029 kuvasz n02104365 schipperke n02105056 groenendael n02105162 malinois n02105251 briard n02105412 kelpie n02105505 komondor n02105641 Old English sheepdog, bobtail n02105855 Shetland sheepdog, Shetland sheep dog, Shetland n02106030 collie n02106166 Border collie n02106382 Bouvier des Flandres, Bouviers des Flandres n02106550 Rottweiler n02106662 German shepherd, German shepherd dog, German police dog, alsatian n02107142 Doberman, Doberman pinscher n02107312 miniature pinscher n02107574 Greater Swiss Mountain dog n02107683 Bernese mountain dog n02107908 Appenzeller n02108000 EntleBucher n02108089 boxer n02108422 bull mastiff n02108551 Tibetan mastiff n02108915 French bulldog n02109047 Great Dane n02109525 Saint Bernard, St Bernard n02109961 Eskimo dog, husky n02110063 malamute, malemute, Alaskan malamute n02110185 Siberian husky n02110341 dalmatian, coach dog, carriage dog n02110627 affenpinscher, monkey pinscher, monkey dog n02110806 basenji n02110958 pug, pug-dog n02111129 Leonberg n02111277 Newfoundland, Newfoundland dog n02111500 Great Pyrenees n02111889 Samoyed, Samoyede n02112018 Pomeranian n02112137 chow, chow chow n02112350 keeshond n02112706 Brabancon griffon n02113023 Pembroke, Pembroke Welsh corgi n02113186 Cardigan, Cardigan Welsh corgi n02113624 toy poodle n02113712 miniature poodle n02113799 standard poodle n02113978 Mexican hairless n02114367 timber wolf, grey wolf, gray wolf, Canis lupus n02114548 white wolf, Arctic wolf, Canis lupus tundrarum n02114712 red wolf, maned wolf, Canis rufus, Canis niger n02114855 coyote, prairie wolf, brush wolf, Canis latrans n02115641 dingo, warrigal, warragal, Canis dingo n02115913 dhole, Cuon alpinus n02116738 African hunting dog, hyena dog, Cape hunting dog, Lycaon pictus n02117135 hyena, hyaena n02119022 red fox, Vulpes vulpes n02119789 kit fox, Vulpes macrotis n02120079 Arctic fox, white fox, Alopex lagopus n02120505 grey fox, gray fox, Urocyon cinereoargenteus n02123045 tabby, tabby cat n02123159 tiger cat n02123394 Persian cat n02123597 Siamese cat, Siamese n02124075 Egyptian cat n02125311 cougar, puma, catamount, mountain lion, painter, panther, Felis concolor n02127052 lynx, catamount n02128385 leopard, Panthera pardus n02128757 snow leopard, ounce, Panthera uncia n02128925 jaguar, panther, Panthera onca, Felis onca n02129165 lion, king of beasts, Panthera leo n02129604 tiger, Panthera tigris n02130308 cheetah, chetah, Acinonyx jubatus n02132136 brown bear, bruin, Ursus arctos n02133161 American black bear, black bear, Ursus americanus, Euarctos americanus n02134084 ice bear, polar bear, Ursus Maritimus, Thalarctos maritimus n02134418 sloth bear, Melursus ursinus, Ursus ursinus n02137549 mongoose n02138441 meerkat, mierkat n02165105 tiger beetle n02165456 ladybug, ladybeetle, lady beetle, ladybird, ladybird beetle n02167151 ground beetle, carabid beetle n02168699 long-horned beetle, longicorn, longicorn beetle n02169497 leaf beetle, chrysomelid n02172182 dung beetle n02174001 rhinoceros beetle n02177972 weevil n02190166 fly n02206856 bee n02219486 ant, emmet, pismire n02226429 grasshopper, hopper n02229544 cricket n02231487 walking stick, walkingstick, stick insect n02233338 cockroach, roach n02236044 mantis, mantid n02256656 cicada, cicala n02259212 leafhopper n02264363 lacewing, lacewing fly n02268443 dragonfly, darning needle, devil's darning needle, sewing needle, snake feeder, snake doctor, mosquito hawk, skeeter hawk n02268853 damselfly n02276258 admiral n02277742 ringlet, ringlet butterfly n02279972 monarch, monarch butterfly, milkweed butterfly, Danaus plexippus n02280649 cabbage butterfly n02281406 sulphur butterfly, sulfur butterfly n02281787 lycaenid, lycaenid butterfly n02317335 starfish, sea star n02319095 sea urchin n02321529 sea cucumber, holothurian n02325366 wood rabbit, cottontail, cottontail rabbit n02326432 hare n02328150 Angora, Angora rabbit n02342885 hamster n02346627 porcupine, hedgehog n02356798 fox squirrel, eastern fox squirrel, Sciurus niger n02361337 marmot n02363005 beaver n02364673 guinea pig, Cavia cobaya n02389026 sorrel n02391049 zebra n02395406 hog, pig, grunter, squealer, Sus scrofa n02396427 wild boar, boar, Sus scrofa n02397096 warthog n02398521 hippopotamus, hippo, river horse, Hippopotamus amphibius n02403003 ox n02408429 water buffalo, water ox, Asiatic buffalo, Bubalus bubalis n02410509 bison n02412080 ram, tup n02415577 bighorn, bighorn sheep, cimarron, Rocky Mountain bighorn, Rocky Mountain sheep, Ovis canadensis n02417914 ibex, Capra ibex n02422106 hartebeest n02422699 impala, Aepyceros melampus n02423022 gazelle n02437312 Arabian camel, dromedary, Camelus dromedarius n02437616 llama n02441942 weasel n02442845 mink n02443114 polecat, fitch, foulmart, foumart, Mustela putorius n02443484 black-footed ferret, ferret, Mustela nigripes n02444819 otter n02445715 skunk, polecat, wood pussy n02447366 badger n02454379 armadillo n02457408 three-toed sloth, ai, Bradypus tridactylus n02480495 orangutan, orang, orangutang, Pongo pygmaeus n02480855 gorilla, Gorilla gorilla n02481823 chimpanzee, chimp, Pan troglodytes n02483362 gibbon, Hylobates lar n02483708 siamang, Hylobates syndactylus, Symphalangus syndactylus n02484975 guenon, guenon monkey n02486261 patas, hussar monkey, Erythrocebus patas n02486410 baboon n02487347 macaque n02488291 langur n02488702 colobus, colobus monkey n02489166 proboscis monkey, Nasalis larvatus n02490219 marmoset n02492035 capuchin, ringtail, Cebus capucinus n02492660 howler monkey, howler n02493509 titi, titi monkey n02493793 spider monkey, Ateles geoffroyi n02494079 squirrel monkey, Saimiri sciureus n02497673 Madagascar cat, ring-tailed lemur, Lemur catta n02500267 indri, indris, Indri indri, Indri brevicaudatus n02504013 Indian elephant, Elephas maximus n02504458 African elephant, Loxodonta africana n02509815 lesser panda, red panda, panda, bear cat, cat bear, Ailurus fulgens n02510455 giant panda, panda, panda bear, coon bear, Ailuropoda melanoleuca n02514041 barracouta, snoek n02526121 eel n02536864 coho, cohoe, coho salmon, blue jack, silver salmon, Oncorhynchus kisutch n02606052 rock beauty, Holocanthus tricolor n02607072 anemone fish n02640242 sturgeon n02641379 gar, garfish, garpike, billfish, Lepisosteus osseus n02643566 lionfish n02655020 puffer, pufferfish, blowfish, globefish n02666196 abacus n02667093 abaya n02669723 academic gown, academic robe, judge's robe n02672831 accordion, piano accordion, squeeze box n02676566 acoustic guitar n02687172 aircraft carrier, carrier, flattop, attack aircraft carrier n02690373 airliner n02692877 airship, dirigible n02699494 altar n02701002 ambulance n02704792 amphibian, amphibious vehicle n02708093 analog clock n02727426 apiary, bee house n02730930 apron n02747177 ashcan, trash can, garbage can, wastebin, ash bin, ash-bin, ashbin, dustbin, trash barrel, trash bin n02749479 assault rifle, assault gun n02769748 backpack, back pack, knapsack, packsack, rucksack, haversack n02776631 bakery, bakeshop, bakehouse n02777292 balance beam, beam n02782093 balloon n02783161 ballpoint, ballpoint pen, ballpen, Biro n02786058 Band Aid n02787622 banjo n02788148 bannister, banister, balustrade, balusters, handrail n02790996 barbell n02791124 barber chair n02791270 barbershop n02793495 barn n02794156 barometer n02795169 barrel, cask n02797295 barrow, garden cart, lawn cart, wheelbarrow n02799071 baseball n02802426 basketball n02804414 bassinet n02804610 bassoon n02807133 bathing cap, swimming cap n02808304 bath towel n02808440 bathtub, bathing tub, bath, tub n02814533 beach wagon, station wagon, wagon, estate car, beach waggon, station waggon, waggon n02814860 beacon, lighthouse, beacon light, pharos n02815834 beaker n02817516 bearskin, busby, shako n02823428 beer bottle n02823750 beer glass n02825657 bell cote, bell cot n02834397 bib n02835271 bicycle-built-for-two, tandem bicycle, tandem n02837789 bikini, two-piece n02840245 binder, ring-binder n02841315 binoculars, field glasses, opera glasses n02843684 birdhouse n02859443 boathouse n02860847 bobsled, bobsleigh, bob n02865351 bolo tie, bolo, bola tie, bola n02869837 bonnet, poke bonnet n02870880 bookcase n02871525 bookshop, bookstore, bookstall n02877765 bottlecap n02879718 bow n02883205 bow tie, bow-tie, bowtie n02892201 brass, memorial tablet, plaque n02892767 brassiere, bra, bandeau n02894605 breakwater, groin, groyne, mole, bulwark, seawall, jetty n02895154 breastplate, aegis, egis n02906734 broom n02909870 bucket, pail n02910353 buckle n02916936 bulletproof vest n02917067 bullet train, bullet n02927161 butcher shop, meat market n02930766 cab, hack, taxi, taxicab n02939185 caldron, cauldron n02948072 candle, taper, wax light n02950826 cannon n02951358 canoe n02951585 can opener, tin opener n02963159 cardigan n02965783 car mirror n02966193 carousel, carrousel, merry-go-round, roundabout, whirligig n02966687 carpenter's kit, tool kit n02971356 carton n02974003 car wheel n02977058 cash machine, cash dispenser, automated teller machine, automatic teller machine, automated teller, automatic teller, ATM n02978881 cassette n02979186 cassette player n02980441 castle n02981792 catamaran n02988304 CD player n02992211 cello, violoncello n02992529 cellular telephone, cellular phone, cellphone, cell, mobile phone n02999410 chain n03000134 chainlink fence n03000247 chain mail, ring mail, mail, chain armor, chain armour, ring armor, ring armour n03000684 chain saw, chainsaw n03014705 chest n03016953 chiffonier, commode n03017168 chime, bell, gong n03018349 china cabinet, china closet n03026506 Christmas stocking n03028079 church, church building n03032252 cinema, movie theater, movie theatre, movie house, picture palace n03041632 cleaver, meat cleaver, chopper n03042490 cliff dwelling n03045698 cloak n03047690 clog, geta, patten, sabot n03062245 cocktail shaker n03063599 coffee mug n03063689 coffeepot n03065424 coil, spiral, volute, whorl, helix n03075370 combination lock n03085013 computer keyboard, keypad n03089624 confectionery, confectionary, candy store n03095699 container ship, containership, container vessel n03100240 convertible n03109150 corkscrew, bottle screw n03110669 cornet, horn, trumpet, trump n03124043 cowboy boot n03124170 cowboy hat, ten-gallon hat n03125729 cradle n03126707 crane n03127747 crash helmet n03127925 crate n03131574 crib, cot n03133878 Crock Pot n03134739 croquet ball n03141823 crutch n03146219 cuirass n03160309 dam, dike, dyke n03179701 desk n03180011 desktop computer n03187595 dial telephone, dial phone n03188531 diaper, nappy, napkin n03196217 digital clock n03197337 digital watch n03201208 dining table, board n03207743 dishrag, dishcloth n03207941 dishwasher, dish washer, dishwashing machine n03208938 disk brake, disc brake n03216828 dock, dockage, docking facility n03218198 dogsled, dog sled, dog sleigh n03220513 dome n03223299 doormat, welcome mat n03240683 drilling platform, offshore rig n03249569 drum, membranophone, tympan n03250847 drumstick n03255030 dumbbell n03259280 Dutch oven n03271574 electric fan, blower n03272010 electric guitar n03272562 electric locomotive n03290653 entertainment center n03291819 envelope n03297495 espresso maker n03314780 face powder n03325584 feather boa, boa n03337140 file, file cabinet, filing cabinet n03344393 fireboat n03345487 fire engine, fire truck n03347037 fire screen, fireguard n03355925 flagpole, flagstaff n03372029 flute, transverse flute n03376595 folding chair n03379051 football helmet n03384352 forklift n03388043 fountain n03388183 fountain pen n03388549 four-poster n03393912 freight car n03394916 French horn, horn n03400231 frying pan, frypan, skillet n03404251 fur coat n03417042 garbage truck, dustcart n03424325 gasmask, respirator, gas helmet n03425413 gas pump, gasoline pump, petrol pump, island dispenser n03443371 goblet n03444034 go-kart n03445777 golf ball n03445924 golfcart, golf cart n03447447 gondola n03447721 gong, tam-tam n03450230 gown n03452741 grand piano, grand n03457902 greenhouse, nursery, glasshouse n03459775 grille, radiator grille n03461385 grocery store, grocery, food market, market n03467068 guillotine n03476684 hair slide n03476991 hair spray n03478589 half track n03481172 hammer n03482405 hamper n03483316 hand blower, blow dryer, blow drier, hair dryer, hair drier n03485407 hand-held computer, hand-held microcomputer n03485794 handkerchief, hankie, hanky, hankey n03492542 hard disc, hard disk, fixed disk n03494278 harmonica, mouth organ, harp, mouth harp n03495258 harp n03496892 harvester, reaper n03498962 hatchet n03527444 holster n03529860 home theater, home theatre n03530642 honeycomb n03532672 hook, claw n03534580 hoopskirt, crinoline n03535780 horizontal bar, high bar n03538406 horse cart, horse-cart n03544143 hourglass n03584254 iPod n03584829 iron, smoothing iron n03590841 jack-o'-lantern n03594734 jean, blue jean, denim n03594945 jeep, landrover n03595614 jersey, T-shirt, tee shirt n03598930 jigsaw puzzle n03599486 jinrikisha, ricksha, rickshaw n03602883 joystick n03617480 kimono n03623198 knee pad n03627232 knot n03630383 lab coat, laboratory coat n03633091 ladle n03637318 lampshade, lamp shade n03642806 laptop, laptop computer n03649909 lawn mower, mower n03657121 lens cap, lens cover n03658185 letter opener, paper knife, paperknife n03661043 library n03662601 lifeboat n03666591 lighter, light, igniter, ignitor n03670208 limousine, limo n03673027 liner, ocean liner n03676483 lipstick, lip rouge n03680355 Loafer n03690938 lotion n03691459 loudspeaker, speaker, speaker unit, loudspeaker system, speaker system n03692522 loupe, jeweler's loupe n03697007 lumbermill, sawmill n03706229 magnetic compass n03709823 mailbag, postbag n03710193 mailbox, letter box n03710637 maillot n03710721 maillot, tank suit n03717622 manhole cover n03720891 maraca n03721384 marimba, xylophone n03724870 mask n03729826 matchstick n03733131 maypole n03733281 maze, labyrinth n03733805 measuring cup n03742115 medicine chest, medicine cabinet n03743016 megalith, megalithic structure n03759954 microphone, mike n03761084 microwave, microwave oven n03763968 military uniform n03764736 milk can n03769881 minibus n03770439 miniskirt, mini n03770679 minivan n03773504 missile n03775071 mitten n03775546 mixing bowl n03776460 mobile home, manufactured home n03777568 Model T n03777754 modem n03781244 monastery n03782006 monitor n03785016 moped n03786901 mortar n03787032 mortarboard n03788195 mosque n03788365 mosquito net n03791053 motor scooter, scooter n03792782 mountain bike, all-terrain bike, off-roader n03792972 mountain tent n03793489 mouse, computer mouse n03794056 mousetrap n03796401 moving van n03803284 muzzle n03804744 nail n03814639 neck brace n03814906 necklace n03825788 nipple n03832673 notebook, notebook computer n03837869 obelisk n03838899 oboe, hautboy, hautbois n03840681 ocarina, sweet potato n03841143 odometer, hodometer, mileometer, milometer n03843555 oil filter n03854065 organ, pipe organ n03857828 oscilloscope, scope, cathode-ray oscilloscope, CRO n03866082 overskirt n03868242 oxcart n03868863 oxygen mask n03871628 packet n03873416 paddle, boat paddle n03874293 paddlewheel, paddle wheel n03874599 padlock n03876231 paintbrush n03877472 pajama, pyjama, pj's, jammies n03877845 palace n03884397 panpipe, pandean pipe, syrinx n03887697 paper towel n03888257 parachute, chute n03888605 parallel bars, bars n03891251 park bench n03891332 parking meter n03895866 passenger car, coach, carriage n03899768 patio, terrace n03902125 pay-phone, pay-station n03903868 pedestal, plinth, footstall n03908618 pencil box, pencil case n03908714 pencil sharpener n03916031 perfume, essence n03920288 Petri dish n03924679 photocopier n03929660 pick, plectrum, plectron n03929855 pickelhaube n03930313 picket fence, paling n03930630 pickup, pickup truck n03933933 pier n03935335 piggy bank, penny bank n03937543 pill bottle n03938244 pillow n03942813 ping-pong ball n03944341 pinwheel n03947888 pirate, pirate ship n03950228 pitcher, ewer n03954731 plane, carpenter's plane, woodworking plane n03956157 planetarium n03958227 plastic bag n03961711 plate rack n03967562 plow, plough n03970156 plunger, plumber's helper n03976467 Polaroid camera, Polaroid Land camera n03976657 pole n03977966 police van, police wagon, paddy wagon, patrol wagon, wagon, black Maria n03980874 poncho n03982430 pool table, billiard table, snooker table n03983396 pop bottle, soda bottle n03991062 pot, flowerpot n03992509 potter's wheel n03995372 power drill n03998194 prayer rug, prayer mat n04004767 printer n04005630 prison, prison house n04008634 projectile, missile n04009552 projector n04019541 puck, hockey puck n04023962 punching bag, punch bag, punching ball, punchball n04026417 purse n04033901 quill, quill pen n04033995 quilt, comforter, comfort, puff n04037443 racer, race car, racing car n04039381 racket, racquet n04040759 radiator n04041544 radio, wireless n04044716 radio telescope, radio reflector n04049303 rain barrel n04065272 recreational vehicle, RV, R.V. n04067472 reel n04069434 reflex camera n04070727 refrigerator, icebox n04074963 remote control, remote n04081281 restaurant, eating house, eating place, eatery n04086273 revolver, six-gun, six-shooter n04090263 rifle n04099969 rocking chair, rocker n04111531 rotisserie n04116512 rubber eraser, rubber, pencil eraser n04118538 rugby ball n04118776 rule, ruler n04120489 running shoe n04125021 safe n04127249 safety pin n04131690 saltshaker, salt shaker n04133789 sandal n04136333 sarong n04141076 sax, saxophone n04141327 scabbard n04141975 scale, weighing machine n04146614 school bus n04147183 schooner n04149813 scoreboard n04152593 screen, CRT screen n04153751 screw n04154565 screwdriver n04162706 seat belt, seatbelt n04179913 sewing machine n04192698 shield, buckler n04200800 shoe shop, shoe-shop, shoe store n04201297 shoji n04204238 shopping basket n04204347 shopping cart n04208210 shovel n04209133 shower cap n04209239 shower curtain n04228054 ski n04229816 ski mask n04235860 sleeping bag n04238763 slide rule, slipstick n04239074 sliding door n04243546 slot, one-armed bandit n04251144 snorkel n04252077 snowmobile n04252225 snowplow, snowplough n04254120 soap dispenser n04254680 soccer ball n04254777 sock n04258138 solar dish, solar collector, solar furnace n04259630 sombrero n04263257 soup bowl n04264628 space bar n04265275 space heater n04266014 space shuttle n04270147 spatula n04273569 speedboat n04275548 spider web, spider's web n04277352 spindle n04285008 sports car, sport car n04286575 spotlight, spot n04296562 stage n04310018 steam locomotive n04311004 steel arch bridge n04311174 steel drum n04317175 stethoscope n04325704 stole n04326547 stone wall n04328186 stopwatch, stop watch n04330267 stove n04332243 strainer n04335435 streetcar, tram, tramcar, trolley, trolley car n04336792 stretcher n04344873 studio couch, day bed n04346328 stupa, tope n04347754 submarine, pigboat, sub, U-boat n04350905 suit, suit of clothes n04355338 sundial n04355933 sunglass n04356056 sunglasses, dark glasses, shades n04357314 sunscreen, sunblock, sun blocker n04366367 suspension bridge n04367480 swab, swob, mop n04370456 sweatshirt n04371430 swimming trunks, bathing trunks n04371774 swing n04372370 switch, electric switch, electrical switch n04376876 syringe n04380533 table lamp n04389033 tank, army tank, armored combat vehicle, armoured combat vehicle n04392985 tape player n04398044 teapot n04399382 teddy, teddy bear n04404412 television, television system n04409515 tennis ball n04417672 thatch, thatched roof n04418357 theater curtain, theatre curtain n04423845 thimble n04428191 thresher, thrasher, threshing machine n04429376 throne n04435653 tile roof n04442312 toaster n04443257 tobacco shop, tobacconist shop, tobacconist n04447861 toilet seat n04456115 torch n04458633 totem pole n04461696 tow truck, tow car, wrecker n04462240 toyshop n04465501 tractor n04467665 trailer truck, tractor trailer, trucking rig, rig, articulated lorry, semi n04476259 tray n04479046 trench coat n04482393 tricycle, trike, velocipede n04483307 trimaran n04485082 tripod n04486054 triumphal arch n04487081 trolleybus, trolley coach, trackless trolley n04487394 trombone n04493381 tub, vat n04501370 turnstile n04505470 typewriter keyboard n04507155 umbrella n04509417 unicycle, monocycle n04515003 upright, upright piano n04517823 vacuum, vacuum cleaner n04522168 vase n04523525 vault n04525038 velvet n04525305 vending machine n04532106 vestment n04532670 viaduct n04536866 violin, fiddle n04540053 volleyball n04542943 waffle iron n04548280 wall clock n04548362 wallet, billfold, notecase, pocketbook n04550184 wardrobe, closet, press n04552348 warplane, military plane n04553703 washbasin, handbasin, washbowl, lavabo, wash-hand basin n04554684 washer, automatic washer, washing machine n04557648 water bottle n04560804 water jug n04562935 water tower n04579145 whiskey jug n04579432 whistle n04584207 wig n04589890 window screen n04590129 window shade n04591157 Windsor tie n04591713 wine bottle n04592741 wing n04596742 wok n04597913 wooden spoon n04599235 wool, woolen, woollen n04604644 worm fence, snake fence, snake-rail fence, Virginia fence n04606251 wreck n04612504 yawl n04613696 yurt n06359193 web site, website, internet site, site n06596364 comic book n06785654 crossword puzzle, crossword n06794110 street sign n06874185 traffic light, traffic signal, stoplight n07248320 book jacket, dust cover, dust jacket, dust wrapper n07565083 menu n07579787 plate n07583066 guacamole n07584110 consomme n07590611 hot pot, hotpot n07613480 trifle n07614500 ice cream, icecream n07615774 ice lolly, lolly, lollipop, popsicle n07684084 French loaf n07693725 bagel, beigel n07695742 pretzel n07697313 cheeseburger n07697537 hotdog, hot dog, red hot n07711569 mashed potato n07714571 head cabbage n07714990 broccoli n07715103 cauliflower n07716358 zucchini, courgette n07716906 spaghetti squash n07717410 acorn squash n07717556 butternut squash n07718472 cucumber, cuke n07718747 artichoke, globe artichoke n07720875 bell pepper n07730033 cardoon n07734744 mushroom n07742313 Granny Smith n07745940 strawberry n07747607 orange n07749582 lemon n07753113 fig n07753275 pineapple, ananas n07753592 banana n07754684 jackfruit, jak, jack n07760859 custard apple n07768694 pomegranate n07802026 hay n07831146 carbonara n07836838 chocolate sauce, chocolate syrup n07860988 dough n07871810 meat loaf, meatloaf n07873807 pizza, pizza pie n07875152 potpie n07880968 burrito n07892512 red wine n07920052 espresso n07930864 cup n07932039 eggnog n09193705 alp n09229709 bubble n09246464 cliff, drop, drop-off n09256479 coral reef n09288635 geyser n09332890 lakeside, lakeshore n09399592 promontory, headland, head, foreland n09421951 sandbar, sand bar n09428293 seashore, coast, seacoast, sea-coast n09468604 valley, vale n09472597 volcano n09835506 ballplayer, baseball player n10148035 groom, bridegroom n10565667 scuba diver n11879895 rapeseed n11939491 daisy n12057211 yellow lady's slipper, yellow lady-slipper, Cypripedium calceolus, Cypripedium parviflorum n12144580 corn n12267677 acorn n12620546 hip, rose hip, rosehip n12768682 buckeye, horse chestnut, conker n12985857 coral fungus n12998815 agaric n13037406 gyromitra n13040303 stinkhorn, carrion fungus n13044778 earthstar n13052670 hen-of-the-woods, hen of the woods, Polyporus frondosus, Grifola frondosa n13054560 bolete n13133613 ear, spike, capitulum n15075141 toilet tissue, toilet paper, bathroom tissue ```
AmazonScience
null
@misc{fitzgerald2022massive, title={MASSIVE: A 1M-Example Multilingual Natural Language Understanding Dataset with 51 Typologically-Diverse Languages}, author={Jack FitzGerald and Christopher Hench and Charith Peris and Scott Mackie and Kay Rottmann and Ana Sanchez and Aaron Nash and Liam Urbach and Vishesh Kakarala and Richa Singh and Swetha Ranganath and Laurie Crist and Misha Britan and Wouter Leeuwis and Gokhan Tur and Prem Natarajan}, year={2022}, eprint={2204.08582}, archivePrefix={arXiv}, primaryClass={cs.CL} } @inproceedings{bastianelli-etal-2020-slurp, title = "{SLURP}: A Spoken Language Understanding Resource Package", author = "Bastianelli, Emanuele and Vanzo, Andrea and Swietojanski, Pawel and Rieser, Verena", booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)", month = nov, year = "2020", address = "Online", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2020.emnlp-main.588", doi = "10.18653/v1/2020.emnlp-main.588", pages = "7252--7262", abstract = "Spoken Language Understanding infers semantic meaning directly from audio data, and thus promises to reduce error propagation and misunderstandings in end-user applications. However, publicly available SLU resources are limited. In this paper, we release SLURP, a new SLU package containing the following: (1) A new challenging dataset in English spanning 18 domains, which is substantially bigger and linguistically more diverse than existing datasets; (2) Competitive baselines based on state-of-the-art NLU and ASR systems; (3) A new transparent metric for entity labelling which enables a detailed error analysis for identifying potential areas of improvement. SLURP is available at https://github.com/pswietojanski/slurp." }
MASSIVE is a parallel dataset of > 1M utterances across 51 languages with annotations for the Natural Language Understanding tasks of intent prediction and slot annotation. Utterances span 60 intents and include 55 slot types. MASSIVE was created by localizing the SLURP dataset, composed of general Intelligent Voice Assistant single-shot interactions.
false
2,014
false
AmazonScience/massive
2022-11-16T15:44:51.000Z
massive
false
ff6bd8e4b27c3543e4f8fe2108f32bb95a6f8740
[]
[ "arxiv:2204.08582", "annotations_creators:expert-generated", "language_creators:found", "license:cc-by-4.0", "multilinguality:af-ZA", "multilinguality:am-ET", "multilinguality:ar-SA", "multilinguality:az-AZ", "multilinguality:bn-BD", "multilinguality:ca-ES", "multilinguality:cy-GB", "multiling...
https://huggingface.co/datasets/AmazonScience/massive/resolve/main/README.md
--- annotations_creators: - expert-generated language_creators: - found license: - cc-by-4.0 multilinguality: - af-ZA - am-ET - ar-SA - az-AZ - bn-BD - ca-ES - cy-GB - da-DK - de-DE - el-GR - en-US - es-ES - fa-IR - fi-FI - fr-FR - he-IL - hi-IN - hu-HU - hy-AM - id-ID - is-IS - it-IT - ja-JP - jv-ID - ka-GE - km-KH - kn-IN - ko-KR - lv-LV - ml-IN - mn-MN - ms-MY - my-MM - nb-NO - nl-NL - pl-PL - pt-PT - ro-RO - ru-RU - sl-SL - sq-AL - sv-SE - sw-KE - ta-IN - te-IN - th-TH - tl-PH - tr-TR - ur-PK - vi-VN - zh-CN - zh-TW size_categories: - 100K<n<1M source_datasets: - original task_categories: - text-classification task_ids: - intent-classification - multi-class-classification paperswithcode_id: massive pretty_name: MASSIVE language_bcp47: - af-ZA - am-ET - ar-SA - az-AZ - bn-BD - ca-ES - cy-GB - da-DK - de-DE - el-GR - en-US - es-ES - fa-IR - fi-FI - fr-FR - he-IL - hi-IN - hu-HU - hy-AM - id-ID - is-IS - it-IT - ja-JP - jv-ID - ka-GE - km-KH - kn-IN - ko-KR - lv-LV - ml-IN - mn-MN - ms-MY - my-MM - nb-NO - nl-NL - pl-PL - pt-PT - ro-RO - ru-RU - sl-SL - sq-AL - sv-SE - sw-KE - ta-IN - te-IN - th-TH - tl-PH - tr-TR - ur-PK - vi-VN - zh-CN - zh-TW tags: - natural-language-understanding --- # MASSIVE 1.1: A 1M-Example Multilingual Natural Language Understanding Dataset with 52 Typologically-Diverse Languages ## Table of Contents - [Dataset Card for [Needs More Information]](#dataset-card-for-needs-more-information) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Initial Data Collection and Normalization](#initial-data-collection-and-normalization) - [Who are the source language producers?](#who-are-the-source-language-producers) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [No Warranty](#no-warranty) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://github.com/alexa/massive - **Repository:** https://github.com/alexa/massive - **Paper:** https://arxiv.org/abs/2204.08582 - **Leaderboard:** https://eval.ai/web/challenges/challenge-page/1697/overview - **Point of Contact:** [GitHub](https://github.com/alexa/massive/issues) ### Dataset Summary MASSIVE 1.1 is a parallel dataset of > 1M utterances across 52 languages with annotations for the Natural Language Understanding tasks of intent prediction and slot annotation. Utterances span 60 intents and include 55 slot types. MASSIVE was created by localizing the SLURP dataset, composed of general Intelligent Voice Assistant single-shot interactions. | Name | Lang | Utt/Lang | Domains | Intents | Slots | |:-------------------------------------------------------------------------------:|:-------:|:--------------:|:-------:|:--------:|:------:| | MASSIVE 1.1 | 52 | 19,521 | 18 | 60 | 55 | | SLURP (Bastianelli et al., 2020) | 1 | 16,521 | 18 | 60 | 55 | | NLU Evaluation Data (Liu et al., 2019) | 1 | 25,716 | 18 | 54 | 56 | | Airline Travel Information System (ATIS) (Price, 1990) | 1 | 5,871 | 1 | 26 | 129 | | ATIS with Hindi and Turkish (Upadhyay et al., 2018) | 3 | 1,315-5,871 | 1 | 26 | 129 | | MultiATIS++ (Xu et al., 2020) | 9 | 1,422-5,897 | 1 | 21-26 | 99-140 | | Snips (Coucke et al., 2018) | 1 | 14,484 | - | 7 | 53 | | Snips with French (Saade et al., 2019) | 2 | 4,818 | 2 | 14-15 | 11-12 | | Task Oriented Parsing (TOP) (Gupta et al., 2018) | 1 | 44,873 | 2 | 25 | 36 | | Multilingual Task-Oriented Semantic Parsing (MTOP) (Li et al., 2021) | 6 | 15,195-22,288 | 11 | 104-113 | 72-75 | | Cross-Lingual Multilingual Task Oriented Dialog (Schuster et al., 2019) | 3 | 5,083-43,323 | 3 | 12 | 11 | | Microsoft Dialog Challenge (Li et al., 2018) | 1 | 38,276 | 3 | 11 | 29 | | Fluent Speech Commands (FSC) (Lugosch et al., 2019) | 1 | 30,043 | - | 31 | - | | Chinese Audio-Textual Spoken Language Understanding (CATSLU) (Zhu et al., 2019) | 1 | 16,258 | 4 | - | 94 | ### Supported Tasks and Leaderboards The dataset can be used to train a model for `natural-language-understanding` (NLU) : - `intent-classification` - `multi-class-classification` - `natural-language-understanding` ### Languages The MASSIVE 1.1 corpora consists of parallel sentences from 52 languages : - `Afrikaans - South Africa (af-ZA)` - `Amharic - Ethiopia (am-ET)` - `Arabic - Saudi Arabia (ar-SA)` - `Azeri - Azerbaijan (az-AZ)` - `Bengali - Bangladesh (bn-BD)` - `Catalan - Spain (ca-ES)` - `Chinese - China (zh-CN)` - `Chinese - Taiwan (zh-TW)` - `Danish - Denmark (da-DK)` - `German - Germany (de-DE)` - `Greek - Greece (el-GR)` - `English - United States (en-US)` - `Spanish - Spain (es-ES)` - `Farsi - Iran (fa-IR)` - `Finnish - Finland (fi-FI)` - `French - France (fr-FR)` - `Hebrew - Israel (he-IL)` - `Hungarian - Hungary (hu-HU)` - `Armenian - Armenia (hy-AM)` - `Indonesian - Indonesia (id-ID)` - `Icelandic - Iceland (is-IS)` - `Italian - Italy (it-IT)` - `Japanese - Japan (ja-JP)` - `Javanese - Indonesia (jv-ID)` - `Georgian - Georgia (ka-GE)` - `Khmer - Cambodia (km-KH)` - `Korean - Korea (ko-KR)` - `Latvian - Latvia (lv-LV)` - `Mongolian - Mongolia (mn-MN)` - `Malay - Malaysia (ms-MY)` - `Burmese - Myanmar (my-MM)` - `Norwegian - Norway (nb-NO)` - `Dutch - Netherlands (nl-NL)` - `Polish - Poland (pl-PL)` - `Portuguese - Portugal (pt-PT)` - `Romanian - Romania (ro-RO)` - `Russian - Russia (ru-RU)` - `Slovanian - Slovania (sl-SL)` - `Albanian - Albania (sq-AL)` - `Swedish - Sweden (sv-SE)` - `Swahili - Kenya (sw-KE)` - `Hindi - India (hi-IN)` - `Kannada - India (kn-IN)` - `Malayalam - India (ml-IN)` - `Tamil - India (ta-IN)` - `Telugu - India (te-IN)` - `Thai - Thailand (th-TH)` - `Tagalog - Philippines (tl-PH)` - `Turkish - Turkey (tr-TR)` - `Urdu - Pakistan (ur-PK)` - `Vietnamese - Vietnam (vi-VN)` - `Welsh - United Kingdom (cy-GB)` ## Load the dataset with HuggingFace ```python from datasets import load_dataset dataset = load_dataset("AmazonScience/massive", "en-US", split='train') print(dataset[0]) ``` ## Dataset Structure ### Data Instances ```json { "id": "0", "locale": "fr-FR", "partition": "test", "scenario": "alarm", "intent": "alarm_set", "utt": "réveille-moi à cinq heures du matin cette semaine", "annot_utt": "réveille-moi à [time : cinq heures du matin] [date : cette semaine]", "worker_id": "22", "slot_method": [ { "slot": "time", "method": "translation" }, { "slot": "date", "method": "translation" } ], "judgments": [ { "worker_id": "22", "intent_score": 1, "slots_score": 1, "grammar_score": 4, "spelling_score": 2, "language_identification": "target" }, { "worker_id": "8", "intent_score": 1, "slots_score": 1, "grammar_score": 4, "spelling_score": 2, "language_identification": "target" }, { "worker_id": "0", "intent_score": 1, "slots_score": 1, "grammar_score": 4, "spelling_score": 2, "language_identification": "target" } ] } ``` ### Data Fields `id`: maps to the original ID in the [SLURP](https://github.com/pswietojanski/slurp) collection. Mapping back to the SLURP en-US utterance, this utterance served as the basis for this localization. `locale`: is the language and country code accoring to ISO-639-1 and ISO-3166. `partition`: is either `train`, `dev`, or `test`, according to the original split in [SLURP](https://github.com/pswietojanski/slurp). `scenario`: is the general domain, aka "scenario" in SLURP terminology, of an utterance `intent`: is the specific intent of an utterance within a domain formatted as `{scenario}_{intent}` `utt`: the raw utterance text without annotations `annot_utt`: the text from `utt` with slot annotations formatted as `[{label} : {entity}]` `worker_id`: The obfuscated worker ID from MTurk of the worker completing the localization of the utterance. Worker IDs are specific to a locale and do *not* map across locales. `slot_method`: for each slot in the utterance, whether that slot was a `translation` (i.e., same expression just in the target language), `localization` (i.e., not the same expression but a different expression was chosen more suitable to the phrase in that locale), or `unchanged` (i.e., the original en-US slot value was copied over without modification). `judgments`: Each judgment collected for the localized utterance has 6 keys. `worker_id` is the obfuscated worker ID from MTurk of the worker completing the judgment. Worker IDs are specific to a locale and do *not* map across locales, but *are* consistent across the localization tasks and the judgment tasks, e.g., judgment worker ID 32 in the example above may appear as the localization worker ID for the localization of a different de-DE utterance, in which case it would be the same worker. ```plain intent_score : "Does the sentence match the intent?" 0: No 1: Yes 2: It is a reasonable interpretation of the goal slots_score : "Do all these terms match the categories in square brackets?" 0: No 1: Yes 2: There are no words in square brackets (utterance without a slot) grammar_score : "Read the sentence out loud. Ignore any spelling, punctuation, or capitalization errors. Does it sound natural?" 0: Completely unnatural (nonsensical, cannot be understood at all) 1: Severe errors (the meaning cannot be understood and doesn't sound natural in your language) 2: Some errors (the meaning can be understood but it doesn't sound natural in your language) 3: Good enough (easily understood and sounds almost natural in your language) 4: Perfect (sounds natural in your language) spelling_score : "Are all words spelled correctly? Ignore any spelling variances that may be due to differences in dialect. Missing spaces should be marked as a spelling error." 0: There are more than 2 spelling errors 1: There are 1-2 spelling errors 2: All words are spelled correctly language_identification : "The following sentence contains words in the following languages (check all that apply)" 1: target 2: english 3: other 4: target & english 5: target & other 6: english & other 7: target & english & other ``` ### Data Splits |Language|Train|Dev|Test| |:---:|:---:|:---:|:---:| |af-ZA|11514|2033|2974| |am-ET|11514|2033|2974| |ar-SA|11514|2033|2974| |az-AZ|11514|2033|2974| |bn-BD|11514|2033|2974| |ca-ES|11514|2033|2974| |cy-GB|11514|2033|2974| |da-DK|11514|2033|2974| |de-DE|11514|2033|2974| |el-GR|11514|2033|2974| |en-US|11514|2033|2974| |es-ES|11514|2033|2974| |fa-IR|11514|2033|2974| |fi-FI|11514|2033|2974| |fr-FR|11514|2033|2974| |he-IL|11514|2033|2974| |hi-IN|11514|2033|2974| |hu-HU|11514|2033|2974| |hy-AM|11514|2033|2974| |id-ID|11514|2033|2974| |is-IS|11514|2033|2974| |it-IT|11514|2033|2974| |ja-JP|11514|2033|2974| |jv-ID|11514|2033|2974| |ka-GE|11514|2033|2974| |km-KH|11514|2033|2974| |kn-IN|11514|2033|2974| |ko-KR|11514|2033|2974| |lv-LV|11514|2033|2974| |ml-IN|11514|2033|2974| |mn-MN|11514|2033|2974| |ms-MY|11514|2033|2974| |my-MM|11514|2033|2974| |nb-NO|11514|2033|2974| |nl-NL|11514|2033|2974| |pl-PL|11514|2033|2974| |pt-PT|11514|2033|2974| |ro-RO|11514|2033|2974| |ru-RU|11514|2033|2974| |sl-SL|11514|2033|2974| |sq-AL|11514|2033|2974| |sv-SE|11514|2033|2974| |sw-KE|11514|2033|2974| |ta-IN|11514|2033|2974| |te-IN|11514|2033|2974| |th-TH|11514|2033|2974| |tl-PH|11514|2033|2974| |tr-TR|11514|2033|2974| |ur-PK|11514|2033|2974| |vi-VN|11514|2033|2974| |zh-CN|11514|2033|2974| |zh-TW|11514|2033|2974| ### Personal and Sensitive Information The corpora is free of personal or sensitive information. ## Additional Information ### Dataset Curators __MASSIVE__: Jack FitzGerald and Christopher Hench and Charith Peris and Scott Mackie and Kay Rottmann and Ana Sanchez and Aaron Nash and Liam Urbach and Vishesh Kakarala and Richa Singh and Swetha Ranganath and Laurie Crist and Misha Britan and Wouter Leeuwis and Gokhan Tur and Prem Natarajan. __SLURP__: Bastianelli, Emanuele and Vanzo, Andrea and Swietojanski, Pawel and Rieser, Verena. __Hugging Face Upload and Integration__: Labrak Yanis (Not affiliated with the original corpus) ### Licensing Information ```plain Copyright Amazon.com Inc. or its affiliates. Attribution 4.0 International ======================================================================= Creative Commons Corporation ("Creative Commons") is not a law firm and does not provide legal services or legal advice. Distribution of Creative Commons public licenses does not create a lawyer-client or other relationship. Creative Commons makes its licenses and related information available on an "as-is" basis. Creative Commons gives no warranties regarding its licenses, any material licensed under their terms and conditions, or any related information. Creative Commons disclaims all liability for damages resulting from their use to the fullest extent possible. Using Creative Commons Public Licenses Creative Commons public licenses provide a standard set of terms and conditions that creators and other rights holders may use to share original works of authorship and other material subject to copyright and certain other rights specified in the public license below. The following considerations are for informational purposes only, are not exhaustive, and do not form part of our licenses. Considerations for licensors: Our public licenses are intended for use by those authorized to give the public permission to use material in ways otherwise restricted by copyright and certain other rights. Our licenses are irrevocable. Licensors should read and understand the terms and conditions of the license they choose before applying it. Licensors should also secure all rights necessary before applying our licenses so that the public can reuse the material as expected. Licensors should clearly mark any material not subject to the license. This includes other CC- licensed material, or material used under an exception or limitation to copyright. More considerations for licensors: wiki.creativecommons.org/Considerations_for_licensors Considerations for the public: By using one of our public licenses, a licensor grants the public permission to use the licensed material under specified terms and conditions. If the licensor's permission is not necessary for any reason--for example, because of any applicable exception or limitation to copyright--then that use is not regulated by the license. Our licenses grant only permissions under copyright and certain other rights that a licensor has authority to grant. Use of the licensed material may still be restricted for other reasons, including because others have copyright or other rights in the material. A licensor may make special requests, such as asking that all changes be marked or described. Although not required by our licenses, you are encouraged to respect those requests where reasonable. More considerations for the public: wiki.creativecommons.org/Considerations_for_licensees ======================================================================= Creative Commons Attribution 4.0 International Public License By exercising the Licensed Rights (defined below), You accept and agree to be bound by the terms and conditions of this Creative Commons Attribution 4.0 International Public License ("Public License"). To the extent this Public License may be interpreted as a contract, You are granted the Licensed Rights in consideration of Your acceptance of these terms and conditions, and the Licensor grants You such rights in consideration of benefits the Licensor receives from making the Licensed Material available under these terms and conditions. Section 1 -- Definitions. a. Adapted Material means material subject to Copyright and Similar Rights that is derived from or based upon the Licensed Material and in which the Licensed Material is translated, altered, arranged, transformed, or otherwise modified in a manner requiring permission under the Copyright and Similar Rights held by the Licensor. For purposes of this Public License, where the Licensed Material is a musical work, performance, or sound recording, Adapted Material is always produced where the Licensed Material is synched in timed relation with a moving image. b. Adapter's License means the license You apply to Your Copyright and Similar Rights in Your contributions to Adapted Material in accordance with the terms and conditions of this Public License. c. Copyright and Similar Rights means copyright and/or similar rights closely related to copyright including, without limitation, performance, broadcast, sound recording, and Sui Generis Database Rights, without regard to how the rights are labeled or categorized. For purposes of this Public License, the rights specified in Section 2(b)(1)-(2) are not Copyright and Similar Rights. d. Effective Technological Measures means those measures that, in the absence of proper authority, may not be circumvented under laws fulfilling obligations under Article 11 of the WIPO Copyright Treaty adopted on December 20, 1996, and/or similar international agreements. e. Exceptions and Limitations means fair use, fair dealing, and/or any other exception or limitation to Copyright and Similar Rights that applies to Your use of the Licensed Material. f. Licensed Material means the artistic or literary work, database, or other material to which the Licensor applied this Public License. g. Licensed Rights means the rights granted to You subject to the terms and conditions of this Public License, which are limited to all Copyright and Similar Rights that apply to Your use of the Licensed Material and that the Licensor has authority to license. h. Licensor means the individual(s) or entity(ies) granting rights under this Public License. i. Share means to provide material to the public by any means or process that requires permission under the Licensed Rights, such as reproduction, public display, public performance, distribution, dissemination, communication, or importation, and to make material available to the public including in ways that members of the public may access the material from a place and at a time individually chosen by them. j. Sui Generis Database Rights means rights other than copyright resulting from Directive 96/9/EC of the European Parliament and of the Council of 11 March 1996 on the legal protection of databases, as amended and/or succeeded, as well as other essentially equivalent rights anywhere in the world. k. You means the individual or entity exercising the Licensed Rights under this Public License. Your has a corresponding meaning. Section 2 -- Scope. a. License grant. 1. Subject to the terms and conditions of this Public License, the Licensor hereby grants You a worldwide, royalty-free, non-sublicensable, non-exclusive, irrevocable license to exercise the Licensed Rights in the Licensed Material to: a. reproduce and Share the Licensed Material, in whole or in part; and b. produce, reproduce, and Share Adapted Material. 2. Exceptions and Limitations. For the avoidance of doubt, where Exceptions and Limitations apply to Your use, this Public License does not apply, and You do not need to comply with its terms and conditions. 3. Term. The term of this Public License is specified in Section 6(a). 4. Media and formats; technical modifications allowed. The Licensor authorizes You to exercise the Licensed Rights in all media and formats whether now known or hereafter created, and to make technical modifications necessary to do so. The Licensor waives and/or agrees not to assert any right or authority to forbid You from making technical modifications necessary to exercise the Licensed Rights, including technical modifications necessary to circumvent Effective Technological Measures. For purposes of this Public License, simply making modifications authorized by this Section 2(a) (4) never produces Adapted Material. 5. Downstream recipients. a. Offer from the Licensor -- Licensed Material. Every recipient of the Licensed Material automatically receives an offer from the Licensor to exercise the Licensed Rights under the terms and conditions of this Public License. b. No downstream restrictions. You may not offer or impose any additional or different terms or conditions on, or apply any Effective Technological Measures to, the Licensed Material if doing so restricts exercise of the Licensed Rights by any recipient of the Licensed Material. 6. No endorsement. Nothing in this Public License constitutes or may be construed as permission to assert or imply that You are, or that Your use of the Licensed Material is, connected with, or sponsored, endorsed, or granted official status by, the Licensor or others designated to receive attribution as provided in Section 3(a)(1)(A)(i). b. Other rights. 1. Moral rights, such as the right of integrity, are not licensed under this Public License, nor are publicity, privacy, and/or other similar personality rights; however, to the extent possible, the Licensor waives and/or agrees not to assert any such rights held by the Licensor to the limited extent necessary to allow You to exercise the Licensed Rights, but not otherwise. 2. Patent and trademark rights are not licensed under this Public License. 3. To the extent possible, the Licensor waives any right to collect royalties from You for the exercise of the Licensed Rights, whether directly or through a collecting society under any voluntary or waivable statutory or compulsory licensing scheme. In all other cases the Licensor expressly reserves any right to collect such royalties. Section 3 -- License Conditions. Your exercise of the Licensed Rights is expressly made subject to the following conditions. a. Attribution. 1. If You Share the Licensed Material (including in modified form), You must: a. retain the following if it is supplied by the Licensor with the Licensed Material: i. identification of the creator(s) of the Licensed Material and any others designated to receive attribution, in any reasonable manner requested by the Licensor (including by pseudonym if designated); ii. a copyright notice; iii. a notice that refers to this Public License; iv. a notice that refers to the disclaimer of warranties; v. a URI or hyperlink to the Licensed Material to the extent reasonably practicable; b. indicate if You modified the Licensed Material and retain an indication of any previous modifications; and c. indicate the Licensed Material is licensed under this Public License, and include the text of, or the URI or hyperlink to, this Public License. 2. You may satisfy the conditions in Section 3(a)(1) in any reasonable manner based on the medium, means, and context in which You Share the Licensed Material. For example, it may be reasonable to satisfy the conditions by providing a URI or hyperlink to a resource that includes the required information. 3. If requested by the Licensor, You must remove any of the information required by Section 3(a)(1)(A) to the extent reasonably practicable. 4. If You Share Adapted Material You produce, the Adapter's License You apply must not prevent recipients of the Adapted Material from complying with this Public License. Section 4 -- Sui Generis Database Rights. Where the Licensed Rights include Sui Generis Database Rights that apply to Your use of the Licensed Material: a. for the avoidance of doubt, Section 2(a)(1) grants You the right to extract, reuse, reproduce, and Share all or a substantial portion of the contents of the database; b. if You include all or a substantial portion of the database contents in a database in which You have Sui Generis Database Rights, then the database in which You have Sui Generis Database Rights (but not its individual contents) is Adapted Material; and c. You must comply with the conditions in Section 3(a) if You Share all or a substantial portion of the contents of the database. For the avoidance of doubt, this Section 4 supplements and does not replace Your obligations under this Public License where the Licensed Rights include other Copyright and Similar Rights. Section 5 -- Disclaimer of Warranties and Limitation of Liability. a. UNLESS OTHERWISE SEPARATELY UNDERTAKEN BY THE LICENSOR, TO THE EXTENT POSSIBLE, THE LICENSOR OFFERS THE LICENSED MATERIAL AS-IS AND AS-AVAILABLE, AND MAKES NO REPRESENTATIONS OR WARRANTIES OF ANY KIND CONCERNING THE LICENSED MATERIAL, WHETHER EXPRESS, IMPLIED, STATUTORY, OR OTHER. THIS INCLUDES, WITHOUT LIMITATION, WARRANTIES OF TITLE, MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE, NON-INFRINGEMENT, ABSENCE OF LATENT OR OTHER DEFECTS, ACCURACY, OR THE PRESENCE OR ABSENCE OF ERRORS, WHETHER OR NOT KNOWN OR DISCOVERABLE. WHERE DISCLAIMERS OF WARRANTIES ARE NOT ALLOWED IN FULL OR IN PART, THIS DISCLAIMER MAY NOT APPLY TO YOU. b. TO THE EXTENT POSSIBLE, IN NO EVENT WILL THE LICENSOR BE LIABLE TO YOU ON ANY LEGAL THEORY (INCLUDING, WITHOUT LIMITATION, NEGLIGENCE) OR OTHERWISE FOR ANY DIRECT, SPECIAL, INDIRECT, INCIDENTAL, CONSEQUENTIAL, PUNITIVE, EXEMPLARY, OR OTHER LOSSES, COSTS, EXPENSES, OR DAMAGES ARISING OUT OF THIS PUBLIC LICENSE OR USE OF THE LICENSED MATERIAL, EVEN IF THE LICENSOR HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH LOSSES, COSTS, EXPENSES, OR DAMAGES. WHERE A LIMITATION OF LIABILITY IS NOT ALLOWED IN FULL OR IN PART, THIS LIMITATION MAY NOT APPLY TO YOU. c. The disclaimer of warranties and limitation of liability provided above shall be interpreted in a manner that, to the extent possible, most closely approximates an absolute disclaimer and waiver of all liability. Section 6 -- Term and Termination. a. This Public License applies for the term of the Copyright and Similar Rights licensed here. However, if You fail to comply with this Public License, then Your rights under this Public License terminate automatically. b. Where Your right to use the Licensed Material has terminated under Section 6(a), it reinstates: 1. automatically as of the date the violation is cured, provided it is cured within 30 days of Your discovery of the violation; or 2. upon express reinstatement by the Licensor. For the avoidance of doubt, this Section 6(b) does not affect any right the Licensor may have to seek remedies for Your violations of this Public License. c. For the avoidance of doubt, the Licensor may also offer the Licensed Material under separate terms or conditions or stop distributing the Licensed Material at any time; however, doing so will not terminate this Public License. d. Sections 1, 5, 6, 7, and 8 survive termination of this Public License. Section 7 -- Other Terms and Conditions. a. The Licensor shall not be bound by any additional or different terms or conditions communicated by You unless expressly agreed. b. Any arrangements, understandings, or agreements regarding the Licensed Material not stated herein are separate from and independent of the terms and conditions of this Public License. Section 8 -- Interpretation. a. For the avoidance of doubt, this Public License does not, and shall not be interpreted to, reduce, limit, restrict, or impose conditions on any use of the Licensed Material that could lawfully be made without permission under this Public License. b. To the extent possible, if any provision of this Public License is deemed unenforceable, it shall be automatically reformed to the minimum extent necessary to make it enforceable. If the provision cannot be reformed, it shall be severed from this Public License without affecting the enforceability of the remaining terms and conditions. c. No term or condition of this Public License will be waived and no failure to comply consented to unless expressly agreed to by the Licensor. d. Nothing in this Public License constitutes or may be interpreted as a limitation upon, or waiver of, any privileges and immunities that apply to the Licensor or You, including from the legal processes of any jurisdiction or authority. ======================================================================= Creative Commons is not a party to its public licenses. Notwithstanding, Creative Commons may elect to apply one of its public licenses to material it publishes and in those instances will be considered the “Licensor.” The text of the Creative Commons public licenses is dedicated to the public domain under the CC0 Public Domain Dedication. Except for the limited purpose of indicating that material is shared under a Creative Commons public license or as otherwise permitted by the Creative Commons policies published at creativecommons.org/policies, Creative Commons does not authorize the use of the trademark "Creative Commons" or any other trademark or logo of Creative Commons without its prior written consent including, without limitation, in connection with any unauthorized modifications to any of its public licenses or any other arrangements, understandings, or agreements concerning use of licensed material. For the avoidance of doubt, this paragraph does not form part of the public licenses. Creative Commons may be contacted at creativecommons.org. ``` ### Citation Information Please cite the following papers when using this dataset. ```latex @misc{fitzgerald2022massive, title={MASSIVE: A 1M-Example Multilingual Natural Language Understanding Dataset with 51 Typologically-Diverse Languages}, author={Jack FitzGerald and Christopher Hench and Charith Peris and Scott Mackie and Kay Rottmann and Ana Sanchez and Aaron Nash and Liam Urbach and Vishesh Kakarala and Richa Singh and Swetha Ranganath and Laurie Crist and Misha Britan and Wouter Leeuwis and Gokhan Tur and Prem Natarajan}, year={2022}, eprint={2204.08582}, archivePrefix={arXiv}, primaryClass={cs.CL} } @inproceedings{bastianelli-etal-2020-slurp, title = "{SLURP}: A Spoken Language Understanding Resource Package", author = "Bastianelli, Emanuele and Vanzo, Andrea and Swietojanski, Pawel and Rieser, Verena", booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)", month = nov, year = "2020", address = "Online", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2020.emnlp-main.588", doi = "10.18653/v1/2020.emnlp-main.588", pages = "7252--7262", abstract = "Spoken Language Understanding infers semantic meaning directly from audio data, and thus promises to reduce error propagation and misunderstandings in end-user applications. However, publicly available SLU resources are limited. In this paper, we release SLURP, a new SLU package containing the following: (1) A new challenging dataset in English spanning 18 domains, which is substantially bigger and linguistically more diverse than existing datasets; (2) Competitive baselines based on state-of-the-art NLU and ASR systems; (3) A new transparent metric for entity labelling which enables a detailed error analysis for identifying potential areas of improvement. SLURP is available at https://github.com/pswietojanski/slurp." } ```
odellus
null
null
null
false
1
false
odellus/beerqa
2022-04-27T21:31:18.000Z
null
false
91d0a0d7cccfff341a3a2806aede1988e9b907c0
[]
[ "license:cc-by-4.0" ]
https://huggingface.co/datasets/odellus/beerqa/resolve/main/README.md
--- license: cc-by-4.0 ---
codeparrot
null
null
null
false
19
false
codeparrot/codeparrot-valid-more-filtering
2022-06-21T17:56:02.000Z
null
false
3a14d4d277b7ad97c668740aa9d0affadbf3fe8d
[]
[]
https://huggingface.co/datasets/codeparrot/codeparrot-valid-more-filtering/resolve/main/README.md
# CodeParrot 🦜 Dataset Cleaned and filtered (validation) ## Dataset Description A dataset of Python files from Github. It is a more filtered version of the validation split [codeparrot-clean-valid](https://huggingface.co/datasets/codeparrot/codeparrot-clean-valid) of [codeparrot-clean](https://huggingface.co/datasets/codeparrot/codeparrot-clean#codeparrot-%F0%9F%A6%9C-dataset-cleaned). The additional filters aim at detecting configuration and test files, as well as outlier files that are unlikely to help the model learn code. The first three filters are applied with a probability of 0.7: - files with a mention of "test file" or "configuration file" or similar in the first 5 lines - files with high occurence of the keywords "test " or "config" - files without a mention of the keywords `def`, `for`, `while` and `class` - files that use the assignment operator ```=``` less than 5 times - files with ratio between number of characters and number of tokens after tokenization < 1.5
codeparrot
null
null
null
false
7
false
codeparrot/codeparrot-train-more-filtering
2022-06-21T17:54:51.000Z
null
false
f6e726786732dbdb82a970091bb6d7ffe2251c80
[]
[]
https://huggingface.co/datasets/codeparrot/codeparrot-train-more-filtering/resolve/main/README.md
# CodeParrot 🦜 Dataset Cleaned and filtered (train) ## Dataset Description A dataset of Python files from Github. It is a more filtered version of the train split [codeparrot-clean-train](https://huggingface.co/datasets/codeparrot/codeparrot-clean-train) of [codeparrot-clean](https://huggingface.co/datasets/codeparrot/codeparrot-clean#codeparrot-%F0%9F%A6%9C-dataset-cleaned). The additional filters aim at detecting configuration and test files, as well as outlier files that are unlikely to help the model learn code. The first three filters are applied with a probability of 0.7: - files with a mention of "test file" or "configuration file" or similar in the first 5 lines - files with high occurence of the keywords "test " or "config" - files without a mention of the keywords `def`, `for`, `while` and `class` - files that use the assignment operator ```=``` less than 5 times - files with ratio between number of characters and number of tokens after tokenization < 1.5
mathigatti
null
null
null
false
5
false
mathigatti/spanish_imdb_synopsis
2022-10-25T10:12:53.000Z
null
false
c8e9269cd96a600bb340ee83ae45f46a02787a12
[]
[ "annotations_creators:no-annotation", "language:es", "license:apache-2.0", "multilinguality:monolingual", "task_categories:summarization", "task_categories:text-generation", "task_categories:text2text-generation" ]
https://huggingface.co/datasets/mathigatti/spanish_imdb_synopsis/resolve/main/README.md
--- annotations_creators: - no-annotation language: - es license: - apache-2.0 multilinguality: - monolingual task_categories: - summarization - text-generation - text2text-generation --- # Dataset Card for Spanish IMDb Synopsis ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Fields](#data-fields) - [Dataset Creation](#dataset-creation) ## Dataset Description 4969 movie synopsis from IMDb in spanish. ### Dataset Summary [N/A] ### Languages All descriptions are in spanish, the other fields have some mix of spanish and english. ## Dataset Structure [N/A] ### Data Fields - `description`: IMDb description for the movie (string), should be spanish - `keywords`: IMDb keywords for the movie (string), mix of spanish and english - `genre`: The genres of the movie (string), mix of spanish and english - `year`: The year the movie was published (float) - `name`: The name of the movie (string), mix of spanish and english - `director`: The name of the main director in the movie, can be empty (string) ## Dataset Creation [This kaggle dataset](https://www.kaggle.com/datasets/komalkhetlani/imdb-dataset) was used as a starting point. Then IMDb was scraped downloading the synopsis of the movies that have more than 5000 votes/reviews and those that did not have a synopsis available in Spanish were discarded.
TalTechNLP
null
null
null
false
7
false
TalTechNLP/VoxLingua107
2022-05-05T10:52:37.000Z
null
false
8ef331f75ebced42eceb7c5f53c47a3dafd8ef20
[]
[ "license:cc-by-nc-4.0" ]
https://huggingface.co/datasets/TalTechNLP/VoxLingua107/resolve/main/README.md
--- license: cc-by-nc-4.0 --- hello
strombergnlp
null
@inproceedings{norregaard-derczynski-2021-danfever, title = "{D}an{FEVER}: claim verification dataset for {D}anish", author = "N{\o}rregaard, Jeppe and Derczynski, Leon", booktitle = "Proceedings of the 23rd Nordic Conference on Computational Linguistics (NoDaLiDa)", month = may # " 31--2 " # jun, year = "2021", address = "Reykjavik, Iceland (Online)", publisher = {Link{\"o}ping University Electronic Press, Sweden}, url = "https://aclanthology.org/2021.nodalida-main.47", pages = "422--428", abstract = "We present a dataset, DanFEVER, intended for multilingual misinformation research. The dataset is in Danish and has the same format as the well-known English FEVER dataset. It can be used for testing methods in multilingual settings, as well as for creating models in production for the Danish language.", }
\
false
1
false
strombergnlp/danfever
2022-10-25T21:42:40.000Z
danfever
false
5d01e3f6a661d48e127ab5d7e3aaa0dc8331438a
[]
[ "annotations_creators:expert-generated", "language_creators:found", "language:da", "license:cc-by-4.0", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "task_categories:text-classification", "task_ids:fact-checking", "task_ids:natural-language-inference", "...
https://huggingface.co/datasets/strombergnlp/danfever/resolve/main/README.md
--- annotations_creators: - expert-generated language_creators: - found language: - da license: - cc-by-4.0 multilinguality: - monolingual size_categories: - 1K<n<10K source_datasets: - original task_categories: - text-classification task_ids: - fact-checking - natural-language-inference paperswithcode_id: danfever pretty_name: DanFEVER tags: - knowledge-verification --- # Dataset Card for DanFEVER ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** [https://github.com/StrombergNLP/danfever](https://github.com/StrombergNLP/danfever) - **Repository:** [https://stromberg.ai/publication/danfever/](https://stromberg.ai/publication/danfever/) - **Paper:** [https://aclanthology.org/2021.nodalida-main.47/](https://aclanthology.org/2021.nodalida-main.47/) - **Leaderboard:** [Needs More Information] - **Point of Contact:** [Leon Derczynski](mailto:leod@itu.dk) - **Size of downloaded dataset files:** 2.82 MiB - **Size of the generated dataset:** 2.80 MiB - **Total amount of disk used:** 5.62 MiB ### Dataset Summary We present a dataset, DanFEVER, intended for multilingual misinformation research. The dataset is in Danish and has the same format as the well-known English FEVER dataset. It can be used for testing methods in multilingual settings, as well as for creating models in production for the Danish language. ### Supported Tasks and Leaderboards This dataset supports the FEVER task, but in Danish. * PwC leaderboard: [Fact Verification on DanFEVER](https://paperswithcode.com/sota/fact-verification-on-danfever) ### Languages This dataset is in Danish; the bcp47 is `da_DK`. ## Dataset Structure ### Data Instances ``` { 'id': '0', 'claim': 'Den 31. oktober 1920 opdagede Walter Baade kometen (944) Hidalgo i det ydre solsystem.', 'label': 0, 'evidence_extract': '(944) Hidalgo (oprindeligt midlertidigt navn: 1920 HZ) er en mørk småplanet med en diameter på ca. 50 km, der befinder sig i det ydre solsystem. Objektet blev opdaget den 31. oktober 1920 af Walter Baade. En asteroide (småplanet, planetoide) er et fast himmellegeme, hvis bane går rundt om Solen (eller en anden stjerne). Pr. 5. maj 2017 kendes mere end 729.626 asteroider og de fleste befinder sig i asteroidebæltet mellem Mars og Jupiter.', 'verifiable': 1, 'evidence': 'wiki_26366, wiki_12289', 'original_id': '1' } ``` ### Data Fields [Needs More Information] ### Data Splits [Needs More Information] ## Dataset Creation ### Curation Rationale A dump of the Danish Wikipedia of 13 February 2020 was stored as well as the relevant articles from Den Store Danske (excerpts only, to comply with copyright laws). Two teams of two people independently sampled evidence, and created and annotated claims from these two sites. ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? The source language is from Wikipedia contributors editors and from dictionary contributors and editors. ### Annotations #### Annotation process Detailed in [this paper](http://www.derczynski.com/papers/danfever.pdf). #### Who are the annotators? The annotators are native Danish speakers and masters students of IT; two female, two male, ages 25-35. ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset The purpose of this dataset is to enable construction of fact-checking systems in Danish. A system that succeeds at this may be able to identify questionable conclusions or inferences. ### Discussion of Biases The data is drawn from relatively formal topics, and so may perform poorly outside these areas. ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information The data here is licensed CC-BY 4.0. If you use this data, you MUST state its origin. ### Citation Information Refer to this work as: > Nørregaard and Derczynski (2021). "DanFEVER: claim verification dataset for Danish", Proceedings of the 23rd Nordic Conference on Computational Linguistics (NoDaLiDa). Bibliographic reference: ```` @inproceedings{norregaard-derczynski-2021-danfever, title = "{D}an{FEVER}: claim verification dataset for {D}anish", author = "N{\o}rregaard, Jeppe and Derczynski, Leon", booktitle = "Proceedings of the 23rd Nordic Conference on Computational Linguistics (NoDaLiDa)", year = "2021", publisher = {Link{\"o}ping University Electronic Press, Sweden}, url = "https://aclanthology.org/2021.nodalida-main.47", pages = "422--428" } ```
JbIPS
null
null
null
false
1
false
JbIPS/stanford-dogs
2022-04-28T09:56:39.000Z
null
false
dad84ecc9f47c2d0701018018903a158bad71867
[]
[ "license:mit" ]
https://huggingface.co/datasets/JbIPS/stanford-dogs/resolve/main/README.md
--- license: mit ---
strombergnlp
null
@inproceedings{derczynski2016broad, title={Broad twitter corpus: A diverse named entity recognition resource}, author={Derczynski, Leon and Bontcheva, Kalina and Roberts, Ian}, booktitle={Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers}, pages={1169--1179}, year={2016} }
This is the Broad Twitter corpus, a dataset of tweets collected over stratified times, places and social uses. The goal is to represent a broad range of activities, giving a dataset more representative of the language used in this hardest of social media formats to process. Further, the BTC is annotated for named entities. For more details see [https://aclanthology.org/C16-1111/](https://aclanthology.org/C16-1111/)
false
213
false
strombergnlp/broad_twitter_corpus
2022-07-01T15:46:36.000Z
broad-twitter-corpus
false
d766cb8a7497d0d507d81f5f681a8d58deedf495
[]
[ "annotations_creators:crowdsourced", "language_creators:found", "language:en", "license:cc-by-4.0", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "task_categories:token-classification", "task_ids:named-entity-recognition" ]
https://huggingface.co/datasets/strombergnlp/broad_twitter_corpus/resolve/main/README.md
--- annotations_creators: - crowdsourced language_creators: - found language: - en license: - cc-by-4.0 multilinguality: - monolingual size_categories: - 100K<n<1M source_datasets: - original task_categories: - token-classification task_ids: - named-entity-recognition paperswithcode_id: broad-twitter-corpus pretty_name: Broad Twitter Corpus --- # Dataset Card for broad_twitter_corpus ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** [https://github.com/GateNLP/broad_twitter_corpus](https://github.com/GateNLP/broad_twitter_corpus) - **Repository:** [https://github.com/GateNLP/broad_twitter_corpus](https://github.com/GateNLP/broad_twitter_corpus) - **Paper:** [http://www.aclweb.org/anthology/C16-1111](http://www.aclweb.org/anthology/C16-1111) - **Leaderboard:** [Named Entity Recognition on Broad Twitter Corpus](https://paperswithcode.com/sota/named-entity-recognition-on-broad-twitter) - **Point of Contact:** [Leon Derczynski](https://github.com/leondz) ### Dataset Summary This is the Broad Twitter corpus, a dataset of tweets collected over stratified times, places and social uses. The goal is to represent a broad range of activities, giving a dataset more representative of the language used in this hardest of social media formats to process. Further, the BTC is annotated for named entities. See the paper, [Broad Twitter Corpus: A Diverse Named Entity Recognition Resource](http://www.aclweb.org/anthology/C16-1111), for details. ### Supported Tasks and Leaderboards * Named Entity Recognition * On PWC: [Named Entity Recognition on Broad Twitter Corpus](https://paperswithcode.com/sota/named-entity-recognition-on-broad-twitter) ### Languages English from UK, US, Australia, Canada, Ireland, New Zealand; `bcp47:en` ## Dataset Structure ### Data Instances Feature |Count ---|---: Documents |9 551 Tokens |165 739 Person entities |5 271 Location entities |3 114 Organization entities |3 732 ### Data Fields Each tweet contains an ID, a list of tokens, and a list of NER tags - `id`: a `string` feature. - `tokens`: a `list` of `strings` - `ner_tags`: a `list` of class IDs (`int`s) representing the NER class: ``` 0: O 1: B-PER 2: I-PER 3: B-ORG 4: I-ORG 5: B-LOC 6: I-LOC ``` ### Data Splits Section|Region|Collection period|Description|Annotators|Tweet count ---|---|---|---|---|---: A | UK| 2012.01| General collection |Expert| 1000 B |UK |2012.01-02 |Non-directed tweets |Expert |2000 E |Global| 2014.07| Related to MH17 disaster| Crowd & expert |200 F |Stratified |2009-2014| Twitterati |Crowd & expert |2000 G |Stratified| 2011-2014| Mainstream news| Crowd & expert| 2351 H |Non-UK| 2014 |General collection |Crowd & expert |2000 The most varied parts of the BTC are sections F and H. However, each of the remaining four sections has some specific readily-identifiable bias. So, we propose that one uses half of section H for evaluation and leaves the other half in the training data. Section H should be partitioned in the order of the JSON-format lines. Note that the CoNLL-format data is readily reconstructible from the JSON format, which is the authoritative data format from which others are derived. **Test**: Section F **Development**: Section H (the paper says "second half of Section H" but ordinality could be ambiguous, so it all goes in. Bonne chance) **Training**: everything else ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information Creative Commons Attribution 4.0 International (CC BY 4.0) ### Citation Information ``` @inproceedings{derczynski2016broad, title={Broad twitter corpus: A diverse named entity recognition resource}, author={Derczynski, Leon and Bontcheva, Kalina and Roberts, Ian}, booktitle={Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers}, pages={1169--1179}, year={2016} } ``` ### Contributions Author-added dataset [@leondz](https://github.com/leondz)
strombergnlp
null
@article{derczynski2015analysis, title={Analysis of named entity recognition and linking for tweets}, author={Derczynski, Leon and Maynard, Diana and Rizzo, Giuseppe and Van Erp, Marieke and Gorrell, Genevieve and Troncy, Rapha{\"e}l and Petrak, Johann and Bontcheva, Kalina}, journal={Information Processing \& Management}, volume={51}, number={2}, pages={32--49}, year={2015}, publisher={Elsevier} }
This data is for the task of named entity recognition and linking/disambiguation over tweets. It comprises the addition of an entity URI layer on top of an NER-annotated tweet dataset. The task is to detect entities and then provide a correct link to them in DBpedia, thus disambiguating otherwise ambiguous entity surface forms; for example, this means linking "Paris" to the correct instance of a city named that (e.g. Paris, France vs. Paris, Texas). The data concentrates on ten types of named entities: company, facility, geographic location, movie, musical artist, person, product, sports team, TV show, and other. The file is tab separated, in CoNLL format, with line breaks between tweets. Data preserves the tokenisation used in the Ritter datasets. PoS labels are not present for all tweets, but where they could be found in the Ritter data, they're given. In cases where a URI could not be agreed, or was not present in DBpedia, there is a NIL. See the paper for a full description of the methodology. For more details see http://www.derczynski.com/papers/ner_single.pdf or https://www.sciencedirect.com/science/article/abs/pii/S0306457314001034
false
5
false
strombergnlp/ipm_nel
2022-10-25T21:41:26.000Z
ipm-nel
false
cc150b1a28983f4796ab486f6e1ef1d1047e523a
[]
[ "annotations_creators:crowdsourced", "language_creators:found", "language:en", "license:cc-by-4.0", "multilinguality:monolingual", "size_categories:1K<n<10K", "task_categories:token-classification", "task_ids:named-entity-recognition", "tags:named-entity-linking" ]
https://huggingface.co/datasets/strombergnlp/ipm_nel/resolve/main/README.md
--- annotations_creators: - crowdsourced language_creators: - found language: - en license: - cc-by-4.0 multilinguality: - monolingual size_categories: - 1K<n<10K source_datasets: [] task_categories: - token-classification task_ids: - named-entity-recognition paperswithcode_id: ipm-nel pretty_name: IPM NEL (Derczynski) tags: - named-entity-linking --- # Dataset Card for "ipm-nel" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** []() - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [http://www.derczynski.com/papers/ner_single.pdf](http://www.derczynski.com/papers/ner_single.pdf) - **Point of Contact:** [Leon Derczynski](https://github.com/leondz) - **Size of downloaded dataset files:** 120 KB - **Size of the generated dataset:** - **Total amount of disk used:** ### Dataset Summary This data is for the task of named entity recognition and linking/disambiguation over tweets. It comprises the addition of an entity URI layer on top of an NER-annotated tweet dataset. The task is to detect entities and then provide a correct link to them in DBpedia, thus disambiguating otherwise ambiguous entity surface forms; for example, this means linking "Paris" to the correct instance of a city named that (e.g. Paris, France vs. Paris, Texas). The data concentrates on ten types of named entities: company, facility, geographic location, movie, musical artist, person, product, sports team, TV show, and other. The file is tab separated, in CoNLL format, with line breaks between tweets. * Data preserves the tokenisation used in the Ritter datasets. * PoS labels are not present for all tweets, but where they could be found in the Ritter data, they're given. * In cases where a URI could not be agreed, or was not present in DBpedia, the linking URI is `NIL`. See the paper, [Analysis of Named Entity Recognition and Linking for Tweets](http://www.derczynski.com/papers/ner_single.pdf) for a full description of the methodology. ### Supported Tasks and Leaderboards * Dataset leaderboard on PWC: [Entity Linking on Derczynski](https://paperswithcode.com/sota/entity-linking-on-derczynski-1) ### Languages English of unknown region (`bcp47:en`) ## Dataset Structure ### Data Instances #### ipm_nel - **Size of downloaded dataset files:** 120 KB - **Size of the generated dataset:** - **Total amount of disk used:** An example of 'train' looks as follows. ``` { 'id': '0', 'tokens': ['#Astros', 'lineup', 'for', 'tonight', '.', 'Keppinger', 'sits', ',', 'Downs', 'plays', '2B', ',', 'CJ', 'bats', '5th', '.', '@alysonfooter', 'http://bit.ly/bHvgCS'], 'ner_tags': [9, 0, 0, 0, 0, 7, 0, 0, 7, 0, 0, 0, 7, 0, 0, 0, 0, 0], 'uris': "['http://dbpedia.org/resource/Houston_Astros', '', '', '', '', 'http://dbpedia.org/resource/Jeff_Keppinger', '', '', 'http://dbpedia.org/resource/Brodie_Downs', '', '', '', 'NIL', '', '', '', '', '']" } ``` ### Data Fields - `id`: a `string` feature. - `tokens`: a `list` of `string` features. - `ner_tags`: a `list` of classification labels (`int`). Full tagset with indices: - `uris`: a `list` of URIs (`string`) that disambiguate entities. Set to `NIL` when an entity has no DBpedia entry, or blank for outside-of-entity tokens. ### Data Splits | name |train| |---------|----:| |ipm_nel|183 sentences| ## Dataset Creation ### Curation Rationale To gather a social media benchmark for named entity linking that is sufficiently different from newswire data. ### Source Data #### Initial Data Collection and Normalization The data is partly harvested from that distributed by [Ritter / Named Entity Recognition in Tweets: An Experimental Study](https://aclanthology.org/D11-1141/), and partly taken from Twitter by the authors. #### Who are the source language producers? English-speaking Twitter users, between October 2011 and September 2013 ### Annotations #### Annotation process The authors were allocated documents and marked them for named entities (where these were not already present) and then attempted to find the best-fitting DBpedia entry for each entity found. Each entity mention was labelled by a random set of three volunteers. The annotation task was mediated using Crowdflower (Biewald, 2012). Our interface design was to show each volunteer the text of the tweet, any URL links contained therein, and a set of candidate targets from DBpedia. The volunteers were encouraged to click on the URL links from the tweet, to gain addition context and thus ensure that the correct DBpedia URI is chosen by them. Candidate entities were shown in random order, using the text from the corresponding DBpedia abstracts (where available) or the actual DBpedia URI otherwise. In addition, the options ‘‘none of the above’’, ‘‘not an entity’’ and ‘‘cannot decide’’ were added, to allow the volunteers to indicate that this entity mention has no corresponding DBpedia URI (none of the above), the highlighted text is not an entity, or that the tweet text (and any links, if available) did not provide sufficient information to reliably disambiguate the entity mention. #### Who are the annotators? The annotators are 10 volunteer NLP researchers, from the authors and the authors' institutions. ### Personal and Sensitive Information The data was public at the time of collection. User names are preserved. ## Considerations for Using the Data ### Social Impact of Dataset There's a risk of user-deleted content being in this data. The data has NOT been vetted for any content, so there's a risk of harmful text. ### Discussion of Biases The data is annotated by NLP researchers; we know that this group has high agreement but low recall on English twitter text [C16-1111](https://aclanthology.org/C16-1111/). ### Other Known Limitations The above limitations apply. ## Additional Information ### Dataset Curators The dataset is curated by the paper's authors. ### Licensing Information The authors distribute this data under Creative Commons attribution license, CC-BY 4.0. You must acknowledge the author if you use this data, but apart from that, you're quite free to do most things. See https://creativecommons.org/licenses/by/4.0/legalcode . ### Citation Information ``` @article{derczynski2015analysis, title={Analysis of named entity recognition and linking for tweets}, author={Derczynski, Leon and Maynard, Diana and Rizzo, Giuseppe and Van Erp, Marieke and Gorrell, Genevieve and Troncy, Rapha{\"e}l and Petrak, Johann and Bontcheva, Kalina}, journal={Information Processing \& Management}, volume={51}, number={2}, pages={32--49}, year={2015}, publisher={Elsevier} } ``` ### Contributions Author-added dataset [@leondz](https://github.com/leondz)
strombergnlp
null
@inproceedings{lillie-etal-2019-joint, title = "Joint Rumour Stance and Veracity Prediction", author = "Lillie, Anders Edelbo and Middelboe, Emil Refsgaard and Derczynski, Leon", booktitle = "Proceedings of the 22nd Nordic Conference on Computational Linguistics", month = sep # "{--}" # oct, year = "2019", address = "Turku, Finland", publisher = {Link{\"o}ping University Electronic Press}, url = "https://aclanthology.org/W19-6122", pages = "208--221", }
This dataset presents a series of stories on Reddit and the conversation around them, annotated for stance. Stories are also annotated for veracity. For more details see https://aclanthology.org/W19-6122/
false
1
false
strombergnlp/dkstance
2022-10-25T21:45:42.000Z
dast
false
1075212523ac84e51b0fe6bc41fef4cd2bf695cc
[]
[ "annotations_creators:expert-generated", "language_creators:found", "language:da", "license:cc-by-4.0", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "task_categories:text-classification", "task_ids:fact-checking", "extra_gated_prompt:Warning: the data in t...
https://huggingface.co/datasets/strombergnlp/dkstance/resolve/main/README.md
--- annotations_creators: - expert-generated language_creators: - found language: - da license: - cc-by-4.0 multilinguality: - monolingual size_categories: - 1K<n<10K source_datasets: - original task_categories: - text-classification task_ids: - fact-checking paperswithcode_id: dast pretty_name: DAST extra_gated_prompt: 'Warning: the data in this repository contains harmful content (misinformative claims).' tags: - stance-detection --- # Dataset Card for "dkstance / DAST" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://stromberg.ai/publication/jointrumourstanceandveracity/](https://stromberg.ai/publication/jointrumourstanceandveracity/) - **Repository:** [https://figshare.com/articles/dataset/Danish_stance-annotated_Reddit_dataset/8217137](https://figshare.com/articles/dataset/Danish_stance-annotated_Reddit_dataset/8217137) - **Paper:** [https://aclanthology.org/W19-6122/](https://aclanthology.org/W19-6122/) - **Point of Contact:** [Leon Derczynski](https://github.com/leondz) - **Size of downloaded dataset files:** - **Size of the generated dataset:** - **Total amount of disk used:** ### Dataset Summary This is an SDQC stance-annotated Reddit dataset for the Danish language generated within a thesis project. The dataset consists of over 5000 comments structured as comment trees and linked to 33 source posts. The dataset is applicable for supervised stance classification and rumour veracity prediction. ### Supported Tasks and Leaderboards * Stance prediction ### Languages ## Dataset Structure ### Data Instances #### DAST / dkstance - **Size of downloaded dataset files:** 4.72 MiB - **Size of the generated dataset:** 3.69 MiB - **Total amount of disk used:** 8.41 MiB An example of 'train' looks as follows. ``` { 'id': '1', 'native_id': 'ebwjq5z', 'text': 'Med de udfordringer som daginstitutionerne har med normeringer, og økonomi i det hele taget, synes jeg det er en vanvittig beslutning at prioritere skattebetalt vegansk kost i daginstitutionerne. Brug dog pengene på noget mere personale, og lad folk selv betale for deres individuelle kostønsker.', 'parent_id': 'a6o3us', 'parent_text': 'Mai Mercado om mad i daginstitutioner: Sund kost rimer ikke på veganer-mad', 'parent_stance': 0, 'source_id': 'a6o3us', 'source_text': 'Mai Mercado om mad i daginstitutioner: Sund kost rimer ikke på veganer-mad', 'source_stance': 0 } ``` ### Data Fields - `id`: a `string` feature. - `native_id`: a `string` feature representing the native ID of the entry. - `text`: a `string` of the comment text in which stance is annotated. - `parent_id`: the `native_id` of this comment's parent. - `parent_text`: a `string` of the parent comment's text. - `parent_stance`: the label of the stance in the comment towards its parent comment. ``` 0: "Supporting", 1: "Denying", 2: "Querying", 3: "Commenting", ``` - `source_id`: the `native_id` of this comment's source / post. - `source_text`: a `string` of the source / post text. - `source_stance`: the label of the stance in the comment towards the original source post. ``` 0: "Supporting", 1: "Denying", 2: "Querying", 3: "Commenting", ``` ### Data Splits | name |size| |---------|----:| |train|3122| |validation|1066| |test|1060| These splits are specified after the original reserach was reported. The splits add an extra level of rigour, in that no source posts' comment tree is spread over more than one partition. ## Dataset Creation ### Curation Rationale Comments around rumourous claims to enable rumour and stance analysis in Danish ### Source Data #### Initial Data Collection and Normalization The data is from Reddit posts that relate to one of a specific set of news stories; these stories are enumerated in the paper. #### Who are the source language producers? Danish-speaking Twitter users. ### Annotations #### Annotation process There was multi-user annotation process mediated through a purpose-built interface for annotating stance in Reddit threads. #### Who are the annotators? * Age: 20-30. * Gender: male. * Race/ethnicity: white northern European. * Native language: Danish. * Socioeconomic status: higher education student. ### Personal and Sensitive Information The data was public at the time of collection. User names are not preserved. ## Considerations for Using the Data ### Social Impact of Dataset There's a risk of user-deleted content being in this data. The data has NOT been vetted for any content, so there's a risk of harmful text. ### Discussion of Biases The source of the text has a strong demographic bias, being mostly young white men who are vocal their opinions. This constrains both the styles of language and discussion contained in the data, as well as the topics discussed and viewpoints held. ### Other Known Limitations The above limitations apply. ## Additional Information ### Dataset Curators The dataset is curated by the paper's authors. ### Licensing Information The authors distribute this data under Creative Commons attribution license, CC-BY 4.0. An NLP data statement is included in the paper describing the work, [https://aclanthology.org/W19-6122.pdf](https://aclanthology.org/W19-6122.pdf) ### Citation Information ``` @inproceedings{lillie-etal-2019-joint, title = "Joint Rumour Stance and Veracity Prediction", author = "Lillie, Anders Edelbo and Middelboe, Emil Refsgaard and Derczynski, Leon", booktitle = "Proceedings of the 22nd Nordic Conference on Computational Linguistics", month = sep # "{--}" # oct, year = "2019", address = "Turku, Finland", publisher = {Link{\"o}ping University Electronic Press}, url = "https://aclanthology.org/W19-6122", pages = "208--221", } ``` ### Contributions Author-added dataset [@leondz](https://github.com/leondz)
strombergnlp
null
@inproceedings{lehmann2019political, title={Political Stance in Danish}, author={Lehmann, Rasmus and Derczynski, Leon}, booktitle={Proceedings of the 22nd Nordic Conference on Computational Linguistics}, pages={197--207}, year={2019} }
Political stance in Danish. Examples represent statements by politicians and are annotated for, against, or neutral to a given topic/article.
false
1
false
strombergnlp/polstance
2022-10-25T21:42:18.000Z
polstance
false
d617a5df1b7ec95b1b290926e2d1f20c55d2c0b9
[]
[ "annotations_creators:expert-generated", "language_creators:found", "language:da", "license:cc-by-4.0", "multilinguality:monolingual", "size_categories:n<1K", "source_datasets:original", "task_categories:text-classification", "task_ids:sentiment-analysis", "tags:stance-detection" ]
https://huggingface.co/datasets/strombergnlp/polstance/resolve/main/README.md
--- annotations_creators: - expert-generated language_creators: - found language: - da license: - cc-by-4.0 multilinguality: - monolingual size_categories: - n<1K source_datasets: - original task_categories: - text-classification task_ids: - sentiment-analysis paperswithcode_id: polstance pretty_name: Political Stance for Danish tags: - stance-detection --- # Dataset Card for "polstance" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://stromberg.ai/publication/politicalstanceindanish/](https://stromberg.ai/publication/politicalstanceindanish/) - **Repository:** [https://github.com/StrombergNLP/Political-Stance-in-Danish/](https://github.com/StrombergNLP/Political-Stance-in-Danish/) - **Paper:** [https://aclanthology.org/W19-6121/](https://aclanthology.org/W19-6121/) - **Point of Contact:** [Leon Derczynski](https://github.com/leondz) - **Size of downloaded dataset files:** 548 KB - **Size of the generated dataset:** 222 KB - **Total amount of disk used:** 770 KB ### Dataset Summary Political stance in Danish. Examples represent statements by politicians and are annotated for, against, or neutral to a given topic/article. ### Supported Tasks and Leaderboards * ### Languages Danish, bcp47: `da-DK` ## Dataset Structure ### Data Instances #### polstance An example of 'train' looks as follows. ``` { 'id': '0', 'topic': 'integration', 'quote': 'Der kunne jeg godt tænke mig, at der stod mere eksplicit, at de (landene, red.) skal bekæmpe menneskesmuglere og tage imod deres egne borgere', 'label': 2, 'quoteID': '516', 'party': 'Det Konservative Folkeparti', 'politician': 'Naser Khader', } ``` ### Data Fields - `id`: a `string` feature. - `topic`: a `string` expressing a topic. - `quote`: a `string` to be classified for its stance to the topic. - `label`: a class label representing the stance the text expresses towards the target. Full tagset with indices: ``` 0: "against", 1: "neutral", 2: "for", ``` - `quoteID`: a `string` of the internal quote ID. - `party`: a `string` describing the party affiliation of the quote utterer at the time of utterance. - `politician`: a `string` naming the politician who uttered the quote. ### Data Splits | name |train| |---------|----:| |polstance|900 sentences| ## Dataset Creation ### Curation Rationale Collection of quotes from politicians to allow detecting how political quotes orient to issues. ### Source Data #### Initial Data Collection and Normalization The data is taken from proceedings of the Danish parliament, the Folketing - [ft.dk](https://ft.dk). #### Who are the source language producers? Danish polticians ### Annotations #### Annotation process Annotators labelled comments for being against, neutral, or for a specified topic #### Who are the annotators? Danish native speakers, 20s, male, studying Software Design. ### Personal and Sensitive Information The data was public at the time of collection and will remain open public record by law in Denmark. ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations The above limitations apply. ## Additional Information ### Dataset Curators The dataset is curated by the paper's authors. ### Licensing Information The authors distribute this data under Creative Commons attribution license, CC-BY 4.0. ### Citation Information ``` @inproceedings{lehmann2019political, title={Political Stance in Danish}, author={Lehmann, Rasmus and Derczynski, Leon}, booktitle={Proceedings of the 22nd Nordic Conference on Computational Linguistics}, pages={197--207}, year={2019} } ``` ### Contributions Author-added dataset [@leondz](https://github.com/leondz)
strombergnlp
null
@inproceedings{derczynski-kjeldsen-2019-bornholmsk, title = "Bornholmsk Natural Language Processing: Resources and Tools", author = "Derczynski, Leon and Kjeldsen, Alex Speed", booktitle = "Proceedings of the 22nd Nordic Conference on Computational Linguistics", month = sep # "{--}" # oct, year = "2019", address = "Turku, Finland", publisher = {Link{\"o}ping University Electronic Press}, url = "https://aclanthology.org/W19-6138", pages = "338--344", abstract = {This paper introduces language processing resources and tools for Bornholmsk, a language spoken on the island of Bornholm, with roots in Danish and closely related to Scanian. This presents an overview of the language and available data, and the first NLP models for this living, minority Nordic language. Sammenfattnijng p{\aa} borrijnholmst: D{\ae}jnna artikkelijn introduserer naturspr{\aa}gsresurser {\aa} varktoi for borrijnholmst, ed spr{\aa}g a d{\ae}r snakkes p{\aa} {\"o}n Borrijnholm me r{\o}dder i danst {\aa} i n{\ae}r familia me sk{\aa}nst. Artikkelijn gjer ed {\^a}uersyn {\^a}uer spr{\aa}ged {\aa} di datan som fijnnes, {\aa} di fosste NLP mod{\ae}llarna for d{\ae}tta l{\ae}wenes nordiska minnret{\^a}lsspr{\aa}ged.}, }
This corpus introduces language processing resources and tools for Bornholmsk, a language spoken on the island of Bornholm, with roots in Danish and closely related to Scanian. Sammenfattnijng på borrijnholmst: Dæjnna artikkelijn introduserer natursprågsresurser å varktoi for borrijnholmst, ed språg a dær snakkes på ön Borrijnholm me rødder i danst å i nær familia me skånst.
false
1
false
strombergnlp/bornholmsk
2022-10-25T21:40:56.000Z
null
false
8ad62edaaf487414e527d3f59edd6f6b52778b41
[]
[ "annotations_creators:no-annotation", "language_creators:found", "language:da", "license:cc-by-4.0", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "task_categories:text-generation", "task_ids:language-modeling", "language_bcp47:da", "language_bcp47:da-bo...
https://huggingface.co/datasets/strombergnlp/bornholmsk/resolve/main/README.md
--- annotations_creators: - no-annotation language_creators: - found language: - da license: - cc-by-4.0 multilinguality: - monolingual size_categories: - 100K<n<1M source_datasets: - original task_categories: - text-generation task_ids: - language-modeling language_bcp47: - da - da-bornholm --- ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://github.com/StrombergNLP/bornholmsk - **Repository:** https://github.com/StrombergNLP/bornholmsk - **Paper:** https://aclanthology.org/W19-6138/ - **Leaderboard:** [Needs More Information] - **Point of Contact:** [Leon Derczynski](https://github.com/leondz) ### Dataset Summary This corpus introduces language processing resources and tools for Bornholmsk, a language spoken on the island of Bornholm, with roots in Danish and closely related to Scanian. Sammenfattnijng på borrijnholmst: Dæjnna artikkelijn introduserer natursprågsresurser å varktoi for borrijnholmst, ed språg a dær snakkes på ön Borrijnholm me rødder i danst å i nær familia me skånst. For more details, see the paper [Bornholmsk Natural Language Processing: Resources and Tools](https://aclanthology.org/W19-6138/). ### Supported Tasks and Leaderboards * ### Languages Bornholmsk, a language variant of Danish spoken on the island of Bornholm. bcp47: `da-bornholm` ## Dataset Structure ### Data Instances 13169 lines, 175 167 words, 801 KB ### Data Fields `id`: the sentence ID, `int` `text`: the Bornholmsk text, `string` ### Data Splits Monolithic ## Dataset Creation ### Curation Rationale To gather as much digital Bornholmsk together as possible ### Source Data #### Initial Data Collection and Normalization From many places - see paper for details. Sources include poems, songs, translations from Danish, folk stories, dictionary entries. #### Who are the source language producers? Native speakers of Bornholmsk who have produced works in their native language, or translated them to Danish. Much of the data is the result of a community of Bornholmsk speakers volunteering their time across the island in an effort to capture this endangered language. ### Annotations #### Annotation process No annotations #### Who are the annotators? No annotations ### Personal and Sensitive Information Unknown, but low risk of presence, given the source material ## Considerations for Using the Data ### Social Impact of Dataset The purpose of this dataset is to capture Bornholmsk digitally and provide a way for NLP systems to interact with it, and perhaps even spark interest in dealing with the language. ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators This collection of Bornholmsk is curated by Leon Derczynski and Alex Speed Kjeldsen ### Licensing Information Creative Commons Attribution 4.0 ### Citation Information ``` @inproceedings{derczynski-kjeldsen-2019-bornholmsk, title = "Bornholmsk Natural Language Processing: Resources and Tools", author = "Derczynski, Leon and Kjeldsen, Alex Speed", booktitle = "Proceedings of the 22nd Nordic Conference on Computational Linguistics", month = sep # "{--}" # oct, year = "2019", address = "Turku, Finland", publisher = {Link{\"o}ping University Electronic Press}, url = "https://aclanthology.org/W19-6138", pages = "338--344", } ```
strombergnlp
null
@inproceedings{derczynski2013twitter, title={Twitter part-of-speech tagging for all: Overcoming sparse and noisy data}, author={Derczynski, Leon and Ritter, Alan and Clark, Sam and Bontcheva, Kalina}, booktitle={Proceedings of the international conference recent advances in natural language processing ranlp 2013}, pages={198--206}, year={2013} }
Part-of-speech information is basic NLP task. However, Twitter text is difficult to part-of-speech tag: it is noisy, with linguistic errors and idiosyncratic style. This data is the vote-constrained bootstrapped data generate to support state-of-the-art results. The data is about 1.5 million English tweets annotated for part-of-speech using Ritter's extension of the PTB tagset. The tweets are from 2012 and 2013, tokenized using the GATE tokenizer and tagged jointly using the CMU ARK tagger and Ritter's T-POS tagger. Only when both these taggers' outputs are completely compatible over a whole tweet, is that tweet added to the dataset. This data is recommend for use a training data **only**, and not evaluation data. For more details see https://gate.ac.uk/wiki/twitter-postagger.html and https://aclanthology.org/R13-1026.pdf
false
4
false
strombergnlp/twitter_pos_vcb
2022-10-25T21:42:56.000Z
twitter-pos-vcb
false
12ff587afc996106440872be6b3656218fad0e82
[]
[ "annotations_creators:machine-generated", "language_creators:found", "language:en", "license:cc-by-4.0", "multilinguality:monolingual", "size_categories:1M<n<10M", "source_datasets:original", "task_categories:token-classification", "task_ids:part-of-speech" ]
https://huggingface.co/datasets/strombergnlp/twitter_pos_vcb/resolve/main/README.md
--- annotations_creators: - machine-generated language_creators: - found language: - en license: - cc-by-4.0 multilinguality: - monolingual size_categories: - 1M<n<10M source_datasets: - original task_categories: - token-classification task_ids: - part-of-speech paperswithcode_id: twitter-pos-vcb pretty_name: Twitter PoS VCB --- # Dataset Card for "twitter-pos-vcb" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://gate.ac.uk/wiki/twitter-postagger.html](https://gate.ac.uk/wiki/twitter-postagger.html) - **Repository:** [https://github.com/GateNLP/gateplugin-Twitter](https://github.com/GateNLP/gateplugin-Twitter) - **Paper:** [https://aclanthology.org/R13-1026.pdf](https://aclanthology.org/R13-1026.pdf) - **Point of Contact:** [Leon Derczynski](https://github.com/leondz) - **Size of downloaded dataset files:** 4.51 MiB - **Size of the generated dataset:** 26.88 MB - **Total amount of disk used:** 31.39 MB ### Dataset Summary Part-of-speech information is basic NLP task. However, Twitter text is difficult to part-of-speech tag: it is noisy, with linguistic errors and idiosyncratic style. This data is the vote-constrained bootstrapped data generate to support state-of-the-art results. The data is about 1.5 million English tweets annotated for part-of-speech using Ritter's extension of the PTB tagset. The tweets are from 2012 and 2013, tokenized using the GATE tokenizer and tagged jointly using the CMU ARK tagger and Ritter's T-POS tagger. Only when both these taggers' outputs are completely compatible over a whole tweet, is that tweet added to the dataset. This data is recommend for use a training data **only**, and not evaluation data. For more details see https://gate.ac.uk/wiki/twitter-postagger.html and https://aclanthology.org/R13-1026.pdf ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages English, non-region-specific. `bcp47:en` ## Dataset Structure ### Data Instances An example of 'train' looks as follows. ``` ``` ### Data Fields The data fields are the same among all splits. #### twitter_pos_vcb - `id`: a `string` feature. - `tokens`: a `list` of `string` features. - `pos_tags`: a `list` of classification labels (`int`). Full tagset with indices: ```python ``` ### Data Splits | name |tokens|sentences| |---------|----:|---------:| |twitter-pos-vcb|1 543 126| 159 492| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information Creative Commons Attribution 4.0 (CC-BY) ### Citation Information ``` @inproceedings{derczynski2013twitter, title={Twitter part-of-speech tagging for all: Overcoming sparse and noisy data}, author={Derczynski, Leon and Ritter, Alan and Clark, Sam and Bontcheva, Kalina}, booktitle={Proceedings of the international conference recent advances in natural language processing ranlp 2013}, pages={198--206}, year={2013} } ``` ### Contributions Author uploaded ([@leondz](https://github.com/leondz))
strombergnlp
null
@inproceedings{dlamini_zulu_stance, title={Bridging the Domain Gap for Stance Detection for the Zulu language}, author={Dlamini, Gcinizwe and Bekkouch, Imad Eddine Ibrahim and Khan, Adil and Derczynski, Leon}, booktitle={Proceedings of IEEE IntelliSys}, year={2022} }
This is a stance detection dataset in the Zulu language. The data is translated to Zulu by Zulu native speakers, from English source texts. Misinformation has become a major concern in recent last years given its spread across our information sources. In the past years, many NLP tasks have been introduced in this area, with some systems reaching good results on English language datasets. Existing AI based approaches for fighting misinformation in literature suggest automatic stance detection as an integral first step to success. Our paper aims at utilizing this progress made for English to transfers that knowledge into other languages, which is a non-trivial task due to the domain gap between English and the target languages. We propose a black-box non-intrusive method that utilizes techniques from Domain Adaptation to reduce the domain gap, without requiring any human expertise in the target language, by leveraging low-quality data in both a supervised and unsupervised manner. This allows us to rapidly achieve similar results for stance detection for the Zulu language, the target language in this work, as are found for English. We also provide a stance detection dataset in the Zulu language.
false
1
false
strombergnlp/zulu_stance
2022-10-25T21:46:14.000Z
zulu-stance
false
9cd7629e8632e2bf2185a7ae2309b8333248d57e
[]
[ "arxiv:2205.03153", "annotations_creators:expert-generated", "language_creators:found", "language:zu", "license:cc-by-4.0", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "task_categories:text-classification", "task_ids:fact-checking", "task_ids:sentiment-...
https://huggingface.co/datasets/strombergnlp/zulu_stance/resolve/main/README.md
--- annotations_creators: - expert-generated language_creators: - found language: - zu license: - cc-by-4.0 multilinguality: - monolingual size_categories: - 1K<n<10K source_datasets: - original task_categories: - text-classification task_ids: - fact-checking - sentiment-classification paperswithcode_id: zulu-stance pretty_name: ZUstance tags: - stance-detection --- # Dataset Card for "zulu-stance" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://arxiv.org/abs/2205.03153](https://arxiv.org/abs/2205.03153) - **Repository:** - **Paper:** [https://arxiv.org/pdf/2205.03153](https://arxiv.org/pdf/2205.03153) - **Point of Contact:** [Leon Derczynski](https://github.com/leondz) - **Size of downloaded dataset files:** 212.54 KiB - **Size of the generated dataset:** 186.76 KiB - **Total amount of disk used:** 399.30KiB ### Dataset Summary This is a stance detection dataset in the Zulu language. The data is translated to Zulu by Zulu native speakers, from English source texts. Our paper aims at utilizing this progress made for English to transfers that knowledge into other languages, which is a non-trivial task due to the domain gap between English and the target languages. We propose a black-box non-intrusive method that utilizes techniques from Domain Adaptation to reduce the domain gap, without requiring any human expertise in the target language, by leveraging low-quality data in both a supervised and unsupervised manner. This allows us to rapidly achieve similar results for stance detection for the Zulu language, the target language in this work, as are found for English. A natively-translated dataset is used for evaluation of domain transfer. ### Supported Tasks and Leaderboards * ### Languages Zulu (`bcp47:zu`) ## Dataset Structure ### Data Instances #### zulu_stance - **Size of downloaded dataset files:** 212.54 KiB - **Size of the generated dataset:** 186.76 KiB - **Total amount of disk used:** 399.30KiB An example of 'train' looks as follows. ``` { 'id': '0', 'text': 'ubukhulu be-islam buba sobala lapho i-smartphone ifaka i-ramayana njengo-ramadan. #semst', 'target': 'Atheism', 'stance': 1} ``` ### Data Fields - `id`: a `string` feature. - `text`: a `string` expressing a stance. - `target`: a `string` of the target/topic annotated here. - `stance`: a class label representing the stance the text expresses towards the target. Full tagset with indices: ``` 0: "FAVOR", 1: "AGAINST", 2: "NONE", ``` ### Data Splits | name |train| |---------|----:| |zulu_stance|1343 sentences| ## Dataset Creation ### Curation Rationale To enable stance detection in Zulu and also to measure domain transfer in translation ### Source Data #### Initial Data Collection and Normalization The original data is taken from [Semeval2016 task 6: Detecting stance in tweets.](https://aclanthology.org/S16-1003/), and then translated manually to Zulu. #### Who are the source language producers? English-speaking Twitter users. ### Annotations #### Annotation process See [Semeval2016 task 6: Detecting stance in tweets.](https://aclanthology.org/S16-1003/); the annotations are taken from there. #### Who are the annotators? See [Semeval2016 task 6: Detecting stance in tweets.](https://aclanthology.org/S16-1003/); the annotations are taken from there. ### Personal and Sensitive Information The data was public at the time of collection. User names are preserved. ## Considerations for Using the Data ### Social Impact of Dataset There's a risk of user-deleted content being in this data. The data has NOT been vetted for any content, so there's a risk of harmful text. ### Discussion of Biases While the data is in Zulu, the source text is not from or about Zulu-speakers, and so still expresses the social biases and topics found in English-speaking Twitter users. Further, some of the topics are USA-specific. The sentiments and ideas in this dataset do not represent Zulu speakers. ### Other Known Limitations The above limitations apply. ## Additional Information ### Dataset Curators The dataset is curated by the paper's authors. ### Licensing Information The authors distribute this data under Creative Commons attribution license, CC-BY 4.0. ### Citation Information ``` @inproceedings{dlamini_zulu_stance, title={Bridging the Domain Gap for Stance Detection for the Zulu language}, author={Dlamini, Gcinizwe and Bekkouch, Imad Eddine Ibrahim and Khan, Adil and Derczynski, Leon}, booktitle={Proceedings of IEEE IntelliSys}, year={2022} } ``` ### Contributions Author-added dataset [@leondz](https://github.com/leondz)
Elfsong
null
null
null
false
1
false
Elfsong/clef_data
2022-08-29T05:41:54.000Z
null
false
00e360ccbfa82f47287311387af1011f28f7e667
[]
[]
https://huggingface.co/datasets/Elfsong/clef_data/resolve/main/README.md
You should know how to use it:) Just in case, you can email me [mingzhe at nus.edu.sg] if you need any help.
tomasmcz
null
null
null
false
1
false
tomasmcz/word2vec_analogy
2022-04-28T14:07:56.000Z
null
false
3fc0666c45d46d03a9bcd43f5e887dda2727328e
[]
[ "license:apache-2.0" ]
https://huggingface.co/datasets/tomasmcz/word2vec_analogy/resolve/main/README.md
--- license: apache-2.0 --- Adapted from https://github.com/nicholas-leonard/word2vec
Zaratahir123
null
null
null
false
1
false
Zaratahir123/23100065
2022-04-28T16:05:20.000Z
null
false
d4b9a7de7eeea7fe4b1e43186b2e9d2b014779b8
[]
[ "license:mit" ]
https://huggingface.co/datasets/Zaratahir123/23100065/resolve/main/README.md
--- license: mit ---
BigScienceBiasEval
null
""" # TODO: Add description of the dataset here # You can copy an official description _DESCRIPTION =
This is a preliminary version of the bias SHADES dataset for evaluating LMs for social biases.
false
102
false
BigScienceBiasEval/bias-shades
2022-10-03T13:49:04.000Z
null
false
788a707f7a58202e1a9dbd012ff986c935d4d113
[]
[ "license:cc-by-sa-4.0" ]
https://huggingface.co/datasets/BigScienceBiasEval/bias-shades/resolve/main/README.md
--- license: cc-by-sa-4.0 --- Possibly a placeholder dataset for the original here: https://huggingface.co/datasets/bigscience-catalogue-data/bias-shades # Data Statement for SHADES > **How to use this document:** > Fill in each section according to the instructions. Give as much detail as you can, but there's no need to extrapolate. The goal is to help people understand your data when they approach it. This could be someone looking at it in ten years, or it could be you yourself looking back at the data in two years. > For full details, the best source is the original Data Statements paper, here: https://www.aclweb.org/anthology/Q18-1041/ . > Instruction fields are given as blockquotes; delete the instructions when you're done, and provide the file with your data, for example as "DATASTATEMENT.md". The lists in some blocks are designed to be filled in, but it's good to also leave a written description of what's happening, as well as the list. It's fine to skip some fields if the information isn't known. > Only blockquoted content should be deleted; the final about statement should be left intact. Data set name: Bias-Shades Citation (if available): TODO. Data set developer(s): This dataset was compiled by dozens of research scientists through the BigScience open science collaboration. Collaborators, representing numerous cultures and languages, joined the project of their own volition. Data statement author(s): Shayne Longpre, Aurélie Névéol, Shanya Sharma[Add name here if you add/edit the data statement :)]. Others who contributed to this document: N/A License: Creative Commons Attribution-ShareAlike 4.0 (CC BY-SA 4.0). ## A. CURATION RATIONALE > *Explanation.* Which texts were included and what were the goals in selecting texts, both in the original collection and in any further sub-selection? This can be especially important in datasets too large to thoroughly inspect by hand. An explicit statement of the curation rationale can help dataset users make inferences about what other kinds of texts systems trained with them could conceivably generalize to. This dataset was curated by hand-crafting stereotype sentences by native speakers from the culture which is being targeted. An initial set of sentences was inferred from stereotypes expressed in the crowS-pairs data set(Nangia et al.). Native speakers first crafted templates for sentences expressing a stereotype. These templates are marked for gender and plurality of the target nouns, so the template can be reused by substituting different targets. Next, the template-target noun pair combinations were annotated for the veracity/reliability of the expressed stereotype. The resulting sentences express common and less common stereotypes in a variety of cultures and languages. ## B. LANGUAGE VARIETY/VARIETIES > *Explanation.* Languages differ from each other in structural ways that can interact with NLP algorithms. Within a language, regional or social dialects can also show great variation (Chambers and Trudgill, 1998). The language and language variety should be described with a language tag from BCP-47 identifying the language variety (e.g., en-US or yue-Hant-HK), and a prose description of the language variety, glossing the BCP-47 tag and also providing further information (e.g., "English as spoken in Palo Alto, California", or "Cantonese written with traditional characters by speakers in Hong Kong who are bilingual in Mandarin"). * BCP-47 language tags: en-US, fr-FR, hi-IN, es-DO, ar-LY, ru-RU, de-DE, nl-NL, ta-IN. * Language variety description: English spoken by native speakers of the United States, native French people from metropolitan France, native Hindi and Tamil speakers from India, Spanish speakers from the Dominican Republic, Arabic speakers from Libya, Russian speakers from Russia, German speakers from Germany, and Dutch speakers from the Netherlands. ## C. CONTRIBUTOR DEMOGRAPHIC > ## C. SPEAKER DEMOGRAPHIC > *Explanation.* Sociolinguistics has found that variation (in pronunciation, prosody, word choice, and grammar) correlates with speaker demographic characteristics (Labov, 1966), as speakers use linguistic variation to construct and project identities (Eckert and Rickford, 2001). Transfer from native languages (L1) can affect the language produced by non-native (L2) speakers (Ellis, 1994, Ch. 8). A further important type of variation is disordered speech (e.g., dysarthria). Specifications include: Participants to the collection project were recruited through the HuggingFace BigScience project, and specifically the Bias and Fairness Evaluation group. Listed below. Speakers: * [ADD YOURSELF!] * Shayne Longpre: English-speaking, male, 28 years old, culturally Canadian. * Aurélie Névéol: French (native), English and Spanish speaking, female, 44 years old, culturally French (also familiar with American culture) * Shanya Sharma: Hindi(native), English speaking, female, 24 years old, culturally Indian * Margaret Mitchell: English, female, mid-30s, U.S.A. * Maraim Masoud: Arabic, English Speaking female. ## D. ANNOTATOR DEMOGRAPHIC > *Explanation.* What are the demographic characteristics of the annotators and annotation guideline developers? Their own “social address” influences their experience with language and thus their perception of what they are annotating. Specifications include: Participants to the collection project were recruited through the HuggingFace BigScience project, and specifically the Bias and Fairness Evaluation group. Speaker and annotator contributors listed in section C. ## E. SPEECH SITUATION N/A ## F. TEXT CHARACTERISTICS > *Explanation.* Both genre and topic influence the vocabulary and structural characteristics of texts (Biber, 1995), and should be specified. Collected data is a collection of offensive stereotyped statements in numerous languages and cultures. They might be upsetting and/or offensive. Along with these stereotyped statements are annotation judgements of how prevalent/real the expressed stereotypes are in the real world. Some statements were created from templates with substituted target nouns, and therefore may express an uncommon or unlikely stereotype. ## G. RECORDING QUALITY N/A ## H. OTHER > *Explanation.* There may be other information of relevance as well. Please use this space to develop any further categories that are relevant for your dataset. ## I. PROVENANCE APPENDIX This initiative is part of the BigScience Workshop: https://bigscience.huggingface.co/. ## About this document A data statement is a characterization of a dataset that provides context to allow developers and users to better understand how experimental results might generalize, how software might be appropriately deployed, and what biases might be reflected in systems built on the software. Data Statements are from the University of Washington. Contact: [datastatements@uw.edu](mailto:datastatements@uw.edu). This document template is licensed as [CC0](https://creativecommons.org/share-your-work/public-domain/cc0/). This version of the markdown Data Statement is from June 4th 2020. The Data Statement template is based on worksheets distributed at the [2020 LREC workshop on Data Statements](https://sites.google.com/uw.edu/data-statements-for-nlp/), by Emily M. Bender, Batya Friedman, and Angelina McMillan-Major. Adapted to community Markdown template by Leon Dercyznski.
smallv0221
null
null
null
false
1
false
smallv0221/dd
2022-04-29T04:34:50.000Z
null
false
a4e4877f5ae4a8df754d2da61fd9b4f71dc3b6c4
[]
[ "license:apache-2.0" ]
https://huggingface.co/datasets/smallv0221/dd/resolve/main/README.md
--- license: apache-2.0 ---
gusevski
null
null
null
false
4
false
gusevski/factrueval2016
2022-04-29T20:34:48.000Z
null
false
96349b9bd7536aac67f892f5cf36197cd49ea722
[]
[ "arxiv:2005.00614" ]
https://huggingface.co/datasets/gusevski/factrueval2016/resolve/main/README.md
# Dataset Card for FactRuEval-2016 ## Dataset Description - **Point of Contact:** [Guskov Sergey](https://gusevski.com) ### Dataset Summary Evaluation of [Named Entity Recognition](https://www.dialog-21.ru/media/3430/starostinaetal.pdf) and Fact Extraction Systems for Russian. ### Supported Tasks and Leaderboards For each of the tasks tagged for this dataset, give a brief description of the tag, metrics, and suggested models (with a link to their HuggingFace implementation if available). Give a similar description of tasks that were not covered by the structured tag set (repace the `task-category-tag` with an appropriate `other:other-task-name`). - `token-classification`: The dataset can be used to train a model for [NER], which consists in [Token Classification]. Success on this task is typically measured by achieving a *high/low* [metric name](https://huggingface.co/metrics/metric_name). The ([model name](https://huggingface.co/model_name) or [model class](https://huggingface.co/transformers/model_doc/model_class.html)) model currently achieves the following score. *[IF A LEADERBOARD IS AVAILABLE]:* This task has an active leaderboard which can be found at [leaderboard url]() and ranks models based on [metric name](https://huggingface.co/metrics/metric_name) while also reporting [other metric name](https://huggingface.co/metrics/other_metric_name). ### Languages RU. ## Dataset Structure ### Data Instances Provide an JSON-formatted example and brief description of a typical instance in the dataset. If available, provide a link to further examples. ``` { 'data': [{'id':'', 'tokens':[], 'ner_tags':[]},...], ... } ``` Provide any additional information that is not covered in the other sections about the data here. In particular describe any relationships between data points and if these relationships are made explicit. ### Data Fields List and describe the fields present in the dataset. Mention their data type, and whether they are used as input or output in any of the tasks the dataset currently supports. If the data has span indices, describe their attributes, such as whether they are at the character level or word level, whether they are contiguous or not, etc. If the datasets contains example IDs, state whether they have an inherent meaning, such as a mapping to other datasets or pointing to relationships between data points. - `id`: order id - `tokens`: list of tokens - `ner_tags`: list of ner tags ### Data Splits Describe and name the splits in the dataset if there are more than one. Describe any criteria for splitting the data, if used. If their are differences between the splits (e.g. if the training annotations are machine-generated and the dev and test ones are created by humans, or if different numbers of annotators contributed to each example), describe them here. Provide the sizes of each split. As appropriate, provide any descriptive statistics for the features, such as average length. For example: | | Tain | Valid | Test | | ----- | ------ | ----- | ---- | | Input Sentences | | | | | Average Sentence Length | | | | ## Dataset Creation ### Curation Rationale What need motivated the creation of this dataset? What are some of the reasons underlying the major choices involved in putting it together? ### Source Data This section describes the source data (e.g. news text and headlines, social media posts, translated sentences,...) #### Initial Data Collection and Normalization Describe the data collection process. Describe any criteria for data selection or filtering. List any key words or search terms used. If possible, include runtime information for the collection process. If data was collected from other pre-existing datasets, link to source here and to their [Hugging Face version](https://huggingface.co/datasets/dataset_name). If the data was modified or normalized after being collected (e.g. if the data is word-tokenized), describe the process and the tools used. #### Who are the source language producers? State whether the data was produced by humans or machine generated. Describe the people or systems who originally created the data. If available, include self-reported demographic or identity information for the source data creators, but avoid inferring this information. Instead state that this information is unknown. See [Larson 2017](https://www.aclweb.org/anthology/W17-1601.pdf) for using identity categories as a variables, particularly gender. Describe the conditions under which the data was created (for example, if the producers were crowdworkers, state what platform was used, or if the data was found, what website the data was found on). If compensation was provided, include that information here. Describe other people represented or mentioned in the data. Where possible, link to references for the information. ### Annotations If the dataset contains annotations which are not part of the initial data collection, describe them in the following paragraphs. #### Annotation process If applicable, describe the annotation process and any tools used, or state otherwise. Describe the amount of data annotated, if not all. Describe or reference annotation guidelines provided to the annotators. If available, provide interannotator statistics. Describe any annotation validation processes. #### Who are the annotators? If annotations were collected for the source data (such as class labels or syntactic parses), state whether the annotations were produced by humans or machine generated. Describe the people or systems who originally created the annotations and their selection criteria if applicable. If available, include self-reported demographic or identity information for the annotators, but avoid inferring this information. Instead state that this information is unknown. See [Larson 2017](https://www.aclweb.org/anthology/W17-1601.pdf) for using identity categories as a variables, particularly gender. Describe the conditions under which the data was annotated (for example, if the annotators were crowdworkers, state what platform was used, or if the data was found, what website the data was found on). If compensation was provided, include that information here. ### Personal and Sensitive Information State whether the dataset uses identity categories and, if so, how the information is used. Describe where this information comes from (i.e. self-reporting, collecting from profiles, inferring, etc.). See [Larson 2017](https://www.aclweb.org/anthology/W17-1601.pdf) for using identity categories as a variables, particularly gender. State whether the data is linked to individuals and whether those individuals can be identified in the dataset, either directly or indirectly (i.e., in combination with other data). State whether the dataset contains other data that might be considered sensitive (e.g., data that reveals racial or ethnic origins, sexual orientations, religious beliefs, political opinions or union memberships, or locations; financial or health data; biometric or genetic data; forms of government identification, such as social security numbers; criminal history). If efforts were made to anonymize the data, describe the anonymization process. ## Considerations for Using the Data ### Social Impact of Dataset Please discuss some of the ways you believe the use of this dataset will impact society. The statement should include both positive outlooks, such as outlining how technologies developed through its use may improve people's lives, and discuss the accompanying risks. These risks may range from making important decisions more opaque to people who are affected by the technology, to reinforcing existing harmful biases (whose specifics should be discussed in the next section), among other considerations. Also describe in this section if the proposed dataset contains a low-resource or under-represented language. If this is the case or if this task has any impact on underserved communities, please elaborate here. ### Discussion of Biases Provide descriptions of specific biases that are likely to be reflected in the data, and state whether any steps were taken to reduce their impact. For Wikipedia text, see for example [Dinan et al 2020 on biases in Wikipedia (esp. Table 1)](https://arxiv.org/abs/2005.00614), or [Blodgett et al 2020](https://www.aclweb.org/anthology/2020.acl-main.485/) for a more general discussion of the topic. If analyses have been run quantifying these biases, please add brief summaries and links to the studies here. ### Other Known Limitations If studies of the datasets have outlined other limitations of the dataset, such as annotation artifacts, please outline and cite them here. ## Additional Information ### Dataset Curators List the people involved in collecting the dataset and their affiliation(s). If funding information is known, include it here. ### Licensing Information MIT
Mim
null
null
null
false
1
false
Mim/autotrain-data-procell-expert
2022-04-29T08:27:03.000Z
null
false
30d1fddbdb897932513e5974736b46c6fe6b6ceb
[]
[ "task_categories:text-classification" ]
https://huggingface.co/datasets/Mim/autotrain-data-procell-expert/resolve/main/README.md
--- task_categories: - text-classification --- # AutoTrain Dataset for project: procell-expert ## Dataset Descritpion This dataset has been automatically processed by AutoTrain for project procell-expert. ### Languages The BCP-47 code for the dataset's language is unk. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json [ { "text": "We studied the antitumor activity and toxicity of ZD1694 (tomudex), a specific inhibitor of thymidyl[...]", "target": 0 }, { "text": "Here we provide data that human prostate cancer cell lines express the platelet-type isoform of 12-L[...]", "target": 0 } ] ``` ### Dataset Fields The dataset has the following fields (also called "features"): ```json { "text": "Value(dtype='string', id=None)", "target": "ClassLabel(num_classes=2, names=['accept', 'reject'], id=None)" } ``` ### Dataset Splits This dataset is split into a train and validation split. The split sizes are as follow: | Split name | Num samples | | ------------ | ------------------- | | train | 155 | | valid | 40 |
plasticfruits
null
null
null
false
2
false
plasticfruits/wikihow_small
2022-04-29T10:24:34.000Z
null
false
474fa5e109d3675a21add2fc4cc0f201159545ec
[]
[ "language:en", "license:mit" ]
https://huggingface.co/datasets/plasticfruits/wikihow_small/resolve/main/README.md
--- language: en license: mit --- # WikiHow Entries Contains wikiHow question, answer and summary in `.json` format.
muibk
null
@inproceedings{freitag-etal-2021-results, title = {Results of the WMT21 Metrics Shared Task: Evaluating Metrics with Expert-based Human Evaluations on TED and News Domain}, author = {Freitag, Markus and Rei, Ricardo and Mathur, Nitika and Lo, Chi-kiu and Stewart, Craig and Foster, George and Lavie, Alon and Bojar, Ondrej}", booktitle = {Proceedings of the Sixth Conference on Machine Translation}, month = {nov}, year = {2021}, address = {Online}, publisher = {Association for Computational Linguistics}, url = {https://aclanthology.org/2021.wmt-1.73}, pages = {733--774} }
This shared task will examine automatic evaluation metrics for machine translation. We will provide you with MT system outputs along with source text and the human reference translations. We are looking for automatic metric scores for translations at the system-level, and segment-level. We will calculate the system-level, and segment-level correlations of your scores with human judgements. We invite submissions of reference-free metrics in addition to reference-based metrics.
false
2
false
muibk/wmt21_metrics_task
2022-07-12T13:13:25.000Z
null
false
0653eeaccbe5f0a9738220e0b1615f791da248b1
[]
[ "annotations_creators:expert-generated", "language_creators:found", "language_creators:machine-generated", "language_creators:expert-generated", "language:bn-hi", "language:cs-en", "language:de-en", "language:de-fr", "language:en-cs", "language:en-de", "language:en-ha", "language:en-is", "la...
https://huggingface.co/datasets/muibk/wmt21_metrics_task/resolve/main/README.md
--- annotations_creators: - expert-generated language_creators: - found - machine-generated - expert-generated language: - bn-hi - cs-en - de-en - de-fr - en-cs - en-de - en-ha - en-is - en-ja - en-ru - en-zh - fr-de - ha-en - hi-bn - is-en - ja-en - ru-en - xh-zh - zh-en - zu-xh license: - unknown multilinguality: - translation paperswithcode_id: null pretty_name: WMT21 Metrics Shared Task size_categories: - 100K<n<1M source_datasets: [] task_categories: - translation task_ids: [] --- # Dataset Card for WMT21 Metrics Task ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [WMT21 Metrics Shared Task](https://www.statmt.org/wmt21/metrics-task.html) - **Repository:** [MT Metrics Eval Github Repository](https://github.com/google-research/mt-metrics-eval) - **Paper:** [Paper](https://aclanthology.org/2021.wmt-1.73/) ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The dataset comprises twenty language pairs: - Bengali-Hindi (`bn-hi`) - Czech-English (`cs-en`) - German-English (`de-en`) - German-French (`de-fr`) - English-Czech (`en-cs`) - English-German (`en-de`) - English-Hausa (`en-ha`) - English-Icelandic (`en-is`) - English-Japanese (`en-ja`) - English-Russian (`en-ru`) - English-Chinese (`en-zh`) - French-German (`fr-de`) - Hausa-English (`ha-en`) - Hindi-Bengali (`hi-bn`) - Icelandic-English (`is-en`) - Japenese-English (`ja-en`) - Russian-English (`ru-en`) - Xhosa-Zulu (`xh-zu`) - Chinese-English (`zh-en`) - Zulu-Xhosa (`zu-xh`) ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@github-username](https://github.com/<github-username>) for adding this dataset.
rish16
null
null
null
false
1
false
rish16/cs4243-database-dict
2022-04-29T12:25:57.000Z
null
false
5f2f2089d47d9d4ba9d20d7cd792703d85b554bc
[]
[ "license:mit" ]
https://huggingface.co/datasets/rish16/cs4243-database-dict/resolve/main/README.md
--- license: mit ---
jamescalam
null
null
null
false
5
false
jamescalam/world-cities-geo
2022-04-29T18:34:46.000Z
null
false
8a8252c8c8e8c9a8f7ecaeb27bae1ac8b2313ab0
[]
[]
https://huggingface.co/datasets/jamescalam/world-cities-geo/resolve/main/README.md
Dataset containing city, country, region, and continents alongside their longitude and latitude co-ordinates. Cartesian coordinates are provided in x, y, z features.
dalle-mini
null
null
null
false
1
false
dalle-mini/vqgan-pairs
2022-10-25T13:56:57.000Z
null
false
03f4a9bf2ec961792476aee76a9fa8722ff8dc1e
[]
[ "license:cc-by-4.0", "license:cc-by-2.0", "license:unknown", "source_datasets:Open Images", "task_categories:other", "tags:super-resolution", "tags:image-enhancement" ]
https://huggingface.co/datasets/dalle-mini/vqgan-pairs/resolve/main/README.md
--- license: - cc-by-4.0 - cc-by-2.0 - unknown source_datasets: - Open Images task_categories: - other task_ids: [] pretty_name: VQGAN Pairs tags: - super-resolution - image-enhancement --- # VQGAN Pairs This dataset contains ~2.4 million image pairs intended for improvement of image quality in VQGAN predictions. Each pair consists of: - A 512x512 crop of an image taken from Open Images. - A 256x256 image encoded and decoded using VQGAN, corresponding to the same image crop as the original. This is the VQGAN implementation that was used for encoding and decoding: https://github.com/patil-suraj/vqgan-jax # License This dataset is created using Open Images, which has the following license: The annotations are licensed by Google LLC under CC BY 4.0 license. The images are listed as having a CC BY 2.0 license. Note: while we tried to identify images that are licensed under a Creative Commons Attribution license, we make no representations or warranties regarding the license status of each image and you should verify the license for each image yourself.
Seledorn
null
null
null
false
1
false
Seledorn/SwissProt-EC
2022-04-30T15:12:33.000Z
null
false
3b03eda1bd275edf2d9caaccf07d32d2c237c0d2
[]
[ "language:protein sequences", "datasets:Swissprot", "tags:Protein", "tags:Enzyme Commission", "tags:EC" ]
https://huggingface.co/datasets/Seledorn/SwissProt-EC/resolve/main/README.md
--- language: - protein sequences datasets: - Swissprot tags: - Protein - Enzyme Commission - EC --- Swissprot is a high quality manually annotated protein database. The dataset contains annotations with the functional properties of the proteins. Here we extract proteins with Enzyme Commission labels. The dataset is ported from Protinfer: https://github.com/google-research/proteinfer. The EC-labels are extracted and indexed, the mapping is provided in `idx_mapping.json`. Proteins without EC tags are removed.
Seledorn
null
null
null
false
1
false
Seledorn/SwissProt-Pfam
2022-04-30T15:15:55.000Z
null
false
0db4e0ea3153c4caa5c49e7387f65b78a8996148
[]
[ "language:protein sequences", "datasets:Swissprot", "tags:Protein", "tags:PFam" ]
https://huggingface.co/datasets/Seledorn/SwissProt-Pfam/resolve/main/README.md
--- language: - protein sequences datasets: - Swissprot tags: - Protein - PFam --- Swissprot is a high quality manually annotated protein database. The dataset contains annotations with the functional properties of the proteins. Here we extract proteins with PFam labels. The dataset is ported from Protinfer: https://github.com/google-research/proteinfer. The Pfam-labels are extracted and indexed, the mapping is provided in `idx_mapping.json`. Proteins without Pfam tags are removed.
Seledorn
null
null
null
false
1
false
Seledorn/SwissProt-GO
2022-04-30T15:16:48.000Z
null
false
f64128a2e9e7a2756daacee8cc00e9b86142e19e
[]
[ "language:protein sequences", "datasets:Swissprot", "tags:Protein", "tags:Gene Ontology", "tags:GO" ]
https://huggingface.co/datasets/Seledorn/SwissProt-GO/resolve/main/README.md
--- language: - protein sequences datasets: - Swissprot tags: - Protein - Gene Ontology - GO --- Swissprot is a high quality manually annotated protein database. The dataset contains annotations with the functional properties of the proteins. Here we extract proteins with Gene Ontology labels. The dataset is ported from Protinfer: https://github.com/google-research/proteinfer. The GO-labels are extracted and indexed, the mapping is provided in `idx_mapping.json`. Proteins without GO tags are removed.
samhellkill
null
null
null
false
1
false
samhellkill/spacekitty-v1
2022-04-30T06:33:09.000Z
null
false
edc48764e7faeea87dbc8b157ddae26d3fb62408
[]
[ "license:other" ]
https://huggingface.co/datasets/samhellkill/spacekitty-v1/resolve/main/README.md
--- license: other ---
charly
null
null
null
false
1
false
charly/test
2022-04-30T17:17:22.000Z
null
false
d59c4caa6cd95db6dea4a389b3195404aeaf5d5d
[]
[ "license:apache-2.0" ]
https://huggingface.co/datasets/charly/test/resolve/main/README.md
--- license: apache-2.0 ---
lightonai
null
null
null
false
1
false
lightonai/SwissProt-EC-leaf
2022-05-05T09:54:42.000Z
null
false
74c9c46cc27003593171ef289c364f42d0f4286b
[]
[ "language:protein sequences", "datasets:Swissprot", "tags:Protein", "tags:Enzyme Commission" ]
https://huggingface.co/datasets/lightonai/SwissProt-EC-leaf/resolve/main/README.md
--- language: - protein sequences datasets: - Swissprot tags: - Protein - Enzyme Commission --- # Dataset Swissprot is a high quality manually annotated protein database. The dataset contains annotations with the functional properties of the proteins. Here we extract proteins with Enzyme Commission labels. The dataset is ported from Protinfer: https://github.com/google-research/proteinfer. The leaf level EC-labels are extracted and indexed, the mapping is provided in `idx_mapping.json`. Proteins without leaf-level-EC tags are removed. ## Example The protein Q87BZ2 have the following EC tags. EC:2.-.-.- (Transferases) EC:2.7.-.- (Transferring phosphorus-containing groups) EC:2.7.1.- (Phosphotransferases with an alcohol group as acceptor) EC:2.7.1.30 (Glycerol kinase) We only extract the leaf level labels, here EC:2.7.1.30, corresponding to glycerol kinase.
osyvokon
null
null
null
false
1
false
osyvokon/wiki-edits-uk
2022-07-02T19:06:25.000Z
null
false
dc89ed1cc59ea92c19958c81c58070a2e95b02ab
[]
[ "annotations_creators:crowdsourced", "language_creators:crowdsourced", "language:uk-UA", "license:cc-by-3.0", "multilinguality:monolingual", "multilinguality:translation", "size_categories:1M<n<10M", "source_datasets:original", "task_categories:other" ]
https://huggingface.co/datasets/osyvokon/wiki-edits-uk/resolve/main/README.md
--- annotations_creators: - crowdsourced language_creators: - crowdsourced language: - uk-UA license: - cc-by-3.0 multilinguality: - monolingual - translation pretty_name: 'Ukrainian Wikipedia edits ' size_categories: - 1M<n<10M source_datasets: - original task_categories: - other task_ids: [] --- # Ukrainian Wikipedia Edits ### Dataset summary A collection of over 5M sentence edits extracted from Ukrainian Wikipedia history revisions. Edits were filtered by edit distance and sentence length. This makes them usable for grammatical error correction (GEC) or spellchecker models pre-training. ### Supported Tasks and Leaderboards * Ukrainian grammatical error correction (GEC) - see [UA-GEC](https://github.com/grammarly/ua-gec) * Ukrainian spelling correction ### Languages Ukrainian ## Dataset Structure ### Data Fields * `src` - sentence before edit * `tgt` - sentence after edit ### Data Splits * `full/train` contains all the data (5,243,376 samples) * `tiny/train` contains a 5000 examples sample. ## Dataset Creation Latest full Ukrainian Wiki dump were used as of 2022-04-30. It was processed with the [wikiedits](https://github.com/snukky/wikiedits) and custom scripts. ### Source Data #### Initial Data Collection and Normalization Wikipedia #### Who are the source language producers? Wikipedia writers ### Annotations #### Annotation process Annotations inferred by comparing two subsequent page revisions. #### Who are the annotators? People who edit Wikipedia pages. ### Personal and Sensitive Information No ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations The data is noisy. In addition to GEC and spelling edits, it contains a good chunk of factual changes and vandalism. More task-specific filters could help. ## Additional Information ### Dataset Curators [Oleksiy Syvokon](https://github.com/asivokon) ### Licensing Information CC-BY-3.0 ### Citation Information ``` @inproceedings{wiked2014, author = {Roman Grundkiewicz and Marcin Junczys-Dowmunt}, title = {The WikEd Error Corpus: A Corpus of Corrective Wikipedia Edits and its Application to Grammatical Error Correction}, booktitle = {Advances in Natural Language Processing -- Lecture Notes in Computer Science}, editor = {Adam Przepiórkowski and Maciej Ogrodniczuk}, publisher = {Springer}, year = {2014}, volume = {8686}, pages = {478--490}, url = {http://emjotde.github.io/publications/pdf/mjd.poltal2014.draft.pdf} } ``` ### Contributions [@snukky](https://github.com/snukky) created tools for dataset processing. [@asivokon](https://github.com/asivokon) generated this dataset.
defector
null
null
null
false
1
false
defector/autotrain-data-company
2022-10-25T10:12:59.000Z
null
false
c517b46f5b0574f716b9e3b173deb85d4db74236
[]
[ "language:en" ]
https://huggingface.co/datasets/defector/autotrain-data-company/resolve/main/README.md
--- language: - en --- # AutoTrain Dataset for project: company ## Dataset Descritpion This dataset has been automatically processed by AutoTrain for project company. ### Languages The BCP-47 code for the dataset's language is en. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json [ { "tokens": [ "sahil", "prasad", "president", "www", "swimcentre", "com", "banik", "baalkrishan", "gandhi", "com", "no", "satish", "nagar", "hisar" ], "tags": [ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ] }, { "tokens": [ "olivia", "wilson", "real", "estate", "agent", "reallygreatsite", "com", "anywhere", "st", "any", "city", "st", "www", "reallygreatsite", "com" ], "tags": [ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ] } ] ``` ### Dataset Fields The dataset has the following fields (also called "features"): ```json { "tokens": "Sequence(feature=Value(dtype='string', id=None), length=-1, id=None)", "tags": "Sequence(feature=ClassLabel(num_classes=2, names=['0', '9'], id=None), length=-1, id=None)" } ``` ### Dataset Splits This dataset is split into a train and validation split. The split sizes are as follow: | Split name | Num samples | | ------------ | ------------------- | | train | 999651 | | valid | 499630 |
Filippo
null
@dataset{osdg_2022_6393942, author = {OSDG and UNDP IICPSD SDG AI Lab and PPMI}, title = {OSDG Community Dataset (OSDG-CD)}, month = apr, year = 2022, note = {{This CSV file uses UTF-8 character encoding. For easy access on MS Excel, open the file using Data → From Text/CSV. Please split CSV data into different columns by using a TAB delimiter.}}, publisher = {Zenodo}, version = {2022.04}, doi = {10.5281/zenodo.6393942}, url = {https://doi.org/10.5281/zenodo.6393942} }
The OSDG Community Dataset (OSDG-CD) is a public dataset of thousands of text excerpts, which were validated by approximately 1,000 OSDG Community Platform (OSDG-CP) citizen scientists from over 110 countries, with respect to the Sustainable Development Goals (SDGs).
false
1
false
Filippo/osdg_cd
2022-10-25T10:13:06.000Z
null
false
bc0f4e55534bfc6c52c3fd86928ccedef7aff585
[]
[ "annotations_creators:crowdsourced", "language_creators:crowdsourced", "language:en", "license:cc-by-4.0", "multilinguality:monolingual", "size_categories:10K<n<100K", "task_categories:text-classification", "task_ids:natural-language-inference" ]
https://huggingface.co/datasets/Filippo/osdg_cd/resolve/main/README.md
--- annotations_creators: - crowdsourced language_creators: - crowdsourced language: - en license: - cc-by-4.0 multilinguality: - monolingual size_categories: - 10K<n<100K task_categories: - text-classification task_ids: - natural-language-inference pretty_name: OSDG Community Dataset (OSDG-CD) --- # Dataset Card for OSDG-CD ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [OSDG-CD homepage](https://doi.org/10.5281/zenodo.6393942) ### Dataset Summary The OSDG Community Dataset (OSDG-CD) is a public dataset of thousands of text excerpts, which were validated by approximately 1,000 OSDG Community Platform (OSDG-CP) citizen scientists from over 110 countries, with respect to the Sustainable Development Goals (SDGs). > **_NOTE:_** There are currently no examples for SDGs 16 and 17. See [this GitHub issue](https://github.com/osdg-ai/osdg-data/issues/3). ### Supported Tasks and Leaderboards TBD ### Languages The language of the dataset is English. ## Dataset Structure ### Data Instances For each instance, there is a string for the text, a string for the SDG, and an integer for the label. ``` {'text': 'Each section states the economic principle, reviews international good practice and discusses the situation in Brazil.', 'label': 5} ``` The average token count for the premises and hypotheses are given below: | Feature | Mean Token Count | | ---------- | ---------------- | | Premise | 14.1 | | Hypothesis | 8.3 | ### Data Fields - `doi`: Digital Object Identifier of the original document - `text_id`: unique text identifier - `text`: text excerpt from the document - `sdg`: the SDG the text is validated against - `label`: an integer from `0` to `17` which corresponds to the `sdg` field - `labels_negative`: the number of volunteers who rejected the suggested SDG label - `labels_positive`: the number of volunteers who accepted the suggested SDG label - `agreement`: agreement score based on the formula ### Data Splits The OSDG-CD dataset has 1 splits: _train_. | Dataset Split | Number of Instances in Split | | ------------- |----------------------------- | | Train | 32,327 | ## Dataset Creation ### Curation Rationale The [The OSDG Community Dataset (OSDG-CD)](https://doi.org/10.5281/zenodo.6393942) was developed as a benchmark for ... with the goal of producing a dataset large enough to train models using neural methodologies. ### Source Data #### Initial Data Collection and Normalization TBD #### Who are the source language producers? TBD ### Annotations #### Annotation process TBD #### Who are the annotators? TBD ### Personal and Sensitive Information The dataset does not contain any personal information about the authors or the crowdworkers. ## Considerations for Using the Data ### Social Impact of Dataset TBD ## Additional Information TBD ### Dataset Curators TBD ### Licensing Information The OSDG Community Dataset (OSDG-CD) is licensed under a [Creative Commons Attribution 4.0 International License](http://creativecommons.org/licenses/by/4.0/). ### Citation Information ``` @dataset{osdg_2022_6393942, author = {OSDG and UNDP IICPSD SDG AI Lab and PPMI}, title = {OSDG Community Dataset (OSDG-CD)}, month = apr, year = 2022, note = {{This CSV file uses UTF-8 character encoding. For easy access on MS Excel, open the file using Data → From Text/CSV. Please split CSV data into different columns by using a TAB delimiter.}}, publisher = {Zenodo}, version = {2022.04}, doi = {10.5281/zenodo.6393942}, url = {https://doi.org/10.5281/zenodo.6393942} } ``` ### Contributions TBD
Anon126
null
@InProceedings{huggingface:dataset, title = {A great new dataset}, author={huggingface, Inc. }, year={2020} }
false
1
false
Anon126/my-raft-submission
2022-05-01T10:50:18.000Z
null
false
4ed58453467ac729dd815696584b8fad8dac4646
[]
[ "benchmark:raft", "type:prediction", "submission_name:none" ]
https://huggingface.co/datasets/Anon126/my-raft-submission/resolve/main/README.md
--- benchmark: raft type: prediction submission_name: none --- # RAFT submissions for my-raft-submission ## Submitting to the leaderboard To make a submission to the [leaderboard](https://huggingface.co/spaces/ought/raft-leaderboard), there are three main steps: 1. Generate predictions on the unlabeled test set of each task 2. Validate the predictions are compatible with the evaluation framework 3. Push the predictions to the Hub! See the instructions below for more details. ### Rules 1. To prevent overfitting to the public leaderboard, we only evaluate **one submission per week**. You can push predictions to the Hub as many times as you wish, but we will only evaluate the most recent commit in a given week. 2. Transfer or meta-learning using other datasets, including further pre-training on other corpora, is allowed. 3. Use of unlabeled test data is allowed, as is it always available in the applied setting. For example, further pre-training using the unlabeled data for a task would be permitted. 4. Systems may be augmented with information retrieved from the internet, e.g. via automated web searches. ### Submission file format For each task in RAFT, you should create a CSV file called `predictions.csv` with your model's predictions on the unlabeled test set. Each file should have exactly 2 columns: * ID (int) * Label (string) See the dummy predictions in the `data` folder for examples with the expected format. Here is a simple example that creates a majority-class baseline: ```python from pathlib import Path import pandas as pd from collections import Counter from datasets import load_dataset, get_dataset_config_names tasks = get_dataset_config_names("ought/raft") for task in tasks: # Load dataset raft_subset = load_dataset("ought/raft", task) # Compute majority class over training set counter = Counter(raft_subset["train"]["Label"]) majority_class = counter.most_common(1)[0][0] # Load predictions file preds = pd.read_csv(f"data/{task}/predictions.csv") # Convert label IDs to label names preds["Label"] = raft_subset["train"].features["Label"].int2str(majority_class) # Save predictions preds.to_csv(f"data/{task}/predictions.csv", index=False) ``` As you can see in the example, each `predictions.csv` file should be stored in the task's subfolder in `data` and at the end you should have something like the following: ``` data ├── ade_corpus_v2 │ ├── predictions.csv │ └── task.json ├── banking_77 │ ├── predictions.csv │ └── task.json ├── neurips_impact_statement_risks │ ├── predictions.csv │ └── task.json ├── one_stop_english │ ├── predictions.csv │ └── task.json ├── overruling │ ├── predictions.csv │ └── task.json ├── semiconductor_org_types │ ├── predictions.csv │ └── task.json ├── systematic_review_inclusion │ ├── predictions.csv │ └── task.json ├── tai_safety_research │ ├── predictions.csv │ └── task.json ├── terms_of_service │ ├── predictions.csv │ └── task.json ├── tweet_eval_hate │ ├── predictions.csv │ └── task.json └── twitter_complaints ├── predictions.csv └── task.json ``` ### Validate your submission To ensure that your submission files are correctly formatted, run the following command from the root of the repository: ``` python cli.py validate ``` If everything is correct, you should see the following message: ``` All submission files validated! ✨ 🚀 ✨ Now you can make a submission 🤗 ``` ### Push your submission to the Hugging Face Hub! The final step is to commit your files and push them to the Hub: ``` python cli.py submit ``` If there are no errors, you should see the following message: ``` Submission successful! 🎉 🥳 🎉 Your submission will be evaulated on Sunday 05 September 2021 ⏳ ``` where the evaluation is run every Sunday and your results will be visible on the leaderboard.
Rodion
null
null
null
false
1
false
Rodion/uno_sustainable_development_goals
2022-05-01T13:54:31.000Z
null
false
869f3ba009c4258ae2d272e664931404de6ec67d
[]
[ "license:afl-3.0" ]
https://huggingface.co/datasets/Rodion/uno_sustainable_development_goals/resolve/main/README.md
--- license: afl-3.0 ---
NazaGara
null
@inproceedings{, title = "", author = "Garagiola, Nazareno", year = "2022", url = "" }
Dataset used to train a NER model
false
1
false
NazaGara/wikiner-es
2022-08-14T15:01:57.000Z
null
false
c3d2d2159db2fb34ca8ef05cbf96f21addeeea8b
[]
[ "license:cc" ]
https://huggingface.co/datasets/NazaGara/wikiner-es/resolve/main/README.md
annotations_creators: - automatic language_creators: - found languages: - es-AR licenses: - cc0-1.0 multilinguality: - monolingual paperswithcode_id: pretty_name: wikiner size_categories: - 100K<n<1M source_datasets: - original task_categories: - token-classification task_ids: - named-entity-recognition --- license: cc --- # Dataset Card for wikiner ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** [Needs More Information] - **Repository:** [Needs More Informatio] - **Paper:** [Learning multilingual named entity recognition from Wikipedia](https://doi.org/10.1016/j.artint.2012.03.006) - **Leaderboard:** [Needs More Information] - **Point of Contact:** [NazaGara](ngaragiola430@mi.unc.edu.ar) ### Dataset Summary Named entities are phrases that contain the names of persons, organizations, locations, times and quantities. Example: [PER Wolff] , currently a journalist in [LOC Argentina] , played with [PER Del Bosque] in the final years of the seventies in [ORG Real Madrid] . ### Supported Tasks and Leaderboards Named Entity Recognition (NER) is a subtask of Information Extraction. Different NER systems were evaluated as a part of the Sixth Message Understanding Conference in 1995 (MUC6). The target language was English. The participating systems performed well. However, many of them used language-specific resources for performing the task and it is unknown how they would have performed on another language than English. After 1995 NER systems have been developed for some European languages and a few Asian languages. There have been at least two studies that have applied one NER system to different languages. Palmer and Day [PD97] have used statistical methods for finding named entities in newswire articles in Chinese, English, French, Japanese, Portuguese and Spanish. They found that the difficulty of the NER task was different for the six languages but that a large part of the task could be performed with simple methods. Cucerzan and Yarowsky [CY99] used both morphological and contextual clues for identifying named entities in English, Greek, Hindi, Rumanian and Turkish. With minimal supervision, they obtained overall F measures between 40 and 70, depending on the languages used. - `named-entity-recognition`: The performance in this task is measured with [F1](https://huggingface.co/metrics/f1) (higher is better). A named entity is correct only if it is an exact match of the corresponding entity in the data. This dataset was used in order to train a Spanish NER model using [BETO](https://huggingface.co/dccuchile/bert-base-spanish-wwm-cased). ### Languages The only supported language is spanish (es). ## Dataset Structure ### Data Fields The dictionary to map the id to the Label names is: { 0: 'O', 1: 'B-PER', 2: 'I-PER', 3: 'B-ORG', 4: 'I-ORG', 5: 'B-LOC', 6: 'I-LOC', 7: 'B-MISC', 8: 'I-MISC' } ### Data Splits The only split is the train split. Number of examples = 128355 ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? Created by Nothman et al. at 2013. ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information [Needs More Information] ### Citation Information [Needs More Information]
hongdijk
null
null
null
false
1
false
hongdijk/kor_nlu_hufsice2
2022-05-23T16:33:29.000Z
null
false
082b1de183522ccd43858713564b51b36ee28f93
[]
[ "license:other" ]
https://huggingface.co/datasets/hongdijk/kor_nlu_hufsice2/resolve/main/README.md
--- license: other ---
hongdijk
null
null
null
false
1
false
hongdijk/kor_nlu_hufs
2022-05-20T10:25:34.000Z
null
false
5cbdab6346ed40a167a4dca0bbd91ef7eeda736d
[]
[ "license:cc-by-sa-4.0" ]
https://huggingface.co/datasets/hongdijk/kor_nlu_hufs/resolve/main/README.md
--- license: cc-by-sa-4.0 ---
kimtaehyuk11
null
@misc{park2021klue, title={KLUE: Korean Language Understanding Evaluation}, author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho}, year={2021}, eprint={2105.09680}, archivePrefix={arXiv}, primaryClass={cs.CL} }
KLUE (Korean Language Understanding Evaluation) Korean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language understanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible to anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain unambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics for every task, thus establishing fair comparison across Korean language models.
false
1
false
kimtaehyuk11/klue
2022-05-01T19:55:19.000Z
null
false
7f9afbbf365b4662fc8a8b677ee94f33d4073752
[]
[ "license:cc-by-sa-4.0" ]
https://huggingface.co/datasets/kimtaehyuk11/klue/resolve/main/README.md
--- license: cc-by-sa-4.0 ---
Diegomejia
null
null
null
false
1
false
Diegomejia/ds1ucb
2022-05-02T02:03:54.000Z
null
false
c44ea85c786b78a24aeefc0fcbe5c5c66648f720
[]
[ "license:mit" ]
https://huggingface.co/datasets/Diegomejia/ds1ucb/resolve/main/README.md
--- license: mit ---
google
null
@article{srinivasan2021wit, title={WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning}, author={Srinivasan, Krishna and Raman, Karthik and Chen, Jiecao and Bendersky, Michael and Najork, Marc}, journal={arXiv preprint arXiv:2103.01913}, year={2021} }
Wikipedia-based Image Text (WIT) Dataset is a large multimodal multilingual dataset. WIT is composed of a curated set of 37.6 million entity rich image-text examples with 11.5 million unique images across 108 Wikipedia languages. Its size enables WIT to be used as a pretraining dataset for multimodal machine learning models.
false
27
false
google/wit
2022-07-04T10:47:07.000Z
wit
false
b52b6486b38d74ddaf95626b15e0f0c50fa5e959
[]
[ "arxiv:2103.01913", "annotations_creators:machine-generated", "language_creators:found", "language:af", "language:ar", "language:ast", "language:azb", "language:be", "language:bg", "language:bn", "language:br", "language:ca", "language:cs", "language:cy", "language:da", "language:de", ...
https://huggingface.co/datasets/google/wit/resolve/main/README.md
--- annotations_creators: - machine-generated language_creators: - found language: - af - ar - ast - azb - be - bg - bn - br - ca - cs - cy - da - de - el - en - eo - es - et - eu - fa - fi - fr - fy - ga - gl - hr - hu - hy - id - it - iw - ja - ka - ko - la - lt - lv - mk - ml - ms - nl - nn - 'no' - pl - pt - ro - ru - sk - sl - sr - sv - th - tr - uk - ur - vi - vo - zh license: - cc-by-sa-3.0 multilinguality: - multilingual paperswithcode_id: wit pretty_name: Wikipedia-based Image Text size_categories: - 10M<n<100M source_datasets: - original - extended|wikipedia task_categories: - text-retrieval - image-to-text task_ids: - text-retrieval-other-text-image-retrieval - image-captioning --- # Dataset Card for WIT ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Dataset Preprocessing](#dataset-preprocessing) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [WIT homepage](https://github.com/google-research-datasets/wit) - **Repository:** [WIT repository](https://github.com/google-research-datasets/wit) - **Paper:** [WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning ](https://arxiv.org/abs/2103.01913) - **Leaderboard:** [WIT leaderboard](https://www.kaggle.com/c/wikipedia-image-caption) - **Point of Contact:** [WIT e-mail](mailto:wit-dataset@google.com) ### Dataset Summary Wikipedia-based Image Text (WIT) Dataset is a large multimodal multilingual dataset. WIT is composed of a curated set of 37.6 million entity rich image-text examples with 11.5 million unique images across 108 Wikipedia languages. Its size enables WIT to be used as a pretraining dataset for multimodal machine learning models. A few unique advantages of WIT: * The largest multimodal dataset (time of this writing) by the number of image-text examples. * A massively multilingual (first of its kind) with coverage for over 100+ languages. * A collection of diverse set of concepts and real world entities. * Brings forth challenging real-world test sets. ### Dataset Preprocessing This dataset doesn't download the images locally by default. Instead, it exposes URLs to the images. To fetch the images, use the following code: ```python from concurrent.futures import ThreadPoolExecutor from functools import partial import io import urllib import PIL.Image from datasets import load_dataset from datasets.utils.file_utils import get_datasets_user_agent def fetch_single_image(image_url, timeout=None, retries=0): for _ in range(retries + 1): try: request = urllib.request.Request( image_url, data=None, headers={"user-agent": get_datasets_user_agent()}, ) with urllib.request.urlopen(request, timeout=timeout) as req: image = PIL.Image.open(io.BytesIO(req.read())) break except Exception: image = None return image def fetch_images(batch, num_threads, timeout=None, retries=0): fetch_single_image_with_args = partial(fetch_single_image, timeout=timeout, retries=retries) with ThreadPoolExecutor(max_workers=num_threads) as executor: batch["image"] = list(executor.map(fetch_single_image_with_args, batch["image_url"])) return batch num_threads = 20 dset = load_dataset("wit") dset = dset.map(fetch_images, batched=True, batch_size=100, fn_kwargs={"num_threads": num_threads}) ``` ### Supported Tasks and Leaderboards - `image-captioning`: This dataset can be used to train a model for image captioning where the goal is to predict a caption given the image. - `text-retrieval`: The goal in this task is to build a model that retrieves the text closest to an image. In these tasks, any combination of the `caption_reference_description`, `caption_attribution_description` and `caption_alt_text_description` fields can be used as the input text/caption. ### Languages The dataset contains examples from all Wikipedia languages, with the following stats: Image-Text | # Lang | Uniq. Images | # Lang ------------ | ------ | ------------- | ------ total > 1M | 9 | images > 1M | 6 total > 500K | 10 | images > 500K | 12 total > 100K | 36 | images > 100K | 35 total > 50K | 15 | images > 50K | 17 total > 14K | 38 | images > 13K | 38 ## Dataset Structure ### Data Instances ``` { 'language': 'en', 'page_url': 'https://en.wikipedia.org/wiki/Oxydactylus', 'image_url': 'https://upload.wikimedia.org/wikipedia/commons/5/5f/Oxydactylus_longipes_fm.jpg', 'page_title': 'Oxydactylus', 'section_title': None, 'hierarchical_section_title': 'Oxydactylus', 'caption_reference_description': None, 'caption_attribution_description': 'English: Mounted skeleton of Oxydactylus longipes in the Field Museum of Natural History.', 'caption_alt_text_description': None, 'mime_type': 'image/jpeg', 'original_height': 3564, 'original_width': 2748, 'is_main_image': True, 'attribution_passes_lang_id': True, 'page_changed_recently': True, 'context_page_description': 'Oxydactylus is an extinct genus of camelid endemic to North America. It lived from the Late Oligocene to the Middle Miocene, existing for approximately 14 million years. The name is from the Ancient Greek οξύς and δάκτυλος.\nThey had very long legs and necks, and were probably adapted to eating high vegetation, much like modern giraffes. Unlike modern camelids, they had hooves, rather than tough sole-pads, and splayed toes.', 'context_section_description': 'Oxydactylus is an extinct genus of camelid endemic to North America. It lived from the Late Oligocene to the Middle Miocene (28.4–13.7 mya), existing for approximately 14 million years. The name is from the Ancient Greek οξύς (oxys, "sharp")and δάκτυλος (daktylos, "finger").\n \nThey had very long legs and necks, and were probably adapted to eating high vegetation, much like modern giraffes. Unlike modern camelids, they had hooves, rather than tough sole-pads, and splayed toes.' } ``` ### Data Fields - `language`: Language code depicting wikipedia language of the page - `page_url`: URL to wikipedia page - `image_url`: URL to wikipedia image - `page_title`: Wikipedia page's title - `section_title`: Section's title - `hierarchical_section_title`: Hierarchical section's title - `caption_reference_description`: This is the caption that is visible on the wiki page directly below the image. - `caption_attribution_description`: This is the text found on the Wikimedia page of the image. This text is common to all occurrences of that image across all Wikipedias and thus can be in a language different to the original page article. - `caption_alt_text_description`: This is the “alt” text associated with the image. While not visible in general, it is commonly used for accessibility / screen readers - `mime_type`: Mime type associated to the image. - `original_height`: Image height - `original_width`: Image width - `is_main_image`: Flag determining if the image is the first image of the page. Usually displayed on the top-right part of the page when using web browsers. - `attribution_passes_lang_id`: Compared `language` field with the attribution language (written in the prefix of the attribution description). - `page_changed_recently`: [More Information Needed] - `context_page_description`: Page description corresponds to the short description of the page. It provides a concise explanation of the scope of the page. - `context_section_description`: Text within the image's section. <p align='center'> <img width='75%' src='https://production-media.paperswithcode.com/datasets/Screenshot_2021-03-04_at_14.26.02.png' alt="Half Dome" /> </br> <b>Figure: WIT annotation example. </b> </p> Details on the field content can be found directly in the [paper, figure 5 and table 12.](https://arxiv.org/abs/2103.01913) ### Data Splits All data is held in `train` split, with a total of 37046386 rows. ## Dataset Creation ### Curation Rationale From the [repository](https://github.com/google-research-datasets/wit#motivation): > Multimodal visio-linguistic models rely on a rich dataset to help them learn to model the relationship between images and texts. Having large image-text datasets can significantly improve performance, as shown by recent works. Furthermore the lack of language coverage in existing datasets (which are mostly only in English) also impedes research in the multilingual multimodal space – we consider this a lost opportunity given the potential shown in leveraging images (as a language-agnostic medium) to help improve our multilingual textual understanding. > > To address these challenges and advance research on multilingual, multimodal learning we created the Wikipedia-based Image Text (WIT) Dataset. WIT is created by extracting multiple different texts associated with an image (e.g., as shown in the above image) from Wikipedia articles and Wikimedia image links. This was accompanied by rigorous filtering to only retain high quality image-text sets. > > The resulting dataset contains over 37.6 million image-text sets – making WIT the largest multimodal dataset (publicly available at the time of this writing) with unparalleled multilingual coverage – with 12K+ examples in each of 108 languages (53 languages have 100K+ image-text pairs). ### Source Data #### Initial Data Collection and Normalization From the [paper, section 3.1](https://arxiv.org/abs/2103.01913): > We started with all Wikipedia content pages (i.e., ignoring other pages that have discussions, comments and such). These number about ∼124M pages across 279 languages. #### Who are the source language producers? Text was extracted from Wikipedia. ### Annotations #### Annotation process WIT was constructed using an automatic process. However it was human-validated. From the [paper, section 3.7](https://arxiv.org/abs/2103.01913): > To further verify the quality of the WIT dataset we performed a study using (crowd-sourced) human annotators. As seen in Fig. 3, we asked raters to answer 3 questions. Given an image and the page title, raters first evaluate the quality of the attribution description and reference description in the first two questions (order randomized). The third question understands the contextual quality of these text descriptions given the page description and caption. Each response is on a 3-point scale: "Yes" if the text perfectly describes the image, "Maybe" if it is sufficiently explanatory and "No" if it is irrelevant or the image is inappropriate. #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases From the [paper, section 3.4](https://arxiv.org/abs/2103.01913): > Lastly we found that certain image-text pairs occurred very frequently. These were often generic images that did not have much to do with the main article page. Common examples included flags, logos, maps, insignia and such. To prevent biasing the data, we heavily under-sampled all such images ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ```bibtex @article{srinivasan2021wit, title={WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning}, author={Srinivasan, Krishna and Raman, Karthik and Chen, Jiecao and Bendersky, Michael and Najork, Marc}, journal={arXiv preprint arXiv:2103.01913}, year={2021} } ``` ### Contributions Thanks to [@thomasw21](https://github.com/thomasw21), [@nateraw](https://github.com/nateraw) and [hassiahk](https://github.com/hassiahk) for adding this dataset.
Davincilee
null
null
null
false
1
false
Davincilee/door_inner_with_SAE
2022-05-02T14:16:14.000Z
null
false
4f517807d7ecf96e4686674ead3a445b22f4b9b1
[]
[ "license:other" ]
https://huggingface.co/datasets/Davincilee/door_inner_with_SAE/resolve/main/README.md
--- license: other ---
shanya
null
@inproceedings{ title = {Storytelling with Dialogue: A Critical Role Dungeons and Dragons Dataset}, author = {Rameshkumar, Revanth and Bailey, Peter}, year = {2020}, publisher = {Association for Computational Linguistics}, conference = {ACL} }
Storytelling with Dialogue: A Critical Role Dungeons and Dragons Dataset. Critical Role is an unscripted, live-streamed show where a fixed group of people play Dungeons and Dragons, an open-ended role-playing game. The dataset is collected from 159 Critical Role episodes transcribed to text dialogues, consisting of 398,682 turns. It also includes corresponding abstractive summaries collected from the Fandom wiki. The dataset is linguistically unique in that the narratives are generated entirely through player collaboration and spoken interaction. For each dialogue, there are a large number of turns, multiple abstractive summaries with varying levels of detail, and semantic ties to the previous dialogues.
false
130
false
shanya/crd3
2022-10-25T10:13:08.000Z
crd3
false
d9a3cfd6830ce040b34c1169d564227de87d5bf8
[]
[ "annotations_creators:no-annotation", "language_creators:crowdsourced", "language:en", "license:cc-by-sa-4.0", "multilinguality:monolingual", "source_datasets:original", "task_categories:summarization", "task_categories:text-generation", "task_categories:fill-mask", "task_ids:dialogue-modeling", ...
https://huggingface.co/datasets/shanya/crd3/resolve/main/README.md
--- pretty_name: CRD3 (Critical Role Dungeons and Dragons Dataset) annotations_creators: - no-annotation language_creators: - crowdsourced language: - en license: - cc-by-sa-4.0 multilinguality: - monolingual source_datasets: - original task_categories: - summarization - text-generation - fill-mask task_ids: - dialogue-modeling size_categories: - 10K<n<100K paperswithcode_id: crd3 --- # Dataset Card for "crd3" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [CRD3 homepage](https://github.com/RevanthRameshkumar/CRD3) - **Repository:** [CRD3 repository](https://github.com/RevanthRameshkumar/CRD3) - **Paper:** [Storytelling with Dialogue: A Critical Role Dungeons and Dragons Dataset](https://www.aclweb.org/anthology/2020.acl-main.459/) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 279.93 MB - **Size of the generated dataset:** 4020.33 MB - **Total amount of disk used:** 4300.25 MB ### Dataset Summary Storytelling with Dialogue: A Critical Role Dungeons and Dragons Dataset. Critical Role is an unscripted, live-streamed show where a fixed group of people play Dungeons and Dragons, an open-ended role-playing game. The dataset is collected from 159 Critical Role episodes transcribed to text dialogues, consisting of 398,682 turns. It also includes corresponding abstractive summaries collected from the Fandom wiki. The dataset is linguistically unique in that the narratives are generated entirely through player collaboration and spoken interaction. For each dialogue, there are a large number of turns, multiple abstractive summaries with varying levels of detail, and semantic ties to the previous dialogues. ### Supported Tasks and Leaderboards `summarization`: The dataset can be used to train a model for abstractive summarization. A [fast abstractive summarization-RL](https://github.com/ChenRocks/fast_abs_rl) model was presented as a baseline, which achieves ROUGE-L-F1 of 25.18. ### Languages The text in the dataset is in English, as spoken by actors on The Critical Role show, which is a weekly unscripted, live-stream of a fixed group of people playing Dungeons and Dragons, a popular role-playing game. ## Dataset Structure We show detailed information for up to 5 configurations of the dataset. ### Data Instances #### default - **Size of downloaded dataset files:** 279.93 MB - **Size of the generated dataset:** 4020.33 MB - **Total amount of disk used:** 4300.25 MB An example of 'train' looks as follows. ``` { "alignment_score": 3.679936647415161, "chunk": "Wish them a Happy Birthday on their Facebook and Twitter pages! Also, as a reminder: D&D Beyond streams their weekly show (\"And Beyond\") every Wednesday on twitch.tv/dndbeyond.", "chunk_id": 1, "turn_end": 6, "turn_num": 4, "turn_start": 4, "turns": { "names": ["SAM"], "utterances": ["Yesterday, guys, was D&D Beyond's first one--", "first one-year anniversary. Take two. Hey guys,", "yesterday was D&D Beyond's one-year anniversary.", "Wish them a happy birthday on their Facebook and", "Twitter pages."] } } ``` ### Data Fields The data fields are the same among all splits. #### default - `chunk`: a `string` feature. - `chunk_id`: a `int32` feature. - `turn_start`: a `int32` feature. - `turn_end`: a `int32` feature. - `alignment_score`: a `float32` feature. - `turn_num`: a `int32` feature. - `turns`: a dictionary feature containing: - `names`: a `string` feature. - `utterances`: a `string` feature. ### Data Splits | name | train |validation| test | |-------|------:|---------:|------:| |default|26,232| 3,470|4,541| ## Dataset Creation ### Curation Rationale Dialogue understanding and abstractive summarization remain both important and challenging problems for computational linguistics. Current paradigms in summarization modeling have specific failures in capturing semantics and pragmatics, content selection, rewriting, and evaluation in the domain of long, story-telling dialogue. CRD3 offers a linguistically rich dataset to explore these domains. ### Source Data #### Initial Data Collection and Normalization Dungeons and Dragons is a popular roleplaying game that is driven by structured storytelling. Critical Role is an unscripted, live-streamed show where a fixed group of people play Dungeons and Dragons. This dataset consists of 159 episodes of the show, where the episodes are transcribed. Inconsistencies (e.g. spelling of speaker names) were manually resolved. The abstractive summaries were collected from the [Critical Role Fandom wiki](https://criticalrole.fandom.com/) #### Who are the source language producers? The language producers are actors on The Critical Role show, which is a weekly unscripted, live-stream of a fixed group of people playing Dungeons and Dragons, a popular role-playing game. ### Annotations #### Annotation process [N/A] #### Who are the annotators? [N/A] ### Personal and Sensitive Information [N/A] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators CRTranscript provided transcripts of the show; contributors of the Critical Role Wiki provided the abstractive summaries. ### Licensing Information This work is licensed under a [Creative Commons Attribution-ShareAlike 4.0 International License][cc-by-sa-4.0]., as corresponding to the Critical Role Wiki https://criticalrole.fandom.com/ ### Citation Information ``` @inproceedings{ title = {Storytelling with Dialogue: A Critical Role Dungeons and Dragons Dataset}, author = {Rameshkumar, Revanth and Bailey, Peter}, year = {2020}, publisher = {Association for Computational Linguistics}, conference = {ACL} } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@lhoestq](https://github.com/lhoestq), [@mariamabarham](https://github.com/mariamabarham), [@lewtun](https://github.com/lewtun) for adding this dataset.
wikimedia
null
null
null
false
85
false
wikimedia/wit_base
2022-11-04T15:09:33.000Z
wit
false
ff6d4fb32fd566d3a1fa20e946cba3234179465e
[]
[ "arxiv:2103.01913", "arxiv:1512.03385", "arxiv:1905.00641", "annotations_creators:machine-generated", "language_creators:found", "language:af", "language:an", "language:ar", "language:arz", "language:ast", "language:az", "language:azb", "language:ba", "language:bar", "language:be", "la...
https://huggingface.co/datasets/wikimedia/wit_base/resolve/main/README.md
--- annotations_creators: - machine-generated language_creators: - found language: - af - an - ar - arz - ast - az - azb - ba - bar - be - bg - bn - br - bs - ca - ce - ceb - ckb - cs - cv - cy - da - de - el - en - eo - es - et - eu - fa - fi - fil - fr - fy - ga - gl - hi - hr - hsb - ht - hu - hy - ia - id - io - is - it - iw - ja - jv - ka - kk - kn - ko - la - lah - lb - lmo - lt - lv - mg - mk - ml - mn - mr - ms - my - nan - nds - ne - nl - nn - 'no' - nv - oc - pa - pl - pt - qu - ro - ru - sco - si - sk - sl - sq - sr - sv - sw - ta - te - tg - th - tr - tt - uk - ur - uz - vec - vi - vo - war - xmf - yue - zh license: - cc-by-sa-4.0 multilinguality: - multilingual size_categories: - 1M<n<10M source_datasets: - original - extended|wikipedia task_categories: - image-to-text - text-retrieval task_ids: - image-captioning paperswithcode_id: wit pretty_name: Wikipedia-based Image Text language_bcp47: - af - an - ar - arz - ast - az - azb - ba - bar - be - be-tarask - bg - bn - br - bs - ca - ce - ceb - ckb - cs - cv - cy - da - de - el - en - eo - es - et - eu - fa - fi - fil - fr - fy - ga - gl - hi - hr - hsb - ht - hu - hy - ia - id - io - is - it - iw - ja - jv - ka - kk - kn - ko - la - lah - lb - lmo - lt - lv - mg - mk - ml - mn - mr - ms - my - nan - nds - ne - nl - nn - 'no' - nv - oc - pa - pl - pt - qu - ro - ru - sco - si - sk - sl - sq - sr - sr-Latn - sv - sw - ta - te - tg - th - tr - tt - uk - ur - uz - vec - vi - vo - war - xmf - yue - zh - zh-TW tags: - text-image-retrieval --- # Dataset Card for WIT ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [WIT homepage](https://github.com/google-research-datasets/wit) - **Paper:** [WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning ](https://arxiv.org/abs/2103.01913) - **Leaderboard:** [WIT leaderboard](https://paperswithcode.com/sota/text-image-retrieval-on-wit) and [WIT Kaggle competition](https://www.kaggle.com/competitions/wikipedia-image-caption/leaderboard) - **Point of Contact:** [Miriam Redi](mailto:miriam@wikimedia.org) ### Dataset Summary Wikimedia's version of the Wikipedia-based Image Text (WIT) Dataset, a large multimodal multilingual dataset. From the [official blog post](https://techblog.wikimedia.org/2021/09/09/the-wikipedia-image-caption-matching-challenge-and-a-huge-release-of-image-data-for-research/): > The core training data is taken from the Wikipedia Image-Text (WIT) Dataset, a large curated set of more than 37 million image-text associations extracted from Wikipedia articles in 108 languages that was recently released by Google Research. > > The WIT dataset offers extremely valuable data about the pieces of text associated with Wikipedia images. However, due to licensing and data volume issues, the Google dataset only provides the image name and corresponding URL for download and not the raw image files. > > Getting easy access to the image files is crucial for participants to successfully develop competitive models. Therefore, today, the Wikimedia Research team is releasing its first large image dataset. It contains more than six million image files from Wikipedia articles in 100+ languages, which correspond to almost [1] all captioned images in the WIT dataset. Image files are provided at a 300-px resolution, a size that is suitable for most of the learning frameworks used to classify and analyze images. > [1] We are publishing all images having a non-null “reference description” in the WIT dataset. For privacy reasons, we are not publishing images where a person is the primary subject, i.e., where a person’s face covers more than 10% of the image surface. To identify faces and their bounding boxes, we use the RetinaFace detector. In addition, to avoid the inclusion of inappropriate images or images that violate copyright constraints, we have removed all images that are candidate for deletion on Commons from the dataset. **Note**: Compared to [Google's version](https://huggingface.co/datasets/google/wit), which has contents of one Wikipedia page per data sample, this version groups contents of all Wikipedia pages available in different languages for the image in one single data sample to avoid duplication of image bytes. ### Supported Tasks and Leaderboards - `image-captioning`: This dataset can be used to train a model for image captioning where the goal is to predict a caption given the image. - `text-retrieval`: The goal in this task is to build a model that retrieves the text (`caption_title_and_reference_description`) closest to an image. The leaderboard for this task can be found [here](https://paperswithcode.com/sota/text-image-retrieval-on-wit). This task also has a competition on [Kaggle](https://www.kaggle.com/c/wikipedia-image-caption). In these tasks, any combination of the `caption_reference_description`, `caption_attribution_description` and `caption_alt_text_description` fields can be used as the input text/caption. ### Languages The dataset contains examples from all Wikipedia languages. ## Dataset Structure ### Data Instances Each instance is an image, its representation in bytes, a pre-computed embedding, and the set of captions attached to the image in Wikipedia. ``` { 'image': <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=300x225 at 0x7F88F3876358>, 'image_url': 'https://upload.wikimedia.org/wikipedia/commons/8/8b/Scolopendra_gigantea.jpg', 'embedding': [1.4784087, 2.8710432, 0.0, 0.51603067, ..., 10.266883, 0.51142216, 0.0, 2.3464653], 'metadata_url': 'http://commons.wikimedia.org/wiki/File:Scolopendra_gigantea.jpg', 'original_height': 3000, 'original_width': 4000, 'mime_type': 'image/jpeg', 'caption_attribution_description': 'English: Puerto Rican Giant Centipede, Scolopendra gigantea; Vieques, Puerto Rico Slovenčina: Stonožka obrovská, Scolopendra gigantea; Vieques, Portoriko', 'wit_features': { 'language': ['ro', 'vi', 'sk', ..., 'nl', 'th', 'lv'], 'page_url': ['https://ro.wikipedia.org/wiki/Scolopendra_gigantea', 'https://vi.wikipedia.org/wiki/Scolopendra_gigantea', 'https://sk.wikipedia.org/wiki/Scolopendra_gigantea', ..., 'https://nl.wikipedia.org/wiki/Scolopendra_gigantea', 'https://th.wikipedia.org/wiki/%E0%B8%95%E0%B8%B0%E0%B8%82%E0%B8%B2%E0%B8%9A%E0%B8%A2%E0%B8%B1%E0%B8%81%E0%B8%A9%E0%B9%8C%E0%B8%82%E0%B8%B2%E0%B9%80%E0%B8%AB%E0%B8%A5%E0%B8%B7%E0%B8%AD%E0%B8%87%E0%B9%80%E0%B8%9B%E0%B8%A3%E0%B8%B9', 'https://lv.wikipedia.org/wiki/Skolopendru_dzimta'], 'attribution_passes_lang_id': [True, True, True, ..., True, True, True], 'caption_alt_text_description': [None, None, None, ..., 'Scolopendra gigantea', None, 'Milzu skolopendra (Scolopendra gigantea)'], 'caption_reference_description': [None, None, None, ..., None, None, 'Milzu skolopendra (Scolopendra gigantea)'], 'caption_title_and_reference_description': [None, 'Scolopendra gigantea [SEP] ', None, ..., 'Scolopendra gigantea [SEP] ', None, 'Skolopendru dzimta [SEP] Milzu skolopendra (Scolopendra gigantea)'], 'context_page_description': ['Scolopendra gigantea este un miriapod din clasa Chilopoda, fiind cel mai mare reprezentant al genului Scolopendra. Adultul poate atinge o lungime de 26 cm, uneori depășind 30 cm. Această specie habitează în regiunile de nord și de vest a Americii de Sud, pe insulele Trinidad, insulele Virgine, Jamaica Hispaniola ș.a. Localnicii denumesc scolopendra chilopodul gigant galben și chilopodul gigant amazonian.', 'Scolopendra gigantea là đại diện lớn nhất của chi Scolopendra nói riêng và cả lớp rết nói chung, thường đạt độ dài 26 cm và có thể vượt quá 30 cm. Sinh sống ở khu vực phía bắc và tây của Nam Mỹ và các đảo Trinidad, Puerto Rico, Saint Thomas, U.S. Virgin Islands, Jamaica, và Hispaniola.', 'Scolopendra gigantea, starší slovenský nazov: štípavica veľká, je živočích z rodu Scolopendra, s veľkosťou do 30 cm.', ..., 'Scolopendra gigantea is een tijgerduizendpoot uit Zuid-Amerika. De soort jaagt onder andere op grote geleedpotigen, amfibieën, reptielen en kleine zoogdieren. Het is voor zover bekend de grootste niet uitgestorven duizendpoot ter wereld.', 'ตะขาบยักษ์ขาเหลืองเปรู หรือ ตะขาบยักษ์อเมซอน เป็นตะขาบชนิดที่มีขนาดใหญ่ที่สุดในสกุล Scolopendra โดยปกติเมื่อโตเต็มที่จะยาว 26 เซนติเมตร แต่บางครั้งก็สามารถโตได้ถึง 30 เซนติเมตร ตะขาบชนิดนี้อาศัยอยู่ทางแถบเหนือและตะวันตกของทวีปอเมริกาใต้ และตามเกาะแก่งของประเทศตรินิแดดและจาไมกา เป็นสัตว์กินเนื้อ โดยกินจิ้งจก, กบ, นก, หนู และแม้แต่ค้างคาวเป็นอาหาร และขึ้นชื่อในเรื่องความดุร้าย', 'Skolpendru dzimta pieder pie simtkāju kārtas. Ap 400 dzimtas sugas sastopamas visā pasaulē, īpaši subtropu un tropu apgabalos. Mitinās augsnē, nobirušās lapās, plaisās, spraugās.'], 'context_section_description': [None, 'Scolopendra gigantea (còn được gọi là Rết chân vàng khổng lồ Peru và Rết khổng lồ Amazon) là đại diện lớn nhất của chi Scolopendra nói riêng và cả lớp rết nói chung, thường đạt độ dài 26\xa0cm (10\xa0in) và có thể vượt quá 30\xa0cm (12\xa0in). Sinh sống ở khu vực phía bắc và tây của Nam Mỹ và các đảo Trinidad, Puerto Rico, Saint Thomas, U.S. Virgin Islands, Jamaica, và Hispaniola.', None, ..., 'Scolopendra gigantea is een tijgerduizendpoot uit Zuid-Amerika. De soort jaagt onder andere op grote geleedpotigen, amfibieën, reptielen en kleine zoogdieren. Het is voor zover bekend de grootste niet uitgestorven duizendpoot ter wereld.', None, 'Skolpendru dzimta (Scolopendridae) pieder pie simtkāju kārtas. Ap 400 dzimtas sugas sastopamas visā pasaulē, īpaši subtropu un tropu apgabalos. Mitinās augsnē, nobirušās lapās, plaisās, spraugās.'], 'hierarchical_section_title': ['Scolopendra gigantea', 'Scolopendra gigantea', 'Scolopendra gigantea', ..., 'Scolopendra gigantea', 'ตะขาบยักษ์ขาเหลืองเปรู', 'Skolopendru dzimta'], 'is_main_image': [True, True, True, ..., True, True, True], 'page_title': ['Scolopendra gigantea', 'Scolopendra gigantea', 'Scolopendra gigantea', ..., 'Scolopendra gigantea', 'ตะขาบยักษ์ขาเหลืองเปรู', 'Skolopendru dzimta'], 'section_title': [None, None, None, ..., None, None, None] } } ``` **Note**: The dataset is stored in Parquet for better performance. This dataset was generated from the original files using [this script](wit_base/blob/main/scripts/wit.py). Additionally, 120 examples from the original files have incorrectly formatted one or more of the following fields: `original_height`, `original_width`, `mime_type` and `caption_attribution_description`. The fixed versions of these examples that were used in the generation script can be found [here](wit_base/blob/main/scripts/corrected_examples.py). ### Data Fields - `image`: A `PIL.Image.Image` object containing the image resized to a width of 300-px while preserving its aspect ratio. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]`. - `image_url`: URL to wikipedia image - `embedding`: Precomputed image embedding. Each image is described with a 2048-dimensional signature extracted from the second-to-last layer of a [ResNet-50](https://arxiv.org/abs/1512.03385) neural network trained with [Imagenet](https://www.image-net.org/) data. These embeddings contain rich information about the image content and layout, in a compact form - `metadata_url`: URL to wikimedia page containing the image and the metadata - `original_height`: Original image height before resizing - `original_width`: Original image width before resizing - `mime_type`: Mime type associated to the image - `caption_attribution_description`: This is the text found on the Wikimedia page of the image. This text is common to all occurrences of that image across all Wikipedias. - `wit_features`: Sequence of captions for the image with language, page URL, information about the page, caption text, etc. - `language`: Language code depicting wikipedia language of the page - `page_url`: URL to wikipedia page - `attribution_passes_lang_id`: Compared `language` field with the attribution language (written in the prefix of the attribution description. - `caption_alt_text_description`: This is the “alt” text associated with the image. While not visible in general, it is commonly used for accessibility / screen readers - `caption_reference_description`: This is the caption that is visible on the wikipedia page directly below the image. - `caption_title_and_reference_description`: Concatenation of `page_title` and `caption_reference_description`. - `context_page_description`: Corresponds to the short description of the page. It provides a concise explanation of the scope of the page. - `context_section_description`: Text within the image's section - `hierarchical_section_title`: Hierarchical section's title - `is_main_image`: Flag determining if the image is the first image of the page. Usually displayed on the top-right part of the page when using web browsers. - `page_changed_recently`: [More Information Needed] - `page_title`: Wikipedia page's title - `section_title`: Section's title <p align='center'> <img width='75%' src='https://production-media.paperswithcode.com/datasets/Screenshot_2021-03-04_at_14.26.02.png' alt="Half Dome" /> </br> <b>Figure: WIT annotation example. </b> </p> Details on the field content can be found directly in the [paper, figure 5 and table 12.](https://arxiv.org/abs/2103.01913) ### Data Splits All data is held in `train` split, with a total of 6477255 examples. ## Dataset Creation ### Curation Rationale From the [official blog post](https://techblog.wikimedia.org/2021/09/09/the-wikipedia-image-caption-matching-challenge-and-a-huge-release-of-image-data-for-research/): > The WIT dataset offers extremely valuable data about the pieces of text associated with Wikipedia images. > Getting easy access to the image files is crucial for participants to successfully develop competitive models. > With this large release of visual data, we aim to help the competition participants—as well as researchers and practitioners who are interested in working with Wikipedia images—find and download the large number of image files associated with the challenge, in a compact form. ### Source Data #### Initial Data Collection and Normalization From the [paper, section 3.1](https://arxiv.org/abs/2103.01913): > We started with all Wikipedia content pages (i.e., ignoring other pages that have discussions, comments and such). These number about ~124M pages across 279 languages. #### Who are the source language producers? Text was extracted from Wikipedia. ### Annotations #### Annotation process WIT was constructed using an automatic process. However it was human-validated. From the [paper, section 3.7](https://arxiv.org/abs/2103.01913): > To further verify the quality of the WIT dataset we performed a study using (crowd-sourced) human annotators. As seen in Fig. 3, we asked raters to answer 3 questions. Given an image and the page title, raters first evaluate the quality of the attribution description and reference description in the first two questions (order randomized). The third question understands the contextual quality of these text descriptions given the page description and caption. Each response is on a 3-point scale: "Yes" if the text perfectly describes the image, "Maybe" if it is sufficiently explanatory and "No" if it is irrelevant or the image is inappropriate. #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information From the [official blog post](https://techblog.wikimedia.org/2021/09/09/the-wikipedia-image-caption-matching-challenge-and-a-huge-release-of-image-data-for-research/#FN1): > For privacy reasons, we are not publishing images where a person is the primary subject, i.e., where a person’s face covers more than 10% of the image surface. To identify faces and their bounding boxes, we use the [RetinaFace](https://arxiv.org/abs/1905.00641) detector. In addition, to avoid the inclusion of inappropriate images or images that violate copyright constraints, we have removed all images that are [candidate for deletion](https://commons.wikimedia.org/wiki/Commons:Deletion_requests) on Commons from the dataset. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases From the [paper, section 3.4](https://arxiv.org/abs/2103.01913): > Lastly we found that certain image-text pairs occurred very frequently. These were often generic images that did not have much to do with the main article page. Common examples included flags, logos, maps, insignia and such. To prevent biasing the data, we heavily under-sampled all such images ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators Miriam Redi, Fabian Kaelin and Tiziano Piccardi. ### Licensing Information [CC BY-SA 4.0 international license](https://creativecommons.org/licenses/by-sa/4.0/) ### Citation Information ```bibtex @article{srinivasan2021wit, title={WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning}, author={Srinivasan, Krishna and Raman, Karthik and Chen, Jiecao and Bendersky, Michael and Najork, Marc}, journal={arXiv preprint arXiv:2103.01913}, year={2021} } ``` ### Contributions Thanks to [@nateraw](https://github.com/nateraw), [yjernite](https://github.com/yjernite) and [mariosasko](https://github.com/mariosasko) for adding this dataset.
pauli31
null
null
null
false
1
false
pauli31/czech-subjectivity-dataset
2022-07-01T15:31:40.000Z
null
false
e1ad894c08c0c58d19d0ca4467dce6b96df042aa
[]
[ "arxiv:2204.13915", "language:cs-CZ", "license:cc-by-nc-sa-4.0", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "task_categories:text-classification", "task_ids:sentiment-classification" ]
https://huggingface.co/datasets/pauli31/czech-subjectivity-dataset/resolve/main/README.md
--- annotations_creators: [] language_creators: [] language: - cs-CZ license: - cc-by-nc-sa-4.0 multilinguality: - monolingual pretty_name: Czech Subjectivity Dataset size_categories: - 1K<n<10K source_datasets: - original task_categories: - text-classification task_ids: - sentiment-classification --- # Dataset Card for Czech Subjectivity Dataset ### Dataset Summary Czech subjectivity dataset (Subj-CS) of 10k manually annotated subjective and objective sentences from movie reviews and descriptions. See the paper description https://arxiv.org/abs/2204.13915 ### Github https://github.com/pauli31/czech-subjectivity-dataset ### Supported Tasks and Leaderboards Subjectivity Analysis ### Languages Czech ### Data Instances train/dev/test ### Licensing Information [Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.](https://creativecommons.org/licenses/by-nc-sa/4.0/) ### Citation Information If you use our dataset or software for academic research, please cite the our [paper](https://arxiv.org/abs/2204.13915) ``` @article{pib2022czech, title={Czech Dataset for Cross-lingual Subjectivity Classification}, author={Pavel Přibáň and Josef Steinberger}, year={2022}, eprint={2204.13915}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` ### Contact pribanp@kiv.zcu.cz ### Contributions Thanks to [@pauli31](https://github.com/pauli31) for adding this dataset.
arbml
null
@misc{alyafeai2021masader, title={Masader: Metadata Sourcing for Arabic Text and Speech Data Resources}, author={Zaid Alyafeai and Maraim Masoud and Mustafa Ghaleb and Maged S. Al-shaibani}, year={2021}, eprint={2110.06744}, archivePrefix={arXiv}, primaryClass={cs.CL} }
Masader is the largest public catalogue for Arabic NLP datasets, which consists of more than 200 datasets annotated with 25 attributes.
false
4,422
false
arbml/masader
2022-07-08T14:45:05.000Z
null
false
d2d59a4ccf942da4f70948219362271f14efc5c8
[]
[ "license:mit" ]
https://huggingface.co/datasets/arbml/masader/resolve/main/README.md
--- license: mit ---
Apo
null
null
null
false
1
false
Apo/ADE20K_sky_13labels
2022-05-03T07:24:40.000Z
null
false
143c5879a8ddd8a45a5ed7c7f429c89b430a80c6
[]
[ "license:afl-3.0" ]
https://huggingface.co/datasets/Apo/ADE20K_sky_13labels/resolve/main/README.md
--- license: afl-3.0 ---
Erwin
null
null
null
false
2
false
Erwin/coffee_reviews_corpus
2022-05-03T08:22:38.000Z
null
false
0d01b67c73ed0a50d03a46e8283b37dab58fe3bc
[]
[ "license:mit" ]
https://huggingface.co/datasets/Erwin/coffee_reviews_corpus/resolve/main/README.md
--- license: mit ---
null
null
@article{tne, author = {Elazar, Yanai and Basmov, Victoria and Goldberg, Yoav and Tsarfaty, Reut}, title = "{Text-based NP Enrichment}", journal = {Transactions of the Association for Computational Linguistics}, year = {2022}, }
TNE is an NLU task, which focus on relations between noun phrases (NPs) that can be mediated via prepositions. The dataset contains 5,497 documents, annotated exhaustively with all possible links between the NPs in each document.
false
440
false
tne
2022-11-03T16:15:29.000Z
null
false
051084916ead68ba57c14143bac8a94f7a94d069
[]
[ "arxiv:2109.12085", "annotations_creators:crowdsourced", "language_creators:found", "language:en", "license:mit", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "task_categories:text-retrieval", "task_ids:document-retrieval" ]
https://huggingface.co/datasets/tne/resolve/main/README.md
--- annotations_creators: - crowdsourced language_creators: - found language: - en license: - mit multilinguality: - monolingual pretty_name: Text-based NP Enrichment size_categories: - 1K<n<10K source_datasets: - original task_categories: - text-retrieval task_ids: - document-retrieval dataset_info: features: - name: id dtype: string - name: text dtype: string - name: tokens sequence: string - name: nps list: - name: text dtype: string - name: first_char dtype: int32 - name: last_char dtype: int32 - name: first_token dtype: int32 - name: last_token dtype: int32 - name: id dtype: string - name: np_relations list: - name: anchor dtype: string - name: complement dtype: string - name: preposition dtype: class_label: names: 0: about 1: for 2: with 3: from 4: among 5: by 6: 'on' 7: at 8: during 9: of 10: member(s) of 11: in 12: after 13: under 14: to 15: into 16: before 17: near 18: outside 19: around 20: between 21: against 22: over 23: inside - name: complement_coref_cluster_id dtype: string - name: coref list: - name: id dtype: string - name: members sequence: string - name: np_type dtype: class_label: names: 0: standard 1: time/date/measurement 2: idiomatic - name: metadata struct: - name: annotators struct: - name: coref_worker dtype: int32 - name: consolidator_worker dtype: int32 - name: np-relations_worker sequence: int32 - name: url dtype: string - name: source dtype: string splits: - name: test num_bytes: 2203716 num_examples: 500 - name: test_ood num_bytes: 2249352 num_examples: 509 - name: train num_bytes: 41308170 num_examples: 3988 - name: validation num_bytes: 5495419 num_examples: 500 download_size: 14194578 dataset_size: 51256657 --- # Dataset Card for Text-based NP Enrichment ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://yanaiela.github.io/TNE/ - **Repository:** https://github.com/yanaiela/TNE - **Paper:** https://arxiv.org/abs/2109.12085 - **Leaderboard:** [TNE OOD](https://leaderboard.allenai.org/tne-ood/submissions/public) [TNE](https://leaderboard.allenai.org/tne/submissions/public) - **Point of Contact:** [Yanai Elazar](mailto:yanaiela@gmail.com) ### Dataset Summary Text-based NP Enrichment (TNE) is a natural language understanding (NLU) task, which focus on relations between noun phrases (NPs) that can be mediated via prepositions. The dataset contains 5,497 documents, annotated exhaustively with all possible links between the NPs in each document. The main data comes from WikiNews, which is used for train/dev/test. We also collected an additional set of 509 documents to serve as out of distribution (OOD) data points, from the Book Corpus, IMDB reviews and Reddit. ### Supported Tasks and Leaderboards The data contain both the main data for the TNE task, as well as coreference resolution data. There are two leaderboards for the TNE data, one for the standard test set, and another one for the OOD test set: - [TNE Leaderboard](https://leaderboard.allenai.org/tne/submissions/public) - [TNE OOD Leaderboard](https://leaderboard.allenai.org/tne-ood/submissions/public) ### Languages The text in the dataset is in English, as spoken in the different domains we include. The associated BCP-47 code is `en`. ## Dataset Structure ### Data Instances The original files are in a jsonl format, containing a dictionary of a single document, in each line. Each document contain a different amount of labels, due to the different amount of NPs. The test and ood splits come without the annotated labels. ### Data Fields A document consists of: * `id`: a unique identifier of a document, beginning with `r` and followed by a number * `text`: the text of the document. The title and subtitles (if exists) are separated with two new lines. The paragraphs are separated by a single new line. * `tokens`: a list of string, containing the tokenized tokens * `nps`: a list of dictionaries, containing the following entries: * `text`: the text of the np * `start_index`: an integer indicating the starting index in the text * `end_index`: an integer indicating the ending index in the text * `start_token`: an integer indicating the first token of the np out of the tokenized tokens * `end_token`: an integer indicating the last token of the np out of the tokenized tokens * `id`: the id of the np * `np_relations`: these are the relation labels of the document. It is a list of dictionaries, where each dictionary contains: * `anchor`: the id of the anchor np * `complement`: the id of the complement np * `preposition`: the preposition that links between the anchor and the complement. This can take one out of 24 pre-defined preposition (23 + member(s)-of) * `complement_coref_cluster_id`: the coreference id, which the complement is part of. * `coref`: the coreference labels. It contains a list of dictionaries, where each dictionary contains: * `id`: the id of the coreference cluster * `members`: the ids of the nps members of such cluster * `np_type`: the type of cluster. It can be either * `standard`: regular coreference cluster * `time/date/measurement`: a time / date / measurement np. These will be singletons. * `idiomatic`: an idiomatic expression * `metadata`: metadata of the document. It contains the following: * `annotators`: a dictionary with anonymized annotators id * `coref_worker`: the coreference worker id * `consolidator_worker`: the consolidator worker id * `np-relations_worker`: the np relations worker id * `url`: the url where the document was taken from (not always existing) * `source`: the original file name where the document was taken from ### Data Splits The dataset is spread across four files, for the four different splits: train, dev, test and test_ood. Additional details on the data statistics can be found in the [paper](https://arxiv.org/abs/2109.12085) ## Dataset Creation ### Curation Rationale TNE was build as a new task for language understanding, focusing on extracting relations between nouns, moderated by prepositions. ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators The dataset was created by Yanai Elazar, Victoria Basmov, Yoav Goldberg, Reut Tsarfaty, during work done at Bar-Ilan University, and AI2. ### Licensing Information The data is released under the MIT license. ### Citation Information ```bibtex @article{tne, author = {Elazar, Yanai and Basmov, Victoria and Goldberg, Yoav and Tsarfaty, Reut}, title = "{Text-based NP Enrichment}", journal = {Transactions of the Association for Computational Linguistics}, year = {2022}, } ``` ### Contributions Thanks to [@yanaiela](https://github.com/yanaiela), who is also the first author of the paper, for adding this dataset.
farazeftekhar
null
null
null
false
1
false
farazeftekhar/geojson
2022-05-03T19:15:48.000Z
null
false
e874dfff456d716a0700c244c36baebd58581ebd
[]
[ "license:other" ]
https://huggingface.co/datasets/farazeftekhar/geojson/resolve/main/README.md
--- license: other ---
orieg
null
@article{Kershaw2020ElsevierOC, title = {Elsevier OA CC-By Corpus}, author = {Daniel James Kershaw and R. Koeling}, journal = {ArXiv}, year = {2020}, volume = {abs/2008.00774}, doi = {https://doi.org/10.48550/arXiv.2008.00774}, url = {https://elsevier.digitalcommonsdata.com/datasets/zm33cdndxs}, keywords = {Science, Natural Language Processing, Machine Learning, Open Dataset}, abstract = {We introduce the Elsevier OA CC-BY corpus. This is the first open corpus of Scientific Research papers which has a representative sample from across scientific disciplines. This corpus not only includes the full text of the article, but also the metadata of the documents, along with the bibliographic information for each reference.} }
Elsevier OA CC-By is a corpus of 40k (40, 091) open access (OA) CC-BY articles from across Elsevier’s journals and include the full text of the article, the metadata, the bibliographic information for each reference, and author highlights.
false
3
false
orieg/elsevier-oa-cc-by
2022-07-01T15:59:58.000Z
elsevier-oa-cc-by
false
77840f2f84038fdf4b608fff764b21b7ef18eb34
[]
[ "arxiv:2008.00774", "annotations_creators:expert-generated", "language_creators:expert-generated", "language:en", "license:cc-by-4.0", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "task_categories:fill-mask", "task_categories:summarization", "task_cate...
https://huggingface.co/datasets/orieg/elsevier-oa-cc-by/resolve/main/README.md
--- annotations_creators: - expert-generated language_creators: - expert-generated language: - en license: - cc-by-4.0 multilinguality: - monolingual pretty_name: Elsevier OA CC-By paperswithcode_id: elsevier-oa-cc-by size_categories: - 10K<n<100K source_datasets: - original task_categories: - fill-mask - summarization - text-classification task_ids: - masked-language-modeling - news-articles-summarization - news-articles-headline-generation --- # Dataset Card for Elsevier OA CC-By ## Table of Contents - [Dataset Card for Elsevier OA CC-By](#dataset-card-for-elsevier-oa-cc-by) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Initial Data Collection and Normalization](#initial-data-collection-and-normalization) - [Who are the source language producers?](#who-are-the-source-language-producers) - [Annotations](#annotations) - [Annotation process](#annotation-process) - [Who are the annotators?](#who-are-the-annotators) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://elsevier.digitalcommonsdata.com/datasets/zm33cdndxs - **Repository:** https://elsevier.digitalcommonsdata.com/datasets/zm33cdndxs - **Paper:** https://arxiv.org/abs/2008.00774 - **Leaderboard:** - **Point of Contact:** [@orieg](https://huggingface.co/orieg) ### Dataset Summary Elsevier OA CC-By: This is a corpus of 40k (40,091) open access (OA) CC-BY articles from across Elsevier’s journals representing a large scale, cross-discipline set of research data to support NLP and ML research. The corpus include full-text articles published in 2014 to 2020 and are categorized in 27 Mid Level ASJC Code (subject classification). ***Distribution of Publication Years*** | Publication Year | Number of Articles | | :---: | :---: | | 2014 | 3018 | | 2015 | 4438 | | 2016 | 5913 | | 2017 | 6419 | | 2018 | 8016 | | 2019 | 10135 | | 2020 | 2159 | ***Distribution of Articles Per Mid Level ASJC Code. Each article can belong to multiple ASJC codes.*** | Discipline | Count | | --- | ---: | | General | 3847 | | Agricultural and Biological Sciences | 4840 | | Arts and Humanities | 982 | | Biochemistry, Genetics and Molecular Biology | 8356 | | Business, Management and Accounting | 937 | | Chemical Engineering | 1878 | | Chemistry | 2490 | | Computer Science | 2039 | | Decision Sciences | 406 | | Earth and Planetary Sciences | 2393 | | Economics, Econometrics and Finance | 976 | | Energy | 2730 | | Engineering | 4778 | | Environmental Science | 6049 | | Immunology and Microbiology | 3211 | | Materials Science | 3477 | | Mathematics | 538 | | Medicine | 7273 | | Neuroscience | 3669 | | Nursing | 308 | | Pharmacology, Toxicology and Pharmaceutics | 2405 | | Physics and Astronomy | 2404 | | Psychology | 1760 | | Social Sciences | 3540 | | Veterinary | 991 | | Dentistry | 40 | | Health Professions | 821 | ### Supported Tasks and Leaderboards [More Information Needed] ### Languages English (`en`). ## Dataset Structure ### Data Instances The original dataset was published with the following json structure: ``` { "docId": <str>, "metadata":{ "title": <str>, "authors": [ { "first": <str>, "initial": <str>, "last": <str>, "email": <str> }, ... ], "issn": <str>, "volume": <str>, "firstpage": <str>, "lastpage": <str>, "pub_year": <int>, "doi": <str>, "pmid": <str>, "openaccess": "Full", "subjareas": [<str>], "keywords": [<str>], "asjc": [<int>], }, "abstract":[ { "sentence": <str>, "startOffset": <int>, "endOffset": <int> }, ... ], "bib_entries":{ "BIBREF0":{ "title":<str>, "authors":[ { "last":<str>, "initial":<str>, "first":<str> }, ... ], "issn": <str>, "volume": <str>, "firstpage": <str>, "lastpage": <str>, "pub_year": <int>, "doi": <str>, "pmid": <str> }, ... }, "body_text":[ { "sentence": <str>, "secId": <str>, "startOffset": <int>, "endOffset": <int>, "title": <str>, "refoffsets": { <str>:{ "endOffset":<int>, "startOffset":<int> } }, "parents": [ { "id": <str>, "title": <str> }, ... ] }, ... ] } ``` ***docId*** The docID is the identifier of the document. This is unique to the document, and can be resolved into a URL for the document through the addition of `https//www.sciencedirect.com/science/pii/<docId>` ***abstract*** This is the author provided abstract for the document ***body_text*** The full text for the document. The text has been split on sentence boundaries, thus making it easier to use across research projects. Each sentence has the title (and ID) of the section which it is from, along with titles (and IDs) of the parent section. The highest-level section takes index 0 in the parents array. If the array is empty then the title of the section for the sentence is the highest level section title. This will allow for the reconstruction of the article structure. References have been extracted from the sentences. The IDs of the extracted reference and their respective offset within the sentence can be found in the “refoffsets” field. The complete list of references are can be found in the “bib_entry” field along with the references’ respective metadata. Some will be missing as we only keep ‘clean’ sentences, ***bib_entities*** All the references from within the document can be found in this section. If the meta data for the reference is available, it has been added against the key for the reference. Where possible information such as the document titles, authors, and relevant identifiers (DOI and PMID) are included. The keys for each reference can be found in the sentence where the reference is used with the start and end offset of where in the sentence that reference was used. ***metadata*** Meta data includes additional information about the article, such as list of authors, relevant IDs (DOI and PMID). Along with a number of classification schemes such as ASJC and Subject Classification. ***author_highlights*** Author highlights were included in the corpus where the author(s) have provided them. The coverage is 61% of all articles. The author highlights, consisting of 4 to 6 sentences, is provided by the author with the aim of summarising the core findings and results in the article. ### Data Fields * ***title***: This is the author provided title for the document. 100% coverage. * ***abstract***: This is the author provided abstract for the document. 99.25% coverage. * ***keywords***: This is the author and publisher provided keywords for the document. 100% coverage. * ***asjc***: This is the disciplines for the document as represented by 334 ASJC (All Science Journal Classification) codes. 100% coverage. * ***subjareas***: This is the Subject Classification for the document as represented by 27 ASJC top-level subject classifications. 100% coverage. * ***body_text***: The full text for the document. 100% coverage. * ***author_highlights***: This is the author provided highlights for the document. 61.31% coverage. ### Data Splits ***Distribution of Publication Years*** | | Train | Test | Validation | | --- | :---: | :---: | :---: | | All Articles | 32072 | 4009 | 4008 | | With Author Highlights | 19644 | 2420 | 2514 | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization Date the data was collected: 2020-06-25T11:00:00.000Z See the [original paper](https://doi.org/10.48550/arXiv.2008.00774) for more detail on the data collection process. #### Who are the source language producers? See `3.1 Data Sampling` in the [original paper](https://doi.org/10.48550/arXiv.2008.00774). ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) ### Citation Information ``` @article{Kershaw2020ElsevierOC, title = {Elsevier OA CC-By Corpus}, author = {Daniel James Kershaw and R. Koeling}, journal = {ArXiv}, year = {2020}, volume = {abs/2008.00774}, doi = {https://doi.org/10.48550/arXiv.2008.00774}, url = {https://elsevier.digitalcommonsdata.com/datasets/zm33cdndxs}, keywords = {Science, Natural Language Processing, Machine Learning, Open Dataset}, abstract = {We introduce the Elsevier OA CC-BY corpus. This is the first open corpus of Scientific Research papers which has a representative sample from across scientific disciplines. This corpus not only includes the full text of the article, but also the metadata of the documents, along with the bibliographic information for each reference.} } ``` ``` @dataset{https://10.17632/zm33cdndxs.3, doi = {10.17632/zm33cdndxs.2}, url = {https://data.mendeley.com/datasets/zm33cdndxs/3}, author = "Daniel Kershaw and Rob Koeling", keywords = {Science, Natural Language Processing, Machine Learning, Open Dataset}, title = {Elsevier OA CC-BY Corpus}, publisher = {Mendeley}, year = {2020}, month = {sep} } ``` ### Contributions Thanks to [@orieg](https://github.com/orieg) for adding this dataset.
allenai
null
null
null
false
3
false
allenai/drug-combo-extraction
2022-05-04T04:12:53.000Z
null
false
8092b28c4f8490d234c1385bb385575ec7408330
[]
[ "license:mit" ]
https://huggingface.co/datasets/allenai/drug-combo-extraction/resolve/main/README.md
--- license: mit ---
Ukhushn
null
null
null
false
1
false
Ukhushn/home-depot
2022-10-25T10:20:53.000Z
null
false
aee7fea371b991a01db75877fd23e37d381379c0
[]
[ "language:en", "language_bcp47:en-US", "license:afl-3.0", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "task_categories:sentence-similarity" ]
https://huggingface.co/datasets/Ukhushn/home-depot/resolve/main/README.md
--- language: - en language_bcp47: - en-US license: - afl-3.0 annotations_creators: - no-annotation language_creators: - found multilinguality: - monolingual pretty_name: Ukhushn/home-depot size_categories: - 10K<n<100K source_datasets: [] task_categories: - sentence-similarity task_ids: [] --- # Dataset Card for Ukhushn/home-depot
nielsr
null
null
null
false
1
false
nielsr/test
2022-05-04T12:09:13.000Z
null
false
ad1898d0b872d98ab4747b07315385a1736ce36b
[]
[]
https://huggingface.co/datasets/nielsr/test/resolve/main/README.md
This is used for the pix2seq model. It contains the preprocessed pixel values of the cats image, useful for converting.
nlpconnect
null
null
null
false
3
false
nlpconnect/DocVQA
2022-05-04T14:24:06.000Z
null
false
fc36ebb3fa15bdf8731049b609e61d1fb5e696a5
[]
[ "license:apache-2.0" ]
https://huggingface.co/datasets/nlpconnect/DocVQA/resolve/main/README.md
--- license: apache-2.0 ---
null
null
@inproceedings{singh2019towards, title={Towards VQA Models That Can Read}, author={Singh, Amanpreet and Natarjan, Vivek and Shah, Meet and Jiang, Yu and Chen, Xinlei and Batra, Dhruv and Parikh, Devi and Rohrbach, Marcus}, booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition}, pages={8317-8326}, year={2019} }
TextVQA requires models to read and reason about text in images to answer questions about them. Specifically, models need to incorporate a new modality of text present in the images and reason over it to answer TextVQA questions. TextVQA dataset contains 45,336 questions over 28,408 images from the OpenImages dataset.
false
153
false
textvqa
2022-11-03T16:07:43.000Z
null
false
383e2193a964bc1eae55ab17fab772d5fad6bb6a
[]
[ "arxiv:1904.08920", "arxiv:2007.00398", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "language:en", "license:cc-by-4.0", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "task_categories:visual-question-answering", "task_ids:visua...
https://huggingface.co/datasets/textvqa/resolve/main/README.md
--- annotations_creators: - crowdsourced language_creators: - crowdsourced language: - en license: - cc-by-4.0 multilinguality: - monolingual pretty_name: TextVQA size_categories: - 10K<n<100K source_datasets: - original task_categories: - visual-question-answering task_ids: - visual-question-answering dataset_info: - config_name: train features: - name: image_id dtype: string - name: question_id dtype: int32 - name: question dtype: string - name: question_tokens sequence: string - name: image dtype: image - name: image_width dtype: int32 - name: image_height dtype: int32 - name: flickr_original_url dtype: string - name: flickr_300k_url dtype: string - name: answers sequence: string - name: image_classes sequence: string - name: set_name dtype: string splits: - name: test num_bytes: 3025046 num_examples: 5734 - name: train num_bytes: 21381310 num_examples: 34602 - name: validation num_bytes: 3077854 num_examples: 5000 download_size: 8070116310 dataset_size: 27484210 - config_name: val features: - name: image_id dtype: string - name: question_id dtype: int32 - name: question dtype: string - name: question_tokens sequence: string - name: image dtype: image - name: image_width dtype: int32 - name: image_height dtype: int32 - name: flickr_original_url dtype: string - name: flickr_300k_url dtype: string - name: answers sequence: string - name: image_classes sequence: string - name: set_name dtype: string splits: - name: test num_bytes: 3025046 num_examples: 5734 - name: train num_bytes: 21381310 num_examples: 34602 - name: validation num_bytes: 3077854 num_examples: 5000 download_size: 8070116310 dataset_size: 27484210 - config_name: test features: - name: image_id dtype: string - name: question_id dtype: int32 - name: question dtype: string - name: question_tokens sequence: string - name: image dtype: image - name: image_width dtype: int32 - name: image_height dtype: int32 - name: flickr_original_url dtype: string - name: flickr_300k_url dtype: string - name: answers sequence: string - name: image_classes sequence: string - name: set_name dtype: string splits: - name: test num_bytes: 3025046 num_examples: 5734 - name: train num_bytes: 21381310 num_examples: 34602 - name: validation num_bytes: 3077854 num_examples: 5000 download_size: 8070116310 dataset_size: 27484210 - config_name: textvqa features: - name: image_id dtype: string - name: question_id dtype: int32 - name: question dtype: string - name: question_tokens sequence: string - name: image dtype: image - name: image_width dtype: int32 - name: image_height dtype: int32 - name: flickr_original_url dtype: string - name: flickr_300k_url dtype: string - name: answers sequence: string - name: image_classes sequence: string - name: set_name dtype: string splits: - name: test num_bytes: 3139726 num_examples: 5734 - name: train num_bytes: 22073350 num_examples: 34602 - name: validation num_bytes: 3177854 num_examples: 5000 download_size: 8070116310 dataset_size: 28390930 --- # Dataset Card for TextVQA ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://textvqa.org - **Repository:** https://github.com/facebookresearch/mmf - **Paper:** https://arxiv.org/abs/1904.08920 - **Leaderboard:** https://eval.ai/web/challenges/challenge-page/874/overview - **Point of Contact:** mailto:amanpreet@nyu.edu ### Dataset Summary TextVQA requires models to read and reason about text in images to answer questions about them. Specifically, models need to incorporate a new modality of text present in the images and reason over it to answer TextVQA questions. TextVQA dataset contains 45,336 questions over 28,408 images from the OpenImages dataset. The dataset uses [VQA accuracy](https://visualqa.org/evaluation.html) metric for evaluation. ### Supported Tasks and Leaderboards - `visual-question-answering`: The dataset can be used for Visual Question Answering tasks where given an image, you have to answer a question based on the image. For the TextVQA dataset specifically, the questions require reading and reasoning about the scene text in the given image. ### Languages The questions in the dataset are in English. ## Dataset Structure ### Data Instances A typical sample mainly contains the question in `question` field, an image object in `image` field, OpenImage image id in `image_id` and lot of other useful metadata. 10 answers per questions are contained in the `answers` attribute. For test set, 10 empty strings are contained in the `answers` field as the answers are not available for it. An example look like below: ``` {'question': 'who is this copyrighted by?', 'image_id': '00685bc495504d61', 'image': <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=384x512 at 0x276021C5EB8>, 'image_classes': ['Vehicle', 'Tower', 'Airplane', 'Aircraft'], 'flickr_original_url': 'https://farm2.staticflickr.com/5067/5620759429_4ea686e643_o.jpg', 'flickr_300k_url': 'https://c5.staticflickr.com/6/5067/5620759429_f43a649fb5_z.jpg', 'image_width': 786, 'image_height': 1024, 'answers': ['simon clancy', 'simon ciancy', 'simon clancy', 'simon clancy', 'the brand is bayard', 'simon clancy', 'simon clancy', 'simon clancy', 'simon clancy', 'simon clancy'], 'question_tokens': ['who', 'is', 'this', 'copyrighted', 'by'], 'question_id': 3, 'set_name': 'train' }, ``` ### Data Fields - `question`: string, the question that is being asked about the image - `image_id`: string, id of the image which is same as the OpenImages id - `image`: A `PIL.Image.Image` object containing the image about which the question is being asked. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]`. - `image_classes`: List[str], The OpenImages classes to which the image belongs to. - `flickr_original_url`: string, URL to original image on Flickr - `flickr_300k_url`: string, URL to resized and low-resolution image on Flickr. - `image_width`: int, Width of the original image. - `image_height`: int, Height of the original image. - `question_tokens`: List[str], A pre-tokenized list of question. - `answers`: List[str], List of 10 human-annotated answers for the question. These 10 answers are collected from 10 different users. The list will contain empty strings for test set for which we don't have the answers. - `question_id`: int, Unique id of the question. - `set_name`: string, the set to which this question belongs. ### Data Splits There are three splits. `train`, `validation` and `test`. The `train` and `validation` sets share images with OpenImages `train` set and have their answers available. For test set answers, we return a list of ten empty strings. To get inference results and numbers on `test` set, you need to go to the [EvalAI leaderboard](https://eval.ai/web/challenges/challenge-page/874/overview) and upload your predictions there. Please see instructions at [https://textvqa.org/challenge/](https://textvqa.org/challenge/). ## Dataset Creation ### Curation Rationale From the paper: > Studies have shown that a dominant class of questions asked by visually impaired users on images of their surroundings involves reading text in the image. But today’s VQA models can not read! Our paper takes a first step towards addressing this problem. First, we introduce a new “TextVQA” dataset to facilitate progress on this important problem. Existing datasets either have a small proportion of questions about text (e.g., the VQA dataset) or are too small (e.g., the VizWiz dataset). TextVQA contains 45,336 questions on 28,408 images that require reasoning about text to answer. ### Source Data #### Initial Data Collection and Normalization The initial images were sourced from [OpenImages](https://storage.googleapis.com/openimages/web/factsfigures_v4.html) v4 dataset. These were first filtered based on automatic heuristics using an OCR system where we only took images which had at least some text detected in them. See [annotation process](#annotation-process) section to understand the next stages. #### Who are the source language producers? English Crowdsource Annotators ### Annotations #### Annotation process After the automatic process of filter the images that contain text, the images were manually verified using human annotators making sure that they had text. In next stage, the annotators were asked to write questions involving scene text for the image. For some images, in this stage, two questions were collected whenever possible. Finally, in the last stage, ten different human annotators answer the questions asked in last stage. #### Who are the annotators? Annotators are from one of the major data collection platforms such as AMT. Exact details are not mentioned in the paper. ### Personal and Sensitive Information The dataset does have similar PII issues as OpenImages and can at some times contain human faces, license plates, and documents. Using provided `image_classes` data field is one option to try to filter out some of this information. ## Considerations for Using the Data ### Social Impact of Dataset The paper helped realize the importance of scene text recognition and reasoning in general purpose machine learning applications and has led to many follow-up works including [TextCaps](https://textvqa.org/textcaps) and [TextOCR](https://textvqa.org/textocr). Similar datasets were introduced over the time which specifically focus on sight-disabled users such as [VizWiz](https://vizwiz.org) or focusing specifically on the same problem as TextVQA like [STVQA](https://paperswithcode.com/dataset/st-vqa), [DocVQA](https://arxiv.org/abs/2007.00398v3) and [OCRVQA](https://ocr-vqa.github.io/). Currently, most methods train on combined dataset from TextVQA and STVQA to achieve state-of-the-art performance on both datasets. ### Discussion of Biases Question-only bias where a model is able to answer the question without even looking at the image is discussed in the [paper](https://arxiv.org/abs/1904.08920) which was a major issue with original VQA dataset. The outlier bias in answers is prevented by collecting 10 different answers which are also taken in consideration by the evaluation metric. ### Other Known Limitations - The dataset is english only but does involve images with non-English latin characters so can involve some multi-lingual understanding. - The performance on the dataset is also dependent on the quality of OCR used as the OCR errors can directly lead to wrong answers. - The metric used for calculating accuracy is same as [VQA accuracy](https://visualqa.org/evaluation.html). This involves one-to-one matching with the given answers and thus doesn't allow analyzing one-off errors through OCR. ## Additional Information ### Dataset Curators - [Amanpreet Singh](https://github.com/apsdehal) - Vivek Natarjan - Meet Shah - Yu Jiang - Xinlei Chen - Dhruv Batra - Devi Parikh - Marcus Rohrbach ### Licensing Information CC by 4.0 ### Citation Information ```bibtex @inproceedings{singh2019towards, title={Towards VQA Models That Can Read}, author={Singh, Amanpreet and Natarjan, Vivek and Shah, Meet and Jiang, Yu and Chen, Xinlei and Batra, Dhruv and Parikh, Devi and Rohrbach, Marcus}, booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition}, pages={8317-8326}, year={2019} } ``` ### Contributions Thanks to [@apsdehal](https://github.com/apsdehal) for adding this dataset.
jordane95
null
null
null
false
1
false
jordane95/msmarco-passage-corpus-10-query-0.4
2022-05-05T09:56:37.000Z
null
false
d8ee804686c79825f75e57849b3c28ddad83715f
[]
[ "license:afl-3.0" ]
https://huggingface.co/datasets/jordane95/msmarco-passage-corpus-10-query-0.4/resolve/main/README.md
--- license: afl-3.0 ---
00data00
null
null
null
false
1
false
00data00/data
2022-05-05T10:48:22.000Z
null
false
42ea7b9b5daa33cc0509a6213c48da0ce8ae13aa
[]
[ "license:afl-3.0" ]
https://huggingface.co/datasets/00data00/data/resolve/main/README.md
--- license: afl-3.0 ---
null
null
@inproceedings{haoyietal-informer-2021, author = {Haoyi Zhou and Shanghang Zhang and Jieqi Peng and Shuai Zhang and Jianxin Li and Hui Xiong and Wancai Zhang}, title = {Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting}, booktitle = {The Thirty-Fifth {AAAI} Conference on Artificial Intelligence, {AAAI} 2021, Virtual Conference}, volume = {35}, number = {12}, pages = {11106--11115}, publisher = {{AAAI} Press}, year = {2021}, }
The data of Electricity Transformers from two separated counties in China collected for two years at hourly and 15-min frequencies. Each data point consists of the target value "oil temperature" and 6 power load features. The train/val/test is 12/4/4 months.
false
58
false
ett
2022-11-03T15:59:15.000Z
null
false
e7fe05d403bfb7cfe5b90e3ad73dec1369667eff
[]
[ "arxiv:2012.07436", "annotations_creators:no-annotation", "language_creators:found", "license:cc-by-4.0", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "task_categories:time-series-forecasting", "task_ids:univariate-time-series-forecasting", "task_ids:multi...
https://huggingface.co/datasets/ett/resolve/main/README.md
--- annotations_creators: - no-annotation language_creators: - found language: [] license: - cc-by-4.0 multilinguality: - monolingual pretty_name: Electricity Transformer Temperature size_categories: - 1K<n<10K source_datasets: - original task_categories: - time-series-forecasting task_ids: - univariate-time-series-forecasting - multivariate-time-series-forecasting dataset_info: - config_name: h1 features: - name: start dtype: timestamp[s] - name: target sequence: float32 - name: feat_static_cat sequence: uint64 - name: feat_dynamic_real sequence: sequence: float32 - name: item_id dtype: string splits: - name: test num_bytes: 77508960 num_examples: 240 - name: train num_bytes: 241978 num_examples: 1 - name: validation num_bytes: 33916080 num_examples: 120 download_size: 2589657 dataset_size: 111667018 - config_name: h2 features: - name: start dtype: timestamp[s] - name: target sequence: float32 - name: feat_static_cat sequence: uint64 - name: feat_dynamic_real sequence: sequence: float32 - name: item_id dtype: string splits: - name: test num_bytes: 77508960 num_examples: 240 - name: train num_bytes: 241978 num_examples: 1 - name: validation num_bytes: 33916080 num_examples: 120 download_size: 2417960 dataset_size: 111667018 - config_name: m1 features: - name: start dtype: timestamp[s] - name: target sequence: float32 - name: feat_static_cat sequence: uint64 - name: feat_dynamic_real sequence: sequence: float32 - name: item_id dtype: string splits: - name: test num_bytes: 1239008640 num_examples: 960 - name: train num_bytes: 967738 num_examples: 1 - name: validation num_bytes: 542089920 num_examples: 480 download_size: 10360719 dataset_size: 1782066298 - config_name: m2 features: - name: start dtype: timestamp[s] - name: target sequence: float32 - name: feat_static_cat sequence: uint64 - name: feat_dynamic_real sequence: sequence: float32 - name: item_id dtype: string splits: - name: test num_bytes: 1239008640 num_examples: 960 - name: train num_bytes: 967738 num_examples: 1 - name: validation num_bytes: 542089920 num_examples: 480 download_size: 9677236 dataset_size: 1782066298 --- # Dataset Card for [Electricity Transformer Temperature](https://github.com/zhouhaoyi/ETDataset) ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Electricity Transformer Dataset](https://github.com/zhouhaoyi/ETDataset) - **Repository:** https://github.com/zhouhaoyi/ETDataset - **Paper:** [Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting](https://arxiv.org/abs/2012.07436) - **Point of Contact:** [Haoyi Zhou](mailto:zhouhy@act.buaa.edu.cn) ### Dataset Summary The electric power distribution problem is the distribution of electricity to different areas depending on its sequential usage. But predicting the future demand of a specific area is difficult, as it varies with weekdays, holidays, seasons, weather, temperatures, etc. However, no existing method can perform a long-term prediction based on super long-term real-world data with high precision. Any false predictions may damage the electrical transformer. So currently, without an efficient method to predict future electric usage, managers have to make decisions based on the empirical number, which is much higher than the real-world demands. It causes unnecessary waste of electric and equipment depreciation. On the other hand, the oil temperatures can reflect the condition of the Transformer. One of the most efficient strategies is to predict how the electrical transformers' oil temperature is safe and avoid unnecessary waste. As a result, to address this problem, the authors and Beijing Guowang Fuda Science & Technology Development Company have provided 2-years worth of data. Specifically, the dataset combines short-term periodical patterns, long-term periodical patterns, long-term trends, and many irregular patterns. The dataset are obtained from 2 Electricity Transformers at 2 stations and come in an `1H` (hourly) or `15T` (15-minute) frequency containing 2 year * 365 days * 24 hours * (4 for 15T) times = 17,520 (70,080 for 15T) data points. The target time series is the **O**il **T**emperature and the dataset comes with the following 6 covariates in the univariate setup: * **H**igh **U**se**F**ul **L**oad * **H**igh **U**se**L**ess **L**oad * **M**iddle **U**se**F**ul **L**oad * **M**iddle **U**se**L**ess **L**oad * **L**ow **U**se**F**ul **L**oad * **L**ow **U**se**L**ess **L**oad ### Dataset Usage To load a particular variant of the dataset just specify its name e.g: ```python load_dataset("ett", "m1", multivariate=False) # univariate 15-min frequency dataset from first transformer ``` or to specify a prediction length: ```python load_dataset("ett", "h2", prediction_length=48) # multivariate dataset from second transformer with prediction length of 48 (hours) ``` ### Supported Tasks and Leaderboards The time series data is split into train/val/test set of 12/4/4 months respectively. Given the prediction length (default: 1 day (24 hours or 24*4 15T)) we create rolling windows of this size for the val/test sets. #### `time-series-forecasting` ##### `univariate-time-series-forecasting` The univariate time series forecasting tasks involves learning the future one dimensional `target` values of a time series in a dataset for some `prediction_length` time steps. The performance of the forecast models can then be validated via the ground truth in the `validation` split and tested via the `test` split. The covriates are stored in the `feat_dynamic_real` key of each time series. ##### `multivariate-time-series-forecasting` The multivariate time series forecasting task involves learning the future vector of `target` values of a time series in a dataset for some `prediction_length` time steps. Similar to the univariate setting the performance of a multivariate model can be validated via the ground truth in the `validation` split and tested via the `test` split. ### Languages ## Dataset Structure ### Data Instances A sample from the training set is provided below: ```python { 'start': datetime.datetime(2012, 1, 1, 0, 0), 'target': [14.0, 18.0, 21.0, 20.0, 22.0, 20.0, ...], 'feat_static_cat': [0], 'feat_dynamic_real': [[0.3, 0.4], [0.1, 0.6], ...], 'item_id': 'OT' } ``` ### Data Fields For the univariate regular time series each series has the following keys: * `start`: a datetime of the first entry of each time series in the dataset * `target`: an array[float32] of the actual target values * `feat_static_cat`: an array[uint64] which contains a categorical identifier of each time series in the dataset * `feat_dynamic_real`: optional array of covariate features * `item_id`: a string identifier of each time series in a dataset for reference For the multivariate time series the `target` is a vector of the multivariate dimension for each time point. ### Data Splits The time series data is split into train/val/test set of 12/4/4 months respectively. ## Dataset Creation ### Curation Rationale Develop time series methods that can perform a long-term prediction based on super long-term real-world data with high precision. ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators * [Haoyi Zhou](mailto:zhouhy@act.buaa.edu.cn) ### Licensing Information [Creative Commons Attribution 4.0 International](https://creativecommons.org/licenses/by/4.0/legalcode) ### Citation Information ```tex @inproceedings{haoyietal-informer-2021, author = {Haoyi Zhou and Shanghang Zhang and Jieqi Peng and Shuai Zhang and Jianxin Li and Hui Xiong and Wancai Zhang}, title = {Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting}, booktitle = {The Thirty-Fifth {AAAI} Conference on Artificial Intelligence, {AAAI} 2021, Virtual Conference}, volume = {35}, number = {12}, pages = {11106--11115}, publisher = {{AAAI} Press}, year = {2021}, } ``` ### Contributions Thanks to [@kashif](https://github.com/kashif) for adding this dataset.
ghomasHudson
null
null
null
false
1
false
ghomasHudson/hotpotExtendedAnoLM
2022-05-05T14:09:26.000Z
null
false
7d5a48f50f02568d74fb4d0ca4c333684addc26d
[]
[]
https://huggingface.co/datasets/ghomasHudson/hotpotExtendedAnoLM/resolve/main/README.md
# hotpotExtendedAno-LM Version of hotpotExtended-Annotated formatted for Language Modelling.
truthisneverlinear
null
null
null
false
1
false
truthisneverlinear/eleventh-doctor-scripts
2022-05-05T14:39:18.000Z
null
false
7d23ca7e27a34288f7083619602a7a902938ead9
[]
[ "language:en", "tags:NLP", "tags:conservation", "tags:dialogue" ]
https://huggingface.co/datasets/truthisneverlinear/eleventh-doctor-scripts/resolve/main/README.md
--- language: en tags: - NLP - conservation - dialogue --- # Doctor Who Dialogues This dataset contains all the script lines of Eleventh Doctor from Doctor Who which is a popular TV series. It can be processed and used for chatbots or relevant stuff.
boli-ai-admin
null
null
null
false
1
false
boli-ai-admin/vishal
2022-05-05T14:59:54.000Z
null
false
bd758693b05d8405157aa662564a93edc53c6be7
[]
[ "license:apache-2.0" ]
https://huggingface.co/datasets/boli-ai-admin/vishal/resolve/main/README.md
--- license: apache-2.0 ---
ablam
null
null
null
false
3
false
ablam/gcode
2022-05-05T19:14:30.000Z
null
false
aa413c82b227dd25308df571e8b9d26e034cf2f7
[]
[]
https://huggingface.co/datasets/ablam/gcode/resolve/main/README.md
# Gcode (Geometric code) ## Details **Usage:** 3D printing <br> **Source:** Printables.com <br> **Slicer:** Prusa <br> **Category:** Art & Design <br> **Subcategory:** Sculptures <br> **Models:** 400 <br> **Sliced files:** 740 (some models have many) <br> **Data format:** txt <br> **Train-test split:** 90/10 <br> **Size:** 11GB
adsabs
null
null
null
false
36
false
adsabs/WIESP2022-NER
2022-11-03T14:49:38.000Z
null
false
5a1e54edeb5345cad0cd899f7b188b9337e16413
[]
[ "annotations_creators:expert-generated", "language_creators:found", "language:en", "license:cc-by-4.0", "multilinguality:monolingual", "size_categories:1K<n<10K", "task_categories:token-classification", "task_ids:named-entity-recognition" ]
https://huggingface.co/datasets/adsabs/WIESP2022-NER/resolve/main/README.md
--- annotations_creators: - expert-generated language_creators: - found language: - en license: - cc-by-4.0 multilinguality: - monolingual pretty_name: 'WIESP2022-NER' size_categories: - 1K<n<10K source_datasets: [] task_categories: - token-classification task_ids: - named-entity-recognition --- # Dataset for the first <a href="https://ui.adsabs.harvard.edu/WIESP/" style="color:blue">Workshop on Information Extraction from Scientific Publications (WIESP/2022)</a>. ## Dataset Description Datasets with text fragments from astrophysics papers, provided by the [NASA Astrophysical Data System](https://ui.adsabs.harvard.edu/) with manually tagged astronomical facilities and other entities of interest (e.g., celestial objects). Datasets are in JSON Lines format (each line is a json dictionary). The datasets are formatted similarly to the CONLL2003 format. Each token is associated with an NER tag. The tags follow the "B-" and "I-" convention from the [IOB2 syntax]("https://en.wikipedia.org/wiki/Inside%E2%80%93outside%E2%80%93beginning_(tagging)") Each entry consists of a dictionary with the following keys: - `"unique_id"`: a unique identifier for this data sample. Must be included in the predictions. - `"tokens"`: the list of tokens (strings) that form the text of this sample. Must be included in the predictions. - `"ner_tags"`: the list of NER tags (in IOB2 format) The following keys are not strictly needed by the participants: - `"ner_ids"`: the pre-computed list of ids corresponding ner_tags, as given by the dictionary in ner_tags.json - `"label_studio_id"`, `"section"`, `"bibcode"`: references for internal NASA/ADS use. ## Instructions for Workshop participants: How to load the data using the Huggingface library: ```python from datasets import load_dataset dataset = load_dataset("adsabs/WIESP2022-NER") ``` How to load the data if you cloned the repository locally: (assuming `./WIESP2022-NER-DEV.jsonl` is in the current directory, change as needed) - python (as list of dictionaries): ```python import json with open("./WIESP2022-NER-DEV.jsonl", 'r') as f:     wiesp_dev_json = [json.loads(l) for l in list(f)] ``` - into Huggingface (as a Huggingface Dataset): ```python from datasets import Dataset wiesp_dev_from_json = Dataset.from_json(path_or_paths="./WIESP2022-NER-DEV.jsonl") ``` How to compute your scores on the training data: 1. format your predictions as a list of dictionaries, each with the same `"unique_id"` and `"tokens"` keys from the dataset, as well as the list of predicted NER tags under the `"pred_ner_tags"` key (see `WIESP2022-NER-DEV-sample-predictions.jsonl` for an example). 2. pass the references and predictions datasets to the `compute_MCC()` and `compute_seqeval()` functions (from the `.py` files with the same names). Requirement to run the scoring scripts: [NumPy](https://numpy.org/install/) [scikit-learn](https://scikit-learn.org/stable/install.html) [seqeval](https://github.com/chakki-works/seqeval#installation) To get scores on the validation data, zip your predictions file (a single `.jsonl' file formatted following the same instructions as above) and upload the `.zip` file to the [Codalabs](https://codalab.lisn.upsaclay.fr/competitions/5062) competition. ## File list ``` ├── WIESP2022-NER-TRAINING.jsonl : 1753 samples for training. ├── WIESP2022-NER-DEV.jsonl : 20 samples for development. ├── WIESP2022-NER-DEV-sample-predictions.jsonl : an example file with properly formatted predictions on the development data. ├── WIESP2022-NER-VALIDATION-NO-LABELS.jsonl : 1366 samples for validation without the NER labels. Used for the WIESP2022 workshop. ├── WIESP2022-NER-VALIDATION.jsonl : 1366 samples for validation ├── WIESP2022-NER-TESTING-NO-LABELS.jsonl : 2505 samples for testing without the NER labels. Used for the WIESP2022 workshop. ├── WIESP2022-NER-TESTING.jsonl : 2505 samples for testing ├── README.MD : this file. ├── tag_definitions.md : short descriptions and examples of the tags used in the task. └── scoring-scripts/ : scripts used to evaluate submissions. ├── compute_MCC.py : computes the Matthews correlation coefficient between two datasets. └── compute_seqeval.py : computes the seqeval scores (precision, recall, f1, overall and for each class) between two datasets. ```
nateraw
null
null
null
false
1
false
nateraw/gradio-guides-files
2022-05-05T21:07:09.000Z
null
false
6ef57a87aea2fddaa26e86f7f3eda48ba3171a8b
[]
[ "license:mit" ]
https://huggingface.co/datasets/nateraw/gradio-guides-files/resolve/main/README.md
--- license: mit ---
BennoKrojer
null
null
null
false
1
false
BennoKrojer/ImageCoDe
2022-05-13T21:26:08.000Z
null
false
b7b7a73dbc74b681b384048716232b243ad25a99
[]
[ "arxiv:2203.15867", "license:afl-3.0" ]
https://huggingface.co/datasets/BennoKrojer/ImageCoDe/resolve/main/README.md
--- license: afl-3.0 --- # Dataset Card for ImageCoDe To get started quickly, load descriptions via: ``` from datasets import load_dataset examples = load_dataset('BennoKrojer/ImageCoDe') ``` And download `image_sets.zip` for all images sets (each directory consisting of 10 images). ## Dataset Description - **Homepage & Leaderboard:** https://mcgill-nlp.github.io/imagecode/ - **Repository:** https://github.com/McGill-NLP/imagecode - **Paper:** https://arxiv.org/abs/2203.15867 - **Point of Contact:** benno DOT krojer ÄT gmail DOT com ### Dataset Summary We introduce ImageCoDe, a vision-and-language benchmark that requires contextual language understanding in the form of pragmatics, temporality, long descriptions and visual nuances. The task: Given a detailed description, retrieve the target image among 10 minimally contrastive images. ImageCoDe contains 21K descriptions and 94K images. THe images are primarily frames based on video datasets. ## Dataset Structure ### Data Instances An instance contains a description, the corresponding image set name, and the target index: ``` {"image_set": "video-storytelling-videowedding_de8dLXvgV-I-shot6_0", "image_index": "8", "description": "The flowers the woman in the teal strapless dress is carrying are completely obscured by the man in the black shirt's head. "} ``` ### Data Splits | Dataset Split | Number of Descriptions in Split | | ------------- |----------------------------- | | Train | 16,594 | | Validation | 2,302 | | Test | 2,306 | ## Dataset Creation ### Curation Rationale The main goal of ImageCoDe is to highlight weaknesses of recent Vision-and-Language models regarding complex language and fine-grained visual representations. In addition, we found that the dataset offers plenty of pragmatic examples and is therefore suitable for studying pragmatics.
nateraw
null
null
null
false
1
false
nateraw/test-imagefolder-metadata
2022-05-06T01:51:57.000Z
null
false
3adb85380e4645c1fcf914a21ba4013410c0fa2b
[]
[]
https://huggingface.co/datasets/nateraw/test-imagefolder-metadata/resolve/main/README.md
# test-imagefolder-metadata
ntt123
null
null
null
false
2
false
ntt123/viet-tts-dataset
2022-05-06T09:03:02.000Z
null
false
e1623289640e76fe2209e753a1b78a2200edc34e
[]
[ "license:cc-by-nc-4.0" ]
https://huggingface.co/datasets/ntt123/viet-tts-dataset/resolve/main/README.md
--- license: cc-by-nc-4.0 --- # Vietnamese Text-To-Speech dataset (VietTTS-v1.1) 🔔🔔🔔 visit https://github.com/NTT123/vietTTS for a vietnamese TTS library (included pretrained models). 🔔🔔🔔 The text is from a collection of novels and short stories from the author "Vu Trong Phung." The text is in public domain. The audio is generated by Google Text-to-Speech offline engine on Android. The audio is NOT for commercial use. Dataset size: `5.4G`. Total audio duration: `35.9 hours`. ### Text-audio samples - Sample 1: + Audio: [file1](https://huggingface.co/datasets/ntt123/viet-tts-dataset/blob/main/000000.wav) + Text: `"Ai" đây tức là một kẻ ăn mày vậy. Anh ta chưa kịp quay đi thì đã thấy mấy con chó vàng chạy xồng xộc ra cứ nhảy xổ vào chân anh.` - Sample 2: + Audio: [file2](https://huggingface.co/datasets/ntt123/viet-tts-dataset/blob/main/022878.wav) + Text: `Ừ, thế mày đã nuôi được bố mẹ mày bữa nào chưa, hay xưa nay vẫn báo hại cơm cha áo mẹ mãi? Mấy hôm thấy ông đơ mặt không thèm nói, mày lại làm già à?` ### Download Get the dataset from here: [link](https://huggingface.co/datasets/ntt123/viet-tts-dataset/blob/main/viet-tts.tar.gz). Or, run the following commands: ``` wget https://huggingface.co/datasets/ntt123/viet-tts-dataset/resolve/main/viet-tts.tar.gz -O viet-tts.tar.gz mkdir -p dataset tar -C dataset -xzf viet-tts.tar.gz ``` `dataset` directory structure: ``` dataset ├── collections.txt ├── meta_data.tsv └── wav ├── 000000.wav ├── 000001.wav ├── 000002.wav ├── 000003.wav ... ``` ### Statistics - Number of clips: 22884 clips. - Shortest audio clip: 0.46 seconds. - Median clip duration: 5.46 seconds. - Mean clip duration: 5.65 seconds. - Longest audio clip: 15.4 seconds. ### Vũ Trọng Phụng's collections - Bệnh Lao Chữa Bằng Mồm Hay Là ... Thầy Lang Bất Hủ, 1934? - Cạm Bẫy Người, 1933. - Cơm Thầy Cơm Cô, 1936. - Đời Là Một Cuộc Chiến Đấu,1939. - Dứt Tình, 1934. - Giông Tố, 1936. - Gương Tống Tiền, N/A. - Hồ Sê Líu, Hồ Líu Sê Sàng, 1936. - Kỹ Nghệ Lấy Tây, 1934. - Làm Đĩ, 1936. - Lấy Nhau Vì Tình, 1937. - Lấy Vợ Xấu, 1937. - Lòng Tự Ái, 1937. - Máu Mê, 1937. - Một Cái Chết, 1931. - Một Con Chó Hay Chim Chuột, 1937. - Một Đồng Bạc, 1939. - Người Có Quyền, 1937. - Sao Mày Không Vỡ Nắp Ơi!, 1934. - Số Đỏ, 1936. - Sư Cụ Triết Lý, 1935. - Trúng Số Độc Đắc, 1938. - Tự Do, 1937. - Từ Lý Thuyết Đến Thực Hành, N/A. - Vỡ Đê, 1936.
searle-j
null
@article{jeon2022user, title={User Guide for KOTE: Korean Online Comments Emotions Dataset}, author={Jeon, Duyoung and Lee, Junho and Kim, Cheongtag}, journal={arXiv preprint arXiv:2205.05300}, year={2022} }
50k Korean online comments labeled for 44 emotion categories.
false
32
false
searle-j/kote
2022-10-20T19:16:24.000Z
null
false
66f0eefe4b675a5d5411c7aa08e2c97fc9a9b17f
[]
[ "annotations_creators:crowdsourced", "language:kor", "license:mit", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "task_categories:text-classification", "task_ids:multi-class-classification", "task_ids:multi-label-classification", "task_ids:text-classific...
https://huggingface.co/datasets/searle-j/kote/resolve/main/README.md
--- annotations_creators: - crowdsourced language: - kor license: - mit multilinguality: - monolingual pretty_name: kote size_categories: - 10K<n<100K source_datasets: - original task_categories: - text-classification task_ids: - multi-class-classification - multi-label-classification - text-classification-other-emotion ---
null
null
@InProceedings{pmlr-v174-pal22a, title = {MedMCQA: A Large-scale Multi-Subject Multi-Choice Dataset for Medical domain Question Answering}, author = {Pal, Ankit and Umapathi, Logesh Kumar and Sankarasubbu, Malaikannan}, booktitle = {Proceedings of the Conference on Health, Inference, and Learning}, pages = {248--260}, year = {2022}, editor = {Flores, Gerardo and Chen, George H and Pollard, Tom and Ho, Joyce C and Naumann, Tristan}, volume = {174}, series = {Proceedings of Machine Learning Research}, month = {07--08 Apr}, publisher = {PMLR}, pdf = {https://proceedings.mlr.press/v174/pal22a/pal22a.pdf}, url = {https://proceedings.mlr.press/v174/pal22a.html}, abstract = {This paper introduces MedMCQA, a new large-scale, Multiple-Choice Question Answering (MCQA) dataset designed to address real-world medical entrance exam questions. More than 194k high-quality AIIMS & NEET PG entrance exam MCQs covering 2.4k healthcare topics and 21 medical subjects are collected with an average token length of 12.77 and high topical diversity. Each sample contains a question, correct answer(s), and other options which requires a deeper language understanding as it tests the 10+ reasoning abilities of a model across a wide range of medical subjects & topics. A detailed explanation of the solution, along with the above information, is provided in this study.} }
MedMCQA is a large-scale, Multiple-Choice Question Answering (MCQA) dataset designed to address real-world medical entrance exam questions. MedMCQA has more than 194k high-quality AIIMS & NEET PG entrance exam MCQs covering 2.4k healthcare topics and 21 medical subjects are collected with an average token length of 12.77 and high topical diversity. The dataset contains questions about the following topics: Anesthesia, Anatomy, Biochemistry, Dental, ENT, Forensic Medicine (FM) Obstetrics and Gynecology (O&G), Medicine, Microbiology, Ophthalmology, Orthopedics Pathology, Pediatrics, Pharmacology, Physiology, Psychiatry, Radiology Skin, Preventive & Social Medicine (PSM) and Surgery
false
273
false
medmcqa
2022-11-03T16:07:43.000Z
medmcqa
false
6f73325ca5f480a097a7201187b7ec7e8620a899
[]
[ "annotations_creators:no-annotation", "language_creators:expert-generated", "language:en", "license:apache-2.0", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "task_categories:question-answering", "task_categories:multiple-choice", "task_ids:multiple-choic...
https://huggingface.co/datasets/medmcqa/resolve/main/README.md
--- annotations_creators: - no-annotation language_creators: - expert-generated language: - en license: - apache-2.0 multilinguality: - monolingual size_categories: - 100K<n<1M source_datasets: - original task_categories: - question-answering - multiple-choice task_ids: - multiple-choice-qa - open-domain-qa paperswithcode_id: medmcqa pretty_name: MedMCQA dataset_info: features: - name: id dtype: string - name: question dtype: string - name: opa dtype: string - name: opb dtype: string - name: opc dtype: string - name: opd dtype: string - name: cop dtype: class_label: names: 0: a 1: b 2: c 3: d - name: choice_type dtype: string - name: exp dtype: string - name: subject_name dtype: string - name: topic_name dtype: string splits: - name: test num_bytes: 1447829 num_examples: 6150 - name: train num_bytes: 131904057 num_examples: 182822 - name: validation num_bytes: 2221468 num_examples: 4183 download_size: 55285460 dataset_size: 135573354 --- # Dataset Card for MedMCQA ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://medmcqa.github.io - **Repository:** https://github.com/medmcqa/medmcqa - **Paper:** [MedMCQA: A Large-scale Multi-Subject Multi-Choice Dataset for Medical domain Question Answering](https://proceedings.mlr.press/v174/pal22a) - **Leaderboard:** https://paperswithcode.com/dataset/medmcqa - **Point of Contact:** [Aaditya Ura](mailto:aadityaura@gmail.com) ### Dataset Summary MedMCQA is a large-scale, Multiple-Choice Question Answering (MCQA) dataset designed to address real-world medical entrance exam questions. MedMCQA has more than 194k high-quality AIIMS & NEET PG entrance exam MCQs covering 2.4k healthcare topics and 21 medical subjects are collected with an average token length of 12.77 and high topical diversity. Each sample contains a question, correct answer(s), and other options which require a deeper language understanding as it tests the 10+ reasoning abilities of a model across a wide range of medical subjects & topics. A detailed explanation of the solution, along with the above information, is provided in this study. MedMCQA provides an open-source dataset for the Natural Language Processing community. It is expected that this dataset would facilitate future research toward achieving better QA systems. The dataset contains questions about the following topics: - Anesthesia - Anatomy - Biochemistry - Dental - ENT - Forensic Medicine (FM) - Obstetrics and Gynecology (O&G) - Medicine - Microbiology - Ophthalmology - Orthopedics - Pathology - Pediatrics - Pharmacology - Physiology - Psychiatry - Radiology - Skin - Preventive & Social Medicine (PSM) - Surgery ### Supported Tasks and Leaderboards multiple-choice-QA, open-domain-QA: The dataset can be used to train a model for multi-choice questions answering, open domain questions answering. Questions in these exams are challenging and generally require deeper domain and language understanding as it tests the 10+ reasoning abilities across a wide range of medical subjects & topics. ### Languages The questions and answers are available in English. ## Dataset Structure ### Data Instances ``` { "question":"A 40-year-old man presents with 5 days of productive cough and fever. Pseudomonas aeruginosa is isolated from a pulmonary abscess. CBC shows an acute effect characterized by marked leukocytosis (50,000 mL) and the differential count reveals a shift to left in granulocytes. Which of the following terms best describes these hematologic findings?", "exp": "Circulating levels of leukocytes and their precursors may occasionally reach very high levels (>50,000 WBC mL). These extreme elevations are sometimes called leukemoid reactions because they are similar to the white cell counts observed in leukemia, from which they must be distinguished. The leukocytosis occurs initially because of the accelerated release of granulocytes from the bone marrow (caused by cytokines, including TNF and IL-1) There is a rise in the number of both mature and immature neutrophils in the blood, referred to as a shift to the left. In contrast to bacterial infections, viral infections (including infectious mononucleosis) are characterized by lymphocytosis Parasitic infestations and certain allergic reactions cause eosinophilia, an increase in the number of circulating eosinophils. Leukopenia is defined as an absolute decrease in the circulating WBC count.", "cop":1, "opa":"Leukemoid reaction", "opb":"Leukopenia", "opc":"Myeloid metaplasia", "opd":"Neutrophilia", "subject_name":"Pathology", "topic_name":"Basic Concepts and Vascular changes of Acute Inflammation", "id":"4e1715fe-0bc3-494e-b6eb-2d4617245aef", "choice_type":"single" } ``` ### Data Fields - `id` : a string question identifier for each example - `question` : question text (a string) - `opa` : Option A - `opb` : Option B - `opc` : Option C - `opd` : Option D - `cop` : Correct option, i.e., 1,2,3,4 - `choice_type` ({"single", "multi"}): Question choice type. - "single": Single-choice question, where each choice contains a single option. - "multi": Multi-choice question, where each choice contains a combination of multiple suboptions. - `exp` : Expert's explanation of the answer - `subject_name` : Medical Subject name of the particular question - `topic_name` : Medical topic name from the particular subject ### Data Splits The goal of MedMCQA is to emulate the rigor of real word medical exams. To enable that, a predefined split of the dataset is provided. The split is by exams instead of the given questions. This also ensures the reusability and generalization ability of the models. The training set of MedMCQA consists of all the collected mock & online test series, whereas the test set consists of all AIIMS PG exam MCQs (years 1991-present). The development set consists of NEET PG exam MCQs (years 2001-present) to approximate real exam evaluation. Similar questions from train , test and dev set were removed based on similarity. The final split sizes are as follow: | | Train | Test | Valid | | ----- | ------ | ----- | ---- | | Question #| 182,822 | 6,150 | 4,183| | Vocab | 94,231 | 11,218 | 10,800 | | Max Ques tokens | 220 | 135| 88 | | Max Ans tokens | 38 | 21 | 25 | ## Dataset Creation ### Curation Rationale Before this attempt, very few works have been done to construct biomedical MCQA datasets (Vilares and Gomez-Rodr, 2019), and they are (1) mostly small, containing up to few thousand questions, and (2) cover a limited number of Medical topics and Subjects. This paper addresses the aforementioned limitations by introducing MedMCQA, a new large-scale, Multiple-Choice Question Answering (MCQA) dataset designed to address real-world medical entrance exam questions. ### Source Data #### Initial Data Collection and Normalization Historical Exam questions from official websites - AIIMS & NEET PG (1991- present) The raw data is collected from open websites and books #### Who are the source language producers? The dataset was created by Ankit Pal, Logesh Kumar Umapathi and Malaikannan Sankarasubbu ### Annotations #### Annotation process The dataset does not contain any additional annotations. #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information [Needs More Information] ### Citation Information If you find this useful in your research, please consider citing the dataset paper ``` @InProceedings{pmlr-v174-pal22a, title = {MedMCQA: A Large-scale Multi-Subject Multi-Choice Dataset for Medical domain Question Answering}, author = {Pal, Ankit and Umapathi, Logesh Kumar and Sankarasubbu, Malaikannan}, booktitle = {Proceedings of the Conference on Health, Inference, and Learning}, pages = {248--260}, year = {2022}, editor = {Flores, Gerardo and Chen, George H and Pollard, Tom and Ho, Joyce C and Naumann, Tristan}, volume = {174}, series = {Proceedings of Machine Learning Research}, month = {07--08 Apr}, publisher = {PMLR}, pdf = {https://proceedings.mlr.press/v174/pal22a/pal22a.pdf}, url = {https://proceedings.mlr.press/v174/pal22a.html}, abstract = {This paper introduces MedMCQA, a new large-scale, Multiple-Choice Question Answering (MCQA) dataset designed to address real-world medical entrance exam questions. More than 194k high-quality AIIMS &amp; NEET PG entrance exam MCQs covering 2.4k healthcare topics and 21 medical subjects are collected with an average token length of 12.77 and high topical diversity. Each sample contains a question, correct answer(s), and other options which requires a deeper language understanding as it tests the 10+ reasoning abilities of a model across a wide range of medical subjects &amp; topics. A detailed explanation of the solution, along with the above information, is provided in this study.} } ``` ### Contributions Thanks to [@monk1337](https://github.com/monk1337) for adding this dataset.
Fhrozen
null
null
null
false
35
false
Fhrozen/FSD50k
2022-05-27T08:50:25.000Z
null
false
67e4d8c2570caef0f90d48fdb756b337875d91db
[]
[ "arxiv:2010.00475", "license:cc-by-4.0", "annotations_creators:unknown", "language_creators:unknown", "size_categories:10K<n<100K", "source_datasets:unknown", "task_categories:audio-classification", "task_ids:other-audio-slot-filling" ]
https://huggingface.co/datasets/Fhrozen/FSD50k/resolve/main/README.md
--- license: cc-by-4.0 annotations_creators: - unknown language_creators: - unknown size_categories: - 10K<n<100K source_datasets: - unknown task_categories: - audio-classification task_ids: - other-audio-slot-filling --- # Freesound Dataset 50k (FSD50K) ## Important **This data set is a copy from the original one located at Zenodo.** ## Dataset Description - **Homepage:** [FSD50K](https://zenodo.org/record/4060432) - **Repository:** [GitHub](https://github.com/edufonseca/FSD50K_baseline) - **Paper:** [FSD50K: An Open Dataset of Human-Labeled Sound Events](https://arxiv.org/abs/2010.00475) - **Leaderboard:** [Paperswithcode Leaderboard](https://paperswithcode.com/dataset/fsd50k) ## Citation If you use the FSD50K dataset, or part of it, please cite our paper: >Eduardo Fonseca, Xavier Favory, Jordi Pons, Frederic Font, Xavier Serra. "FSD50K: an Open Dataset of Human-Labeled Sound Events", arXiv 2020. ### Data curators Eduardo Fonseca, Xavier Favory, Jordi Pons, Mercedes Collado, Ceren Can, Rachit Gupta, Javier Arredondo, Gary Avendano and Sara Fernandez ### Contact You are welcome to contact Eduardo Fonseca should you have any questions at eduardo.fonseca@upf.edu. ## About FSD50K Freesound Dataset 50k (or **FSD50K** for short) is an open dataset of human-labeled sound events containing 51,197 <a href="https://freesound.org/">Freesound</a> clips unequally distributed in 200 classes drawn from the <a href="https://research.google.com/audioset/ontology/index.html">AudioSet Ontology</a> [1]. FSD50K has been created at the <a href="https://www.upf.edu/web/mtg">Music Technology Group of Universitat Pompeu Fabra</a>. What follows is a brief summary of FSD50K's most important characteristics. Please have a look at our paper (especially Section 4) to extend the basic information provided here with relevant details for its usage, as well as discussion, limitations, applications and more. **Basic characteristics:** - FSD50K is composed mainly of sound events produced by physical sound sources and production mechanisms. - Following AudioSet Ontology’s main families, the FSD50K vocabulary encompasses mainly *Human sounds*, *Sounds of things*, *Animal*, *Natural sounds* and *Music*. - The dataset has 200 sound classes (144 leaf nodes and 56 intermediate nodes) hierarchically organized with a subset of the AudioSet Ontology. The vocabulary can be inspected in `vocabulary.csv` (see Files section below). - FSD50K contains 51,197 audio clips totalling 108.3 hours of audio. - The audio content has been manually labeled by humans following a data labeling process using the <a href="https://annotator.freesound.org/">Freesound Annotator</a> platform [2]. - Clips are of variable length from 0.3 to 30s, due to the diversity of the sound classes and the preferences of Freesound users when recording sounds. - Ground truth labels are provided at the clip-level (i.e., weak labels). - The dataset poses mainly a multi-label sound event classification problem (but also allows a variety of sound event research tasks, see Sec. 4D). - All clips are provided as uncompressed PCM 16 bit 44.1 kHz mono audio files. - The audio clips are grouped into a development (*dev*) set and an evaluation (*eval*) set such that they do not have clips from the same Freesound uploader. **Dev set:** - 40,966 audio clips totalling 80.4 hours of audio - Avg duration/clip: 7.1s - 114,271 smeared labels (i.e., labels propagated in the upwards direction to the root of the ontology) - Labels are correct but could be occasionally incomplete - A train/validation split is provided (Sec. 3H). If a different split is used, it should be specified for reproducibility and fair comparability of results (see Sec. 5C of our paper) **Eval set:** - 10,231 audio clips totalling 27.9 hours of audio - Avg duration/clip: 9.8s - 38,596 smeared labels - Eval set is labeled exhaustively (labels are correct and complete for the considered vocabulary) **NOTE:** All classes in FSD50K are represented in AudioSet, except `Crash cymbal`, `Human group actions`, `Human voice`, `Respiratory sounds`, and `Domestic sounds, home sounds`. ## License All audio clips in FSD50K are released under Creative Commons (CC) licenses. Each clip has its own license as defined by the clip uploader in Freesound, some of them requiring attribution to their original authors and some forbidding further commercial reuse. For attribution purposes and to facilitate attribution of these files to third parties, we include a mapping from the audio clips to their corresponding licenses. The licenses are specified in the files `dev_clips_info_FSD50K.json` and `eval_clips_info_FSD50K.json`. These licenses are CC0, CC-BY, CC-BY-NC and CC Sampling+. In addition, FSD50K as a whole is the result of a curation process and it has an additional license: FSD50K is released under <a href="https://creativecommons.org/licenses/by/4.0/">CC-BY</a>. This license is specified in the `LICENSE-DATASET` file downloaded with the `FSD50K.doc` zip file. ## Files FSD50K can be downloaded as a series of zip files with the following directory structure: <div class="highlight"><pre><span></span>root │ └───clips/ Audio clips │ │ │ └─── dev/ Audio clips in the dev set │ │ │ └─── eval/ Audio clips in the eval set │ └───labels/ Files for FSD50K's ground truth │ │ │ └─── dev.csv Ground truth for the dev set │ │ │ └─── eval.csv Ground truth for the eval set │ │ │ └─── vocabulary.csv List of 200 sound classes in FSD50K │ └───metadata/ Files for additional metadata │ │ │ └─── class_info_FSD50K.json Metadata about the sound classes │ │ │ └─── dev_clips_info_FSD50K.json Metadata about the dev clips │ │ │ └─── eval_clips_info_FSD50K.json Metadata about the eval clips │ │ │ └─── pp_pnp_ratings_FSD50K.json PP/PNP ratings │ │ │ └─── collection/ Files for the *sound collection* format │ │ └───README.md The dataset description file that you are reading │ └───LICENSE-DATASET License of the FSD50K dataset as an entity </pre></div> Each row (i.e. audio clip) of `dev.csv` contains the following information: - `fname`: the file name without the `.wav` extension, e.g., the fname `64760` corresponds to the file `64760.wav` in disk. This number is the Freesound id. We always use Freesound ids as filenames. - `labels`: the class labels (i.e., the ground truth). Note these class labels are *smeared*, i.e., the labels have been propagated in the upwards direction to the root of the ontology. More details about the label smearing process can be found in Appendix D of our paper. - `mids`: the Freebase identifiers corresponding to the class labels, as defined in the <a href="https://github.com/audioset/ontology/blob/master/ontology.json">AudioSet Ontology specification</a> - `split`: whether the clip belongs to *train* or *val* (see paper for details on the proposed split) Rows in `eval.csv` follow the same format, except that there is no `split` column. **NOTE:** We use a slightly different format than AudioSet for the naming of class labels in order to avoid potential problems with spaces, commas, etc. Example: we use `Accelerating_and_revving_and_vroom` instead of the original `Accelerating, revving, vroom`. You can go back to the original AudioSet naming using the information provided in `vocabulary.csv` (class label and mid for the 200 classes of FSD50K) and the <a href="https://github.com/audioset/ontology/blob/master/ontology.json">AudioSet Ontology specification</a>. ### Files with additional metadata (metadata/) To allow a variety of analysis and approaches with FSD50K, we provide the following metadata: 1. `class_info_FSD50K.json`: python dictionary where each entry corresponds to one sound class and contains: `FAQs` utilized during the annotation of the class, `examples` (representative audio clips), and `verification_examples` (audio clips presented to raters during annotation as a quality control mechanism). Audio clips are described by the Freesound id. **NOTE:** It may be that some of these examples are not included in the FSD50K release. 2. `dev_clips_info_FSD50K.json`: python dictionary where each entry corresponds to one dev clip and contains: title, description, tags, clip license, and the uploader name. All these metadata are provided by the uploader. 3. `eval_clips_info_FSD50K.json`: same as before, but with eval clips. 4. `pp_pnp_ratings.json`: python dictionary where each entry corresponds to one clip in the dataset and contains the PP/PNP ratings for the labels associated with the clip. More specifically, these ratings are gathered for the labels validated in **the validation task** (Sec. 3 of paper). This file includes 59,485 labels for the 51,197 clips in FSD50K. Out of these labels: - 56,095 labels have inter-annotator agreement (PP twice, or PNP twice). Each of these combinations can be occasionally accompanied by other (non-positive) ratings. - 3390 labels feature other rating configurations such as *i)* only one PP rating and one PNP rating (and nothing else). This can be considered inter-annotator agreement at the ``Present” level; *ii)* only one PP rating (and nothing else); *iii)* only one PNP rating (and nothing else). Ratings' legend: PP=1; PNP=0.5; U=0; NP=-1. **NOTE:** The PP/PNP ratings have been provided in the *validation* task. Subsequently, a subset of these clips corresponding to the eval set was exhaustively labeled in the *refinement* task, hence receiving additional labels in many cases. For these eval clips, you might want to check their labels in `eval.csv` in order to have more info about their audio content (see Sec. 3 for details). 5. `collection/`: This folder contains metadata for what we call the ***sound collection format***. This format consists of the raw annotations gathered, featuring all generated class labels without any restriction. We provide the *collection* format to make available some annotations that do not appear in the FSD50K *ground truth* release. This typically happens in the case of classes for which we gathered human-provided annotations, but that were discarded in the FSD50K release due to data scarcity (more specifically, they were merged with their parents). In other words, the main purpose of the `collection` format is to make available annotations for tiny classes. The format of these files in analogous to that of the files in `FSD50K.ground_truth/`. A couple of examples show the differences between **collection** and **ground truth** formats: `clip`: `labels_in_collection` -- `labels_in_ground_truth` `51690`: `Owl` -- `Bird,Wild_Animal,Animal` `190579`: `Toothbrush,Electric_toothbrush` -- `Domestic_sounds_and_home_sounds` In the first example, raters provided the label `Owl`. However, due to data scarcity, `Owl` labels were merged into their parent `Bird`. Then, labels `Wild_Animal,Animal` were added via label propagation (smearing). The second example shows one of the most extreme cases, where raters provided the labels `Electric_toothbrush,Toothbrush`, which both had few data. Hence, they were merged into Toothbrush's parent, which unfortunately is `Domestic_sounds_and_home_sounds` (a rather vague class containing a variety of children sound classes). **NOTE:** Labels in the collection format are not smeared. **NOTE:** While in FSD50K's ground truth the vocabulary encompasses 200 classes (common for dev and eval), since the *collection* format is composed of raw annotations, the vocabulary here is much larger (over 350 classes), and it is slightly different in dev and eval. For further questions, please contact eduardo.fonseca@upf.edu, or join the <a href="https://groups.google.com/g/freesound-annotator">freesound-annotator Google Group</a>. ## Download Clone this repository: ``` git clone https://huggingface.co/Fhrozen/FSD50k ``` ## Baseline System Several baseline systems for FSD50K are available at <a href="https://github.com/edufonseca/FSD50K_baseline">https://github.com/edufonseca/FSD50K_baseline</a>. The experiments are described in Sec 5 of our paper. ## References and links [1] Jort F Gemmeke, Daniel PW Ellis, Dylan Freedman, Aren Jansen, Wade Lawrence, R Channing Moore, Manoj Plakal, and Marvin Ritter. "Audio set: An ontology and human-labeled dataset for audio events." In Proceedings of the International Conference on Acoustics, Speech and Signal Processing, 2017. [<a href="https://ai.google/research/pubs/pub45857">PDF</a>] [2] Eduardo Fonseca, Jordi Pons, Xavier Favory, Frederic Font, Dmitry Bogdanov, Andres Ferraro, Sergio Oramas, Alastair Porter, and Xavier Serra. "Freesound Datasets: A Platform for the Creation of Open Audio Datasets." In Proceedings of the International Conference on Music Information Retrieval, 2017. [<a href="https://repositori.upf.edu/bitstream/handle/10230/33299/fonseca_ismir17_freesound.pdf">PDF</a>] Companion site for FSD50K: <a href="https://annotator.freesound.org/fsd/release/FSD50K/">https://annotator.freesound.org/fsd/release/FSD50K/</a> Freesound Annotator: <a href="https://annotator.freesound.org/">https://annotator.freesound.org/</a> Freesound: <a href="https://freesound.org">https://freesound.org</a> Eduardo Fonseca's personal website: <a href="http://www.eduardofonseca.net/">http://www.eduardofonseca.net/</a> More datasets collected by us: <a href="http://www.eduardofonseca.net/datasets/">http://www.eduardofonseca.net/datasets/</a> ## Acknowledgments The authors would like to thank everyone who contributed to FSD50K with annotations, and especially Mercedes Collado, Ceren Can, Rachit Gupta, Javier Arredondo, Gary Avendano and Sara Fernandez for their commitment and perseverance. The authors would also like to thank Daniel P.W. Ellis and Manoj Plakal from Google Research for valuable discussions. This work is partially supported by the European Union’s Horizon 2020 research and innovation programme under grant agreement No 688382 <a href="https://www.audiocommons.org/">AudioCommons</a>, and two Google Faculty Research Awards <a href="https://ai.googleblog.com/2018/03/google-faculty-research-awards-2017.html">2017</a> and <a href="https://ai.googleblog.com/2019/03/google-faculty-research-awards-2018.html">2018</a>, and the Maria de Maeztu Units of Excellence Programme (MDM-2015-0502).
filwsyl
null
@article{lecun2010mnist, title={MNIST handwritten digit database}, author={LeCun, Yann and Cortes, Corinna and Burges, CJ}, journal={ATT Labs [Online]. Available: http://yann.lecun.com/exdb/mnist}, volume={2}, year={2010} }
The MNIST dataset consists of 70,000 28x28 black-and-white images in 10 classes (one for each digits), with 7,000 images per class. There are 60,000 training images and 10,000 test images.
false
1
false
filwsyl/video_tags
2022-10-25T10:13:17.000Z
mnist
false
a9d58d45d5363ececbe0485f26350fff6835f611
[]
[ "annotations_creators:expert-generated", "language_creators:found", "language:enx", "license:mit", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:extended|other-nist", "task_categories:image-classification", "task_ids:multi-class-image-classification" ]
https://huggingface.co/datasets/filwsyl/video_tags/resolve/main/README.md
--- annotations_creators: - expert-generated language_creators: - found language: - enx license: - mit multilinguality: - monolingual size_categories: - 10K<n<100K source_datasets: - extended|other-nist task_categories: - image-classification task_ids: - multi-class-image-classification paperswithcode_id: mnist pretty_name: MNIST --- # Dataset Card for MNIST ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** http://yann.lecun.com/exdb/mnist/ - **Repository:** - **Paper:** MNIST handwritten digit database by Yann LeCun, Corinna Cortes, and CJ Burges - **Leaderboard:** - **Point of Contact:** ### Dataset Summary The MNIST dataset consists of 70,000 28x28 black-and-white images of handwritten digits extracted from two NIST databases. There are 60,000 images in the training dataset and 10,000 images in the validation dataset, one class per digit so a total of 10 classes, with 7,000 images (6,000 train images and 1,000 test images) per class. Half of the image were drawn by Census Bureau employees and the other half by high school students (this split is evenly distributed in the training and testing sets). ### Supported Tasks and Leaderboards - `image-classification`: The goal of this task is to classify a given image of a handwritten digit into one of 10 classes representing integer values from 0 to 9, inclusively. The leaderboard is available [here](https://paperswithcode.com/sota/image-classification-on-mnist). ### Languages English ## Dataset Structure ### Data Instances A data point comprises an image and its label: ``` { 'image': <PIL.PngImagePlugin.PngImageFile image mode=L size=28x28 at 0x276021F6DD8>, 'label': 5 } ``` ### Data Fields - `image`: A `PIL.Image.Image` object containing the 28x28 image. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]` - `label`: an integer between 0 and 9 representing the digit. ### Data Splits The data is split into training and test set. All the images in the test set were drawn by different individuals than the images in the training set. The training set contains 60,000 images and the test set 10,000 images. ## Dataset Creation ### Curation Rationale The MNIST database was created to provide a testbed for people wanting to try pattern recognition methods or machine learning algorithms while spending minimal efforts on preprocessing and formatting. Images of the original dataset (NIST) were in two groups, one consisting of images drawn by Census Bureau employees and one consisting of images drawn by high school students. In NIST, the training set was built by grouping all the images of the Census Bureau employees, and the test set was built by grouping the images form the high school students. The goal in building MNIST was to have a training and test set following the same distributions, so the training set contains 30,000 images drawn by Census Bureau employees and 30,000 images drawn by high school students, and the test set contains 5,000 images of each group. The curators took care to make sure all the images in the test set were drawn by different individuals than the images in the training set. ### Source Data #### Initial Data Collection and Normalization The original images from NIST were size normalized to fit a 20x20 pixel box while preserving their aspect ratio. The resulting images contain grey levels (i.e., pixels don't simply have a value of black and white, but a level of greyness from 0 to 255) as a result of the anti-aliasing technique used by the normalization algorithm. The images were then centered in a 28x28 image by computing the center of mass of the pixels, and translating the image so as to position this point at the center of the 28x28 field. #### Who are the source language producers? Half of the source images were drawn by Census Bureau employees, half by high school students. According to the dataset curator, the images from the first group are more easily recognizable. ### Annotations #### Annotation process The images were not annotated after their creation: the image creators annotated their images with the corresponding label after drawing them. #### Who are the annotators? Same as the source data creators. ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators Chris Burges, Corinna Cortes and Yann LeCun ### Licensing Information MIT Licence ### Citation Information ``` @article{lecun2010mnist, title={MNIST handwritten digit database}, author={LeCun, Yann and Cortes, Corinna and Burges, CJ}, journal={ATT Labs [Online]. Available: http://yann.lecun.com/exdb/mnist}, volume={2}, year={2010} } ``` ### Contributions Thanks to [@sgugger](https://github.com/sgugger) for adding this dataset.
taln-ls2n
null
@inproceedings{gallina-etal-2019-kptimes, title = "{KPT}imes: A Large-Scale Dataset for Keyphrase Generation on News Documents", author = "Gallina, Ygor and Boudin, Florian and Daille, Beatrice", booktitle = "Proceedings of the 12th International Conference on Natural Language Generation", month = oct # "{--}" # nov, year = "2019", address = "Tokyo, Japan", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/W19-8617", doi = "10.18653/v1/W19-8617", pages = "130--135", abstract = "Keyphrase generation is the task of predicting a set of lexical units that conveys the main content of a source text. Existing datasets for keyphrase generation are only readily available for the scholarly domain and include non-expert annotations. In this paper we present KPTimes, a large-scale dataset of news texts paired with editor-curated keyphrases. Exploring the dataset, we show how editors tag documents, and how their annotations differ from those found in existing datasets. We also train and evaluate state-of-the-art neural keyphrase generation models on KPTimes to gain insights on how well they perform on the news domain. The dataset is available online at https:// github.com/ygorg/KPTimes.", }
KPTimes benchmark dataset for keyphrase extraction an generation.
false
14
false
taln-ls2n/kptimes
2022-09-23T07:38:28.000Z
null
false
36d51f10c05d1598552a0374b04d7b8e58efddbc
[]
[ "annotations_creators:unknown", "language_creators:unknown", "language:en", "license:cc-by-4.0", "multilinguality:monolingual", "task_categories:text-generation", "task_ids:keyphrase-generation", "task_ids:keyphrase-extraction", "size_categories:100K<n<1M" ]
https://huggingface.co/datasets/taln-ls2n/kptimes/resolve/main/README.md
--- annotations_creators: - unknown language_creators: - unknown language: - en license: - cc-by-4.0 multilinguality: - monolingual task_categories: - text-mining - text-generation task_ids: - keyphrase-generation - keyphrase-extraction size_categories: - 100K<n<1M pretty_name: KPTimes --- # KPTimes Benchmark Dataset for Keyphrase Generation ## About KPTimes is a dataset for benchmarking keyphrase extraction and generation models. The dataset is composed of 290K news articles in English collected from the [New York Times](https://www.nytimes.com/) and the [Japan Times](https://www.japantimes.co.jp/). Keyphrases were annotated by editors in a semi-automated manner (that is, editors revise a set of keyphrases proposed by an algorithm and provide additional keyphrases). Details about the dataset can be found in the original paper [(Gallina et al., 2019)][gallina-2019]. Reference (indexer-assigned) keyphrases are also categorized under the PRMU (<u>P</u>resent-<u>R</u>eordered-<u>M</u>ixed-<u>U</u>nseen) scheme as proposed in [(Boudin and Gallina, 2021)][boudin-2021]. Text pre-processing (tokenization) is carried out using `spacy` (`en_core_web_sm` model) with a special rule to avoid splitting words with hyphens (e.g. graph-based is kept as one token). Stemming (Porter's stemmer implementation provided in `nltk`) is applied before reference keyphrases are matched against the source text. Details about the process can be found in `prmu.py`. <u>P</u>resent keyphrases are ordered according to their first occurrence position in the text. ## Content and statistics The dataset contains the following test split: | Split | # documents | #words | # keyphrases | % Present | % Reordered | % Mixed | % Unseen | | :--------- | ----------: | -----: | -----------: | --------: | ----------: | ------: | -------: | | Train | 259,923 | 921 | 5.03 | 45.61 | 15.57 | 29.63 | 9.19 | | Validation | 10,000 | 921 | 5.02 | 45.22 | 15.78 | 29.60 | 9.41 | | Test | 20,000 | 648 | 5.03 | 60.64 | 8.90 | 18.95 | 11.51 | The following data fields are available : - **id**: unique identifier of the document. - **title**: title of the document. - **abstract**: abstract of the document. - **keyphrases**: list of reference keyphrases. - **prmu**: list of <u>P</u>resent-<u>R</u>eordered-<u>M</u>ixed-<u>U</u>nseen categories for reference keyphrases. - **date**: publishing date (YYYY/MM/DD) - **categories**: categories of the article (1 or 2 categories) ## References - (Gallina et al., 2019) Ygor Gallina, Florian Boudin, and Beatrice Daille. 2019. [KPTimes: A Large-Scale Dataset for Keyphrase Generation on News Documents][gallina-2019]. In Proceedings of the 12th International Conference on Natural Language Generation, pages 130–135, Tokyo, Japan. Association for Computational Linguistics. - (Boudin and Gallina, 2021) Florian Boudin and Ygor Gallina. 2021. [Redefining Absent Keyphrases and their Effect on Retrieval Effectiveness][boudin-2021]. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 4185–4193, Online. Association for Computational Linguistics. [gallina-2019]: https://aclanthology.org/W19-8617/ [boudin-2021]: https://aclanthology.org/2021.naacl-main.330/
swcrazyfan
null
null
null
false
1
false
swcrazyfan/net-kjv
2022-05-06T10:05:48.000Z
null
false
da97033c65ab45c0f6735cfa5b9c18ff8e9f1bde
[]
[]
https://huggingface.co/datasets/swcrazyfan/net-kjv/resolve/main/README.md
languages: - en task_categories: - translation licenses: - unknown # Dataset Card for [Needs More Information] ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** [Needs More Information] - **Repository:** [Needs More Information] - **Paper:** [Needs More Information] - **Leaderboard:** [Needs More Information] - **Point of Contact:** [Needs More Information] ### Dataset Summary This is a dataset made up of two Bible translations-- NET and KJV. ### Supported Tasks and Leaderboards [Needs More Information] ### Languages English ## Dataset Structure ### Data Instances [Needs More Information] ### Data Fields [Needs More Information] ### Data Splits [Needs More Information] ## Dataset Creation ### Curation Rationale The original intention is to use the dataset to "translate" between modern and 17th-century English. By doing so, we can potentially read understand things from that period more clearly. ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations Before the 18th and 19th centuries, English spelling was inconsistent. Because of this, the model often does not recognize spellings different from those in the KJV. The model was trained on a relatively small amount of data, so it will not be as accurate as a model trained on a larger data set. ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information [Needs More Information] ### Citation Information [Needs More Information]
vehnem
null
null
null
false
1
false
vehnem/DBpedia
2022-05-06T11:01:55.000Z
null
false
b603503ce82cf503a03ffa407feb2c17d4f4c27a
[]
[ "license:cc-by-sa-3.0" ]
https://huggingface.co/datasets/vehnem/DBpedia/resolve/main/README.md
--- license: cc-by-sa-3.0 ---