id
stringlengths
2
115
lastModified
stringlengths
24
24
tags
list
author
stringlengths
2
42
description
stringlengths
0
68.7k
citation
stringlengths
0
10.7k
cardData
null
likes
int64
0
3.55k
downloads
int64
0
10.1M
card
stringlengths
0
1.01M
lcw99/wikipedia-korean-20221001
2022-10-10T03:55:17.000Z
[ "language:ko", "region:us" ]
lcw99
null
null
null
3
1,417
--- language: - ko ---
osunlp/ConflictQA
2023-06-15T18:45:52.000Z
[ "task_categories:question-answering", "size_categories:10K<n<100K", "language:en", "license:apache-2.0", "arxiv:2305.13300", "region:us" ]
osunlp
data for ConflictQA.
@article{xie2023adaptive, title={Adaptive Chameleon or Stubborn Sloth: Unraveling the Behavior of Large Language Models in Knowledge Conflicts}, author={Xie, Jian and Zhang, Kai and Chen, Jiangjie and Lou, Renze and Su, Yu}, journal={arXiv preprint arXiv:2305.13300}, year={2023} }
null
4
1,417
--- license: apache-2.0 task_categories: - question-answering language: - en pretty_name: conflictQA size_categories: - 10K<n<100K --- # Dataset Card for ConflcitQA ## Dataset Description - **Repository:** https://github.com/OSU-NLP-Group/LLM-Knowledge-Conflict - **Paper:** https://arxiv.org/abs/2305.13300 - **Point of Contact:** Point of Contact: [Jian Xie](mailto:jianx0321@gmail.com) ## Citation If our paper or related resources prove valuable to your research, we kindly ask for citation. Please feel free to contact us with any inquiries. ```bib @article{xie2023adaptive, title={Adaptive Chameleon or Stubborn Sloth: Unraveling the Behavior of Large Language Models in Knowledge Conflicts}, author={Xie, Jian and Zhang, Kai and Chen, Jiangjie and Lou, Renze and Su, Yu}, journal={arXiv preprint arXiv:2305.13300}, url={arxiv.org/abs/2305.13300}, year={2023} } ``` # ConflcitQA We provide the conflictQA GPT-4 (ChatGPT) version, which utilizes GPT-4 (ChatGPT) guided parametric memory. ```json {"question": "What is George Rankin's occupation?", "popularity": 142, "ground_truth": ["politician", "political leader", "political figure", "polit.", "pol"], "memory_answer": "George Rankin's occupation is a professional photographer.", "parametric_memory": "As a professional photographer, George Rankin...", "counter_answer": "George Rankin's occupation is political figure.", "counter_memory": "George Rankin has been actively involved in politics for over a decade...", "parametric_memory_aligned_evidence": "George Rankin has a website showcasing his photography portfolio...", "counter_memory_aligned_evidence": "George Rankin Major General George James Rankin..."} ``` # Data Fields - "question": The question in natural language - "popularity": The monthly page views on Wikipedia for the given question - "ground_truth": The factual answer to the question, which may include multiple possible answers - "memory_answer": The answer provided by the LLM to the question - "parametric_memory": The supportive evidence from LLM's parametric memory for the answer - "counter_answer": The answer contradicting the "memory_answer" - "counter_memory": The generation-based evidence supporting the counter_answer - "parametric_memory_aligned_evidence": Additional evidence supporting the "memory_answer", which could be generated or derived from Wikipedia/human annotation - "counter_memory_aligned_evidence": Additional evidence supporting the "counter_answer", either generated or sourced from Wikipedia/human annotation
lucadiliello/naturalquestionsshortqa
2023-06-06T08:35:50.000Z
[ "region:us" ]
lucadiliello
null
null
null
1
1,413
--- dataset_info: features: - name: context dtype: string - name: question dtype: string - name: answers sequence: string - name: key dtype: string - name: labels list: - name: end sequence: int64 - name: start sequence: int64 splits: - name: train num_bytes: 100706304 num_examples: 104071 - name: validation num_bytes: 12941478 num_examples: 12836 download_size: 61870589 dataset_size: 113647782 --- # Dataset Card for "naturalquestionsshortqa" Split taken from the MRQA 2019 Shared Task, formatted and filtered for Question Answering. For the original dataset, have a look [here](https://huggingface.co/datasets/mrqa).
keremberke/csgo-object-detection
2023-01-27T13:39:19.000Z
[ "task_categories:object-detection", "roboflow", "roboflow2huggingface", "region:us" ]
keremberke
null
@misc{ wlots_dataset, title = { wlots Dataset }, type = { Open Source Dataset }, author = { asd }, howpublished = { \\url{ https://universe.roboflow.com/asd-culfr/wlots } }, url = { https://universe.roboflow.com/asd-culfr/wlots }, journal = { Roboflow Universe }, publisher = { Roboflow }, year = { 2022 }, month = { may }, note = { visited on 2023-01-27 }, }
null
4
1,411
--- task_categories: - object-detection tags: - roboflow - roboflow2huggingface --- <div align="center"> <img width="640" alt="keremberke/csgo-object-detection" src="https://huggingface.co/datasets/keremberke/csgo-object-detection/resolve/main/thumbnail.jpg"> </div> ### Dataset Labels ``` ['ct', 'cthead', 't', 'thead'] ``` ### Number of Images ```json {'train': 3879, 'valid': 383, 'test': 192} ``` ### How to Use - Install [datasets](https://pypi.org/project/datasets/): ```bash pip install datasets ``` - Load the dataset: ```python from datasets import load_dataset ds = load_dataset("keremberke/csgo-object-detection", name="full") example = ds['train'][0] ``` ### Roboflow Dataset Page [https://universe.roboflow.com/asd-culfr/wlots/dataset/1](https://universe.roboflow.com/asd-culfr/wlots/dataset/1?ref=roboflow2huggingface) ### Citation ``` @misc{ wlots_dataset, title = { wlots Dataset }, type = { Open Source Dataset }, author = { asd }, howpublished = { \\url{ https://universe.roboflow.com/asd-culfr/wlots } }, url = { https://universe.roboflow.com/asd-culfr/wlots }, journal = { Roboflow Universe }, publisher = { Roboflow }, year = { 2022 }, month = { may }, note = { visited on 2023-01-27 }, } ``` ### License CC BY 4.0 ### Dataset Summary This dataset was exported via roboflow.com on December 28, 2022 at 8:08 PM GMT Roboflow is an end-to-end computer vision platform that helps you * collaborate with your team on computer vision projects * collect & organize images * understand unstructured image data * annotate, and create datasets * export, train, and deploy computer vision models * use active learning to improve your dataset over time It includes 4454 images. Ct-cthead-t-thead are annotated in COCO format. The following pre-processing was applied to each image: * Auto-orientation of pixel data (with EXIF-orientation stripping) * Resize to 416x416 (Fill (with center crop)) The following augmentation was applied to create 3 versions of each source image: * Random brigthness adjustment of between -15 and +15 percent
sahil2801/CodeAlpaca-20k
2023-10-03T11:46:04.000Z
[ "task_categories:text-generation", "size_categories:10K<n<100K", "language:en", "license:cc-by-4.0", "code", "region:us" ]
sahil2801
null
null
null
108
1,393
--- license: cc-by-4.0 task_categories: - text-generation tags: - code pretty_name: CodeAlpaca 20K size_categories: - 10K<n<100K language: - en ---
banghua/hh_reward_model_labeled
2023-08-06T02:03:27.000Z
[ "region:us" ]
banghua
null
null
null
0
1,392
--- dataset_info: features: - name: prompt dtype: string - name: response dtype: string - name: chosen dtype: string - name: rejected dtype: string splits: - name: train num_bytes: 225756769 num_examples: 124503 download_size: 136142109 dataset_size: 225756769 --- # Dataset Card for "hh_reward_model_labeled" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
ola13/small-the_pile
2022-11-24T11:40:52.000Z
[ "region:us" ]
ola13
null
null
null
3
1,387
--- dataset_info: features: - name: text dtype: string - name: meta struct: - name: perplexity_score dtype: float64 - name: pile_set_name dtype: string splits: - name: train num_bytes: 606056668 num_examples: 100000 download_size: 328667964 dataset_size: 606056668 --- # Dataset Card for "small-the_pile" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
iamtarun/python_code_instructions_18k_alpaca
2023-07-27T15:51:36.000Z
[ "task_categories:question-answering", "task_categories:text2text-generation", "task_categories:text-generation", "size_categories:10K<n<100K", "code", "region:us" ]
iamtarun
null
null
null
12
1,377
--- dataset_info: features: - name: instruction dtype: string - name: input dtype: string - name: output dtype: string - name: prompt dtype: string splits: - name: train num_bytes: 25180782 num_examples: 18612 download_size: 11357076 dataset_size: 25180782 configs: - config_name: default data_files: - split: train path: data/train-* task_categories: - question-answering - text2text-generation - text-generation tags: - code size_categories: - 10K<n<100K --- # Dataset Card for python_code_instructions_18k_alpaca The dataset contains problem descriptions and code in python language. This dataset is taken from [sahil2801/code_instructions_120k](https://huggingface.co/datasets/sahil2801/code_instructions_120k), which adds a prompt column in alpaca style. Refer to the source [here](https://huggingface.co/datasets/sahil2801/code_instructions_120k).
guardian_authorship
2023-04-05T10:06:55.000Z
[ "task_categories:text-classification", "task_ids:multi-class-classification", "task_ids:topic-classification", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:unknown", "region:us" ]
null
A dataset cross-topic authorship attribution. The dataset is provided by Stamatatos 2013. 1- The cross-topic scenarios are based on Table-4 in Stamatatos 2017 (Ex. cross_topic_1 => row 1:P S U&W ). 2- The cross-genre scenarios are based on Table-5 in the same paper. (Ex. cross_genre_1 => row 1:B P S&U&W). 3- The same-topic/genre scenario is created by grouping all the datasts as follows. For ex., to use same_topic and split the data 60-40 use: train_ds = load_dataset('guardian_authorship', name="cross_topic_<<#>>", split='train[:60%]+validation[:60%]+test[:60%]') tests_ds = load_dataset('guardian_authorship', name="cross_topic_<<#>>", split='train[-40%:]+validation[-40%:]+test[-40%:]') IMPORTANT: train+validation+test[:60%] will generate the wrong splits because the data is imbalanced * See https://huggingface.co/docs/datasets/splits.html for detailed/more examples
@article{article, author = {Stamatatos, Efstathios}, year = {2013}, month = {01}, pages = {421-439}, title = {On the robustness of authorship attribution based on character n-gram features}, volume = {21}, journal = {Journal of Law and Policy} } @inproceedings{stamatatos2017authorship, title={Authorship attribution using text distortion}, author={Stamatatos, Efstathios}, booktitle={Proc. of the 15th Conf. of the European Chapter of the Association for Computational Linguistics}, volume={1} pages={1138--1149}, year={2017} }
null
2
1,372
--- annotations_creators: - found language_creators: - found language: - en license: - unknown multilinguality: - monolingual size_categories: - 1K<n<10K source_datasets: - original task_categories: - text-classification task_ids: - multi-class-classification - topic-classification pretty_name: GuardianAuthorship dataset_info: - config_name: cross_topic_1 features: - name: author dtype: class_label: names: '0': catherinebennett '1': georgemonbiot '2': hugoyoung '3': jonathanfreedland '4': martinkettle '5': maryriddell '6': nickcohen '7': peterpreston '8': pollytoynbee '9': royhattersley '10': simonhoggart '11': willhutton '12': zoewilliams - name: topic dtype: class_label: names: '0': Politics '1': Society '2': UK '3': World '4': Books - name: article dtype: string splits: - name: train num_bytes: 677054 num_examples: 112 - name: test num_bytes: 1283126 num_examples: 207 - name: validation num_bytes: 374390 num_examples: 62 download_size: 3100749 dataset_size: 2334570 - config_name: cross_genre_1 features: - name: author dtype: class_label: names: '0': catherinebennett '1': georgemonbiot '2': hugoyoung '3': jonathanfreedland '4': martinkettle '5': maryriddell '6': nickcohen '7': peterpreston '8': pollytoynbee '9': royhattersley '10': simonhoggart '11': willhutton '12': zoewilliams - name: topic dtype: class_label: names: '0': Politics '1': Society '2': UK '3': World '4': Books - name: article dtype: string splits: - name: train num_bytes: 406144 num_examples: 63 - name: test num_bytes: 1657512 num_examples: 269 - name: validation num_bytes: 677054 num_examples: 112 download_size: 3100749 dataset_size: 2740710 - config_name: cross_topic_2 features: - name: author dtype: class_label: names: '0': catherinebennett '1': georgemonbiot '2': hugoyoung '3': jonathanfreedland '4': martinkettle '5': maryriddell '6': nickcohen '7': peterpreston '8': pollytoynbee '9': royhattersley '10': simonhoggart '11': willhutton '12': zoewilliams - name: topic dtype: class_label: names: '0': Politics '1': Society '2': UK '3': World '4': Books - name: article dtype: string splits: - name: train num_bytes: 677054 num_examples: 112 - name: test num_bytes: 1104764 num_examples: 179 - name: validation num_bytes: 552752 num_examples: 90 download_size: 3100749 dataset_size: 2334570 - config_name: cross_topic_3 features: - name: author dtype: class_label: names: '0': catherinebennett '1': georgemonbiot '2': hugoyoung '3': jonathanfreedland '4': martinkettle '5': maryriddell '6': nickcohen '7': peterpreston '8': pollytoynbee '9': royhattersley '10': simonhoggart '11': willhutton '12': zoewilliams - name: topic dtype: class_label: names: '0': Politics '1': Society '2': UK '3': World '4': Books - name: article dtype: string splits: - name: train num_bytes: 677054 num_examples: 112 - name: test num_bytes: 927138 num_examples: 152 - name: validation num_bytes: 730378 num_examples: 117 download_size: 3100749 dataset_size: 2334570 - config_name: cross_topic_4 features: - name: author dtype: class_label: names: '0': catherinebennett '1': georgemonbiot '2': hugoyoung '3': jonathanfreedland '4': martinkettle '5': maryriddell '6': nickcohen '7': peterpreston '8': pollytoynbee '9': royhattersley '10': simonhoggart '11': willhutton '12': zoewilliams - name: topic dtype: class_label: names: '0': Politics '1': Society '2': UK '3': World '4': Books - name: article dtype: string splits: - name: train num_bytes: 374390 num_examples: 62 - name: test num_bytes: 1283126 num_examples: 207 - name: validation num_bytes: 677054 num_examples: 112 download_size: 3100749 dataset_size: 2334570 - config_name: cross_topic_5 features: - name: author dtype: class_label: names: '0': catherinebennett '1': georgemonbiot '2': hugoyoung '3': jonathanfreedland '4': martinkettle '5': maryriddell '6': nickcohen '7': peterpreston '8': pollytoynbee '9': royhattersley '10': simonhoggart '11': willhutton '12': zoewilliams - name: topic dtype: class_label: names: '0': Politics '1': Society '2': UK '3': World '4': Books - name: article dtype: string splits: - name: train num_bytes: 374390 num_examples: 62 - name: test num_bytes: 1407428 num_examples: 229 - name: validation num_bytes: 552752 num_examples: 90 download_size: 3100749 dataset_size: 2334570 - config_name: cross_topic_6 features: - name: author dtype: class_label: names: '0': catherinebennett '1': georgemonbiot '2': hugoyoung '3': jonathanfreedland '4': martinkettle '5': maryriddell '6': nickcohen '7': peterpreston '8': pollytoynbee '9': royhattersley '10': simonhoggart '11': willhutton '12': zoewilliams - name: topic dtype: class_label: names: '0': Politics '1': Society '2': UK '3': World '4': Books - name: article dtype: string splits: - name: train num_bytes: 374390 num_examples: 62 - name: test num_bytes: 1229802 num_examples: 202 - name: validation num_bytes: 730378 num_examples: 117 download_size: 3100749 dataset_size: 2334570 - config_name: cross_topic_7 features: - name: author dtype: class_label: names: '0': catherinebennett '1': georgemonbiot '2': hugoyoung '3': jonathanfreedland '4': martinkettle '5': maryriddell '6': nickcohen '7': peterpreston '8': pollytoynbee '9': royhattersley '10': simonhoggart '11': willhutton '12': zoewilliams - name: topic dtype: class_label: names: '0': Politics '1': Society '2': UK '3': World '4': Books - name: article dtype: string splits: - name: train num_bytes: 552752 num_examples: 90 - name: test num_bytes: 1104764 num_examples: 179 - name: validation num_bytes: 677054 num_examples: 112 download_size: 3100749 dataset_size: 2334570 - config_name: cross_topic_8 features: - name: author dtype: class_label: names: '0': catherinebennett '1': georgemonbiot '2': hugoyoung '3': jonathanfreedland '4': martinkettle '5': maryriddell '6': nickcohen '7': peterpreston '8': pollytoynbee '9': royhattersley '10': simonhoggart '11': willhutton '12': zoewilliams - name: topic dtype: class_label: names: '0': Politics '1': Society '2': UK '3': World '4': Books - name: article dtype: string splits: - name: train num_bytes: 552752 num_examples: 90 - name: test num_bytes: 1407428 num_examples: 229 - name: validation num_bytes: 374390 num_examples: 62 download_size: 3100749 dataset_size: 2334570 - config_name: cross_topic_9 features: - name: author dtype: class_label: names: '0': catherinebennett '1': georgemonbiot '2': hugoyoung '3': jonathanfreedland '4': martinkettle '5': maryriddell '6': nickcohen '7': peterpreston '8': pollytoynbee '9': royhattersley '10': simonhoggart '11': willhutton '12': zoewilliams - name: topic dtype: class_label: names: '0': Politics '1': Society '2': UK '3': World '4': Books - name: article dtype: string splits: - name: train num_bytes: 552752 num_examples: 90 - name: test num_bytes: 1051440 num_examples: 174 - name: validation num_bytes: 730378 num_examples: 117 download_size: 3100749 dataset_size: 2334570 - config_name: cross_topic_10 features: - name: author dtype: class_label: names: '0': catherinebennett '1': georgemonbiot '2': hugoyoung '3': jonathanfreedland '4': martinkettle '5': maryriddell '6': nickcohen '7': peterpreston '8': pollytoynbee '9': royhattersley '10': simonhoggart '11': willhutton '12': zoewilliams - name: topic dtype: class_label: names: '0': Politics '1': Society '2': UK '3': World '4': Books - name: article dtype: string splits: - name: train num_bytes: 730378 num_examples: 117 - name: test num_bytes: 927138 num_examples: 152 - name: validation num_bytes: 677054 num_examples: 112 download_size: 3100749 dataset_size: 2334570 - config_name: cross_topic_11 features: - name: author dtype: class_label: names: '0': catherinebennett '1': georgemonbiot '2': hugoyoung '3': jonathanfreedland '4': martinkettle '5': maryriddell '6': nickcohen '7': peterpreston '8': pollytoynbee '9': royhattersley '10': simonhoggart '11': willhutton '12': zoewilliams - name: topic dtype: class_label: names: '0': Politics '1': Society '2': UK '3': World '4': Books - name: article dtype: string splits: - name: train num_bytes: 730378 num_examples: 117 - name: test num_bytes: 1229802 num_examples: 202 - name: validation num_bytes: 374390 num_examples: 62 download_size: 3100749 dataset_size: 2334570 - config_name: cross_topic_12 features: - name: author dtype: class_label: names: '0': catherinebennett '1': georgemonbiot '2': hugoyoung '3': jonathanfreedland '4': martinkettle '5': maryriddell '6': nickcohen '7': peterpreston '8': pollytoynbee '9': royhattersley '10': simonhoggart '11': willhutton '12': zoewilliams - name: topic dtype: class_label: names: '0': Politics '1': Society '2': UK '3': World '4': Books - name: article dtype: string splits: - name: train num_bytes: 730378 num_examples: 117 - name: test num_bytes: 1051440 num_examples: 174 - name: validation num_bytes: 552752 num_examples: 90 download_size: 3100749 dataset_size: 2334570 - config_name: cross_genre_2 features: - name: author dtype: class_label: names: '0': catherinebennett '1': georgemonbiot '2': hugoyoung '3': jonathanfreedland '4': martinkettle '5': maryriddell '6': nickcohen '7': peterpreston '8': pollytoynbee '9': royhattersley '10': simonhoggart '11': willhutton '12': zoewilliams - name: topic dtype: class_label: names: '0': Politics '1': Society '2': UK '3': World '4': Books - name: article dtype: string splits: - name: train num_bytes: 406144 num_examples: 63 - name: test num_bytes: 1960176 num_examples: 319 - name: validation num_bytes: 374390 num_examples: 62 download_size: 3100749 dataset_size: 2740710 - config_name: cross_genre_3 features: - name: author dtype: class_label: names: '0': catherinebennett '1': georgemonbiot '2': hugoyoung '3': jonathanfreedland '4': martinkettle '5': maryriddell '6': nickcohen '7': peterpreston '8': pollytoynbee '9': royhattersley '10': simonhoggart '11': willhutton '12': zoewilliams - name: topic dtype: class_label: names: '0': Politics '1': Society '2': UK '3': World '4': Books - name: article dtype: string splits: - name: train num_bytes: 406144 num_examples: 63 - name: test num_bytes: 1781814 num_examples: 291 - name: validation num_bytes: 552752 num_examples: 90 download_size: 3100749 dataset_size: 2740710 - config_name: cross_genre_4 features: - name: author dtype: class_label: names: '0': catherinebennett '1': georgemonbiot '2': hugoyoung '3': jonathanfreedland '4': martinkettle '5': maryriddell '6': nickcohen '7': peterpreston '8': pollytoynbee '9': royhattersley '10': simonhoggart '11': willhutton '12': zoewilliams - name: topic dtype: class_label: names: '0': Politics '1': Society '2': UK '3': World '4': Books - name: article dtype: string splits: - name: train num_bytes: 406144 num_examples: 63 - name: test num_bytes: 1604188 num_examples: 264 - name: validation num_bytes: 730378 num_examples: 117 download_size: 3100749 dataset_size: 2740710 --- # Dataset Card for "guardian_authorship" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [http://www.icsd.aegean.gr/lecturers/stamatatos/papers/JLP2013.pdf](http://www.icsd.aegean.gr/lecturers/stamatatos/papers/JLP2013.pdf) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 49.61 MB - **Size of the generated dataset:** 38.98 MB - **Total amount of disk used:** 88.59 MB ### Dataset Summary A dataset cross-topic authorship attribution. The dataset is provided by Stamatatos 2013. 1- The cross-topic scenarios are based on Table-4 in Stamatatos 2017 (Ex. cross_topic_1 => row 1:P S U&W ). 2- The cross-genre scenarios are based on Table-5 in the same paper. (Ex. cross_genre_1 => row 1:B P S&U&W). 3- The same-topic/genre scenario is created by grouping all the datasts as follows. For ex., to use same_topic and split the data 60-40 use: train_ds = load_dataset('guardian_authorship', name="cross_topic_<<#>>", split='train[:60%]+validation[:60%]+test[:60%]') tests_ds = load_dataset('guardian_authorship', name="cross_topic_<<#>>", split='train[-40%:]+validation[-40%:]+test[-40%:]') IMPORTANT: train+validation+test[:60%] will generate the wrong splits because the data is imbalanced * See https://huggingface.co/docs/datasets/splits.html for detailed/more examples ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### cross_genre_1 - **Size of downloaded dataset files:** 3.10 MB - **Size of the generated dataset:** 2.74 MB - **Total amount of disk used:** 5.84 MB An example of 'train' looks as follows. ``` { "article": "File 1a\n", "author": 0, "topic": 4 } ``` #### cross_genre_2 - **Size of downloaded dataset files:** 3.10 MB - **Size of the generated dataset:** 2.74 MB - **Total amount of disk used:** 5.84 MB An example of 'validation' looks as follows. ``` { "article": "File 1a\n", "author": 0, "topic": 1 } ``` #### cross_genre_3 - **Size of downloaded dataset files:** 3.10 MB - **Size of the generated dataset:** 2.74 MB - **Total amount of disk used:** 5.84 MB An example of 'validation' looks as follows. ``` { "article": "File 1a\n", "author": 0, "topic": 2 } ``` #### cross_genre_4 - **Size of downloaded dataset files:** 3.10 MB - **Size of the generated dataset:** 2.74 MB - **Total amount of disk used:** 5.84 MB An example of 'validation' looks as follows. ``` { "article": "File 1a\n", "author": 0, "topic": 3 } ``` #### cross_topic_1 - **Size of downloaded dataset files:** 3.10 MB - **Size of the generated dataset:** 2.34 MB - **Total amount of disk used:** 5.43 MB An example of 'validation' looks as follows. ``` { "article": "File 1a\n", "author": 0, "topic": 1 } ``` ### Data Fields The data fields are the same among all splits. #### cross_genre_1 - `author`: a classification label, with possible values including `catherinebennett` (0), `georgemonbiot` (1), `hugoyoung` (2), `jonathanfreedland` (3), `martinkettle` (4). - `topic`: a classification label, with possible values including `Politics` (0), `Society` (1), `UK` (2), `World` (3), `Books` (4). - `article`: a `string` feature. #### cross_genre_2 - `author`: a classification label, with possible values including `catherinebennett` (0), `georgemonbiot` (1), `hugoyoung` (2), `jonathanfreedland` (3), `martinkettle` (4). - `topic`: a classification label, with possible values including `Politics` (0), `Society` (1), `UK` (2), `World` (3), `Books` (4). - `article`: a `string` feature. #### cross_genre_3 - `author`: a classification label, with possible values including `catherinebennett` (0), `georgemonbiot` (1), `hugoyoung` (2), `jonathanfreedland` (3), `martinkettle` (4). - `topic`: a classification label, with possible values including `Politics` (0), `Society` (1), `UK` (2), `World` (3), `Books` (4). - `article`: a `string` feature. #### cross_genre_4 - `author`: a classification label, with possible values including `catherinebennett` (0), `georgemonbiot` (1), `hugoyoung` (2), `jonathanfreedland` (3), `martinkettle` (4). - `topic`: a classification label, with possible values including `Politics` (0), `Society` (1), `UK` (2), `World` (3), `Books` (4). - `article`: a `string` feature. #### cross_topic_1 - `author`: a classification label, with possible values including `catherinebennett` (0), `georgemonbiot` (1), `hugoyoung` (2), `jonathanfreedland` (3), `martinkettle` (4). - `topic`: a classification label, with possible values including `Politics` (0), `Society` (1), `UK` (2), `World` (3), `Books` (4). - `article`: a `string` feature. ### Data Splits | name |train|validation|test| |-------------|----:|---------:|---:| |cross_genre_1| 63| 112| 269| |cross_genre_2| 63| 62| 319| |cross_genre_3| 63| 90| 291| |cross_genre_4| 63| 117| 264| |cross_topic_1| 112| 62| 207| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @article{article, author = {Stamatatos, Efstathios}, year = {2013}, month = {01}, pages = {421-439}, title = {On the robustness of authorship attribution based on character n-gram features}, volume = {21}, journal = {Journal of Law and Policy} } @inproceedings{stamatatos2017authorship, title={Authorship attribution using text distortion}, author={Stamatatos, Efstathios}, booktitle={Proc. of the 15th Conf. of the European Chapter of the Association for Computational Linguistics}, volume={1} pages={1138--1149}, year={2017} } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@eltoto1219](https://github.com/eltoto1219), [@malikaltakrori](https://github.com/malikaltakrori) for adding this dataset.
cs_restaurants
2022-11-18T19:49:56.000Z
[ "task_categories:text2text-generation", "task_categories:text-generation", "task_categories:fill-mask", "task_ids:dialogue-modeling", "task_ids:language-modeling", "task_ids:masked-language-modeling", "annotations_creators:found", "language_creators:expert-generated", "language_creators:machine-generated", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:extended|other-san-francisco-restaurants", "language:cs", "license:cc-by-4.0", "intent-to-text", "arxiv:1910.05298", "region:us" ]
null
This is a dataset for NLG in task-oriented spoken dialogue systems with Czech as the target language. It originated as a translation of the English San Francisco Restaurants dataset by Wen et al. (2015).
@article{DBLP:journals/corr/abs-1910-05298, author = {Ondrej Dusek and Filip Jurcicek}, title = {Neural Generation for Czech: Data and Baselines}, journal = {CoRR}, volume = {abs/1910.05298}, year = {2019}, url = {http://arxiv.org/abs/1910.05298}, archivePrefix = {arXiv}, eprint = {1910.05298}, timestamp = {Wed, 16 Oct 2019 16:25:53 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-1910-05298.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} }
null
1
1,369
--- annotations_creators: - found language_creators: - expert-generated - machine-generated language: - cs license: - cc-by-4.0 multilinguality: - monolingual size_categories: - 1K<n<10K source_datasets: - extended|other-san-francisco-restaurants task_categories: - text2text-generation - text-generation - fill-mask task_ids: - dialogue-modeling - language-modeling - masked-language-modeling paperswithcode_id: czech-restaurant-information pretty_name: Czech Restaurant tags: - intent-to-text dataset_info: features: - name: dialogue_act dtype: string - name: delexicalized_dialogue_act dtype: string - name: text dtype: string - name: delexicalized_text dtype: string config_name: CSRestaurants splits: - name: train num_bytes: 654071 num_examples: 3569 - name: validation num_bytes: 181528 num_examples: 781 - name: test num_bytes: 191334 num_examples: 842 download_size: 1463019 dataset_size: 1026933 --- # Dataset Card for Czech Restaurant ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Repository:** [Czech restaurants homepage](https://github.com/UFAL-DSG/cs_restaurant_dataset) - **Paper:** [Czech restaurants on Arxiv](https://arxiv.org/abs/1910.05298) ### Dataset Summary This is a dataset for NLG in task-oriented spoken dialogue systems with Czech as the target language. It originated as a translation of the [English San Francisco Restaurants dataset](https://www.repository.cam.ac.uk/handle/1810/251304) by Wen et al. (2015). The domain is restaurant information in Prague, with random/fictional values. It includes input dialogue acts and the corresponding outputs in Czech. ### Supported Tasks and Leaderboards - `other-intent-to-text`: The dataset can be used to train a model for data-to-text generation: from a desired dialogue act, the model must produce textual output that conveys this intention. ### Languages The entire dataset is in Czech, translated from the English San Francisco dataset by professional translators. ## Dataset Structure ### Data Instances Example of a data instance: ``` { "da": "?request(area)", "delex_da": "?request(area)", "text": "Jakou lokalitu hledáte ?", "delex_text": "Jakou lokalitu hledáte ?" } ``` ### Data Fields - `da`: input dialogue act - `delex_da`: input dialogue act, delexicalized - `text`: output text - `delex_text`: output text, delexicalized ### Data Splits The order of the instances is random; the split is roughly 3:1:1 between train, development, and test, ensuring that the different sections don't share the same DAs (so the generators need to generalize to unseen DAs), but they share as many generic different DA types as possible (e.g., confirm, inform_only_match etc.). DA types that only have a single corresponding DA (e.g., bye()) are included in the training set. The training, development, and test set contain 3569, 781, and 842 instances, respectively. ## Dataset Creation ### Curation Rationale While most current neural NLG systems do not explicitly contain language-specific components and are thus capable of multilingual generation in principle, there has been little work to test these capabilities experimentally. This goes hand in hand with the scarcity of non-English training datasets for NLG – the only data-to-text NLG set known to us is a small sportscasting Korean dataset (Chenet al., 2010), which only contains a limited number of named entities, reducing the need for their inflection. Since most generators are only tested on English, they do not need to handle grammar complexities not present in English. A prime example is the delexicalization technique used by most current generators. We create a novel dataset for Czech delexicalized generation; this extends the typical task of data-to-text NLG by requiring attribute value inflection. We choose Czech as an example of a morphologically complex language with a large set of NLP tools readily available. ### Source Data #### Initial Data Collection and Normalization The original data was collected from the [English San Francisco Restaurants dataset](https://www.repository.cam.ac.uk/handle/1810/251304) by Wen et al. (2015). #### Who are the source language producers? The original data was produced in interactions between Amazon Mechanical Turk workers and themed around San Francisco restaurants. This data was then translated into Czech and localized to Prague restaurants by professional translators. ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information This data does not contain personal information. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators Ondřej Dušek, Filip Jurčíček, Josef Dvořák, Petra Grycová, Matěj Hejda, Jana Olivová, Michal Starý, Eva Štichová, Charles University. This work was funded by the Ministry of Education, Youth and Sports of the Czech Republic under the grant agreement LK11221 and core research funding, SVV project 260 333, and GAUK grant 2058214 of Charles University in Prague. It used language resources stored and distributed by the LINDAT/CLARIN project of the Ministry of Education, Youth and Sports of the Czech Republic (project LM2015071). ### Licensing Information [Creative Commons 4.0 BY-SA](https://creativecommons.org/licenses/by-sa/4.0/) ### Citation Information ``` @article{DBLP:journals/corr/abs-1910-05298, author = {Ondrej Dusek and Filip Jurcicek}, title = {Neural Generation for Czech: Data and Baselines}, journal = {CoRR}, volume = {abs/1910.05298}, year = {2019}, url = {http://arxiv.org/abs/1910.05298}, archivePrefix = {arXiv}, eprint = {1910.05298}, timestamp = {Wed, 16 Oct 2019 16:25:53 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-1910-05298.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ``` ### Contributions Thanks to [@TevenLeScao](https://github.com/TevenLeScao) for adding this dataset.
ccdv/arxiv-summarization
2022-12-08T06:58:05.000Z
[ "task_categories:summarization", "task_categories:text-generation", "multilinguality:monolingual", "size_categories:100K<n<1M", "language:en", "conditional-text-generation", "region:us" ]
ccdv
Arxiv dataset for summarization. From paper: A Discourse-Aware Attention Model for Abstractive Summarization of Long Documents" by A. Cohan et al. See: https://aclanthology.org/N18-2097.pdf See: https://github.com/armancohan/long-summarization
@inproceedings{cohan-etal-2018-discourse, title = "A Discourse-Aware Attention Model for Abstractive Summarization of Long Documents", author = "Cohan, Arman and Dernoncourt, Franck and Kim, Doo Soon and Bui, Trung and Kim, Seokhwan and Chang, Walter and Goharian, Nazli", booktitle = "Proceedings of the 2018 Conference of the North {A}merican Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers)", month = jun, year = "2018", address = "New Orleans, Louisiana", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/N18-2097", doi = "10.18653/v1/N18-2097", pages = "615--621", abstract = "Neural abstractive summarization models have led to promising results in summarizing relatively short documents. We propose the first model for abstractive summarization of single, longer-form documents (e.g., research papers). Our approach consists of a new hierarchical encoder that models the discourse structure of a document, and an attentive discourse-aware decoder to generate the summary. Empirical results on two large-scale datasets of scientific papers show that our model significantly outperforms state-of-the-art models.", }
null
31
1,359
--- language: - en multilinguality: - monolingual size_categories: - 100K<n<1M task_categories: - summarization - text-generation task_ids: [] tags: - conditional-text-generation train-eval-index: - config: document task: summarization task_id: summarization splits: eval_split: test col_mapping: article: text abstract: target --- # Arxiv dataset for summarization Dataset for summarization of long documents.\ Adapted from this [repo](https://github.com/armancohan/long-summarization).\ Note that original data are pre-tokenized so this dataset returns " ".join(text) and add "\n" for paragraphs. \ This dataset is compatible with the [`run_summarization.py`](https://github.com/huggingface/transformers/tree/master/examples/pytorch/summarization) script from Transformers if you add this line to the `summarization_name_mapping` variable: ```python "ccdv/arxiv-summarization": ("article", "abstract") ``` ### Data Fields - `id`: paper id - `article`: a string containing the body of the paper - `abstract`: a string containing the abstract of the paper ### Data Splits This dataset has 3 splits: _train_, _validation_, and _test_. \ Token counts are white space based. | Dataset Split | Number of Instances | Avg. tokens | | ------------- | --------------------|:----------------------| | Train | 203,037 | 6038 / 299 | | Validation | 6,436 | 5894 / 172 | | Test | 6,440 | 5905 / 174 | # Cite original article ``` @inproceedings{cohan-etal-2018-discourse, title = "A Discourse-Aware Attention Model for Abstractive Summarization of Long Documents", author = "Cohan, Arman and Dernoncourt, Franck and Kim, Doo Soon and Bui, Trung and Kim, Seokhwan and Chang, Walter and Goharian, Nazli", booktitle = "Proceedings of the 2018 Conference of the North {A}merican Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers)", month = jun, year = "2018", address = "New Orleans, Louisiana", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/N18-2097", doi = "10.18653/v1/N18-2097", pages = "615--621", abstract = "Neural abstractive summarization models have led to promising results in summarizing relatively short documents. We propose the first model for abstractive summarization of single, longer-form documents (e.g., research papers). Our approach consists of a new hierarchical encoder that models the discourse structure of a document, and an attentive discourse-aware decoder to generate the summary. Empirical results on two large-scale datasets of scientific papers show that our model significantly outperforms state-of-the-art models.", } ```
masakhane/masakhaner2
2023-09-11T18:00:07.000Z
[ "task_categories:token-classification", "task_ids:named-entity-recognition", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:multilingual", "size_categories:1K<n<10K", "source_datasets:original", "language:bm", "language:bbj", "language:ee", "language:fon", "language:ha", "language:ig", "language:rw", "language:lg", "language:luo", "language:mos", "language:ny", "language:pcm", "language:sn", "language:sw", "language:tn", "language:tw", "language:wo", "language:xh", "language:yo", "language:zu", "license:afl-3.0", "ner", "masakhaner", "masakhane", "arxiv:2103.11811", "arxiv:2210.12391", "region:us" ]
masakhane
MasakhaNER 2.0 is the largest publicly available high-quality dataset for named entity recognition (NER) in 20 African languages. Named entities are phrases that contain the names of persons, organizations, locations, times and quantities. Example: [PER Wolff] , currently a journalist in [LOC Argentina] , played with [PER Del Bosque] in the final years of the seventies in [ORG Real Madrid] . MasakhaNER is a named entity dataset consisting of PER, ORG, LOC, and DATE entities annotated by Masakhane for 20 African languages: - Bambara (bam) - Ghomala (bbj) - Ewe (ewe) - Fon (fon) - Hausa (hau) - Igbo (ibo) - Kinyarwanda (kin) - Luganda (lug) - Dholuo (luo) - Mossi (mos) - Chichewa (nya) - Nigerian Pidgin - chShona (sna) - Kiswahili (swą) - Setswana (tsn) - Twi (twi) - Wolof (wol) - isiXhosa (xho) - Yorùbá (yor) - isiZulu (zul) The train/validation/test sets are available for all the ten languages. For more details see https://arxiv.org/abs/2103.11811
@article{Adelani2022MasakhaNER2A, title={MasakhaNER 2.0: Africa-centric Transfer Learning for Named Entity Recognition}, author={David Ifeoluwa Adelani and Graham Neubig and Sebastian Ruder and Shruti Rijhwani and Michael Beukman and Chester Palen-Michel and Constantine Lignos and Jesujoba Oluwadara Alabi and Shamsuddeen Hassan Muhammad and Peter Nabende and Cheikh M. Bamba Dione and Andiswa Bukula and Rooweither Mabuya and Bonaventure F. P. Dossou and Blessing K. Sibanda and Happy Buzaaba and Jonathan Mukiibi and Godson Kalipe and Derguene Mbaye and Amelia Taylor and Fatoumata Kabore and Chris C. Emezue and Anuoluwapo Aremu and Perez Ogayo and Catherine W. Gitau and Edwin Munkoh-Buabeng and Victoire Memdjokam Koagne and Allahsera Auguste Tapo and Tebogo Macucwa and Vukosi Marivate and Elvis Mboning and Tajuddeen R. Gwadabe and Tosin P. Adewumi and Orevaoghene Ahia and Joyce Nakatumba-Nabende and Neo L. Mokono and Ignatius M Ezeani and Chiamaka Ijeoma Chukwuneke and Mofetoluwa Adeyemi and Gilles Hacheme and Idris Abdulmumin and Odunayo Ogundepo and Oreen Yousuf and Tatiana Moteu Ngoli and Dietrich Klakow}, journal={ArXiv}, year={2022}, volume={abs/2210.12391} }
null
8
1,351
--- annotations_creators: - expert-generated language: - bm - bbj - ee - fon - ha - ig - rw - lg - luo - mos - ny - pcm - sn - sw - tn - tw - wo - xh - yo - zu language_creators: - expert-generated license: - afl-3.0 multilinguality: - multilingual pretty_name: masakhaner2.0 size_categories: - 1K<n<10K source_datasets: - original tags: - ner - masakhaner - masakhane task_categories: - token-classification task_ids: - named-entity-recognition --- # Dataset Card for [Dataset Name] ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [homepage](https://github.com/masakhane-io/masakhane-ner) - **Repository:** [github](https://github.com/masakhane-io/masakhane-ner) - **Paper:** [paper](https://arxiv.org/abs/2103.11811) - **Point of Contact:** [Masakhane](https://www.masakhane.io/) or didelani@lsv.uni-saarland.de ### Dataset Summary MasakhaNER 2.0 is the largest publicly available high-quality dataset for named entity recognition (NER) in 20 African languages created by the Masakhane community. Named entities are phrases that contain the names of persons, organizations, locations, times and quantities. Example: [PER Wolff] , currently a journalist in [LOC Argentina] , played with [PER Del Bosque] in the final years of the seventies in [ORG Real Madrid] . MasakhaNER 2.0 is a named entity dataset consisting of PER, ORG, LOC, and DATE entities annotated by Masakhane for 20 African languages The train/validation/test sets are available for all the 20 languages. For more details see https://arxiv.org/abs/2210.12391 ### Supported Tasks and Leaderboards [More Information Needed] - `named-entity-recognition`: The performance in this task is measured with [F1](https://huggingface.co/metrics/f1) (higher is better). A named entity is correct only if it is an exact match of the corresponding entity in the data. ### Languages There are 20 languages available : - Bambara (bam) - Ghomala (bbj) - Ewe (ewe) - Fon (fon) - Hausa (hau) - Igbo (ibo) - Kinyarwanda (kin) - Luganda (lug) - Dholuo (luo) - Mossi (mos) - Chichewa (nya) - Nigerian Pidgin - chShona (sna) - Kiswahili (swą) - Setswana (tsn) - Twi (twi) - Wolof (wol) - isiXhosa (xho) - Yorùbá (yor) - isiZulu (zul) ## Dataset Structure ### Data Instances The examples look like this for Yorùbá: ``` from datasets import load_dataset data = load_dataset('masakhane/masakhaner2', 'yor') # Please, specify the language code # A data point consists of sentences seperated by empty line and tab-seperated tokens and tags. {'id': '0', 'ner_tags': [B-DATE, I-DATE, 0, 0, 0, 0, 0, B-PER, I-PER, I-PER, O, O, O, O], 'tokens': ['Wákàtí', 'méje', 'ti', 'ré', 'kọjá', 'lọ', 'tí', 'Luis', 'Carlos', 'Díaz', 'ti', 'di', 'awati', '.'] } ``` ### Data Fields - `id`: id of the sample - `tokens`: the tokens of the example text - `ner_tags`: the NER tags of each token The NER tags correspond to this list: ``` "O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC", "B-DATE", "I-DATE", ``` In the NER tags, a B denotes the first item of a phrase and an I any non-initial word. There are four types of phrases: person names (PER), organizations (ORG), locations (LOC) and dates & time (DATE). It is assumed that named entities are non-recursive and non-overlapping. In case a named entity is embedded in another named entity usually, only the top level entity is marked. ### Data Splits For all languages, there are three splits. The original splits were named `train`, `dev` and `test` and they correspond to the `train`, `validation` and `test` splits. The splits have the following sizes : | Language | train | validation | test | |-----------------|------:|-----------:|------:| | Bambara | 4463 | 638 | 1274 | | Ghomala | 3384 | 483 | 966 | | Ewe | 3505 | 501 | 1001 | | Fon. | 4343 | 621 | 1240 | | Hausa | 5716 | 816 | 1633 | | Igbo | 7634 | 1090 | 2181 | | Kinyarwanda | 7825 | 1118 | 2235 | | Luganda | 4942 | 706 | 1412 | | Luo | 5161 | 737 | 1474 | | Mossi | 4532 | 648 | 1613 | | Nigerian-Pidgin | 5646 | 806 | 1294 | | Chichewa | 6250 | 893 | 1785 | | chiShona | 6207 | 887 | 1773 | | Kiswahili | 6593 | 942 | 1883 | | Setswana | 3289 | 499 | 996 | | Akan/Twi | 4240 | 605 | 1211 | | Wolof | 4593 | 656 | 1312 | | isiXhosa | 5718 | 817 | 1633 | | Yoruba | 6877 | 983 | 1964 | | isiZulu | 5848 | 836 | 1670 | ## Dataset Creation ### Curation Rationale The dataset was introduced to introduce new resources to 20 languages that were under-served for natural language processing. [More Information Needed] ### Source Data The source of the data is from the news domain, details can be found here https://arxiv.org/abs/2210.12391 #### Initial Data Collection and Normalization The articles were word-tokenized, information on the exact pre-processing pipeline is unavailable. #### Who are the source language producers? The source language was produced by journalists and writers employed by the news agency and newspaper mentioned above. ### Annotations #### Annotation process Details can be found here https://arxiv.org/abs/2103.11811 #### Who are the annotators? Annotators were recruited from [Masakhane](https://www.masakhane.io/) ### Personal and Sensitive Information The data is sourced from newspaper source and only contains mentions of public figures or individuals ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations Users should keep in mind that the dataset only contains news text, which might limit the applicability of the developed systems to other domains. ## Additional Information ### Dataset Curators ### Licensing Information The licensing status of the data is CC 4.0 Non-Commercial ### Citation Information Provide the [BibTex](http://www.bibtex.org/)-formatted reference for the dataset. For example: ``` @article{Adelani2022MasakhaNER2A, title={MasakhaNER 2.0: Africa-centric Transfer Learning for Named Entity Recognition}, author={David Ifeoluwa Adelani and Graham Neubig and Sebastian Ruder and Shruti Rijhwani and Michael Beukman and Chester Palen-Michel and Constantine Lignos and Jesujoba Oluwadara Alabi and Shamsuddeen Hassan Muhammad and Peter Nabende and Cheikh M. Bamba Dione and Andiswa Bukula and Rooweither Mabuya and Bonaventure F. P. Dossou and Blessing K. Sibanda and Happy Buzaaba and Jonathan Mukiibi and Godson Kalipe and Derguene Mbaye and Amelia Taylor and Fatoumata Kabore and Chris C. Emezue and Anuoluwapo Aremu and Perez Ogayo and Catherine W. Gitau and Edwin Munkoh-Buabeng and Victoire Memdjokam Koagne and Allahsera Auguste Tapo and Tebogo Macucwa and Vukosi Marivate and Elvis Mboning and Tajuddeen R. Gwadabe and Tosin P. Adewumi and Orevaoghene Ahia and Joyce Nakatumba-Nabende and Neo L. Mokono and Ignatius M Ezeani and Chiamaka Ijeoma Chukwuneke and Mofetoluwa Adeyemi and Gilles Hacheme and Idris Abdulmumin and Odunayo Ogundepo and Oreen Yousuf and Tatiana Moteu Ngoli and Dietrich Klakow}, journal={ArXiv}, year={2022}, volume={abs/2210.12391} } ``` ### Contributions Thanks to [@dadelani](https://github.com/dadelani) for adding this dataset.
Dahoas/cot_gsm8k
2023-05-31T13:01:00.000Z
[ "region:us" ]
Dahoas
null
null
null
4
1,351
--- dataset_info: features: - name: question dtype: string - name: answer dtype: string - name: prompt dtype: string - name: response dtype: string splits: - name: train num_bytes: 7710945 num_examples: 7217 - name: val num_bytes: 267770 num_examples: 256 - name: test num_bytes: 1436697 num_examples: 1319 download_size: 5472201 dataset_size: 9415412 --- # Dataset Card for "cot_gsm8k" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
conll2012_ontonotesv5
2023-01-25T15:03:49.000Z
[ "task_categories:token-classification", "task_ids:named-entity-recognition", "task_ids:part-of-speech", "task_ids:coreference-resolution", "task_ids:parsing", "task_ids:lemmatization", "task_ids:word-sense-disambiguation", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:multilingual", "size_categories:10K<n<100K", "source_datasets:original", "language:ar", "language:en", "language:zh", "license:cc-by-nc-nd-4.0", "semantic-role-labeling", "region:us" ]
null
OntoNotes v5.0 is the final version of OntoNotes corpus, and is a large-scale, multi-genre, multilingual corpus manually annotated with syntactic, semantic and discourse information. This dataset is the version of OntoNotes v5.0 extended and is used in the CoNLL-2012 shared task. It includes v4 train/dev and v9 test data for English/Chinese/Arabic and corrected version v12 train/dev/test data (English only). The source of data is the Mendeley Data repo [ontonotes-conll2012](https://data.mendeley.com/datasets/zmycy7t9h9), which seems to be as the same as the official data, but users should use this dataset on their own responsibility. See also summaries from paperwithcode, [OntoNotes 5.0](https://paperswithcode.com/dataset/ontonotes-5-0) and [CoNLL-2012](https://paperswithcode.com/dataset/conll-2012-1) For more detailed info of the dataset like annotation, tag set, etc., you can refer to the documents in the Mendeley repo mentioned above.
@inproceedings{pradhan-etal-2013-towards, title = "Towards Robust Linguistic Analysis using {O}nto{N}otes", author = {Pradhan, Sameer and Moschitti, Alessandro and Xue, Nianwen and Ng, Hwee Tou and Bj{\"o}rkelund, Anders and Uryupina, Olga and Zhang, Yuchen and Zhong, Zhi}, booktitle = "Proceedings of the Seventeenth Conference on Computational Natural Language Learning", month = aug, year = "2013", address = "Sofia, Bulgaria", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/W13-3516", pages = "143--152", } Ralph Weischedel, Martha Palmer, Mitchell Marcus, Eduard Hovy, Sameer Pradhan, Lance Ramshaw, Nianwen Xue, Ann Taylor, Jeff Kaufman, Michelle Franchini, Mohammed El-Bachouti, Robert Belvin, Ann Houston. OntoNotes Release 5.0 LDC2013T19. Web Download. Philadelphia: Linguistic Data Consortium, 2013.
null
23
1,343
--- annotations_creators: - expert-generated language_creators: - found language: - ar - en - zh license: - cc-by-nc-nd-4.0 multilinguality: - multilingual size_categories: - 10K<n<100K source_datasets: - original task_categories: - token-classification task_ids: - named-entity-recognition - part-of-speech - coreference-resolution - parsing - lemmatization - word-sense-disambiguation paperswithcode_id: ontonotes-5-0 pretty_name: CoNLL2012 shared task data based on OntoNotes 5.0 tags: - semantic-role-labeling dataset_info: - config_name: english_v4 features: - name: document_id dtype: string - name: sentences list: - name: part_id dtype: int32 - name: words sequence: string - name: pos_tags sequence: class_label: names: '0': XX '1': '``' '2': $ '3': '''''' '4': ',' '5': -LRB- '6': -RRB- '7': . '8': ':' '9': ADD '10': AFX '11': CC '12': CD '13': DT '14': EX '15': FW '16': HYPH '17': IN '18': JJ '19': JJR '20': JJS '21': LS '22': MD '23': NFP '24': NN '25': NNP '26': NNPS '27': NNS '28': PDT '29': POS '30': PRP '31': PRP$ '32': RB '33': RBR '34': RBS '35': RP '36': SYM '37': TO '38': UH '39': VB '40': VBD '41': VBG '42': VBN '43': VBP '44': VBZ '45': WDT '46': WP '47': WP$ '48': WRB - name: parse_tree dtype: string - name: predicate_lemmas sequence: string - name: predicate_framenet_ids sequence: string - name: word_senses sequence: float32 - name: speaker dtype: string - name: named_entities sequence: class_label: names: '0': O '1': B-PERSON '2': I-PERSON '3': B-NORP '4': I-NORP '5': B-FAC '6': I-FAC '7': B-ORG '8': I-ORG '9': B-GPE '10': I-GPE '11': B-LOC '12': I-LOC '13': B-PRODUCT '14': I-PRODUCT '15': B-DATE '16': I-DATE '17': B-TIME '18': I-TIME '19': B-PERCENT '20': I-PERCENT '21': B-MONEY '22': I-MONEY '23': B-QUANTITY '24': I-QUANTITY '25': B-ORDINAL '26': I-ORDINAL '27': B-CARDINAL '28': I-CARDINAL '29': B-EVENT '30': I-EVENT '31': B-WORK_OF_ART '32': I-WORK_OF_ART '33': B-LAW '34': I-LAW '35': B-LANGUAGE '36': I-LANGUAGE - name: srl_frames list: - name: verb dtype: string - name: frames sequence: string - name: coref_spans sequence: sequence: int32 length: 3 splits: - name: train num_bytes: 112246121 num_examples: 1940 - name: validation num_bytes: 14116925 num_examples: 222 - name: test num_bytes: 14709044 num_examples: 222 download_size: 193644139 dataset_size: 141072090 - config_name: chinese_v4 features: - name: document_id dtype: string - name: sentences list: - name: part_id dtype: int32 - name: words sequence: string - name: pos_tags sequence: class_label: names: '0': X '1': AD '2': AS '3': BA '4': CC '5': CD '6': CS '7': DEC '8': DEG '9': DER '10': DEV '11': DT '12': ETC '13': FW '14': IJ '15': INF '16': JJ '17': LB '18': LC '19': M '20': MSP '21': NN '22': NR '23': NT '24': OD '25': 'ON' '26': P '27': PN '28': PU '29': SB '30': SP '31': URL '32': VA '33': VC '34': VE '35': VV - name: parse_tree dtype: string - name: predicate_lemmas sequence: string - name: predicate_framenet_ids sequence: string - name: word_senses sequence: float32 - name: speaker dtype: string - name: named_entities sequence: class_label: names: '0': O '1': B-PERSON '2': I-PERSON '3': B-NORP '4': I-NORP '5': B-FAC '6': I-FAC '7': B-ORG '8': I-ORG '9': B-GPE '10': I-GPE '11': B-LOC '12': I-LOC '13': B-PRODUCT '14': I-PRODUCT '15': B-DATE '16': I-DATE '17': B-TIME '18': I-TIME '19': B-PERCENT '20': I-PERCENT '21': B-MONEY '22': I-MONEY '23': B-QUANTITY '24': I-QUANTITY '25': B-ORDINAL '26': I-ORDINAL '27': B-CARDINAL '28': I-CARDINAL '29': B-EVENT '30': I-EVENT '31': B-WORK_OF_ART '32': I-WORK_OF_ART '33': B-LAW '34': I-LAW '35': B-LANGUAGE '36': I-LANGUAGE - name: srl_frames list: - name: verb dtype: string - name: frames sequence: string - name: coref_spans sequence: sequence: int32 length: 3 splits: - name: train num_bytes: 77195698 num_examples: 1391 - name: validation num_bytes: 10828169 num_examples: 172 - name: test num_bytes: 9585138 num_examples: 166 download_size: 193644139 dataset_size: 97609005 - config_name: arabic_v4 features: - name: document_id dtype: string - name: sentences list: - name: part_id dtype: int32 - name: words sequence: string - name: pos_tags sequence: string - name: parse_tree dtype: string - name: predicate_lemmas sequence: string - name: predicate_framenet_ids sequence: string - name: word_senses sequence: float32 - name: speaker dtype: string - name: named_entities sequence: class_label: names: '0': O '1': B-PERSON '2': I-PERSON '3': B-NORP '4': I-NORP '5': B-FAC '6': I-FAC '7': B-ORG '8': I-ORG '9': B-GPE '10': I-GPE '11': B-LOC '12': I-LOC '13': B-PRODUCT '14': I-PRODUCT '15': B-DATE '16': I-DATE '17': B-TIME '18': I-TIME '19': B-PERCENT '20': I-PERCENT '21': B-MONEY '22': I-MONEY '23': B-QUANTITY '24': I-QUANTITY '25': B-ORDINAL '26': I-ORDINAL '27': B-CARDINAL '28': I-CARDINAL '29': B-EVENT '30': I-EVENT '31': B-WORK_OF_ART '32': I-WORK_OF_ART '33': B-LAW '34': I-LAW '35': B-LANGUAGE '36': I-LANGUAGE - name: srl_frames list: - name: verb dtype: string - name: frames sequence: string - name: coref_spans sequence: sequence: int32 length: 3 splits: - name: train num_bytes: 42017761 num_examples: 359 - name: validation num_bytes: 4859292 num_examples: 44 - name: test num_bytes: 4900664 num_examples: 44 download_size: 193644139 dataset_size: 51777717 - config_name: english_v12 features: - name: document_id dtype: string - name: sentences list: - name: part_id dtype: int32 - name: words sequence: string - name: pos_tags sequence: class_label: names: '0': XX '1': '``' '2': $ '3': '''''' '4': '*' '5': ',' '6': -LRB- '7': -RRB- '8': . '9': ':' '10': ADD '11': AFX '12': CC '13': CD '14': DT '15': EX '16': FW '17': HYPH '18': IN '19': JJ '20': JJR '21': JJS '22': LS '23': MD '24': NFP '25': NN '26': NNP '27': NNPS '28': NNS '29': PDT '30': POS '31': PRP '32': PRP$ '33': RB '34': RBR '35': RBS '36': RP '37': SYM '38': TO '39': UH '40': VB '41': VBD '42': VBG '43': VBN '44': VBP '45': VBZ '46': VERB '47': WDT '48': WP '49': WP$ '50': WRB - name: parse_tree dtype: string - name: predicate_lemmas sequence: string - name: predicate_framenet_ids sequence: string - name: word_senses sequence: float32 - name: speaker dtype: string - name: named_entities sequence: class_label: names: '0': O '1': B-PERSON '2': I-PERSON '3': B-NORP '4': I-NORP '5': B-FAC '6': I-FAC '7': B-ORG '8': I-ORG '9': B-GPE '10': I-GPE '11': B-LOC '12': I-LOC '13': B-PRODUCT '14': I-PRODUCT '15': B-DATE '16': I-DATE '17': B-TIME '18': I-TIME '19': B-PERCENT '20': I-PERCENT '21': B-MONEY '22': I-MONEY '23': B-QUANTITY '24': I-QUANTITY '25': B-ORDINAL '26': I-ORDINAL '27': B-CARDINAL '28': I-CARDINAL '29': B-EVENT '30': I-EVENT '31': B-WORK_OF_ART '32': I-WORK_OF_ART '33': B-LAW '34': I-LAW '35': B-LANGUAGE '36': I-LANGUAGE - name: srl_frames list: - name: verb dtype: string - name: frames sequence: string - name: coref_spans sequence: sequence: int32 length: 3 splits: - name: train num_bytes: 174173192 num_examples: 10539 - name: validation num_bytes: 24264804 num_examples: 1370 - name: test num_bytes: 18254144 num_examples: 1200 download_size: 193644139 dataset_size: 216692140 --- # Dataset Card for CoNLL2012 shared task data based on OntoNotes 5.0 ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [CoNLL-2012 Shared Task](https://conll.cemantix.org/2012/data.html), [Author's page](https://cemantix.org/data/ontonotes.html) - **Repository:** [Mendeley](https://data.mendeley.com/datasets/zmycy7t9h9) - **Paper:** [Towards Robust Linguistic Analysis using OntoNotes](https://aclanthology.org/W13-3516/) - **Leaderboard:** - **Point of Contact:** ### Dataset Summary OntoNotes v5.0 is the final version of OntoNotes corpus, and is a large-scale, multi-genre, multilingual corpus manually annotated with syntactic, semantic and discourse information. This dataset is the version of OntoNotes v5.0 extended and is used in the CoNLL-2012 shared task. It includes v4 train/dev and v9 test data for English/Chinese/Arabic and corrected version v12 train/dev/test data (English only). The source of data is the Mendeley Data repo [ontonotes-conll2012](https://data.mendeley.com/datasets/zmycy7t9h9), which seems to be as the same as the official data, but users should use this dataset on their own responsibility. See also summaries from paperwithcode, [OntoNotes 5.0](https://paperswithcode.com/dataset/ontonotes-5-0) and [CoNLL-2012](https://paperswithcode.com/dataset/conll-2012-1) For more detailed info of the dataset like annotation, tag set, etc., you can refer to the documents in the Mendeley repo mentioned above. ### Supported Tasks and Leaderboards - [Named Entity Recognition on Ontonotes v5 (English)](https://paperswithcode.com/sota/named-entity-recognition-ner-on-ontonotes-v5) - [Coreference Resolution on OntoNotes](https://paperswithcode.com/sota/coreference-resolution-on-ontonotes) - [Semantic Role Labeling on OntoNotes](https://paperswithcode.com/sota/semantic-role-labeling-on-ontonotes) - ... ### Languages V4 data for Arabic, Chinese, English, and V12 data for English ## Dataset Structure ### Data Instances ``` { {'document_id': 'nw/wsj/23/wsj_2311', 'sentences': [{'part_id': 0, 'words': ['CONCORDE', 'trans-Atlantic', 'flights', 'are', '$', '2, 'to', 'Paris', 'and', '$', '3, 'to', 'London', '.']}, 'pos_tags': [25, 18, 27, 43, 2, 12, 17, 25, 11, 2, 12, 17, 25, 7], 'parse_tree': '(TOP(S(NP (NNP CONCORDE) (JJ trans-Atlantic) (NNS flights) )(VP (VBP are) (NP(NP(NP ($ $) (CD 2,400) )(PP (IN to) (NP (NNP Paris) ))) (CC and) (NP(NP ($ $) (CD 3,200) )(PP (IN to) (NP (NNP London) ))))) (. .) ))', 'predicate_lemmas': [None, None, None, 'be', None, None, None, None, None, None, None, None, None, None], 'predicate_framenet_ids': [None, None, None, '01', None, None, None, None, None, None, None, None, None, None], 'word_senses': [None, None, None, None, None, None, None, None, None, None, None, None, None, None], 'speaker': None, 'named_entities': [7, 6, 0, 0, 0, 15, 0, 5, 0, 0, 15, 0, 5, 0], 'srl_frames': [{'frames': ['B-ARG1', 'I-ARG1', 'I-ARG1', 'B-V', 'B-ARG2', 'I-ARG2', 'I-ARG2', 'I-ARG2', 'I-ARG2', 'I-ARG2', 'I-ARG2', 'I-ARG2', 'I-ARG2', 'O'], 'verb': 'are'}], 'coref_spans': [], {'part_id': 0, 'words': ['In', 'a', 'Centennial', 'Journal', 'article', 'Oct.', '5', ',', 'the', 'fares', 'were', 'reversed', '.']}]} 'pos_tags': [17, 13, 25, 25, 24, 25, 12, 4, 13, 27, 40, 42, 7], 'parse_tree': '(TOP(S(PP (IN In) (NP (DT a) (NML (NNP Centennial) (NNP Journal) ) (NN article) ))(NP (NNP Oct.) (CD 5) ) (, ,) (NP (DT the) (NNS fares) )(VP (VBD were) (VP (VBN reversed) )) (. .) ))', 'predicate_lemmas': [None, None, None, None, None, None, None, None, None, None, None, 'reverse', None], 'predicate_framenet_ids': [None, None, None, None, None, None, None, None, None, None, None, '01', None], 'word_senses': [None, None, None, None, None, None, None, None, None, None, None, None, None], 'speaker': None, 'named_entities': [0, 0, 4, 22, 0, 12, 30, 0, 0, 0, 0, 0, 0], 'srl_frames': [{'frames': ['B-ARGM-LOC', 'I-ARGM-LOC', 'I-ARGM-LOC', 'I-ARGM-LOC', 'I-ARGM-LOC', 'B-ARGM-TMP', 'I-ARGM-TMP', 'O', 'B-ARG1', 'I-ARG1', 'O', 'B-V', 'O'], 'verb': 'reversed'}], 'coref_spans': [], } ``` ### Data Fields - **`document_id`** (*`str`*): This is a variation on the document filename - **`sentences`** (*`List[Dict]`*): All sentences of the same document are in a single example for the convenience of concatenating sentences. Every element in `sentences` is a *`Dict`* composed of the following data fields: - **`part_id`** (*`int`*) : Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. - **`words`** (*`List[str]`*) : - **`pos_tags`** (*`List[ClassLabel]` or `List[str]`*) : This is the Penn-Treebank-style part of speech. When parse information is missing, all parts of speech except the one for which there is some sense or proposition annotation are marked with a XX tag. The verb is marked with just a VERB tag. - tag set : Note tag sets below are founded by scanning all the data, and I found it seems to be a little bit different from officially stated tag sets. See official documents in the [Mendeley repo](https://data.mendeley.com/datasets/zmycy7t9h9) - arabic : str. Because pos tag in Arabic is compounded and complex, hard to represent it by `ClassLabel` - chinese v4 : `datasets.ClassLabel(num_classes=36, names=["X", "AD", "AS", "BA", "CC", "CD", "CS", "DEC", "DEG", "DER", "DEV", "DT", "ETC", "FW", "IJ", "INF", "JJ", "LB", "LC", "M", "MSP", "NN", "NR", "NT", "OD", "ON", "P", "PN", "PU", "SB", "SP", "URL", "VA", "VC", "VE", "VV",])`, where `X` is for pos tag missing - english v4 : `datasets.ClassLabel(num_classes=49, names=["XX", "``", "$", "''", ",", "-LRB-", "-RRB-", ".", ":", "ADD", "AFX", "CC", "CD", "DT", "EX", "FW", "HYPH", "IN", "JJ", "JJR", "JJS", "LS", "MD", "NFP", "NN", "NNP", "NNPS", "NNS", "PDT", "POS", "PRP", "PRP$", "RB", "RBR", "RBS", "RP", "SYM", "TO", "UH", "VB", "VBD", "VBG", "VBN", "VBP", "VBZ", "WDT", "WP", "WP$", "WRB",])`, where `XX` is for pos tag missing, and `-LRB-`/`-RRB-` is "`(`" / "`)`". - english v12 : `datasets.ClassLabel(num_classes=51, names="english_v12": ["XX", "``", "$", "''", "*", ",", "-LRB-", "-RRB-", ".", ":", "ADD", "AFX", "CC", "CD", "DT", "EX", "FW", "HYPH", "IN", "JJ", "JJR", "JJS", "LS", "MD", "NFP", "NN", "NNP", "NNPS", "NNS", "PDT", "POS", "PRP", "PRP$", "RB", "RBR", "RBS", "RP", "SYM", "TO", "UH", "VB", "VBD", "VBG", "VBN", "VBP", "VBZ", "VERB", "WDT", "WP", "WP$", "WRB",])`, where `XX` is for pos tag missing, and `-LRB-`/`-RRB-` is "`(`" / "`)`". - **`parse_tree`** (*`Optional[str]`*) : An serialized NLTK Tree representing the parse. It includes POS tags as pre-terminal nodes. When the parse information is missing, the parse will be `None`. - **`predicate_lemmas`** (*`List[Optional[str]]`*) : The predicate lemma of the words for which we have semantic role information or word sense information. All other indices are `None`. - **`predicate_framenet_ids`** (*`List[Optional[int]]`*) : The PropBank frameset ID of the lemmas in predicate_lemmas, or `None`. - **`word_senses`** (*`List[Optional[float]]`*) : The word senses for the words in the sentence, or None. These are floats because the word sense can have values after the decimal, like 1.1. - **`speaker`** (*`Optional[str]`*) : This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. When it is not available, it will be `None`. - **`named_entities`** (*`List[ClassLabel]`*) : The BIO tags for named entities in the sentence. - tag set : `datasets.ClassLabel(num_classes=37, names=["O", "B-PERSON", "I-PERSON", "B-NORP", "I-NORP", "B-FAC", "I-FAC", "B-ORG", "I-ORG", "B-GPE", "I-GPE", "B-LOC", "I-LOC", "B-PRODUCT", "I-PRODUCT", "B-DATE", "I-DATE", "B-TIME", "I-TIME", "B-PERCENT", "I-PERCENT", "B-MONEY", "I-MONEY", "B-QUANTITY", "I-QUANTITY", "B-ORDINAL", "I-ORDINAL", "B-CARDINAL", "I-CARDINAL", "B-EVENT", "I-EVENT", "B-WORK_OF_ART", "I-WORK_OF_ART", "B-LAW", "I-LAW", "B-LANGUAGE", "I-LANGUAGE",])` - **`srl_frames`** (*`List[{"word":str, "frames":List[str]}]`*) : A dictionary keyed by the verb in the sentence for the given Propbank frame labels, in a BIO format. - **`coref spans`** (*`List[List[int]]`*) : The spans for entity mentions involved in coreference resolution within the sentence. Each element is a tuple composed of (cluster_id, start_index, end_index). Indices are inclusive. ### Data Splits Each dataset (arabic_v4, chinese_v4, english_v4, english_v12) has 3 splits: _train_, _validation_, and _test_ ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ``` @inproceedings{pradhan-etal-2013-towards, title = "Towards Robust Linguistic Analysis using {O}nto{N}otes", author = {Pradhan, Sameer and Moschitti, Alessandro and Xue, Nianwen and Ng, Hwee Tou and Bj{\"o}rkelund, Anders and Uryupina, Olga and Zhang, Yuchen and Zhong, Zhi}, booktitle = "Proceedings of the Seventeenth Conference on Computational Natural Language Learning", month = aug, year = "2013", address = "Sofia, Bulgaria", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/W13-3516", pages = "143--152", } ``` ### Contributions Thanks to [@richarddwang](https://github.com/richarddwang) for adding this dataset.
TIGER-Lab/MathInstruct
2023-10-07T01:40:07.000Z
[ "task_categories:text-generation", "size_categories:100K<n<1M", "language:en", "license:mit", "arxiv:2309.05653", "region:us" ]
TIGER-Lab
null
null
null
87
1,340
--- license: mit task_categories: - text-generation language: - en pretty_name: MathInstruct size_categories: - 100K<n<1M --- # 🦣 MAmmoTH: Building Math Generalist Models through Hybrid Instruction Tuning MathInstruct is a meticulously curated instruction tuning dataset that is lightweight yet generalizable. MathInstruct is compiled from 13 math rationale datasets, six of which are newly curated by this work. It uniquely focuses on the hybrid use of chain-of-thought (CoT) and program-of-thought (PoT) rationales, and ensures extensive coverage of diverse mathematical fields. Project Page: [https://tiger-ai-lab.github.io/MAmmoTH/](https://tiger-ai-lab.github.io/MAmmoTH/) Paper: [https://arxiv.org/pdf/2309.05653.pdf](https://arxiv.org/pdf/2309.05653.pdf) Code: [https://github.com/TIGER-AI-Lab/MAmmoTH](https://github.com/TIGER-AI-Lab/MAmmoTH) Models: | | **Base Model: Llama-2** | **Base Model: Code Llama** | |-----|---------------------------------------------------------------|--------------------------------------------------------------------------| | 7B | 🦣 [MAmmoTH-7B](https://huggingface.co/TIGER-Lab/MAmmoTH-7B) | 🦣 [MAmmoTH-Coder-7B](https://huggingface.co/TIGER-Lab/MAmmoTH-Coder-7B) | | 13B | 🦣 [MAmmoTH-13B](https://huggingface.co/TIGER-Lab/MAmmoTH-13B) | 🦣 [MAmmoTH-Coder-13B](https://huggingface.co/TIGER-Lab/MAmmoTH-Coder-13B)| | 34B | - | 🦣 [MAmmoTH-Coder-34B](https://huggingface.co/TIGER-Lab/MAmmoTH-Coder-34B)| | 70B | 🦣 [MAmmoTH-70B](https://huggingface.co/TIGER-Lab/MAmmoTH-70B) | - | ## **License** Please check out the license of each subset in our curated dataset MathInstruct. | Dataset Name | License Type | |--------------|----------------| | GSM8K | MIT | | GSM8K-RFT | Non listed | | AQuA-RAT | Apache 2.0 | | MATH | MIT | | TheoremQA | MIT | | Camel-Math | Attribution-NonCommercial 4.0 International | | NumGLUE | Apache-2.0 | | CrowdSourced (Lila) | Attribution 4.0 International | | MathQA | Apache-2.0 | | Our Curated | MIT | ## **Citation** Please cite our paper if you use our data, model or code. Please also kindly cite the original dataset papers. ``` @article{yue2023mammoth, title={MAmmoTH: Building Math Generalist Models through Hybrid Instruction Tuning}, author={Xiang Yue, Xingwei Qu, Ge Zhang, Yao Fu, Wenhao Huang, Huan Sun, Yu Su, Wenhu Chen}, journal={arXiv preprint arXiv:2309.05653}, year={2023} } ```
bigcode/guanaco-commits
2023-06-28T08:54:47.000Z
[ "region:us" ]
bigcode
null
null
null
3
1,339
--- dataset_info: features: - name: prompt dtype: string - name: completion dtype: string splits: - name: train num_bytes: 17347601.0 num_examples: 12958 - name: test num_bytes: 827046.0 num_examples: 629 download_size: 10948498 dataset_size: 18174647.0 --- # Dataset Card for "guanaco-commits" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
pie/brat
2023-09-20T16:04:35.000Z
[ "region:us" ]
pie
null
null
null
0
1,328
Entry not found
lhoestq/test
2022-07-01T15:26:34.000Z
[ "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:n<1K", "source_datasets:original", "language:en", "license:mit", "region:us" ]
lhoestq
This is a test dataset.
\
null
0
1,324
--- type: test annotations_creators: - expert-generated language_creators: - found language: - en license: - mit multilinguality: - monolingual size_categories: - n<1K source_datasets: - original task_categories: - other-test task_ids: - other-test paperswithcode_id: null pretty_name: Test Dataset --- This is a test dataset
HuggingFaceH4/test-dataset-all-splits
2023-04-25T22:09:49.000Z
[ "region:us" ]
HuggingFaceH4
null
null
null
0
1,320
--- dataset_info: features: - name: chosen list: - name: content dtype: string - name: role dtype: string - name: rejected list: - name: content dtype: string - name: role dtype: string - name: prompt dtype: string - name: messages list: - name: content dtype: string - name: role dtype: string splits: - name: train_ift num_bytes: 230850 num_examples: 100 - name: train_rl num_bytes: 369068 num_examples: 100 - name: train_rm num_bytes: 369068 num_examples: 100 - name: test_rm num_bytes: 312141 num_examples: 100 - name: test_rl num_bytes: 312141 num_examples: 100 - name: test_ift num_bytes: 218856 num_examples: 100 download_size: 1071322 dataset_size: 1812124 --- # Dataset Card for "test-dataset-all-splits" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
tner/bc5cdr
2022-07-18T00:43:04.000Z
[ "task_categories:token-classification", "task_ids:named-entity-recognition", "multilinguality:monolingual", "size_categories:10K<n<100K", "language:en", "license:other", "region:us" ]
tner
[Bio Creative 5 CDR NER dataset](https://academic.oup.com/database/article/doi/10.1093/database/baw032/2630271?login=true)
@article{wei2016assessing, title={Assessing the state of the art in biomedical relation extraction: overview of the BioCreative V chemical-disease relation (CDR) task}, author={Wei, Chih-Hsuan and Peng, Yifan and Leaman, Robert and Davis, Allan Peter and Mattingly, Carolyn J and Li, Jiao and Wiegers, Thomas C and Lu, Zhiyong}, journal={Database}, volume={2016}, year={2016}, publisher={Oxford Academic} }
null
1
1,319
--- language: - en license: - other multilinguality: - monolingual size_categories: - 10K<n<100K task_categories: - token-classification task_ids: - named-entity-recognition pretty_name: BioCreative V CDR --- # Dataset Card for "tner/bc5cdr" ## Dataset Description - **Repository:** [T-NER](https://github.com/asahi417/tner) - **Paper:** [https://academic.oup.com/database/article/doi/10.1093/database/baw032/2630271?login=true](https://academic.oup.com/database/article/doi/10.1093/database/baw032/2630271?login=true) - **Dataset:** BioCreative V CDR - **Domain:** Biomedical - **Number of Entity:** 2 ### Dataset Summary BioCreative V CDR NER dataset formatted in a part of [TNER](https://github.com/asahi417/tner) project. The original dataset consists of long documents which cannot be fed on LM because of the length, so we split them into sentences to reduce their size. - Entity Types: `Chemical`, `Disease` ## Dataset Structure ### Data Instances An example of `train` looks as follows. ``` { 'tags': [2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2, 0, 0], 'tokens': ['Fasciculations', 'in', 'six', 'areas', 'of', 'the', 'body', 'were', 'scored', 'from', '0', 'to', '3', 'and', 'summated', 'as', 'a', 'total', 'fasciculation', 'score', '.'] } ``` ### Label ID The label2id dictionary can be found at [here](https://huggingface.co/datasets/tner/bc5cdr/raw/main/dataset/label.json). ```python { "O": 0, "B-Chemical": 1, "B-Disease": 2, "I-Disease": 3, "I-Chemical": 4 } ``` ### Data Splits | name |train|validation|test| |---------|----:|---------:|---:| |bc5cdr|5228| 5330|5865| ### Citation Information ``` @article{wei2016assessing, title={Assessing the state of the art in biomedical relation extraction: overview of the BioCreative V chemical-disease relation (CDR) task}, author={Wei, Chih-Hsuan and Peng, Yifan and Leaman, Robert and Davis, Allan Peter and Mattingly, Carolyn J and Li, Jiao and Wiegers, Thomas C and Lu, Zhiyong}, journal={Database}, volume={2016}, year={2016}, publisher={Oxford Academic} } ```
Cohere/wikipedia-22-12-en-embeddings
2023-03-22T16:51:57.000Z
[ "task_categories:text-retrieval", "task_ids:document-retrieval", "annotations_creators:expert-generated", "multilinguality:multilingual", "language:en", "license:apache-2.0", "region:us" ]
Cohere
null
null
null
34
1,315
--- annotations_creators: - expert-generated language: - en multilinguality: - multilingual size_categories: [] source_datasets: [] tags: [] task_categories: - text-retrieval license: - apache-2.0 task_ids: - document-retrieval --- # Wikipedia (en) embedded with cohere.ai `multilingual-22-12` encoder We encoded [Wikipedia (en)](https://en.wikipedia.org) using the [cohere.ai](https://txt.cohere.ai/multilingual/) `multilingual-22-12` embedding model. To get an overview how this dataset was created and pre-processed, have a look at [Cohere/wikipedia-22-12](https://huggingface.co/datasets/Cohere/wikipedia-22-12). ## Embeddings We compute for `title+" "+text` the embeddings using our `multilingual-22-12` embedding model, a state-of-the-art model that works for semantic search in 100 languages. If you want to learn more about this model, have a look at [cohere.ai multilingual embedding model](https://txt.cohere.ai/multilingual/). ## Further languages We provide embeddings of Wikipedia in many different languages: [ar](https://huggingface.co/datasets/Cohere/wikipedia-22-12-ar-embeddings), [de](https://huggingface.co/datasets/Cohere/wikipedia-22-12-de-embeddings), [en](https://huggingface.co/datasets/Cohere/wikipedia-22-12-en-embeddings), [es](https://huggingface.co/datasets/Cohere/wikipedia-22-12-es-embeddings), [fr](https://huggingface.co/datasets/Cohere/wikipedia-22-12-fr-embeddings), [hi](https://huggingface.co/datasets/Cohere/wikipedia-22-12-hi-embeddings), [it](https://huggingface.co/datasets/Cohere/wikipedia-22-12-it-embeddings), [ja](https://huggingface.co/datasets/Cohere/wikipedia-22-12-ja-embeddings), [ko](https://huggingface.co/datasets/Cohere/wikipedia-22-12-ko-embeddings), [simple english](https://huggingface.co/datasets/Cohere/wikipedia-22-12-simple-embeddings), [zh](https://huggingface.co/datasets/Cohere/wikipedia-22-12-zh-embeddings), You can find the Wikipedia datasets without embeddings at [Cohere/wikipedia-22-12](https://huggingface.co/datasets/Cohere/wikipedia-22-12). ## Loading the dataset You can either load the dataset like this: ```python from datasets import load_dataset docs = load_dataset(f"Cohere/wikipedia-22-12-en-embeddings", split="train") ``` Or you can also stream it without downloading it before: ```python from datasets import load_dataset docs = load_dataset(f"Cohere/wikipedia-22-12-en-embeddings", split="train", streaming=True) for doc in docs: docid = doc['id'] title = doc['title'] text = doc['text'] emb = doc['emb'] ``` ## Search A full search example: ```python #Run: pip install cohere datasets from datasets import load_dataset import torch import cohere co = cohere.Client(f"<<COHERE_API_KEY>>") # Add your cohere API key from www.cohere.com #Load at max 1000 documents + embeddings max_docs = 1000 docs_stream = load_dataset(f"Cohere/wikipedia-22-12-en-embeddings", split="train", streaming=True) docs = [] doc_embeddings = [] for doc in docs_stream: docs.append(doc) doc_embeddings.append(doc['emb']) if len(docs) >= max_docs: break doc_embeddings = torch.tensor(doc_embeddings) query = 'Who founded Youtube' response = co.embed(texts=[query], model='multilingual-22-12') query_embedding = response.embeddings query_embedding = torch.tensor(query_embedding) # Compute dot score between query embedding and document embeddings dot_scores = torch.mm(query_embedding, doc_embeddings.transpose(0, 1)) top_k = torch.topk(dot_scores, k=3) # Print results print("Query:", query) for doc_id in top_k.indices[0].tolist(): print(docs[doc_id]['title']) print(docs[doc_id]['text'], "\n") ``` ## Performance You can find performance on the MIRACL dataset (a semantic search evaluation dataset) here: [miracl-en-queries-22-12#performance](https://huggingface.co/datasets/Cohere/miracl-en-queries-22-12#performance)
ccdv/mediasum
2022-10-25T10:56:04.000Z
[ "task_categories:summarization", "task_categories:text2text-generation", "multilinguality:monolingual", "size_categories:100K<n<1M", "language:en", "conditional-text-generation", "region:us" ]
ccdv
MediaSum dataset for summarization. From paper: "MediaSum: A Large-scale Media Interview Dataset for Dialogue Summarization" by C. Zhu et al."
@article{zhu2021mediasum, title={MediaSum: A Large-scale Media Interview Dataset for Dialogue Summarization}, author={Zhu, Chenguang and Liu, Yang and Mei, Jie and Zeng, Michael}, journal={arXiv preprint arXiv:2103.06410}, year={2021} }
null
5
1,309
--- language: - en multilinguality: - monolingual size_categories: - 100K<n<1M task_categories: - summarization - text2text-generation task_ids: [] tags: - conditional-text-generation --- # MediaSum dataset for summarization Summarization dataset copied from [MediaSum: A Large-scale Media Interview Dataset for Dialogue Summarization](https://github.com/zcgzcgzcg1/MediaSum) This dataset is compatible with the [`run_summarization.py`](https://github.com/huggingface/transformers/tree/master/examples/pytorch/summarization) script from Transformers if you add this line to the `summarization_name_mapping` variable: ```python "ccdv/mediasum": ("document", "summary") ``` # Configs 4 possibles configs: - `roberta` will concatenate documents with "\</s\>" - `newline` will concatenate documents with "\n" - `bert` will concatenate documents with "[SEP]" - `list` will return the list of documents instead of a single string Add `_prepended` to config name to prepend the speaker name before each dialogue: `speaker: text` \ Default is `roberta_prepended` (compatible with BART). ### Data Fields - `id`: paper id - `document`: a string/list containing the body of a set of documents - `summary`: a string containing the abstract of the set ### Data Splits This dataset has 3 splits: _train_, _validation_, and _test_. \ | Dataset Split | Number of Instances | | ------------- | --------------------| | Train | 443596 | | Validation | 10000 | | Test | 10000 | # Cite original article ``` @article{zhu2021mediasum, title={MediaSum: A Large-scale Media Interview Dataset for Dialogue Summarization}, author={Zhu, Chenguang and Liu, Yang and Mei, Jie and Zeng, Michael}, journal={arXiv preprint arXiv:2103.06410}, year={2021} } ```
bigbio/med_qa
2023-09-26T13:00:32.000Z
[ "multilinguality:multilingual", "language:en", "language:zh", "license:unknown", "region:us" ]
bigbio
In this work, we present the first free-form multiple-choice OpenQA dataset for solving medical problems, MedQA, collected from the professional medical board exams. It covers three languages: English, simplified Chinese, and traditional Chinese, and contains 12,723, 34,251, and 14,123 questions for the three languages, respectively. Together with the question data, we also collect and release a large-scale corpus from medical textbooks from which the reading comprehension models can obtain necessary knowledge for answering the questions.
@article{jin2021disease, title={What disease does this patient have? a large-scale open domain question answering dataset from medical exams}, author={Jin, Di and Pan, Eileen and Oufattole, Nassim and Weng, Wei-Hung and Fang, Hanyi and Szolovits, Peter}, journal={Applied Sciences}, volume={11}, number={14}, pages={6421}, year={2021}, publisher={MDPI} }
null
20
1,308
--- language: - en - zh bigbio_language: - English - Chinese (Simplified) - Chinese (Traditional, Taiwan) license: unknown multilinguality: multilingual bigbio_license_shortname: UNKNOWN pretty_name: MedQA homepage: https://github.com/jind11/MedQA bigbio_pubmed: False bigbio_public: True bigbio_tasks: - QUESTION_ANSWERING --- # Dataset Card for MedQA ## Dataset Description - **Homepage:** https://github.com/jind11/MedQA - **Pubmed:** False - **Public:** True - **Tasks:** QA In this work, we present the first free-form multiple-choice OpenQA dataset for solving medical problems, MedQA, collected from the professional medical board exams. It covers three languages: English, simplified Chinese, and traditional Chinese, and contains 12,723, 34,251, and 14,123 questions for the three languages, respectively. Together with the question data, we also collect and release a large-scale corpus from medical textbooks from which the reading comprehension models can obtain necessary knowledge for answering the questions. ## Citation Information ``` @article{jin2021disease, title={What disease does this patient have? a large-scale open domain question answering dataset from medical exams}, author={Jin, Di and Pan, Eileen and Oufattole, Nassim and Weng, Wei-Hung and Fang, Hanyi and Szolovits, Peter}, journal={Applied Sciences}, volume={11}, number={14}, pages={6421}, year={2021}, publisher={MDPI} } ```
PolyAI/banking77
2022-10-25T10:12:22.000Z
[ "task_categories:text-classification", "task_ids:intent-classification", "task_ids:multi-class-classification", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:cc-by-4.0", "arxiv:2003.04807", "region:us" ]
PolyAI
BANKING77 dataset provides a very fine-grained set of intents in a banking domain. It comprises 13,083 customer service queries labeled with 77 intents. It focuses on fine-grained single-domain intent detection.
@inproceedings{Casanueva2020, author = {I{\~{n}}igo Casanueva and Tadas Temcinas and Daniela Gerz and Matthew Henderson and Ivan Vulic}, title = {Efficient Intent Detection with Dual Sentence Encoders}, year = {2020}, month = {mar}, note = {Data available at https://github.com/PolyAI-LDN/task-specific-datasets}, url = {https://arxiv.org/abs/2003.04807}, booktitle = {Proceedings of the 2nd Workshop on NLP for ConvAI - ACL 2020} }
null
17
1,305
--- annotations_creators: - expert-generated extended: - original language_creators: - expert-generated language: - en license: - cc-by-4.0 multilinguality: - monolingual size_categories: - 10K<n<100K source_datasets: - original task_categories: - text-classification task_ids: - intent-classification - multi-class-classification paperswithcode_id: null pretty_name: BANKING77 --- # Dataset Card for BANKING77 ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Github](https://github.com/PolyAI-LDN/task-specific-datasets) - **Repository:** [Github](https://github.com/PolyAI-LDN/task-specific-datasets) - **Paper:** [ArXiv](https://arxiv.org/abs/2003.04807) - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Dataset composed of online banking queries annotated with their corresponding intents. BANKING77 dataset provides a very fine-grained set of intents in a banking domain. It comprises 13,083 customer service queries labeled with 77 intents. It focuses on fine-grained single-domain intent detection. ### Supported Tasks and Leaderboards Intent classification, intent detection ### Languages English ## Dataset Structure ### Data Instances An example of 'train' looks as follows: ``` { 'label': 11, # integer label corresponding to "card_arrival" intent 'text': 'I am still waiting on my card?' } ``` ### Data Fields - `text`: a string feature. - `label`: One of classification labels (0-76) corresponding to unique intents. Intent names are mapped to `label` in the following way: | label | intent (category) | |---:|:-------------------------------------------------| | 0 | activate_my_card | | 1 | age_limit | | 2 | apple_pay_or_google_pay | | 3 | atm_support | | 4 | automatic_top_up | | 5 | balance_not_updated_after_bank_transfer | | 6 | balance_not_updated_after_cheque_or_cash_deposit | | 7 | beneficiary_not_allowed | | 8 | cancel_transfer | | 9 | card_about_to_expire | | 10 | card_acceptance | | 11 | card_arrival | | 12 | card_delivery_estimate | | 13 | card_linking | | 14 | card_not_working | | 15 | card_payment_fee_charged | | 16 | card_payment_not_recognised | | 17 | card_payment_wrong_exchange_rate | | 18 | card_swallowed | | 19 | cash_withdrawal_charge | | 20 | cash_withdrawal_not_recognised | | 21 | change_pin | | 22 | compromised_card | | 23 | contactless_not_working | | 24 | country_support | | 25 | declined_card_payment | | 26 | declined_cash_withdrawal | | 27 | declined_transfer | | 28 | direct_debit_payment_not_recognised | | 29 | disposable_card_limits | | 30 | edit_personal_details | | 31 | exchange_charge | | 32 | exchange_rate | | 33 | exchange_via_app | | 34 | extra_charge_on_statement | | 35 | failed_transfer | | 36 | fiat_currency_support | | 37 | get_disposable_virtual_card | | 38 | get_physical_card | | 39 | getting_spare_card | | 40 | getting_virtual_card | | 41 | lost_or_stolen_card | | 42 | lost_or_stolen_phone | | 43 | order_physical_card | | 44 | passcode_forgotten | | 45 | pending_card_payment | | 46 | pending_cash_withdrawal | | 47 | pending_top_up | | 48 | pending_transfer | | 49 | pin_blocked | | 50 | receiving_money | | 51 | Refund_not_showing_up | | 52 | request_refund | | 53 | reverted_card_payment? | | 54 | supported_cards_and_currencies | | 55 | terminate_account | | 56 | top_up_by_bank_transfer_charge | | 57 | top_up_by_card_charge | | 58 | top_up_by_cash_or_cheque | | 59 | top_up_failed | | 60 | top_up_limits | | 61 | top_up_reverted | | 62 | topping_up_by_card | | 63 | transaction_charged_twice | | 64 | transfer_fee_charged | | 65 | transfer_into_account | | 66 | transfer_not_received_by_recipient | | 67 | transfer_timing | | 68 | unable_to_verify_identity | | 69 | verify_my_identity | | 70 | verify_source_of_funds | | 71 | verify_top_up | | 72 | virtual_card_not_working | | 73 | visa_or_mastercard | | 74 | why_verify_identity | | 75 | wrong_amount_of_cash_received | | 76 | wrong_exchange_rate_for_cash_withdrawal | ### Data Splits | Dataset statistics | Train | Test | | --- | --- | --- | | Number of examples | 10 003 | 3 080 | | Average character length | 59.5 | 54.2 | | Number of intents | 77 | 77 | | Number of domains | 1 | 1 | ## Dataset Creation ### Curation Rationale Previous intent detection datasets such as Web Apps, Ask Ubuntu, the Chatbot Corpus or SNIPS are limited to small number of classes (<10), which oversimplifies the intent detection task and does not emulate the true environment of commercial systems. Although there exist large scale *multi-domain* datasets ([HWU64](https://github.com/xliuhw/NLU-Evaluation-Data) and [CLINC150](https://github.com/clinc/oos-eval)), the examples per each domain may not sufficiently capture the full complexity of each domain as encountered "in the wild". This dataset tries to fill the gap and provides a very fine-grained set of intents in a *single-domain* i.e. **banking**. Its focus on fine-grained single-domain intent detection makes it complementary to the other two multi-domain datasets. ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process The dataset does not contain any additional annotations. #### Who are the annotators? [N/A] ### Personal and Sensitive Information [N/A] ## Considerations for Using the Data ### Social Impact of Dataset The purpose of this dataset it to help develop better intent detection systems. Any comprehensive intent detection evaluation should involve both coarser-grained multi-domain datasets and a fine-grained single-domain dataset such as BANKING77. ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [PolyAI](https://github.com/PolyAI-LDN) ### Licensing Information Creative Commons Attribution 4.0 International ### Citation Information ``` @inproceedings{Casanueva2020, author = {I{\~{n}}igo Casanueva and Tadas Temcinas and Daniela Gerz and Matthew Henderson and Ivan Vulic}, title = {Efficient Intent Detection with Dual Sentence Encoders}, year = {2020}, month = {mar}, note = {Data available at https://github.com/PolyAI-LDN/task-specific-datasets}, url = {https://arxiv.org/abs/2003.04807}, booktitle = {Proceedings of the 2nd Workshop on NLP for ConvAI - ACL 2020} } ``` ### Contributions Thanks to [@dkajtoch](https://github.com/dkajtoch) for adding this dataset.
ChilleD/SVAMP
2023-04-24T07:55:08.000Z
[ "task_categories:text-generation", "size_categories:n<1K", "language:en", "license:mit", "region:us" ]
ChilleD
null
null
null
1
1,297
--- license: mit task_categories: - text-generation language: - en size_categories: - n<1K ---
alzoubi36/piextract
2023-06-25T07:11:15.000Z
[ "region:us" ]
alzoubi36
null
null
null
0
1,297
--- dataset_info: features: - name: COLLECT struct: - name: subtask dtype: string - name: tags sequence: string - name: tokens sequence: string - name: NOT_COLLECT struct: - name: subtask dtype: string - name: tags sequence: string - name: tokens sequence: string - name: NOT_SHARE struct: - name: subtask dtype: string - name: tags sequence: string - name: tokens sequence: string - name: SHARE struct: - name: subtask dtype: string - name: tags sequence: string - name: tokens sequence: string splits: - name: train num_bytes: 3453408 num_examples: 2579 - name: test num_bytes: 1580498 num_examples: 1029 - name: validation num_bytes: 662810 num_examples: 456 download_size: 1013894 dataset_size: 5696716 --- # Dataset for the PI-Extract task in the [PrivacyGLUE](https://github.com/infsys-lab/privacy-glue) dataset
stingning/ultrachat
2023-07-04T10:19:58.000Z
[ "task_categories:conversational", "task_categories:text-generation", "size_categories:1M<n<10M", "language:en", "license:cc-by-nc-4.0", "region:us" ]
stingning
null
null
null
121
1,291
--- license: cc-by-nc-4.0 task_categories: - conversational - text-generation language: - en size_categories: - 1M<n<10M pretty_name: UltraChat --- # Dataset Card for Dataset Name ## Dataset Description An open-source, large-scale, and multi-round dialogue data powered by Turbo APIs. In consideration of factors such as safeguarding privacy, **we do not directly use any data available on the Internet as prompts**. To ensure generation quality, two separate ChatGPT Turbo APIs are adopted in generation, where one plays the role of the user to generate queries and the other generates the response. We instruct the user model with carefully designed prompts to mimic human user behavior and call the two APIs iteratively. The generated dialogues undergo further post-processing and filtering. ULtraChat is composed of three sectors: - 🌏 **Questions about the World**: The dialogue data in this sector is derived from a wide range of inquiries related to concepts, entities, and objects from the real world. The topics covered are extensive, spanning areas such as technology, art, and entrepreneurship. - ✍🏻 **Writing and Creation**: The dialogue data in this sector is driven by the demands for writing/creation from scratch, and encompasses any tasks that an AI assistant may aid within the creative process, spanning from email composition to crafting narratives and plays, and beyond. - 📋 **Assistance on Existent Materials**: The dialogue data in this sector is generated based on existing materials, including but not limited to rewriting, continuation, summarization, and inference, covering a diverse range of topics. - Repository: [UltraChat](https://github.com/thunlp/UltraChat) - Explorer: [plain-explorer](http://39.101.77.220/), [Nomic-AI-Atlas-Explorer](https://atlas.nomic.ai/map/0ce65783-c3a9-40b5-895d-384933f50081/a7b46301-022f-45d8-bbf4-98107eabdbac) ## Dataset Structure Each line in the downloaded data file is a json dict containing the data id and dialogue data in a list format. Below is an example line. ``` { "id": "0", "data": [ "How can cross training benefit groups like runners, swimmers, or weightlifters?", "Cross training can benefit groups like runners, swimmers, or weightlifters in the following ways: ...", "That makes sense. I've been wanting to improve my running time, but I never thought about incorporating strength training. Do you have any recommendations for specific exercises?", "Sure, here are some strength training exercises that can benefit runners: ...", "Hmm, I'm not really a fan of weightlifting though. Can I incorporate other forms of exercise into my routine to improve my running time?", "Yes, absolutely! ...", "..." ] } ``` ### Citation Information ```bibtex @misc{UltraChat, author = {Ding, Ning and Chen, Yulin and Xu, Bokai and Hu, Shengding and Qin, Yujia and Liu, Zhiyuan and Sun, Maosong and Zhou, Bowen}, title = {UltraChat: A Large-scale Auto-generated Multi-round Dialogue Data}, year = {2023}, publisher = {GitHub}, journal = {GitHub repository}, howpublished = {\url{https://github.com/thunlp/ultrachat}}, } ```
lm1b
2023-06-27T15:36:19.000Z
[ "task_categories:text-generation", "task_categories:fill-mask", "task_ids:language-modeling", "task_ids:masked-language-modeling", "language:en", "arxiv:1312.3005", "region:us" ]
null
A benchmark corpus to be used for measuring progress in statistical language modeling. This has almost one billion words in the training data.
@article{DBLP:journals/corr/ChelbaMSGBK13, author = {Ciprian Chelba and Tomas Mikolov and Mike Schuster and Qi Ge and Thorsten Brants and Phillipp Koehn}, title = {One Billion Word Benchmark for Measuring Progress in Statistical Language Modeling}, journal = {CoRR}, volume = {abs/1312.3005}, year = {2013}, url = {http://arxiv.org/abs/1312.3005}, archivePrefix = {arXiv}, eprint = {1312.3005}, timestamp = {Mon, 13 Aug 2018 16:46:16 +0200}, biburl = {https://dblp.org/rec/bib/journals/corr/ChelbaMSGBK13}, bibsource = {dblp computer science bibliography, https://dblp.org} }
null
8
1,286
--- pretty_name: One Billion Word Language Model Benchmark paperswithcode_id: billion-word-benchmark dataset_info: features: - name: text dtype: string config_name: plain_text splits: - name: train num_bytes: 4238206516 num_examples: 30301028 - name: test num_bytes: 42942045 num_examples: 306688 download_size: 1792209805 dataset_size: 4281148561 task_categories: - text-generation - fill-mask task_ids: - language-modeling - masked-language-modeling language: - en --- # Dataset Card for One Billion Word Language Model Benchmark ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [statmt](http://www.statmt.org/lm-benchmark/) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [arxiv](https://arxiv.org/pdf/1312.3005v3.pdf) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 1.79 GB - **Size of the generated dataset:** 4.28 GB - **Total amount of disk used:** 6.07 GB ### Dataset Summary A benchmark corpus to be used for measuring progress in statistical language modeling. This has almost one billion words in the training data. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### plain_text - **Size of downloaded dataset files:** 1.79 GB - **Size of the generated dataset:** 4.28 GB - **Total amount of disk used:** 6.07 GB An example of 'train' looks as follows. ``` This example was too long and was cropped: { "text": "While athletes in different professions dealt with doping scandals and other controversies , Woods continued to do what he did best : dominate the field of professional golf and rake in endorsements ." } ``` ### Data Fields The data fields are the same among all splits. #### plain_text - `text`: a `string` feature. ### Data Splits | name | train | test | |------------|----------|--------| | plain_text | 30301028 | 306688 | ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations The dataset doesn't contain annotations. ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needeate this repository accordingly. ### Citation Information ```bibtex @misc{chelba2014billion, title={One Billion Word Benchmark for Measuring Progress in Statistical Language Modeling}, author={Ciprian Chelba and Tomas Mikolov and Mike Schuster and Qi Ge and Thorsten Brants and Phillipp Koehn and Tony Robinson}, year={2014}, eprint={1312.3005}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` ### Contributions Thanks to [@patrickvonplaten](https://github.com/patrickvonplaten), [@lewtun](https://github.com/lewtun), [@jplu](https://github.com/jplu), [@thomwolf](https://github.com/thomwolf) for adding this dataset.
proteinea/secondary_structure_prediction
2023-03-02T22:42:31.000Z
[ "doi:10.57967/hf/1104", "region:us" ]
proteinea
null
null
null
1
1,281
Entry not found
visual_genome
2023-06-29T15:23:59.000Z
[ "task_categories:image-to-text", "task_categories:object-detection", "task_categories:visual-question-answering", "task_ids:image-captioning", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:en", "license:cc-by-4.0", "region:us" ]
null
Visual Genome enable to model objects and relationships between objects. They collect dense annotations of objects, attributes, and relationships within each image. Specifically, the dataset contains over 108K images where each image has an average of 35 objects, 26 attributes, and 21 pairwise relationships between objects.
@article{Krishna2016VisualGC, title={Visual Genome: Connecting Language and Vision Using Crowdsourced Dense Image Annotations}, author={Ranjay Krishna and Yuke Zhu and Oliver Groth and Justin Johnson and Kenji Hata and Joshua Kravitz and Stephanie Chen and Yannis Kalantidis and Li-Jia Li and David A. Shamma and Michael S. Bernstein and Li Fei-Fei}, journal={International Journal of Computer Vision}, year={2017}, volume={123}, pages={32-73}, url={https://doi.org/10.1007/s11263-016-0981-7}, doi={10.1007/s11263-016-0981-7} }
null
29
1,278
--- annotations_creators: - found language_creators: - found language: - en license: - cc-by-4.0 multilinguality: - monolingual size_categories: - 100K<n<1M source_datasets: - original task_categories: - image-to-text - object-detection - visual-question-answering task_ids: - image-captioning paperswithcode_id: visual-genome pretty_name: VisualGenome dataset_info: features: - name: image dtype: image - name: image_id dtype: int32 - name: url dtype: string - name: width dtype: int32 - name: height dtype: int32 - name: coco_id dtype: int64 - name: flickr_id dtype: int64 - name: regions list: - name: region_id dtype: int32 - name: image_id dtype: int32 - name: phrase dtype: string - name: x dtype: int32 - name: y dtype: int32 - name: width dtype: int32 - name: height dtype: int32 config_name: region_descriptions_v1.0.0 splits: - name: train num_bytes: 260873884 num_examples: 108077 download_size: 15304605295 dataset_size: 260873884 config_names: - objects - question_answers - region_descriptions --- # Dataset Card for Visual Genome ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Dataset Preprocessing](#dataset-preprocessing) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://homes.cs.washington.edu/~ranjay/visualgenome/ - **Repository:** - **Paper:** https://doi.org/10.1007/s11263-016-0981-7 - **Leaderboard:** - **Point of Contact:** ranjaykrishna [at] gmail [dot] com ### Dataset Summary Visual Genome is a dataset, a knowledge base, an ongoing effort to connect structured image concepts to language. From the paper: > Despite progress in perceptual tasks such as image classification, computers still perform poorly on cognitive tasks such as image description and question answering. Cognition is core to tasks that involve not just recognizing, but reasoning about our visual world. However, models used to tackle the rich content in images for cognitive tasks are still being trained using the same datasets designed for perceptual tasks. To achieve success at cognitive tasks, models need to understand the interactions and relationships between objects in an image. When asked “What vehicle is the person riding?”, computers will need to identify the objects in an image as well as the relationships riding(man, carriage) and pulling(horse, carriage) to answer correctly that “the person is riding a horse-drawn carriage.” Visual Genome has: - 108,077 image - 5.4 Million Region Descriptions - 1.7 Million Visual Question Answers - 3.8 Million Object Instances - 2.8 Million Attributes - 2.3 Million Relationships From the paper: > Our dataset contains over 108K images where each image has an average of 35 objects, 26 attributes, and 21 pairwise relationships between objects. We canonicalize the objects, attributes, relationships, and noun phrases in region descriptions and questions answer pairs to WordNet synsets. ### Dataset Preprocessing ### Supported Tasks and Leaderboards ### Languages All of annotations use English as primary language. ## Dataset Structure ### Data Instances When loading a specific configuration, users has to append a version dependent suffix: ```python from datasets import load_dataset load_dataset("visual_genome", "region_description_v1.2.0") ``` #### region_descriptions An example of looks as follows. ``` { "image": <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=800x600 at 0x7F2F60698610>, "image_id": 1, "url": "https://cs.stanford.edu/people/rak248/VG_100K_2/1.jpg", "width": 800, "height": 600, "coco_id": null, "flickr_id": null, "regions": [ { "region_id": 1382, "image_id": 1, "phrase": "the clock is green in colour", "x": 421, "y": 57, "width": 82, "height": 139 }, ... ] } ``` #### objects An example of looks as follows. ``` { "image": <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=800x600 at 0x7F2F60698610>, "image_id": 1, "url": "https://cs.stanford.edu/people/rak248/VG_100K_2/1.jpg", "width": 800, "height": 600, "coco_id": null, "flickr_id": null, "objects": [ { "object_id": 1058498, "x": 421, "y": 91, "w": 79, "h": 339, "names": [ "clock" ], "synsets": [ "clock.n.01" ] }, ... ] } ``` #### attributes An example of looks as follows. ``` { "image": <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=800x600 at 0x7F2F60698610>, "image_id": 1, "url": "https://cs.stanford.edu/people/rak248/VG_100K_2/1.jpg", "width": 800, "height": 600, "coco_id": null, "flickr_id": null, "attributes": [ { "object_id": 1058498, "x": 421, "y": 91, "w": 79, "h": 339, "names": [ "clock" ], "synsets": [ "clock.n.01" ], "attributes": [ "green", "tall" ] }, ... } ] ``` #### relationships An example of looks as follows. ``` { "image": <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=800x600 at 0x7F2F60698610>, "image_id": 1, "url": "https://cs.stanford.edu/people/rak248/VG_100K_2/1.jpg", "width": 800, "height": 600, "coco_id": null, "flickr_id": null, "relationships": [ { "relationship_id": 15927, "predicate": "ON", "synsets": "['along.r.01']", "subject": { "object_id": 5045, "x": 119, "y": 338, "w": 274, "h": 192, "names": [ "shade" ], "synsets": [ "shade.n.01" ] }, "object": { "object_id": 5046, "x": 77, "y": 328, "w": 714, "h": 262, "names": [ "street" ], "synsets": [ "street.n.01" ] } } ... } ] ``` #### question_answers An example of looks as follows. ``` { "image": <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=800x600 at 0x7F2F60698610>, "image_id": 1, "url": "https://cs.stanford.edu/people/rak248/VG_100K_2/1.jpg", "width": 800, "height": 600, "coco_id": null, "flickr_id": null, "qas": [ { "qa_id": 986768, "image_id": 1, "question": "What color is the clock?", "answer": "Green.", "a_objects": [], "q_objects": [] }, ... } ] ``` ### Data Fields When loading a specific configuration, users has to append a version dependent suffix: ```python from datasets import load_dataset load_dataset("visual_genome", "region_description_v1.2.0") ``` #### region_descriptions - `image`: A `PIL.Image.Image` object containing the image. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]` - `image_id`: Unique numeric ID of the image. - `url`: URL of source image. - `width`: Image width. - `height`: Image height. - `coco_id`: Id mapping to MSCOCO indexing. - `flickr_id`: Id mapping to Flicker indexing. - `regions`: Holds a list of `Region` dataclasses: - `region_id`: Unique numeric ID of the region. - `image_id`: Unique numeric ID of the image. - `x`: x coordinate of bounding box's top left corner. - `y`: y coordinate of bounding box's top left corner. - `width`: Bounding box width. - `height`: Bounding box height. #### objects - `image`: A `PIL.Image.Image` object containing the image. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]` - `image_id`: Unique numeric ID of the image. - `url`: URL of source image. - `width`: Image width. - `height`: Image height. - `coco_id`: Id mapping to MSCOCO indexing. - `flickr_id`: Id mapping to Flicker indexing. - `objects`: Holds a list of `Object` dataclasses: - `object_id`: Unique numeric ID of the object. - `x`: x coordinate of bounding box's top left corner. - `y`: y coordinate of bounding box's top left corner. - `w`: Bounding box width. - `h`: Bounding box height. - `names`: List of names associated with the object. This field can hold multiple values in the sense the multiple names are considered as acceptable. For example: ['monitor', 'computer'] at https://cs.stanford.edu/people/rak248/VG_100K/3.jpg - `synsets`: List of `WordNet synsets`. #### attributes - `image`: A `PIL.Image.Image` object containing the image. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]` - `image_id`: Unique numeric ID of the image. - `url`: URL of source image. - `width`: Image width. - `height`: Image height. - `coco_id`: Id mapping to MSCOCO indexing. - `flickr_id`: Id mapping to Flicker indexing. - `attributes`: Holds a list of `Object` dataclasses: - `object_id`: Unique numeric ID of the region. - `x`: x coordinate of bounding box's top left corner. - `y`: y coordinate of bounding box's top left corner. - `w`: Bounding box width. - `h`: Bounding box height. - `names`: List of names associated with the object. This field can hold multiple values in the sense the multiple names are considered as acceptable. For example: ['monitor', 'computer'] at https://cs.stanford.edu/people/rak248/VG_100K/3.jpg - `synsets`: List of `WordNet synsets`. - `attributes`: List of attributes associated with the object. #### relationships - `image`: A `PIL.Image.Image` object containing the image. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]` - `image_id`: Unique numeric ID of the image. - `url`: URL of source image. - `width`: Image width. - `height`: Image height. - `coco_id`: Id mapping to MSCOCO indexing. - `flickr_id`: Id mapping to Flicker indexing. - `relationships`: Holds a list of `Relationship` dataclasses: - `relationship_id`: Unique numeric ID of the object. - `predicate`: Predicate defining relationship between a subject and an object. - `synsets`: List of `WordNet synsets`. - `subject`: Object dataclass. See subsection on `objects`. - `object`: Object dataclass. See subsection on `objects`. #### question_answers - `image`: A `PIL.Image.Image` object containing the image. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]` - `image_id`: Unique numeric ID of the image. - `url`: URL of source image. - `width`: Image width. - `height`: Image height. - `coco_id`: Id mapping to MSCOCO indexing. - `flickr_id`: Id mapping to Flicker indexing. - `qas`: Holds a list of `Question-Answering` dataclasses: - `qa_id`: Unique numeric ID of the question-answer pair. - `image_id`: Unique numeric ID of the image. - `question`: Question. - `answer`: Answer. - `q_objects`: List of object dataclass associated with `question` field. See subsection on `objects`. - `a_objects`: List of object dataclass associated with `answer` field. See subsection on `objects`. ### Data Splits All the data is contained in training set. ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? From the paper: > We used Amazon Mechanical Turk (AMT) as our primary source of annotations. Overall, a total of over 33, 000 unique workers contributed to the dataset. The dataset was collected over the course of 6 months after 15 months of experimentation and iteration on the data representation. Approximately 800, 000 Human Intelligence Tasks (HITs) were launched on AMT, where each HIT involved creating descriptions, questions and answers, or region graphs. Each HIT was designed such that workers manage to earn anywhere between $6-$8 per hour if they work continuously, in line with ethical research standards on Mechanical Turk (Salehi et al., 2015). Visual Genome HITs achieved a 94.1% retention rate, meaning that 94.1% of workers who completed one of our tasks went ahead to do more. [...] 93.02% of workers contributed from the United States. The majority of our workers were between the ages of 25 and 34 years old. Our youngest contributor was 18 years and the oldest was 68 years old. We also had a near-balanced split of 54.15% male and 45.85% female workers. ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Visual Genome by Ranjay Krishna is licensed under a Creative Commons Attribution 4.0 International License. ### Citation Information ```bibtex @article{Krishna2016VisualGC, title={Visual Genome: Connecting Language and Vision Using Crowdsourced Dense Image Annotations}, author={Ranjay Krishna and Yuke Zhu and Oliver Groth and Justin Johnson and Kenji Hata and Joshua Kravitz and Stephanie Chen and Yannis Kalantidis and Li-Jia Li and David A. Shamma and Michael S. Bernstein and Li Fei-Fei}, journal={International Journal of Computer Vision}, year={2017}, volume={123}, pages={32-73}, url={https://doi.org/10.1007/s11263-016-0981-7}, doi={10.1007/s11263-016-0981-7} } ``` ### Contributions Due to limitation of the dummy_data creation, we provide a `fix_generated_dummy_data.py` script that fix the dataset in-place. Thanks to [@thomasw21](https://github.com/thomasw21) for adding this dataset.
big_patent
2023-06-01T14:59:54.000Z
[ "task_categories:summarization", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "size_categories:10K<n<100K", "size_categories:1M<n<10M", "source_datasets:original", "language:en", "license:cc-by-4.0", "patent-summarization", "arxiv:1906.03741", "region:us" ]
null
BIGPATENT, consisting of 1.3 million records of U.S. patent documents along with human written abstractive summaries. Each US patent application is filed under a Cooperative Patent Classification (CPC) code. There are nine such classification categories: A (Human Necessities), B (Performing Operations; Transporting), C (Chemistry; Metallurgy), D (Textiles; Paper), E (Fixed Constructions), F (Mechanical Engineering; Lightning; Heating; Weapons; Blasting), G (Physics), H (Electricity), and Y (General tagging of new or cross-sectional technology) There are two features: - description: detailed description of patent. - abstract: Patent abastract.
@misc{sharma2019bigpatent, title={BIGPATENT: A Large-Scale Dataset for Abstractive and Coherent Summarization}, author={Eva Sharma and Chen Li and Lu Wang}, year={2019}, eprint={1906.03741}, archivePrefix={arXiv}, primaryClass={cs.CL} }
null
24
1,276
--- annotations_creators: - no-annotation language_creators: - found language: - en license: - cc-by-4.0 multilinguality: - monolingual size_categories: - 100K<n<1M - 10K<n<100K - 1M<n<10M source_datasets: - original task_categories: - summarization task_ids: [] paperswithcode_id: bigpatent pretty_name: Big Patent tags: - patent-summarization dataset_info: - config_name: all features: - name: description dtype: string - name: abstract dtype: string splits: - name: train num_bytes: 38367048389 num_examples: 1207222 - name: validation num_bytes: 2115827002 num_examples: 67068 - name: test num_bytes: 2129505280 num_examples: 67072 download_size: 10142923776 dataset_size: 42612380671 - config_name: a features: - name: description dtype: string - name: abstract dtype: string splits: - name: train num_bytes: 5683460620 num_examples: 174134 - name: validation num_bytes: 313324505 num_examples: 9674 - name: test num_bytes: 316633277 num_examples: 9675 download_size: 10142923776 dataset_size: 6313418402 - config_name: b features: - name: description dtype: string - name: abstract dtype: string splits: - name: train num_bytes: 4236070976 num_examples: 161520 - name: validation num_bytes: 234425138 num_examples: 8973 - name: test num_bytes: 231538734 num_examples: 8974 download_size: 10142923776 dataset_size: 4702034848 - config_name: c features: - name: description dtype: string - name: abstract dtype: string splits: - name: train num_bytes: 4506249306 num_examples: 101042 - name: validation num_bytes: 244684775 num_examples: 5613 - name: test num_bytes: 252566793 num_examples: 5614 download_size: 10142923776 dataset_size: 5003500874 - config_name: d features: - name: description dtype: string - name: abstract dtype: string splits: - name: train num_bytes: 264717412 num_examples: 10164 - name: validation num_bytes: 14560482 num_examples: 565 - name: test num_bytes: 14403430 num_examples: 565 download_size: 10142923776 dataset_size: 293681324 - config_name: e features: - name: description dtype: string - name: abstract dtype: string splits: - name: train num_bytes: 881101433 num_examples: 34443 - name: validation num_bytes: 48646158 num_examples: 1914 - name: test num_bytes: 48586429 num_examples: 1914 download_size: 10142923776 dataset_size: 978334020 - config_name: f features: - name: description dtype: string - name: abstract dtype: string splits: - name: train num_bytes: 2146383473 num_examples: 85568 - name: validation num_bytes: 119632631 num_examples: 4754 - name: test num_bytes: 119596303 num_examples: 4754 download_size: 10142923776 dataset_size: 2385612407 - config_name: g features: - name: description dtype: string - name: abstract dtype: string splits: - name: train num_bytes: 8877854206 num_examples: 258935 - name: validation num_bytes: 492581177 num_examples: 14385 - name: test num_bytes: 496324853 num_examples: 14386 download_size: 10142923776 dataset_size: 9866760236 - config_name: h features: - name: description dtype: string - name: abstract dtype: string splits: - name: train num_bytes: 8075621958 num_examples: 257019 - name: validation num_bytes: 447602356 num_examples: 14279 - name: test num_bytes: 445460513 num_examples: 14279 download_size: 10142923776 dataset_size: 8968684827 - config_name: y features: - name: description dtype: string - name: abstract dtype: string splits: - name: train num_bytes: 3695589005 num_examples: 124397 - name: validation num_bytes: 200369780 num_examples: 6911 - name: test num_bytes: 204394948 num_examples: 6911 download_size: 10142923776 dataset_size: 4100353733 config_names: - a - all - b - c - d - e - f - g - h - y --- # Dataset Card for Big Patent ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Big Patent](https://evasharma.github.io/bigpatent/) - **Repository:** - **Paper:** [BIGPATENT: A Large-Scale Dataset for Abstractive and Coherent Summarization](https://arxiv.org/abs/1906.03741) - **Leaderboard:** - **Point of Contact:** [Lu Wang](mailto:wangluxy@umich.edu) ### Dataset Summary BIGPATENT, consisting of 1.3 million records of U.S. patent documents along with human written abstractive summaries. Each US patent application is filed under a Cooperative Patent Classification (CPC) code. There are nine such classification categories: - a: Human Necessities - b: Performing Operations; Transporting - c: Chemistry; Metallurgy - d: Textiles; Paper - e: Fixed Constructions - f: Mechanical Engineering; Lightning; Heating; Weapons; Blasting - g: Physics - h: Electricity - y: General tagging of new or cross-sectional technology Current defaults are 2.1.2 version (fix update to cased raw strings) and 'all' CPC codes: ```python from datasets import load_dataset ds = load_dataset("big_patent") # default is 'all' CPC codes ds = load_dataset("big_patent", "all") # the same as above ds = load_dataset("big_patent", "a") # only 'a' CPC codes ds = load_dataset("big_patent", codes=["a", "b"]) ``` To use 1.0.0 version (lower cased tokenized words), pass both parameters `codes` and `version`: ```python ds = load_dataset("big_patent", codes="all", version="1.0.0") ds = load_dataset("big_patent", codes="a", version="1.0.0") ds = load_dataset("big_patent", codes=["a", "b"], version="1.0.0") ``` ### Supported Tasks and Leaderboards [More Information Needed] ### Languages English ## Dataset Structure ### Data Instances Each instance contains a pair of `description` and `abstract`. `description` is extracted from the Description section of the Patent while `abstract` is extracted from the Abstract section. ``` { 'description': 'FIELD OF THE INVENTION \n [0001] This invention relates to novel calcium phosphate-coated implantable medical devices and processes of making same. The unique calcium-phosphate coated implantable medical devices minimize...', 'abstract': 'This invention relates to novel calcium phosphate-coated implantable medical devices...' } ``` ### Data Fields - `description`: detailed description of patent. - `abstract`: Patent abastract. ### Data Splits | | train | validation | test | |:----|------------------:|-------------:|-------:| | all | 1207222 | 67068 | 67072 | | a | 174134 | 9674 | 9675 | | b | 161520 | 8973 | 8974 | | c | 101042 | 5613 | 5614 | | d | 10164 | 565 | 565 | | e | 34443 | 1914 | 1914 | | f | 85568 | 4754 | 4754 | | g | 258935 | 14385 | 14386 | | h | 257019 | 14279 | 14279 | | y | 124397 | 6911 | 6911 | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ```bibtex @article{DBLP:journals/corr/abs-1906-03741, author = {Eva Sharma and Chen Li and Lu Wang}, title = {{BIGPATENT:} {A} Large-Scale Dataset for Abstractive and Coherent Summarization}, journal = {CoRR}, volume = {abs/1906.03741}, year = {2019}, url = {http://arxiv.org/abs/1906.03741}, eprinttype = {arXiv}, eprint = {1906.03741}, timestamp = {Wed, 26 Jun 2019 07:14:58 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-1906-03741.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ``` ### Contributions Thanks to [@mattbui](https://github.com/mattbui) for adding this dataset.
bigbio/meqsum
2022-12-22T15:45:35.000Z
[ "multilinguality:monolingual", "language:en", "license:unknown", "region:us" ]
bigbio
Dataset for medical question summarization introduced in the ACL 2019 paper "On the Summarization of Consumer Health Questions". Question understanding is one of the main challenges in question answering. In real world applications, users often submit natural language questions that are longer than needed and include peripheral information that increases the complexity of the question, leading to substantially more false positives in answer retrieval. In this paper, we study neural abstractive models for medical question summarization. We introduce the MeQSum corpus of 1,000 summarized consumer health questions.
@inproceedings{ben-abacha-demner-fushman-2019-summarization, title = "On the Summarization of Consumer Health Questions", author = "Ben Abacha, Asma and Demner-Fushman, Dina", booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics", month = jul, year = "2019", address = "Florence, Italy", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/P19-1215", doi = "10.18653/v1/P19-1215", pages = "2228--2234", abstract = "Question understanding is one of the main challenges in question answering. In real world applications, users often submit natural language questions that are longer than needed and include peripheral information that increases the complexity of the question, leading to substantially more false positives in answer retrieval. In this paper, we study neural abstractive models for medical question summarization. We introduce the MeQSum corpus of 1,000 summarized consumer health questions. We explore data augmentation methods and evaluate state-of-the-art neural abstractive models on this new task. In particular, we show that semantic augmentation from question datasets improves the overall performance, and that pointer-generator networks outperform sequence-to-sequence attentional models on this task, with a ROUGE-1 score of 44.16{\%}. We also present a detailed error analysis and discuss directions for improvement that are specific to question summarization.", }
null
0
1,269
--- language: - en bigbio_language: - English license: unknown multilinguality: monolingual bigbio_license_shortname: UNKNOWN pretty_name: MeQSum homepage: https://github.com/abachaa/MeQSum bigbio_pubmed: False bigbio_public: True bigbio_tasks: - SUMMARIZATION --- # Dataset Card for MeQSum ## Dataset Description - **Homepage:** https://github.com/abachaa/MeQSum - **Pubmed:** False - **Public:** True - **Tasks:** SUM Dataset for medical question summarization introduced in the ACL 2019 paper "On the Summarization of Consumer Health Questions". Question understanding is one of the main challenges in question answering. In real world applications, users often submit natural language questions that are longer than needed and include peripheral information that increases the complexity of the question, leading to substantially more false positives in answer retrieval. In this paper, we study neural abstractive models for medical question summarization. We introduce the MeQSum corpus of 1,000 summarized consumer health questions. ## Citation Information ``` @inproceedings{ben-abacha-demner-fushman-2019-summarization, title = "On the Summarization of Consumer Health Questions", author = "Ben Abacha, Asma and Demner-Fushman, Dina", booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics", month = jul, year = "2019", address = "Florence, Italy", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/P19-1215", doi = "10.18653/v1/P19-1215", pages = "2228--2234", abstract = "Question understanding is one of the main challenges in question answering. In real world applications, users often submit natural language questions that are longer than needed and include peripheral information that increases the complexity of the question, leading to substantially more false positives in answer retrieval. In this paper, we study neural abstractive models for medical question summarization. We introduce the MeQSum corpus of 1,000 summarized consumer health questions. We explore data augmentation methods and evaluate state-of-the-art neural abstractive models on this new task. In particular, we show that semantic augmentation from question datasets improves the overall performance, and that pointer-generator networks outperform sequence-to-sequence attentional models on this task, with a ROUGE-1 score of 44.16{\%}. We also present a detailed error analysis and discuss directions for improvement that are specific to question summarization.", } ```
tyqiangz/multilingual-sentiments
2023-05-23T15:01:51.000Z
[ "task_categories:text-classification", "task_ids:sentiment-analysis", "task_ids:sentiment-classification", "multilinguality:monolingual", "multilinguality:multilingual", "size_categories:100K<n<1M", "size_categories:1M<n<10M", "language:de", "language:en", "language:es", "language:fr", "language:ja", "language:zh", "language:id", "language:ar", "language:hi", "language:it", "language:ms", "language:pt", "license:apache-2.0", "region:us" ]
tyqiangz
null
null
null
18
1,255
--- language: - de - en - es - fr - ja - zh - id - ar - hi - it - ms - pt license: apache-2.0 multilinguality: - monolingual - multilingual size_categories: - 100K<n<1M - 1M<n<10M task_categories: - text-classification task_ids: - sentiment-analysis - sentiment-classification --- # Multilingual Sentiments Dataset A collection of multilingual sentiments datasets grouped into 3 classes -- positive, neutral, negative. Most multilingual sentiment datasets are either 2-class positive or negative, 5-class ratings of products reviews (e.g. Amazon multilingual dataset) or multiple classes of emotions. However, to an average person, sometimes positive, negative and neutral classes suffice and are more straightforward to perceive and annotate. Also, a positive/negative classification is too naive, most of the text in the world is actually neutral in sentiment. Furthermore, most multilingual sentiment datasets don't include Asian languages (e.g. Malay, Indonesian) and are dominated by Western languages (e.g. English, German). Git repo: https://github.com/tyqiangz/multilingual-sentiment-datasets ## Dataset Description - **Webpage:** https://github.com/tyqiangz/multilingual-sentiment-datasets
Joanne/Unified_Benchmark_for_Metaphor_Identification
2023-03-13T17:32:19.000Z
[ "region:us" ]
Joanne
[Unified Benchmark for Metaphor Identification]
null
null
0
1,254
Entry not found
mstz/acute_inflammation
2023-04-15T11:37:39.000Z
[ "task_categories:tabular-classification", "size_categories:100<n<1K", "language:en", "acute_inflammation", "tabular_classification", "binary_classification", "multiclass_classification", "UCI", "region:us" ]
mstz
null
@misc{misc_acute_inflammations_184, author = {Czerniak,Jacek}, title = {{Acute Inflammations}}, year = {2009}, howpublished = {UCI Machine Learning Repository}, note = {{DOI}: \\url{10.24432/C5V59S}} }
null
0
1,246
--- language: - en tags: - acute_inflammation - tabular_classification - binary_classification - multiclass_classification - UCI pretty_name: Acute Inflammation size_categories: - 100<n<1K task_categories: - tabular-classification configs: - inflammation - nephritis - bladder --- # Acute Inflammation The [Acute Inflammation dataset](https://archive.ics.uci.edu/ml/datasets/Acute+Inflammations) from the [UCI ML repository](https://archive-beta.ics.uci.edu). Predict whether the patient has an acute inflammation. # Configurations and tasks | **Configuration** | **Task** | Description | |-------------------|---------------------------|---------------------------------------------------------------| | inflammation | Binary classification | Does the patient have an acute inflammation? | | nephritis | Binary classification | Does the patient have a nephritic pelvis? | | bladder | Binary classification | Does the patient have bladder inflammation? | nephritis # Usage ```python from datasets import load_dataset dataset = load_dataset("mstz/acute_inflammation", "inflammation")["train"] ``` # Features Target feature changes according to the selected configuration and is always in last position in the dataset. | **Feature** | **Type** | |---------------------------------------|---------------| | `temperature` | `[float64]` | | `has_nausea` | `[bool]` | | `has_lumbar_pain` | `[bool]` | | `has_urine_pushing` | `[bool]` | | `has_micturition_pains` | `[bool]` | | `has_burnt_urethra` | `[bool]` | | `has_inflammed_bladder` | `[bool]` | | `has_nephritis_of_renal_pelvis` | `[bool]` | | `has_acute_inflammation` | `[int8]` |
lamini/taylor_swift
2023-07-24T03:47:45.000Z
[ "region:us" ]
lamini
null
null
null
1
1,223
--- dataset_info: features: - name: question dtype: string - name: answer dtype: string - name: input_ids sequence: int32 - name: attention_mask sequence: int8 - name: labels sequence: int64 splits: - name: train num_bytes: 850749.3 num_examples: 783 - name: test num_bytes: 94527.7 num_examples: 87 download_size: 303257 dataset_size: 945277.0 --- # Dataset Card for "taylor_swift" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
heegyu/hh-rlhf-vicuna-format
2023-09-06T03:07:11.000Z
[ "region:us" ]
heegyu
null
null
null
1
1,223
--- configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* dataset_info: features: - name: chosen struct: - name: from dtype: string - name: value dtype: string - name: rejected struct: - name: from dtype: string - name: value dtype: string - name: source dtype: string - name: context list: - name: from dtype: string - name: value dtype: string - name: instruction struct: - name: from dtype: string - name: value dtype: string splits: - name: train num_bytes: 210193081 num_examples: 160800 - name: test num_bytes: 11248194 num_examples: 8552 download_size: 128509060 dataset_size: 221441275 --- Original data is [Antrophic/hh-rlhf](https://huggingface.co/datasets/Anthropic/hh-rlhf). It was converted to Vicuna format sample ``` {'chosen': {'from': 'gpt', 'value': 'A horseshoe is usually made out of metal and is about 3 to 3.5 inches long and around 1 inch thick. The horseshoe should also have a 2 inch by 3 inch flat at the bottom where the rubber meets the metal. We also need two stakes and six horseshoes.'}, 'rejected': {'from': 'gpt', 'value': 'Horseshoes are either metal or plastic discs. The horseshoes come in different weights, and the lighter ones are easier to throw, so they are often the standard for beginning players.'}, 'source': 'helpful-base', 'context': [{'from': 'human', 'value': 'Hi, I want to learn to play horseshoes. Can you teach me?'}, {'from': 'gpt', 'value': 'I can, but maybe I should begin by telling you that a typical game consists of 2 players and 6 or 8 horseshoes.'}], 'instruction': {'from': 'human', 'value': 'Okay. What else is needed to play, and what are the rules?'}} ``` source columns has 4 values ``` {'harmless-base', 'helpful-base', 'helpful-online', 'helpful-rejection-sampled'} ``` In context, chosen, rejected column, 'from' key is either 'human' or 'gpt'.
allenai/mslr2022
2022-11-18T21:16:10.000Z
[ "task_categories:summarization", "task_categories:text2text-generation", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:extended|other-MS^2", "source_datasets:extended|other-Cochrane", "language:en", "license:apache-2.0", "region:us" ]
allenai
The Multidocument Summarization for Literature Review (MSLR) Shared Task aims to study how medical evidence from different clinical studies are summarized in literature reviews. Reviews provide the highest quality of evidence for clinical care, but are expensive to produce manually. (Semi-)automation via NLP may facilitate faster evidence synthesis without sacrificing rigor. The MSLR shared task uses two datasets to assess the current state of multidocument summarization for this task, and to encourage the development of modeling contributions, scaffolding tasks, methods for model interpretability, and improved automated evaluation methods in this domain.
@inproceedings{DeYoung2021MS2MS, title = {MSˆ2: Multi-Document Summarization of Medical Studies}, author = {Jay DeYoung and Iz Beltagy and Madeleine van Zuylen and Bailey Kuehl and Lucy Lu Wang}, booktitle = {EMNLP}, year = {2021} } @article{Wallace2020GeneratingN, title = {Generating (Factual?) Narrative Summaries of RCTs: Experiments with Neural Multi-Document Summarization}, author = {Byron C. Wallace and Sayantani Saha and Frank Soboczenski and Iain James Marshall}, year = 2020, journal = {AMIA Annual Symposium}, volume = {abs/2008.11293} }
null
5
1,217
--- annotations_creators: - expert-generated language_creators: - expert-generated language: - en license: - apache-2.0 multilinguality: - monolingual size_categories: - 10K<n<100K source_datasets: - extended|other-MS^2 - extended|other-Cochrane task_categories: - summarization - text2text-generation paperswithcode_id: multi-document-summarization pretty_name: MSLR Shared Task --- # Dataset Card for MSLR2022 ## Table of Contents - [Dataset Card for MSLR2022](#dataset-card-for-mslr2022) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Initial Data Collection and Normalization](#initial-data-collection-and-normalization) - [Who are the source language producers?](#who-are-the-source-language-producers) - [Annotations](#annotations) - [Annotation process](#annotation-process) - [Who are the annotators?](#who-are-the-annotators) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://github.com/allenai/mslr-shared-task - **Repository:** https://github.com/allenai/mslr-shared-task - **Paper:** https://aclanthology.org/2021.emnlp-main.594 - **Leaderboard:** https://github.com/allenai/mslr-shared-task#leaderboard - **Point of Contact:** https://github.com/allenai/mslr-shared-task#contact-us ### Dataset Summary The Multidocument Summarization for Literature Review (MSLR) Shared Task aims to study how medical evidence from different clinical studies are summarized in literature reviews. Reviews provide the highest quality of evidence for clinical care, but are expensive to produce manually. (Semi-)automation via NLP may facilitate faster evidence synthesis without sacrificing rigor. The MSLR shared task uses two datasets to assess the current state of multidocument summarization for this task, and to encourage the development of modeling contributions, scaffolding tasks, methods for model interpretability, and improved automated evaluation methods in this domain. ### Supported Tasks and Leaderboards This dataset is used for the MSLR2022 Shared Task. For information on the shared task leaderboard, please refer [here](https://github.com/allenai/mslr-shared-task#leaderboard). ### Languages English ## Dataset Structure More information on dataset structure [here](https://github.com/allenai/mslr-shared-task#data-structure). ### Data Instances __MS^2__ ```json { "review_id": "30760312", "pmid": [ "22776744", "25271670", "3493740", "1863023", "16291984", "23984728", "23996433", "18466198", "12151469", "27400308", "16053970", "22922316", "11897647", "11597664", "4230647" ], "title": [ "Improved Cell Survival and Paracrine Capacity of Human Embryonic Stem Cell-Derived Mesenchymal Stem Cells Promote Therapeutic Potential for Pulmonary Arterial Hypertension", "Adipose-derived stem cells attenuate pulmonary arterial hypertension and ameliorate pulmonary arterial remodeling in monocrotaline-induced pulmonary hypertensive rats", "Effect of bone marrow mesenchymal stem cells on experimental pulmonary arterial hypertension", "Survival in patients with primary pulmonary hypertension. Results from a national prospective registry.", "Sildenafil citrate therapy for pulmonary arterial hypertension.", "Macitentan and morbidity and mortality in pulmonary arterial hypertension.", "Long-term research of stem cells in monocrotaline-induced pulmonary arterial hypertension", "Safety and efficacy of autologous endothelial progenitor cells transplantation in children with idiopathic pulmonary arterial hypertension: open-label pilot study.", "Inhaled iloprost for severe pulmonary hypertension.", "Sildenafil reduces pulmonary vascular resistance in single ventricular physiology.", "Ambrisentan therapy for pulmonary arterial hypertension.", "Mesenchymal stem cell prevention of vascular remodeling in high flow-induced pulmonary hypertension through a paracrine mechanism.", "Continuous subcutaneous infusion of treprostinil, a prostacyclin analogue, in patients with pulmonary arterial hypertension: a double-blind, randomized, placebo-controlled trial.", "Effects of the dual endothelin-receptor antagonist bosentan in patients with pulmonary hypertension: a randomised placebocontrolled study", "SYRCLE\\u2019s risk of bias tool for animal studies" ], "abstract": [ "Although transplantation of adult bone marrow mesenchymal stem cells ( BM-MSCs ) holds promise in the treatment for pulmonary arterial hypertension ( PAH ) , the poor survival and differentiation potential of adult BM-MSCs have limited their therapeutic efficiency . Here , we compared the therapeutic efficacy of human embryonic stem cell-derived MSCs ( hESC-MSCs ) with adult BM-MSCs for the treatment of PAH in an animal model . One week following monocrotaline (MCT)-induced PAH , mice were r and omly assigned to receive phosphate-buffered saline ( MCT group ) ; 3.0 \\u00d7 106 human BM-derived MSCs ( BM-MSCs group ) or 3.0 \\u00d7 106 hESC-derived MSCs ( hESC-MSCs group ) via tail vein injection . At 3 weeks posttransplantation , the right ventricular systolic pressure ( RVSP ) , degree of RV hypertrophy , and medial wall thickening of pulmonary arteries were lower= , and pulmonary capillary density was higher in the hESC-MSC group as compared with BM-MSC and MCT groups ( all p < 0.05 ) . At 1 week posttransplantation , the number of engrafted MSCs in the lungs was found significantly higher in the hESC-MSC group than in the BM-MSC group ( all p < 0.01 ) . At 3 weeks posttransplantation , implanted BM-MSCs were undetectable whereas hESC-MSCs were not only engrafted in injured pulmonary arteries but had also undergone endothelial differentiation . In addition , protein profiling of hESC-MSC- and BM-MSC-conditioned medium revealed a differential paracrine capacity . Classification of these factors into bioprocesses revealed that secreted factors from hESC-MSCs were preferentially involved in early embryonic development and tissue differentiation , especially blood vessel morphogenesis . We concluded that improved cell survival and paracrine capacity of hESC-MSCs provide better therapeutic efficacy than BM-MSCs in the treatment for PAH", "Abstract We investigated the effect of adipose-derived stem cells ( ADSCs ) transplantation effects on structural remodeling and pulmonary artery pressure in monocrotaline (MCT)-induced pulmonary hypertensive rats . In the first experiment , 32 male Sprague-Dawley ( SD ) rats were r and omly divided into four groups ( n = 8/group ) : 3 ADSCs treated groups and normal control ( Ctrl ) . ADSCs were administered through the left jugular vein at 105 , 106 and 107 cells , respectively , and a cell density of 106cells/ml was shown to be optimal . The GFP-tagged ADSCs were identified in the lungs and differentiated into endothelial-like cells . In the second experiment , 96 male SD rats were r and omly divided into three groups ( n = 32/group ) : Ctrl , MCT-induced pulmonary arterial hypertension ( PAH ) , and PAH treated with ADSCs ( ADSCs ) . Two weeks post-MCT administration , the ADSCs group received 1 \\u00d7 106 ADSCs via the external jugular vein . Compared to PAH rats , mean pulmonary arterial pressure was decreased in rats at 1 , 2 , and 3 weeks after ADSCs-treatment ( 18.63 \\u00b1 2.15 mmHg versus 24.53 \\u00b1 2.90 mmHg ; 23.07 \\u00b1 2.84 mmHg versus 33.18 \\u00b1 2.30 mmHg ; 22.98 \\u00b1 2.34 mmHg versus 36.38 \\u00b1 3.28 mmHg , p < 0.05 ) . Meanwhile , the right heart hypertrophy index ( 36.2 1 \\u00b1 4.27 % versus 41.01 \\u00b1 1.29 % ; 39.47 \\u00b1 4.02 % versus 48.75 \\u00b1 2 .13 % ; 41.02 \\u00b1 0.9 % versus 50.52 \\u00b1 1.49 % , p < 0.05 , respectively ) , ratio of wall/lumen thickness , as well as the wall/lumen area were significantly reduced in PAH rats at these time points following ADSCs-treatment , as compared with untreated PAH rats . In summary , ADSCs may colonize the pulmonary arteries , attenuate pulmonary arterial hypertension and ameliorate pulmonary arterial remodeling", "The aim of the present study was to investigate the effect of bone marrow mesenchymal stem cell ( BMSC ) transp1antation on lung and heart damage in a rat model of monocrotaline (MCT)-induced pulmonary arterial hypertension ( PAH ) . The animals were r and omly divided into 3 groups : control , PAH and BMSC implantation groups . Structural changes in the pulmonary vascular wall , such as the pulmonary artery lumen area ( VA ) and vascular area ( TAA ) were measured by hematoxylin and eosin ( H&E ) staining , and the hemodynamics were detected by echocardiography . Two weeks post-operation , our results demonstrated that sublingual vein injection of BMSCs significantly attenuated the pulmonary vascular structural and hemodynamic changes caused by pulmonary arterial hypertension . The mechanism may be executed via paracrine effects", "OBJECTIVE To characterize mortality in persons diagnosed with primary pulmonary hypertension and to investigate factors associated with survival . DESIGN Registry with prospect i ve follow-up . SETTING Thirty-two clinical centers in the United States participating in the Patient Registry for the Characterization of Primary Pulmonary Hypertension supported by the National Heart , Lung , and Blood Institute . PATIENTS Patients ( 194 ) diagnosed at clinical centers between 1 July 1981 and 31 December 1985 and followed through 8 August 1988 . MEASUREMENTS At diagnosis , measurements of hemodynamic variables , pulmonary function , and gas exchange variables were taken in addition to information on demographic variables , medical history , and life-style . Patients were followed for survival at 6-month intervals . MAIN RESULTS The estimated median survival of these patients was 2.8 years ( 95 % Cl , 1.9 to 3.7 years ) . Estimated single-year survival rates were as follows : at 1 year , 68 % ( Cl , 61 % to 75 % ) ; at 3 years , 48 % ( Cl , 41 % to 55 % ) ; and at 5 years , 34 % ( Cl , 24 % to 44 % ) . Variables associated with poor survival included a New York Heart Association ( NYHA ) functional class of III or IV , presence of Raynaud phenomenon , elevated mean right atrial pressure , elevated mean pulmonary artery pressure , decreased cardiac index , and decreased diffusing capacity for carbon monoxide ( DLCO ) . Drug therapy at entry or discharge was not associated with survival duration . CONCLUSIONS Mortality was most closely associated with right ventricular hemodynamic function and can be characterized by means of an equation using three variables : mean pulmonary artery pressure , mean right atrial pressure , and cardiac index . Such an equation , once vali date d prospect ively , could be used as an adjunct in planning treatment strategies and allocating medical re sources", "BACKGROUND Sildenafil inhibits phosphodiesterase type 5 , an enzyme that metabolizes cyclic guanosine monophosphate , thereby enhancing the cyclic guanosine monophosphate-mediated relaxation and growth inhibition of vascular smooth-muscle cells , including those in the lung . METHODS In this double-blind , placebo-controlled study , we r and omly assigned 278 patients with symptomatic pulmonary arterial hypertension ( either idiopathic or associated with connective-tissue disease or with repaired congenital systemic-to-pulmonary shunts ) to placebo or sildenafil ( 20 , 40 , or 80 mg ) orally three times daily for 12 weeks . The primary end point was the change from baseline to week 12 in the distance walked in six minutes . The change in mean pulmonary-artery pressure and World Health Organization ( WHO ) functional class and the incidence of clinical worsening were also assessed , but the study was not powered to assess mortality . Patients completing the 12-week r and omized study could enter a long-term extension study . RESULTS The distance walked in six minutes increased from baseline in all sildenafil groups ; the mean placebo-corrected treatment effects were 45 m ( + 13.0 percent ) , 46 m ( + 13.3 percent ) , and 50 m ( + 14.7 percent ) for 20 , 40 , and 80 mg of sildenafil , respectively ( P<0.001 for all comparisons ) . All sildenafil doses reduced the mean pulmonary-artery pressure ( P=0.04 , P=0.01 , and P<0.001 , respectively ) , improved the WHO functional class ( P=0.003 , P<0.001 , and P<0.001 , respectively ) , and were associated with side effects such as flushing , dyspepsia , and diarrhea . The incidence of clinical worsening did not differ significantly between the patients treated with sildenafil and those treated with placebo . Among the 222 patients completing one year of treatment with sildenafil monotherapy , the improvement from baseline at one year in the distance walked in six minutes was 51 m. CONCLUSIONS Sildenafil improves exercise capacity , WHO functional class , and hemodynamics in patients with symptomatic pulmonary arterial hypertension", "BACKGROUND Current therapies for pulmonary arterial hypertension have been adopted on the basis of short-term trials with exercise capacity as the primary end point . We assessed the efficacy of macitentan , a new dual endothelin-receptor antagonist , using a primary end point of morbidity and mortality in a long-term trial . METHODS We r and omly assigned patients with symptomatic pulmonary arterial hypertension to receive placebo once daily , macitentan at a once-daily dose of 3 mg , or macitentan at a once-daily dose of 10 mg . Stable use of oral or inhaled therapy for pulmonary arterial hypertension , other than endothelin-receptor antagonists , was allowed at study entry . The primary end point was the time from the initiation of treatment to the first occurrence of a composite end point of death , atrial septostomy , lung transplantation , initiation of treatment with intravenous or subcutaneous prostanoids , or worsening of pulmonary arterial hypertension . RESULTS A total of 250 patients were r and omly assigned to placebo , 250 to the 3-mg macitentan dose , and 242 to the 10-mg macitentan dose . The primary end point occurred in 46.4 % , 38.0 % , and 31.4 % of the patients in these groups , respectively . The hazard ratio for the 3-mg macitentan dose as compared with placebo was 0.70 ( 97.5 % confidence interval [ CI ] , 0.52 to 0.96 ; P=0.01 ) , and the hazard ratio for the 10-mg macitentan dose as compared with placebo was 0.55 ( 97.5 % CI , 0.39 to 0.76 ; P<0.001 ) . Worsening of pulmonary arterial hypertension was the most frequent primary end-point event . The effect of macitentan on this end point was observed regardless of whether the patient was receiving therapy for pulmonary arterial hypertension at baseline . Adverse events more frequently associated with macitentan than with placebo were headache , nasopharyngitis , and anemia . CONCLUSIONS Macitentan significantly reduced morbidity and mortality among patients with pulmonary arterial hypertension in this event-driven study . ( Funded by Actelion Pharmaceuticals ; SERAPHIN Clinical Trials.gov number , NCT00660179 . )", "Our previous studies have shown that bone marrow mesenchymal stem cells ( BMSCs ) can inhibit the progression of pulmonary artery hypertension ( PAH ) in the monocrotaline ( MCT ) model in the short term . The aim of this study was to further investigate the long-term effect of BMSCs on PAH and to explore the mechanism of the protective effect including the pulmonary vascular remodeling and cell differentiation . PAH model was established by subcutaneous injection of 50 mg/kg MCT as previously study . Postoperatively , the animals were r and omly divided into three groups ( n = 10 in each group ) : control , PAH group , and BMSCs implantation group . Six months after injection , immunology and immunohistochemistry analysis indicated the MCT-induced intima-media thickness in muscular arteries was reduced ( P < 0.05 ) ; the area of collagen fibers in lung tissue was lower ( P < 0.05 ) , and the proliferating cell nuclear antigen level in pulmonary artery smooth muscle cells was decreased ( P < 0.05 ) . Immunofluorescence showed that the cells have the ability to differentiate between von Willebr and factor and vascular endothelial growth factor . Six months after intravenous injection , BMSCs could significantly improve pulmonary function by inhibiting the ventricular remodeling and the effect of cell differentiation", "Experimental data suggest that transplantation of EPCs attenuates monocrotaline-induced pulmonary hypertension in rats and dogs . In addition , our previous studies suggested that autologous EPC transplantation was feasible , safe , and might have beneficial effects on exercise capacity and pulmonary hemodynamics in adults with IPAH . Thus , we hypothesized that transplantation of EPCs would improve exercise capacity and pulmonary hemodynamics in children with IPAH . Thirteen children with IPAH received intravenous infusion of autologous EPCs . The right-sided heart catheterization and 6-MWD test were performed at baseline and at the time of 12 wk after cell infusion . At the time of 12 wk , mPAP decreased by 6.4 mmHg from 70.3 + /- 19.0 to 63.9 + /- 19.3 mmHg ( p = 0.015 ) . PVR decreased by approximately 19 % from 1118 + /- 537 to 906 + /- 377 dyn s/cm(5 ) ( p = 0.047 ) . CO increased from 3.39 + /- 0.79 to 3.85 + /- 0.42 L/min ( p = 0.048 ) . The 6-MWD increased by 39 m from 359 + /- 82 to 399 + /- 74 m ( p = 0.012 ) . NYHA functional class also improved . There were no severe adverse events with cell infusion . The small pilot study suggested that intravenous infusion of autologous EPCs was feasible , safe , and associated with significant improvements in exercise capacity , NYHA functional class , and pulmonary hemodynamics in children with IPAH . Confirmation of these results in a r and omized controlled trial are essential", "BACKGROUND Uncontrolled studies suggested that aerosolized iloprost , a stable analogue of prostacyclin , causes selective pulmonary vasodilatation and improves hemodynamics and exercise capacity in patients with pulmonary hypertension . METHODS We compared repeated daily inhalations of 2.5 or 5.0 microg of iloprost ( six or nine times per day ; median inhaled dose , 30 microg per day ) with inhalation of placebo . A total of 203 patients with selected forms of severe pulmonary arterial hypertension and chronic thromboembolic pulmonary hypertension ( New York Heart Association [ NYHA ] functional class III or IV ) were included . The primary end point was met if , after week 12 , the NYHA class and distance walked in six minutes were improved by at least one class and at least 10 percent , respectively , in the absence of clinical deterioration according to predefined criteria and death . RESULTS The combined clinical end point was met by 16.8 percent of the patients receiving iloprost , as compared with 4.9 percent of the patients receiving placebo ( P=0.007 ) . There were increases in the distance walked in six minutes of 36.4 m in the iloprost group as a whole ( P=0.004 ) and of 58.8 m in the subgroup of patients with primary pulmonary hypertension . Overall , 4.0 percent of patients in the iloprost group ( including one who died ) and 13.7 percent of those in the placebo group ( including four who died ) did not complete the study ( P=0.024 ) ; the most common reason for withdrawal was clinical deterioration . As compared with base-line values , hemodynamic values were significantly improved at 12 weeks when measured after iloprost inhalation ( P<0.001 ) , were largely unchanged when measured before iloprost inhalation , and were significantly worse in the placebo group . Further significant beneficial effects of iloprost treatment included an improvement in the NYHA class ( P=0.03 ) , dyspnea ( P=0.015 ) , and quality of life ( P=0.026 ) . Syncope occurred with similar frequency in the two groups but was more frequently rated as serious in the iloprost group , although this adverse effect was not associated with clinical deterioration . CONCLUSIONS Inhaled iloprost is an effective therapy for patients with severe pulmonary hypertension", "BACKGROUND High pulmonary vascular resistance ( PVR ) may be a risk factor for early and late mortality in both Glen shunt and Fontan operation patients . Furthermore , PVR may increase long after the Fontan operation . Whether pulmonary vasodilators such as phosphodiesterase 5 inhibitors can decrease PVR in patients with single ventricular physiology remains undetermined . METHODS AND RESULTS This was a prospect i ve , multicenter study . Patients with single ventricular physiology who have a PVR index higher than 2.5 Wood units \\u00b7 \\u33a1 ( WU ) were enrolled . Cardiac catheterization was performed before and after administration of sildenafil in all patients . After the Fontan operation , a six minute walk test ( 6MWT ) was also performed . A total of 42 patients were enrolled . PVR was significantly decreased in each stage of single ventricular physiology after sildenafil administration : from 4.3\\u00b11.5WU to 2.1\\u00b10.6WU ( p<0.01 ) in patients before a Glenn shunt , from 3.2\\u00b10.5WU to 1.6\\u00b10.6WU ( p<0.001 ) in patients after a Glenn shunt , and from 3.9\\u00b11.7WU to 2.3\\u00b10.8WU ( p<0.001 ) in patients after Fontan . In patients after Fontan , the 6MWT increased from 416\\u00b174 m to 485\\u00b172 m ( p<0.01 ) , and NYHA functional class improved significantly ( p<0.05 ) after sildenafil administration . No major side effects were observed in any patients . CONCLUSIONS Sildenafil reduced PVR in patients with single ventricle physiology . Sildenafil increased exercise capacity and improved NYHA functional class in patients after a Fontan operation . This implies that pulmonary vasodilation is a potential therapeutic target in selected patients with elevated PVR with single ventricle physiology . Long-term clinical significance warrants further study", "OBJECTIVES The purpose of this study was to examine the efficacy and safety of four doses of ambrisentan , an oral endothelin type A receptor-selective antagonist , in patients with pulmonary arterial hypertension ( PAH ) . BACKGROUND Pulmonary arterial hypertension is a life-threatening and progressive disease with limited treatment options . Endothelin is a vasoconstrictor and smooth muscle cell mitogen that plays a critical role in the pathogenesis and progression of PAH . METHODS In this double-blind , dose-ranging study , 64 patients with idiopathic PAH or PAH associated with collagen vascular disease , anorexigen use , or human immunodeficiency virus infection were r and omized to receive 1 , 2.5 , 5 , or 10 mg of ambrisentan once daily for 12 weeks followed by 12 weeks of open-label ambrisentan . The primary end point was an improvement from baseline in 6-min walk distance ( 6MWD ) ; secondary end points included Borg dyspnea index , World Health Organization ( WHO ) functional class , a subject global assessment , and cardiopulmonary hemodynamics . RESULTS At 12 weeks , ambrisentan increased 6MWD ( + 36.1 m , p < 0.0001 ) with similar and statistically significant increases for each dose group ( range , + 33.9 to + 38.1 m ) . Improvements were also observed in Borg dyspnea index , WHO functional class , subject global assessment , mean pulmonary arterial pressure ( -5.2 mm Hg , p < 0.0001 ) , and cardiac index ( + 0.33 l/min/m2 , p < 0.0008 ) . Adverse events were mild and unrelated to dose , including the incidence of elevated serum aminotransferase concentrations > 3 times the upper limit of normal ( 3.1 % ) . CONCLUSIONS Ambrisentan appears to improve exercise capacity , symptoms , and hemodynamics in patients with PAH . The incidence and severity of liver enzyme abnormalities appear to be low", "UNLABELLED Pulmonary arterial hypertension ( PAH ) is characterized by functional and structural changes in the pulmonary vasculature , and despite the drug treatment that made significant progress , the prognosis of patients with advanced PH remains extremely poor . In the present study , we investigated the early effect of bone marrow mesenchymal stem cells ( BMSCs ) on experimental high blood flow-induced PAH model rats and discussed the mechanism . BMSCs were isolated , cultured from bone marrow of Sprague-Dawley ( SD ) rat . The animal model of PAH was created by surgical methods to produce a left-to-right shunt . Following the successful establishment of the PAH model , rats were r and omly assigned to three groups ( n=20 in each group ) : sham group ( control ) , PAH group , and BMSC group ( received a sublingual vein injection of 1 - 5 \\u00d7 10(6 ) BMSCs ) . Two weeks after the administration , BMSCs significantly reduced the vascular remodeling , improved the hemodynamic data , and deceased the right ventricle weight ratio to left ventricular plus septal weight ( RV/LV+S ) ( P<0.05 ) . Real-time reverse transcription-polymerase chain reaction ( RT-PCR ) and immunohistochemistry analysis results indicated that the inflammation factors such as interleukin-1\\u03b2 ( IL-1\\u03b2 ) , IL-6 , and tumor necrosis factor-\\u03b1 ( TNF-\\u03b1 ) were reduced ( P<0.05 ) ; the expression of matrix metallo proteinase-9 ( MMP-9 ) was lower ( P<0.05 ) ; vascular endothelial growth factor ( VEGF ) was higher in BMSC group than those in PAH group ( P<0.05 ) . CONCLUSION Sublingual vein injection of BMSCs for 2 weeks , significantly improved the lung and heart injury caused by left-to-right shunt-induced PAH ; decreased pulmonary vascular remodeling and inflammation ; and enhanced angiogenesis", "Pulmonary arterial hypertension is a life-threatening disease for which continuous intravenous prostacyclin has proven to be effective . However , this treatment requires a permanent central venous catheter with the associated risk of serious complications such as sepsis , thromboembolism , or syncope . Treprostinil , a stable prostacyclin analogue , can be administered by a continuous subcutaneous infusion , avoiding these risks . We conducted a 12-week , double-blind , placebo-controlled multicenter trial in 470 patients with pulmonary arterial hypertension , either primary or associated with connective tissue disease or congenital systemic-to-pulmonary shunts . Exercise capacity improved with treprostinil and was unchanged with placebo ; the between treatment group difference in median six-minute walking distance was 16 m ( p = 0.006 ) . Improvement in exercise capacity was greater in the sicker patients and was dose-related , but independent of disease etiology . Concomitantly , treprostinil significantly improved indices of dyspnea , signs and symptoms of pulmonary hypertension , and hemodynamics . The most common side effect attributed to treprostinil was infusion site pain ( 85 % ) leading to premature discontinuation from the study in 8 % of patients . Three patients in the treprostinil treatment group presented with an episode of gastrointestinal hemorrhage . We conclude that chronic subcutaneous infusion of treprostinil is an effective treatment with an acceptable safety profile in patients with pulmonary arterial hypertension", "BACKGROUND Endothelin 1 , a powerful endogenous vasoconstrictor and mitogen , might be a cause of pulmonary hypertension . We describe the efficacy and safety of bosentan , a dual endothelin-receptor antagonist that can be taken orally , in patients with severe pulmonary hypertension . METHODS In this double-blind , placebo-controlled study , 32 patients with pulmonary hypertension ( primary or associated with scleroderma ) were r and omly assigned to bosentan ( 62.5 mg taken twice daily for 4 weeks then 125 mg twice daily ) or placebo for a minimum of 12 weeks . The primary endpoint was change in exercise capacity . Secondary endpoints included changes in cardiopulmonary haemodynamics , Borg dyspnoea index , WHO functional class , and withdrawal due to clinical worsening . Analysis was by intention to treat . FINDINGS In patients given bosentan , the distance walked in 6 min improved by 70 m at 12 weeks compared with baseline , whereas it worsened by 6 m in those on placebo ( difference 76 m [ 95 % CI 12 - 139 ] , p=0.021 ) . The improvement was maintained for at least 20 weeks . The cardiac index was 1.0 L min(-1 ) m(-2 ) ( 95 % CI 0.6 - 1.4 , p<0.0001 ) greater in patients given bosentan than in those given placebo . Pulmonary vascular resistance decreased by 223 dyn s cm(-)(5 ) with bosentan , but increased by 191 dyn s cm(-5 ) with placebo ( difference -415 [ -608 to -221 ] , p=0.0002 ) . Patients given bosentan had a reduced Borg dyspnoea index and an improved WHO functional class . All three withdrawals from clinical worsening were in the placebo group ( p=0.033 ) . The number and nature of adverse events did not differ between the two groups . INTERPRETATION Bosentan increases exercise capacity and improves haemodynamics in patients with pulmonary hypertension , suggesting that endothelin has an important role in pulmonary hypertension", "Background Systematic Review s ( SRs ) of experimental animal studies are not yet common practice , but awareness of the merits of conducting such SRs is steadily increasing . As animal intervention studies differ from r and omized clinical trials ( RCT ) in many aspects , the methodology for SRs of clinical trials needs to be adapted and optimized for animal intervention studies . The Cochrane Collaboration developed a Risk of Bias ( RoB ) tool to establish consistency and avoid discrepancies in assessing the method ological quality of RCTs . A similar initiative is warranted in the field of animal experimentation . Methods We provide an RoB tool for animal intervention studies ( SYRCLE \\u2019s RoB tool ) . This tool is based on the Cochrane RoB tool and has been adjusted for aspects of bias that play a specific role in animal intervention studies . To enhance transparency and applicability , we formulated signalling questions to facilitate judgment . Results The result ing RoB tool for animal studies contains 10 entries . These entries are related to selection bias , performance bias , detection bias , attrition bias , reporting bias and other biases . Half these items are in agreement with the items in the Cochrane RoB tool . Most of the variations between the two tools are due to differences in design between RCTs and animal studies . Shortcomings in , or unfamiliarity with , specific aspects of experimental design of animal studies compared to clinical studies also play a role . Conclusions SYRCLE \\u2019s RoB tool is an adapted version of the Cochrane RoB tool . Widespread adoption and implementation of this tool will facilitate and improve critical appraisal of evidence from animal studies . This may subsequently enhance the efficiency of translating animal research into clinical practice and increase awareness of the necessity of improving the method ological quality of animal studies" ], "target": "Conclusions SC therapy is effective for PAH in pre clinical studies .\\nThese results may help to st and ardise pre clinical animal studies and provide a theoretical basis for clinical trial design in the future .", "background": "Background Despite significant progress in drug treatment , the prognosis of patients with advanced pulmonary arterial hypertension ( PAH ) remains extremely poor .\\nMany pre clinical studies have reported the efficacy of stem cell ( SC ) therapy for PAH ; however , this approach remains controversial .\\nThe aim of this systematic review and meta- analysis is to assess the potential efficacy of SC therapy for PAH .", "reviews_info": "Background Despite significant progress in drug treatment , the prognosis of patients with advanced pulmonary arterial hypertension ( PAH ) remains extremely poor .\\nMany pre clinical studies have reported the efficacy of stem cell ( SC ) therapy for PAH ; however , this approach remains controversial .\\nThe aim of this systematic review and meta- analysis is to assess the potential efficacy of SC therapy for PAH ." } ``` __Cochrane__ ```json { "review_id": "CD007697", "pmid": [ "16394043" ], "title": [ "Aggressive surgical effort and improved survival in advanced-stage ovarian cancer." ], "abstract": [ "Residual disease after initial surgery for ovarian cancer is the strongest prognostic factor for survival. However, the extent of surgical resection required to achieve optimal cytoreduction is controversial. Our goal was to estimate the effect of aggressive surgical resection on ovarian cancer patient survival.\\n A retrospective cohort study of consecutive patients with International Federation of Gynecology and Obstetrics stage IIIC ovarian cancer undergoing primary surgery was conducted between January 1, 1994, and December 31, 1998. The main outcome measures were residual disease after cytoreduction, frequency of radical surgical resection, and 5-year disease-specific survival.\\n The study comprised 194 patients, including 144 with carcinomatosis. The mean patient age and follow-up time were 64.4 and 3.5 years, respectively. After surgery, 131 (67.5%) of the 194 patients had less than 1 cm of residual disease (definition of optimal cytoreduction). Considering all patients, residual disease was the only independent predictor of survival; the need to perform radical procedures to achieve optimal cytoreduction was not associated with a decrease in survival. For the subgroup of patients with carcinomatosis, residual disease and the performance of radical surgical procedures were the only independent predictors. Disease-specific survival was markedly improved for patients with carcinomatosis operated on by surgeons who most frequently used radical procedures compared with those least likely to use radical procedures (44% versus 17%, P < .001).\\n Overall, residual disease was the only independent predictor of survival. Minimizing residual disease through aggressive surgical resection was beneficial, especially in patients with carcinomatosis.\\n II-2." ], "target": "We found only low quality evidence comparing ultra-radical and standard surgery in women with advanced ovarian cancer and carcinomatosis. The evidence suggested that ultra-radical surgery may result in better survival.\\u00a0 It was unclear whether there were any differences in progression-free survival, QoL and morbidity between the two groups. The cost-effectiveness of this intervention has not been investigated. We are, therefore, unable to reach definite conclusions about the relative benefits and adverse effects of the two types of surgery.\\nIn order to determine the role of ultra-radical surgery in the management of advanced stage ovarian cancer, a sufficiently powered randomised controlled trial comparing ultra-radical and standard surgery or well-designed non-randomised studies would be required." } ``` ### Data Fields __MS^2__ - `"review_id"`: The PubMed ID of the review. - `"pmid"`: The PubMed IDs of the included studies. - `"title"`: The titles of the included studies. - `"abstract"`: The abstracts of the included studies. - `"target"`: The conclusions, taken from the abstract of the review, that serve as the summarization target. - `"background"`: A description of the reviews objective. __Cochrane__ - `"review_id"`: The PubMed ID of the review. - `"pmid"`: The PubMed IDs of the included studies. - `"title"`: The titles of the included studies. - `"abstract"`: The abstracts of the included studies. - `"target"`: The conclusions, taken from the abstract of the review, that serve as the summarization target. ### Data Splits Each dataset is split into training, validation and test partitions __MS^2__ | train | validation | test | |------:|-----------:|-----:| | 14188 | 2021 | 1667 | __Cochrane__ | train | validation | test | |------:|-----------:|-----:| | 3752 | 470 | 470 | ## Dataset Creation Please refer to the following papers for details about dataset curation: [MSˆ2: A Dataset for Multi-Document Summarization of Medical Studies](https://aclanthology.org/2021.emnlp-main.594.pdf) [Generating (Factual?) Narrative Summaries of RCTs: Experiments with Neural Multi-Document Summarization](https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8378607/) ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information Licensing information can be found [here](https://github.com/allenai/mslr-shared-task/blob/main/LICENSE). ### Citation Information **DeYoung, Jay, Iz Beltagy, Madeleine van Zuylen, Bailey Kuehl and Lucy Lu Wang. "MS2: A Dataset for Multi-Document Summarization of Medical Studies." EMNLP (2021).** ```bibtex @inproceedings{DeYoung2021MS2MS, title={MSˆ2: Multi-Document Summarization of Medical Studies}, author={Jay DeYoung and Iz Beltagy and Madeleine van Zuylen and Bailey Kuehl and Lucy Lu Wang}, booktitle={EMNLP}, year={2021} } ``` **Byron C. Wallace, Sayantani Saha, Frank Soboczenski, and Iain James Marshall. (2020). "Generating (factual?) narrative summaries of RCTs: Experiments with neural multi-document summarization." AMIA Annual Symposium.** ```bibtex @article{Wallace2020GeneratingN, title={Generating (Factual?) Narrative Summaries of RCTs: Experiments with Neural Multi-Document Summarization}, author={Byron C. Wallace and Sayantani Saha and Frank Soboczenski and Iain James Marshall}, journal={AMIA Annual Symposium}, year={2020}, volume={abs/2008.11293} } ```
adv_glue
2023-06-01T14:57:45.000Z
[ "task_categories:text-classification", "task_ids:natural-language-inference", "task_ids:sentiment-classification", "annotations_creators:other", "language_creators:machine-generated", "multilinguality:monolingual", "size_categories:n<1K", "source_datasets:extended|glue", "language:en", "license:cc-by-sa-4.0", "paraphrase-identification", "qa-nli", "arxiv:2111.02840", "region:us" ]
null
Adversarial GLUE Benchmark (AdvGLUE) is a comprehensive robustness evaluation benchmark that focuses on the adversarial robustness evaluation of language models. It covers five natural language understanding tasks from the famous GLUE tasks and is an adversarial version of GLUE benchmark.
@article{Wang2021AdversarialGA, title={Adversarial GLUE: A Multi-Task Benchmark for Robustness Evaluation of Language Models}, author={Boxin Wang and Chejian Xu and Shuohang Wang and Zhe Gan and Yu Cheng and Jianfeng Gao and Ahmed Hassan Awadallah and B. Li}, journal={ArXiv}, year={2021}, volume={abs/2111.02840} }
null
4
1,214
--- annotations_creators: - other language_creators: - machine-generated language: - en license: - cc-by-sa-4.0 multilinguality: - monolingual size_categories: - n<1K source_datasets: - extended|glue task_categories: - text-classification task_ids: - natural-language-inference - sentiment-classification pretty_name: Adversarial GLUE tags: - paraphrase-identification - qa-nli dataset_info: - config_name: adv_sst2 features: - name: sentence dtype: string - name: label dtype: class_label: names: '0': negative '1': positive - name: idx dtype: int32 splits: - name: validation num_bytes: 16595 num_examples: 148 download_size: 40662 dataset_size: 16595 - config_name: adv_qqp features: - name: question1 dtype: string - name: question2 dtype: string - name: label dtype: class_label: names: '0': not_duplicate '1': duplicate - name: idx dtype: int32 splits: - name: validation num_bytes: 9926 num_examples: 78 download_size: 40662 dataset_size: 9926 - config_name: adv_mnli features: - name: premise dtype: string - name: hypothesis dtype: string - name: label dtype: class_label: names: '0': entailment '1': neutral '2': contradiction - name: idx dtype: int32 splits: - name: validation num_bytes: 23736 num_examples: 121 download_size: 40662 dataset_size: 23736 - config_name: adv_mnli_mismatched features: - name: premise dtype: string - name: hypothesis dtype: string - name: label dtype: class_label: names: '0': entailment '1': neutral '2': contradiction - name: idx dtype: int32 splits: - name: validation num_bytes: 40982 num_examples: 162 download_size: 40662 dataset_size: 40982 - config_name: adv_qnli features: - name: question dtype: string - name: sentence dtype: string - name: label dtype: class_label: names: '0': entailment '1': not_entailment - name: idx dtype: int32 splits: - name: validation num_bytes: 34877 num_examples: 148 download_size: 40662 dataset_size: 34877 - config_name: adv_rte features: - name: sentence1 dtype: string - name: sentence2 dtype: string - name: label dtype: class_label: names: '0': entailment '1': not_entailment - name: idx dtype: int32 splits: - name: validation num_bytes: 25998 num_examples: 81 download_size: 40662 dataset_size: 25998 config_names: - adv_mnli - adv_mnli_mismatched - adv_qnli - adv_qqp - adv_rte - adv_sst2 --- # Dataset Card for Adversarial GLUE ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://adversarialglue.github.io/ - **Repository:** - **Paper:** [arXiv](https://arxiv.org/pdf/2111.02840.pdf) - **Leaderboard:** - **Point of Contact:** - **Size of downloaded dataset files:** 202.75 kB ### Dataset Summary Adversarial GLUE Benchmark (AdvGLUE) is a comprehensive robustness evaluation benchmark that focuses on the adversarial robustness evaluation of language models. It covers five natural language understanding tasks from the famous GLUE tasks and is an adversarial version of GLUE benchmark. AdvGLUE considers textual adversarial attacks from different perspectives and hierarchies, including word-level transformations, sentence-level manipulations, and human-written adversarial examples, which provide comprehensive coverage of various adversarial linguistic phenomena. ### Supported Tasks and Leaderboards Leaderboard available on the homepage: [https://adversarialglue.github.io/](https://adversarialglue.github.io/). ### Languages AdvGLUE deviates from the GLUE dataset, which has a base language of English. ## Dataset Structure ### Data Instances #### default - **Size of downloaded dataset files:** 202.75 kB - **Example**: ```python >>> datasets.load_dataset('adv_glue', 'adv_sst2')['validation'][0] {'sentence': "it 's an uneven treat that bores fun at the democratic exercise while also examining its significance for those who take part .", 'label': 1, 'idx': 0} ``` ### Data Fields The data fields are the same as in the GLUE dataset, which differ by task. The data fields are the same among all splits. #### adv_mnli - `premise`: a `string` feature. - `hypothesis`: a `string` feature. - `label`: a classification label, with possible values including `entailment` (0), `neutral` (1), `contradiction` (2). - `idx`: a `int32` feature. #### adv_mnli_matched - `premise`: a `string` feature. - `hypothesis`: a `string` feature. - `label`: a classification label, with possible values including `entailment` (0), `neutral` (1), `contradiction` (2). - `idx`: a `int32` feature. #### adv_mnli_mismatched - `premise`: a `string` feature. - `hypothesis`: a `string` feature. - `label`: a classification label, with possible values including `entailment` (0), `neutral` (1), `contradiction` (2). - `idx`: a `int32` feature. #### adv_qnli [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### adv_qqp [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### adv_rte [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### adv_sst2 [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Data Splits Adversarial GLUE provides only a 'dev' split. ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information The dataset is distributed under the [CC BY-SA 4.0](http://creativecommons.org/licenses/by-sa/4.0/legalcode) license. ### Citation Information ```bibtex @article{Wang2021AdversarialGA, title={Adversarial GLUE: A Multi-Task Benchmark for Robustness Evaluation of Language Models}, author={Boxin Wang and Chejian Xu and Shuohang Wang and Zhe Gan and Yu Cheng and Jianfeng Gao and Ahmed Hassan Awadallah and B. Li}, journal={ArXiv}, year={2021}, volume={abs/2111.02840} } ``` ### Contributions Thanks to [@jxmorris12](https://github.com/jxmorris12) for adding this dataset.
Babelscape/wikineural
2022-11-13T07:52:46.000Z
[ "task_categories:token-classification", "task_ids:named-entity-recognition", "annotations_creators:machine-generated", "language_creators:machine-generated", "multilinguality:multilingual", "source_datasets:original", "language:de", "language:en", "language:es", "language:fr", "language:it", "language:nl", "language:pl", "language:pt", "language:ru", "license:cc-by-nc-sa-4.0", "structure-prediction", "arxiv:1810.04805", "region:us" ]
Babelscape
null
null
null
14
1,207
--- annotations_creators: - machine-generated language_creators: - machine-generated language: - de - en - es - fr - it - nl - pl - pt - ru license: - cc-by-nc-sa-4.0 multilinguality: - multilingual source_datasets: - original task_categories: - token-classification task_ids: - named-entity-recognition pretty_name: wikineural-dataset tags: - structure-prediction --- ## Table of Contents - [Description](#description) - [Dataset Structure](#dataset-structure) - [Additional Information](#additional-information) ## Dataset Card for WikiNEuRal dataset ## Dataset Description - **Summary:** Training data for NER in 9 languages. - **Repository:** [https://github.com/Babelscape/wikineural](https://github.com/Babelscape/wikineural) - **Paper:** [https://aclanthology.org/wikineural](https://aclanthology.org/2021.findings-emnlp.215/) - **Point of Contact:** [tedeschi@babelscape.com](tedeschi@babelscape.com) ## Description - **Summary:** In a nutshell, WikiNEuRal consists in a novel technique which builds upon a multilingual lexical knowledge base (i.e., [BabelNet](https://babelnet.org/)) and transformer-based architectures (i.e., [BERT](https://arxiv.org/abs/1810.04805)) to produce high-quality annotations for multilingual NER. It shows consistent improvements of up to 6 span-based F1-score points against state-of-the-art alternative data production methods on common benchmarks for NER. We used this methodology to automatically generate training data for NER in 9 languages. - **Repository:** [https://github.com/Babelscape/wikineural](https://github.com/Babelscape/wikineural) - **Paper:** [https://aclanthology.org/wikineural](https://aclanthology.org/2021.findings-emnlp.215/) - **Point of Contact:** [tedeschi@babelscape.com](tedeschi@babelscape.com) ## Dataset Structure The data fields are the same among all splits. - `tokens`: a `list` of `string` features. - `ner_tags`: a `list` of classification labels (`int`). Full tagset with indices: ```python {'O': 0, 'B-PER': 1, 'I-PER': 2, 'B-ORG': 3, 'I-ORG': 4, 'B-LOC': 5, 'I-LOC': 6, 'B-MISC': 7, 'I-MISC': 8} ``` - `lang`: a `string` feature. Full list of language: Dutch (nl), English (en), French (fr), German (de), Italian (it), Polish (pl), Portugues (pt), Russian (ru), Spanish (es). ## Dataset Statistics The table below shows the number of sentences, number of tokens and number of instances per class, for each of the 9 languages. | Dataset Version | Sentences | Tokens | PER | ORG | LOC | MISC | OTHER | | :------------- | -------------: | -------------: | -------------: | -------------: | -------------: | -------------: | -------------: | | WikiNEuRal EN | 116k | 2.73M | 51k | 31k | 67k | 45k | 2.40M | | WikiNEuRal ES | 95k | 2.33M | 43k | 17k | 68k | 25k | 2.04M | | WikiNEuRal NL | 107k | 1.91M | 46k | 22k | 61k | 24k | 1.64M | | WikiNEuRal DE | 124k | 2.19M | 60k | 32k | 59k | 25k | 1.87M | | WikiNEuRal RU | 123k | 2.39M | 40k | 26k | 89k | 25k | 2.13M | | WikiNEuRal IT | 111k | 2.99M | 67k | 22k | 97k | 26k | 2.62M | | WikiNEuRal FR | 127k | 3.24M | 76k | 25k | 101k | 29k | 2.83M | | WikiNEuRal PL | 141k | 2.29M | 59k | 34k | 118k | 22k | 1.91M | | WikiNEuRal PT | 106k | 2.53M | 44k | 17k | 112k | 25k | 2.20M | ## Additional Information - **Licensing Information**: Contents of this repository are restricted to only non-commercial research purposes under the [Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License (CC BY-NC-SA 4.0)](https://creativecommons.org/licenses/by-nc-sa/4.0/). Copyright of the dataset contents belongs to the original copyright holders. - **Citation Information**: Please consider citing our work if you use data and/or code from this repository. ```bibtex @inproceedings{tedeschi-etal-2021-wikineural-combined, title = "{W}iki{NE}u{R}al: {C}ombined Neural and Knowledge-based Silver Data Creation for Multilingual {NER}", author = "Tedeschi, Simone and Maiorca, Valentino and Campolungo, Niccol{\`o} and Cecconi, Francesco and Navigli, Roberto", booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2021", month = nov, year = "2021", address = "Punta Cana, Dominican Republic", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2021.findings-emnlp.215", pages = "2521--2533", abstract = "Multilingual Named Entity Recognition (NER) is a key intermediate task which is needed in many areas of NLP. In this paper, we address the well-known issue of data scarcity in NER, especially relevant when moving to a multilingual scenario, and go beyond current approaches to the creation of multilingual silver data for the task. We exploit the texts of Wikipedia and introduce a new methodology based on the effective combination of knowledge-based approaches and neural models, together with a novel domain adaptation technique, to produce high-quality training corpora for NER. We evaluate our datasets extensively on standard benchmarks for NER, yielding substantial improvements up to 6 span-based F1-score points over previous state-of-the-art systems for data creation.", } ``` - **Contributions**: Thanks to [@sted97](https://github.com/sted97) for adding this dataset.
zxvix/pubmed_subset_new
2023-08-23T09:04:37.000Z
[ "region:us" ]
zxvix
null
null
null
0
1,202
--- dataset_info: features: - name: MedlineCitation struct: - name: PMID dtype: int32 - name: DateCompleted struct: - name: Year dtype: int32 - name: Month dtype: int32 - name: Day dtype: int32 - name: NumberOfReferences dtype: int32 - name: DateRevised struct: - name: Year dtype: int32 - name: Month dtype: int32 - name: Day dtype: int32 - name: Article struct: - name: Abstract struct: - name: AbstractText dtype: string - name: ArticleTitle dtype: string - name: AuthorList struct: - name: Author sequence: - name: LastName dtype: string - name: ForeName dtype: string - name: Initials dtype: string - name: CollectiveName dtype: string - name: Language dtype: string - name: GrantList struct: - name: Grant sequence: - name: GrantID dtype: string - name: Agency dtype: string - name: Country dtype: string - name: PublicationTypeList struct: - name: PublicationType sequence: string - name: MedlineJournalInfo struct: - name: Country dtype: string - name: ChemicalList struct: - name: Chemical sequence: - name: RegistryNumber dtype: string - name: NameOfSubstance dtype: string - name: CitationSubset dtype: string - name: MeshHeadingList struct: - name: MeshHeading sequence: - name: DescriptorName dtype: string - name: QualifierName dtype: string - name: PubmedData struct: - name: ArticleIdList sequence: - name: ArticleId sequence: string - name: PublicationStatus dtype: string - name: History struct: - name: PubMedPubDate sequence: - name: Year dtype: int32 - name: Month dtype: int32 - name: Day dtype: int32 - name: ReferenceList sequence: - name: Citation dtype: string - name: CitationId dtype: int32 - name: text dtype: string - name: title dtype: string splits: - name: train num_bytes: 3033204166.457245 num_examples: 1000000 - name: test num_bytes: 3033204.166457245 num_examples: 1000 download_size: 1638343655 dataset_size: 3036237370.623702 --- # Dataset Card for "pubmed_subset_new" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
mteb/banking77
2022-09-27T19:15:02.000Z
[ "language:en", "region:us" ]
mteb
null
null
null
0
1,200
--- language: - en ---
JeanKaddour/minipile
2023-06-20T10:08:26.000Z
[ "task_categories:text-generation", "task_categories:fill-mask", "task_ids:language-modeling", "task_ids:masked-language-modeling", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:monolingual", "size_categories:1M<n<10M", "source_datasets:original", "language:en", "license:other", "arxiv:2304.08442", "arxiv:2201.07311", "region:us" ]
JeanKaddour
null
null
null
34
1,199
--- dataset_info: features: - name: text dtype: string splits: - name: train num_bytes: 5906108510 num_examples: 1000000 - name: validation num_bytes: 2779386 num_examples: 500 - name: test num_bytes: 58558191 num_examples: 10000 download_size: 3177432813 dataset_size: 5967446087 annotations_creators: - no-annotation language_creators: - found language: - en license: other multilinguality: - monolingual pretty_name: MiniPile size_categories: - 1M<n<10M source_datasets: - original task_categories: - text-generation - fill-mask task_ids: - language-modeling - masked-language-modeling paperswithcode_id: minipile --- # Dataset Card for MiniPile ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Languages](#languages) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description [The MiniPile Challenge for Data-Efficient Language Models](https://arxiv.org/abs/2304.08442) ### Dataset Summary MiniPile is a 6GB subset of the [deduplicated The Pile corpus](https://huggingface.co/datasets/EleutherAI/the_pile_deduplicated). To curate MiniPile, we perform a simple, three-step data filtering process: we (1) infer embeddings for all documents of the Pile, (2) cluster the embedding space using k-means, and (3) filter out low-quality clusters. The primary motivation for curating MiniPile is that (i) diverse pre-training datasets (like the Pile) are often too large for academic budgets and (ii) most smaller-scale datasets are fairly homogeneous and thereby unrepresentative of contemporary general-purpose language models. MiniPile aims to fill this gap and thereby facilitate data-efficient research on model architectures, training procedures, optimizers, etc. More details on the MiniPile curation procedure and some pre-training results be found in the [MiniPile paper](https://arxiv.org/abs/2304.08442). For more details on the Pile corpus, we refer the reader to [the Pile datasheet](https://arxiv.org/abs/2201.07311). ### Languages English (`EN`) ## Additional Information ### Dataset Curators MiniPile is a subset of the Pile, curated by Jean Kaddour. The Pile was created by Leo Gao, Stella Biderman, Sid Black, Laurence Golding, Travis Hoppe, Charles Foster, Jason Phang, Horace He, Anish Thite, Noa Nabeshima, Shawn Presser, Connor Leahy. ### Licensing Information Since MiniPile is a subset of the Pile, the same MIT License holds. ### Citation Information ``` @article{kaddour2023minipile, title={The MiniPile Challenge for Data-Efficient Language Models}, author={Kaddour, Jean}, journal={arXiv preprint arXiv:2304.08442}, year={2023} } @article{gao2020pile, title={The {P}ile: An 800{GB} dataset of diverse text for language modeling}, author={Gao, Leo and Biderman, Stella and Black, Sid and Golding, Laurence and Hoppe, Travis and Foster, Charles and Phang, Jason and He, Horace and Thite, Anish and Nabeshima, Noa and others}, journal={arXiv preprint arXiv:2101.00027}, year={2020} } ```
CShorten/ML-ArXiv-Papers
2022-06-27T12:15:11.000Z
[ "license:afl-3.0", "region:us" ]
CShorten
null
null
null
13
1,195
--- license: afl-3.0 --- This dataset contains the subset of ArXiv papers with the "cs.LG" tag to indicate the paper is about Machine Learning. The core dataset is filtered from the full ArXiv dataset hosted on Kaggle: https://www.kaggle.com/datasets/Cornell-University/arxiv. The original dataset contains roughly 2 million papers. This dataset contains roughly 100,000 papers following the category filtering. The dataset is maintained by with requests to the ArXiv API. The current iteration of the dataset only contains the title and abstract of the paper. The ArXiv dataset contains additional features that we may look to include in future releases. We have highlighted the top two features on the roadmap for integration: <ul> <li> <b>authors</b> </li> <li> <b>update_date</b> </li> <li> Submitter </li> <li> Comments </li> <li> Journal-ref </li> <li> doi </li> <li> report-no </li> <li> categories </li> <li> license </li> <li> versions </li> <li> authors_parsed </li> </ul>
gbharti/finance-alpaca
2023-09-26T04:13:35.000Z
[ "language:en", "region:us" ]
gbharti
null
null
null
44
1,194
--- language: - en --- This dataset is a combination of Stanford's Alpaca (https://github.com/tatsu-lab/stanford_alpaca) and FiQA (https://sites.google.com/view/fiqa/) with another 1.3k pairs custom generated using GPT3.5 Script for tuning through Kaggle's (https://www.kaggle.com) free resources using PEFT/LoRa: https://www.kaggle.com/code/gbhacker23/wealth-alpaca-lora GitHub repo with performance analyses, training and data generation scripts, and inference notebooks: https://github.com/gaurangbharti1/wealth-alpaca Cleaner dataset: https://huggingface.co/datasets/gbharti/wealth-alpaca_lora (no major changes, just cleaned up) CSV format: https://huggingface.co/datasets/gbharti/finance-alpaca-csv
spider
2022-11-03T16:31:49.000Z
[ "task_categories:text2text-generation", "annotations_creators:expert-generated", "language_creators:expert-generated", "language_creators:machine-generated", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:cc-by-4.0", "text-to-sql", "region:us" ]
null
Spider is a large-scale complex and cross-domain semantic parsing and text-toSQL dataset annotated by 11 college students
@article{yu2018spider, title={Spider: A large-scale human-labeled dataset for complex and cross-domain semantic parsing and text-to-sql task}, author={Yu, Tao and Zhang, Rui and Yang, Kai and Yasunaga, Michihiro and Wang, Dongxu and Li, Zifan and Ma, James and Li, Irene and Yao, Qingning and Roman, Shanelle and others}, journal={arXiv preprint arXiv:1809.08887}, year={2018} }
null
55
1,190
--- annotations_creators: - expert-generated language_creators: - expert-generated - machine-generated language: - en license: - cc-by-4.0 multilinguality: - monolingual size_categories: - 1K<n<10K source_datasets: - original task_categories: - text2text-generation task_ids: [] paperswithcode_id: spider-1 pretty_name: Spider tags: - text-to-sql dataset_info: features: - name: db_id dtype: string - name: query dtype: string - name: question dtype: string - name: query_toks sequence: string - name: query_toks_no_value sequence: string - name: question_toks sequence: string config_name: spider splits: - name: train num_bytes: 4743786 num_examples: 7000 - name: validation num_bytes: 682090 num_examples: 1034 download_size: 99736136 dataset_size: 5425876 --- # Dataset Card for Spider ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://yale-lily.github.io/spider - **Repository:** https://github.com/taoyds/spider - **Paper:** https://www.aclweb.org/anthology/D18-1425/ - **Point of Contact:** [Yale LILY](https://yale-lily.github.io/) ### Dataset Summary Spider is a large-scale complex and cross-domain semantic parsing and text-to-SQL dataset annotated by 11 Yale students The goal of the Spider challenge is to develop natural language interfaces to cross-domain databases ### Supported Tasks and Leaderboards The leaderboard can be seen at https://yale-lily.github.io/spider ### Languages The text in the dataset is in English. ## Dataset Structure ### Data Instances **What do the instances that comprise the dataset represent?** Each instance is natural language question and the equivalent SQL query **How many instances are there in total?** **What data does each instance consist of?** [More Information Needed] ### Data Fields * **db_id**: Database name * **question**: Natural language to interpret into SQL * **query**: Target SQL query * **query_toks**: List of tokens for the query * **query_toks_no_value**: List of tokens for the query * **question_toks**: List of tokens for the question ### Data Splits **train**: 7000 questions and SQL query pairs **dev**: 1034 question and SQL query pairs [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? [More Information Needed] ### Annotations The dataset was annotated by 11 college students at Yale University #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases [More Information Needed] ### Other Known Limitations ## Additional Information The listed authors in the homepage are maintaining/supporting the dataset. ### Dataset Curators [More Information Needed] ### Licensing Information The spider dataset is licensed under the [CC BY-SA 4.0](https://creativecommons.org/licenses/by-sa/4.0/legalcode) [More Information Needed] ### Citation Information ``` @article{yu2018spider, title={Spider: A large-scale human-labeled dataset for complex and cross-domain semantic parsing and text-to-sql task}, author={Yu, Tao and Zhang, Rui and Yang, Kai and Yasunaga, Michihiro and Wang, Dongxu and Li, Zifan and Ma, James and Li, Irene and Yao, Qingning and Roman, Shanelle and others}, journal={arXiv preprint arXiv:1809.08887}, year={2018} } ``` ### Contributions Thanks to [@olinguyen](https://github.com/olinguyen) for adding this dataset.
squad_kor_v1
2023-06-15T15:25:29.000Z
[ "task_categories:question-answering", "task_ids:extractive-qa", "annotations_creators:crowdsourced", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:ko", "license:cc-by-nd-4.0", "arxiv:1909.07005", "region:us" ]
null
KorQuAD 1.0 is a large-scale Korean dataset for machine reading comprehension task consisting of human generated questions for Wikipedia articles. We benchmark the data collecting process of SQuADv1.0 and crowdsourced 70,000+ question-answer pairs. 1,637 articles and 70,079 pairs of question answers were collected. 1,420 articles are used for the training set, 140 for the dev set, and 77 for the test set. 60,407 question-answer pairs are for the training set, 5,774 for the dev set, and 3,898 for the test set.
@article{lim2019korquad1, title={Korquad1. 0: Korean qa dataset for machine reading comprehension}, author={Lim, Seungyoung and Kim, Myungji and Lee, Jooyoul}, journal={arXiv preprint arXiv:1909.07005}, year={2019} }
null
9
1,190
--- annotations_creators: - crowdsourced language_creators: - found language: - ko license: - cc-by-nd-4.0 multilinguality: - monolingual size_categories: - 10K<n<100K source_datasets: - original task_categories: - question-answering task_ids: - extractive-qa paperswithcode_id: korquad pretty_name: The Korean Question Answering Dataset dataset_info: features: - name: id dtype: string - name: title dtype: string - name: context dtype: string - name: question dtype: string - name: answers sequence: - name: text dtype: string - name: answer_start dtype: int32 config_name: squad_kor_v1 splits: - name: train num_bytes: 83380337 num_examples: 60407 - name: validation num_bytes: 8261729 num_examples: 5774 download_size: 42408533 dataset_size: 91642066 --- # Dataset Card for KorQuAD v1.0 ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://korquad.github.io/KorQuad%201.0/ - **Repository:** https://github.com/korquad/korquad.github.io/tree/master/dataset - **Paper:** https://arxiv.org/abs/1909.07005 ### Dataset Summary KorQuAD 1.0 is a large-scale question-and-answer dataset constructed for Korean machine reading comprehension, and investigate the dataset to understand the distribution of answers and the types of reasoning required to answer the question. This dataset benchmarks the data generating process of SQuAD v1.0 to meet the standard. ### Supported Tasks and Leaderboards `question-answering` ### Languages Korean ## Dataset Structure Follows the standars SQuAD format. ### Data Instances An example from the data set looks as follows: ``` {'answers': {'answer_start': [54], 'text': ['교향곡']}, 'context': '1839년 바그너는 괴테의 파우스트을 처음 읽고 그 내용에 마음이 끌려 이를 소재로 해서 하나의 교향곡을 쓰려는 뜻을 갖는다. 이 시기 바그너는 1838년에 빛 독촉으로 산전수전을 다 걲은 상황이라 좌절과 실망에 가득했으며 메피스토펠레스를 만나는 파우스트의 심경에 공감했다고 한다. 또한 파리에서 아브네크의 지휘로 파리 음악원 관현악단이 연주하는 베토벤의 교향곡 9번을 듣고 깊은 감명을 받았는데, 이것이 이듬해 1월에 파우스트의 서곡으로 쓰여진 이 작품에 조금이라도 영향을 끼쳤으리라는 것은 의심할 여지가 없다. 여기의 라단조 조성의 경우에도 그의 전기에 적혀 있는 것처럼 단순한 정신적 피로나 실의가 반영된 것이 아니라 베토벤의 합창교향곡 조성의 영향을 받은 것을 볼 수 있다. 그렇게 교향곡 작곡을 1839년부터 40년에 걸쳐 파리에서 착수했으나 1악장을 쓴 뒤에 중단했다. 또한 작품의 완성과 동시에 그는 이 서곡(1악장)을 파리 음악원의 연주회에서 연주할 파트보까지 준비하였으나, 실제로는 이루어지지는 않았다. 결국 초연은 4년 반이 지난 후에 드레스덴에서 연주되었고 재연도 이루어졌지만, 이후에 그대로 방치되고 말았다. 그 사이에 그는 리엔치와 방황하는 네덜란드인을 완성하고 탄호이저에도 착수하는 등 분주한 시간을 보냈는데, 그런 바쁜 생활이 이 곡을 잊게 한 것이 아닌가 하는 의견도 있다.', 'id': '6566495-0-0', 'question': '바그너는 괴테의 파우스트를 읽고 무엇을 쓰고자 했는가?', 'title': '파우스트_서곡'} ``` ### Data Fields ``` {'id': Value(dtype='string', id=None), 'title': Value(dtype='string', id=None), 'context': Value(dtype='string', id=None), 'question': Value(dtype='string', id=None), 'answers': Sequence(feature={'text': Value(dtype='string', id=None), 'answer_start': Value(dtype='int32', id=None)}, length=-1, id=None)} ``` ### Data Splits - Train: 60407 - Validation: 5774 ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data Wikipedia #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [CC BY-ND 2.0 KR](https://creativecommons.org/licenses/by-nd/2.0/kr/deed.en) ### Citation Information ``` @article{lim2019korquad1, title={Korquad1. 0: Korean qa dataset for machine reading comprehension}, author={Lim, Seungyoung and Kim, Myungji and Lee, Jooyoul}, journal={arXiv preprint arXiv:1909.07005}, year={2019} ``` ### Contributions Thanks to [@cceyda](https://github.com/cceyda) for adding this dataset.
dalle-mini/YFCC100M_OpenAI_subset
2021-08-26T17:56:01.000Z
[ "arxiv:1503.01817", "region:us" ]
dalle-mini
The YFCC100M is one of the largest publicly and freely useable multimedia collection, containing the metadata of around 99.2 million photos and 0.8 million videos from Flickr, all of which were shared under one of the various Creative Commons licenses. This version is a subset defined in openai/CLIP.
@article{thomee2016yfcc100m, author = "Bart Thomee and David A. Shamma and Gerald Friedland and Benjamin Elizalde and Karl Ni and Douglas Poland and Damian Borth and Li-Jia Li", title = "{YFCC100M}: The New Data in Multimedia Research", journal = "Communications of the {ACM}", volume = "59", number = "2", pages = "64--73", year = "2016", url = "http://cacm.acm.org/magazines/2016/2/197425-yfcc100m/fulltext", }
null
7
1,188
# YFCC100M subset from OpenAI Subset of [YFCC100M](https://arxiv.org/abs/1503.01817) used by OpenAI for [CLIP](https://github.com/openai/CLIP/blob/main/data/yfcc100m.md), filtered to contain only the images that we could retrieve. | Split | train | validation | | --- | --- | --- | | Number of samples | 14,808,859 | 16,374 | | Size | 1.9 TB | 2.1 GB | Features: * from the original dataset: `title`, `description`, `photoid`, `uid`, `unickname`, `datetaken`, `dateuploaded`, `capturedevice`, `usertags`, `machinetags`, `longitude`, `latitude`, `accuracy`, `pageurl`, `downloadurl`, `licensename`, `licenseurl`, `serverid`, `farmid`, `secret`, `secretoriginal`, `ext`, `marker`, `key` * `img`: image content, can be loaded with `PIL.Image.open(io.BytesIO(item['img']))` * `title_clean` and `description_clean`: derived from `title` and `description` using `clean_text` function detailed below ```python def clean_text(text): # decode url text = urllib.parse.unquote_plus(text) # remove html tags text = re.sub('<[^<]+?>', '', text) # remove multiple spaces + "\r" + "\n" + "\t" text = " ".join(text.split()) return text ```
nlpai-lab/kullm-v2
2023-06-01T05:45:04.000Z
[ "task_categories:text-generation", "size_categories:10K<n<100K", "language:ko", "license:apache-2.0", "region:us" ]
nlpai-lab
null
null
null
37
1,184
--- license: apache-2.0 task_categories: - text-generation language: - ko pretty_name: kullm size_categories: - 10K<n<100K --- # Dataset Card for "KULLM-v2" ## Dataset Summary Korean translation of GPT4ALL, Dolly, and Vicuna data. repository: [nlpai-lab/KULLM](https://github.com/nlpai-lab/KULLM) huggingface: [nlpai-lab/kullm-v2](https://huggingface.co/nlpai-lab/kullm-polyglot-12.8b-v2) #### Translate dataset Translated 'instruction', 'input', and 'output' in the dataset via the DeepL API ## Lisence Apache-2.0 ```python >>> from datasets import load_dataset >>> ds = load_dataset("nlpai-lab/kullm-v2", split="train") >>> ds DatasetDict({ train: Dataset({ features: ['id', 'instruction', 'input', 'output'], num_rows: 152630 }) }) ``` ```python >>> ds[0] {'id': 'alpaca_{idx}', 'instruction': '3원색이란 무엇인가요?', 'input': '', 'output': '세 가지 기본 색은 빨강, 파랑, 노랑입니다. 이 색은 다른 색을 혼합하여 만들 수 없고 다른 모든 색은 다양한 비율로 조합하여 만들 수 있기 때문에 원색이라고 부릅니다. 빛에 사용되는 첨가제 색상 시스템에서 원색은 빨강, 녹색, 파랑(RGB)입니다.'} ```
eduagarcia/portuguese_benchmark
2023-07-09T06:31:26.000Z
[ "region:us" ]
eduagarcia
null
null
null
2
1,182
Entry not found
Tevatron/wikipedia-nq
2021-11-22T05:32:24.000Z
[ "region:us" ]
Tevatron
null
@inproceedings{karpukhin-etal-2020-dense, title = "Dense Passage Retrieval for Open-Domain Question Answering", author = "Karpukhin, Vladimir and Oguz, Barlas and Min, Sewon and Lewis, Patrick and Wu, Ledell and Edunov, Sergey and Chen, Danqi and Yih, Wen-tau", booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)", month = nov, year = "2020", address = "Online", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/2020.emnlp-main.550", doi = "10.18653/v1/2020.emnlp-main.550", pages = "6769--6781", }
null
2
1,175
Entry not found
drop
2023-04-05T10:05:02.000Z
[ "task_categories:question-answering", "task_categories:text2text-generation", "task_ids:extractive-qa", "task_ids:abstractive-qa", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:cc-by-sa-4.0", "region:us" ]
null
DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs. . DROP is a crowdsourced, adversarially-created, 96k-question benchmark, in which a system must resolve references in a question, perhaps to multiple input positions, and perform discrete operations over them (such as addition, counting, or sorting). These operations require a much more comprehensive understanding of the content of paragraphs than what was necessary for prior datasets.
@inproceedings{Dua2019DROP, author={Dheeru Dua and Yizhong Wang and Pradeep Dasigi and Gabriel Stanovsky and Sameer Singh and Matt Gardner}, title={DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs}, booktitle={Proc. of NAACL}, year={2019} }
null
9
1,174
--- pretty_name: DROP annotations_creators: - crowdsourced language_creators: - crowdsourced language: - en license: - cc-by-sa-4.0 multilinguality: - monolingual size_categories: - 10K<n<100K source_datasets: - original task_categories: - question-answering - text2text-generation task_ids: - extractive-qa - abstractive-qa paperswithcode_id: drop dataset_info: features: - name: section_id dtype: string - name: query_id dtype: string - name: passage dtype: string - name: question dtype: string - name: answers_spans sequence: - name: spans dtype: string - name: types dtype: string splits: - name: train num_bytes: 105572762 num_examples: 77400 - name: validation num_bytes: 11737787 num_examples: 9535 download_size: 8308692 dataset_size: 117310549 --- # Dataset Card for "drop" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://allennlp.org/drop](https://allennlp.org/drop) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 8.30 MB - **Size of the generated dataset:** 110.91 MB - **Total amount of disk used:** 119.21 MB ### Dataset Summary DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs. . DROP is a crowdsourced, adversarially-created, 96k-question benchmark, in which a system must resolve references in a question, perhaps to multiple input positions, and perform discrete operations over them (such as addition, counting, or sorting). These operations require a much more comprehensive understanding of the content of paragraphs than what was necessary for prior datasets. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### default - **Size of downloaded dataset files:** 8.30 MB - **Size of the generated dataset:** 110.91 MB - **Total amount of disk used:** 119.21 MB An example of 'validation' looks as follows. ``` This example was too long and was cropped: { "answers_spans": { "spans": ["Chaz Schilens"] }, "passage": "\" Hoping to rebound from their loss to the Patriots, the Raiders stayed at home for a Week 16 duel with the Houston Texans. Oak...", "question": "Who scored the first touchdown of the game?" } ``` ### Data Fields The data fields are the same among all splits. #### default - `passage`: a `string` feature. - `question`: a `string` feature. - `answers_spans`: a dictionary feature containing: - `spans`: a `string` feature. ### Data Splits | name |train|validation| |-------|----:|---------:| |default|77409| 9536| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @inproceedings{Dua2019DROP, author={Dheeru Dua and Yizhong Wang and Pradeep Dasigi and Gabriel Stanovsky and Sameer Singh and Matt Gardner}, title={ {DROP}: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs}, booktitle={Proc. of NAACL}, year={2019} } ``` ### Contributions Thanks to [@patrickvonplaten](https://github.com/patrickvonplaten), [@thomwolf](https://github.com/thomwolf), [@mariamabarham](https://github.com/mariamabarham), [@lewtun](https://github.com/lewtun) for adding this dataset.
tau/sled
2022-10-25T07:33:44.000Z
[ "task_categories:question-answering", "task_categories:summarization", "task_categories:text-generation", "task_ids:multiple-choice-qa", "task_ids:natural-language-inference", "language:en", "license:mit", "multi-hop-question-answering", "query-based-summarization", "long-texts", "arxiv:2208.00748", "arxiv:2201.03533", "arxiv:2104.02112", "arxiv:2104.07091", "arxiv:2104.05938", "arxiv:1712.07040", "arxiv:2105.03011", "arxiv:2112.08608", "arxiv:2110.01799", "arxiv:1606.05250", "arxiv:1809.09600", "region:us" ]
tau
Efficient Long-Text Understanding with Short-Text Models. Our SLiding-Encoder and Decoder uses any pretrained encoder-decoder model, to independtly encode overlapping chunks of the inputs, and perform fusion-in-decoder to achieve linear-memory requirment for long-range natural language understanding.
@inproceedings{Ivgi2022EfficientLU, title={Efficient Long-Text Understanding with Short-Text Models}, author={Maor Ivgi and Uri Shaham and Jonathan Berant}, year={2022} } Note that each SLED dataset has its own citation. Please see the source to get the correct citation for each contained dataset (and also cite the SCROLLS dataset on which it is based).
null
3
1,172
--- language: - en license: - mit task_categories: - question-answering - summarization - text-generation task_ids: - multiple-choice-qa - natural-language-inference configs: - gov_report - summ_screen_fd - qmsum - qasper - narrative_qa - quality - contract_nli - squad - squad_shuffled_distractors - squad_ordered_distractors - hotpotqa - hotpotqa_second_only tags: - multi-hop-question-answering - query-based-summarization - long-texts --- ## Dataset Description - **Repository:** [SLED Github repository](https://github.com/Mivg/SLED) - **Paper:** [Efficient Long-Text Understanding with Short-Text Models ](https://arxiv.org/pdf/2208.00748.pdf) # Dataset Card for SCROLLS ## Overview This dataset is based on the [SCROLLS](https://huggingface.co/datasets/tau/scrolls) dataset ([paper](https://arxiv.org/pdf/2201.03533.pdf)), the [SQuAD 1.1](https://huggingface.co/datasets/squad) dataset and the [HotpotQA](https://huggingface.co/datasets/hotpot_qa) dataset. It doesn't contain any unpblished data, but includes the configuration needed for the [Efficient Long-Text Understanding with Short-Text Models ](https://arxiv.org/pdf/2208.00748.pdf) paper. ## Tasks The tasks included are: #### GovReport ([Huang et al., 2021](https://arxiv.org/pdf/2104.02112.pdf)) GovReport is a summarization dataset of reports addressing various national policy issues published by the Congressional Research Service and the U.S. Government Accountability Office, where each document is paired with a hand-written executive summary. The reports and their summaries are longer than their equivalents in other popular long-document summarization datasets; for example, GovReport's documents are approximately 1.5 and 2.5 times longer than the documents in Arxiv and PubMed, respectively. #### SummScreenFD ([Chen et al., 2021](https://arxiv.org/pdf/2104.07091.pdf)) SummScreenFD is a summarization dataset in the domain of TV shows (e.g. Friends, Game of Thrones). Given a transcript of a specific episode, the goal is to produce the episode's recap. The original dataset is divided into two complementary subsets, based on the source of its community contributed transcripts. For SCROLLS, we use the ForeverDreaming (FD) subset, as it incorporates 88 different shows, making it a more diverse alternative to the TV MegaSite (TMS) subset, which has only 10 shows. Community-authored recaps for the ForeverDreaming transcripts were collected from English Wikipedia and TVMaze. #### QMSum ([Zhong et al., 2021](https://arxiv.org/pdf/2104.05938.pdf)) QMSum is a query-based summarization dataset, consisting of 232 meetings transcripts from multiple domains. The corpus covers academic group meetings at the International Computer Science Institute and their summaries, industrial product meetings for designing a remote control, and committee meetings of the Welsh and Canadian Parliaments, dealing with a variety of public policy issues. Annotators were tasked with writing queries about the broad contents of the meetings, as well as specific questions about certain topics or decisions, while ensuring that the relevant text for answering each query spans at least 200 words or 10 turns. #### NarrativeQA ([Kočiský et al., 2021](https://arxiv.org/pdf/1712.07040.pdf)) NarrativeQA (Kočiský et al., 2021) is an established question answering dataset over entire books from Project Gutenberg and movie scripts from different websites. Annotators were given summaries of the books and scripts obtained from Wikipedia, and asked to generate question-answer pairs, resulting in about 30 questions and answers for each of the 1,567 books and scripts. They were encouraged to use their own words rather then copying, and avoid asking yes/no questions or ones about the cast. Each question was then answered by an additional annotator, providing each question with two reference answers (unless both answers are identical). #### Qasper ([Dasigi et al., 2021](https://arxiv.org/pdf/2105.03011.pdf)) Qasper is a question answering dataset over NLP papers filtered from the Semantic Scholar Open Research Corpus (S2ORC). Questions were written by NLP practitioners after reading only the title and abstract of the papers, while another set of NLP practitioners annotated the answers given the entire document. Qasper contains abstractive, extractive, and yes/no questions, as well as unanswerable ones. #### QuALITY ([Pang et al., 2021](https://arxiv.org/pdf/2112.08608.pdf)) QuALITY is a multiple-choice question answering dataset over articles and stories sourced from Project Gutenberg, the Open American National Corpus, and more. Experienced writers wrote questions and distractors, and were incentivized to write answerable, unambiguous questions such that in order to correctly answer them, human annotators must read large portions of the given document. Reference answers were then calculated using the majority vote between of the annotators and writer's answers. To measure the difficulty of their questions, Pang et al. conducted a speed validation process, where another set of annotators were asked to answer questions given only a short period of time to skim through the document. As a result, 50% of the questions in QuALITY are labeled as hard, i.e. the majority of the annotators in the speed validation setting chose the wrong answer. #### ContractNLI ([Koreeda and Manning, 2021](https://arxiv.org/pdf/2110.01799.pdf)) Contract NLI is a natural language inference dataset in the legal domain. Given a non-disclosure agreement (the premise), the task is to predict whether a particular legal statement (the hypothesis) is entailed, not entailed (neutral), or cannot be entailed (contradiction) from the contract. The NDAs were manually picked after simple filtering from the Electronic Data Gathering, Analysis, and Retrieval system (EDGAR) and Google. The dataset contains a total of 607 contracts and 17 unique hypotheses, which were combined to produce the dataset's 10,319 examples. #### SQuAD 1.1 ([Rajpurkar et al., 2016](https://arxiv.org/pdf/1606.05250.pdf)) Stanford Question Answering Dataset (SQuAD) is a reading comprehension \ dataset, consisting of questions posed by crowdworkers on a set of Wikipedia \ articles, where the answer to every question is a segment of text, or span, \ from the corresponding reading passage, or the question might be unanswerable. #### HotpotQA ([Yang et al., 2018](https://arxiv.org/pdf/1809.09600.pdf)) HotpotQA is a new dataset with 113k Wikipedia-based question-answer pairs with four key features: (1) the questions require finding and reasoning over multiple supporting documents to answer; (2) the questions are diverse and not constrained to any pre-existing knowledge bases or knowledge schemas; (3) we provide sentence-level supporting facts required for reasoning, allowingQA systems to reason with strong supervisionand explain the predictions; (4) we offer a new type of factoid comparison questions to testQA systems’ ability to extract relevant facts and perform necessary comparison. ## Data Fields All the datasets in the benchmark are in the same input-output format - `input`: a `string` feature. The input document. - `input_prefix`: an optional `string` feature, for the datasets containing prefix (e.g. question) - `output`: a `string` feature. The target. - `id`: a `string` feature. Unique per input. - `pid`: a `string` feature. Unique per input-output pair (can differ from 'id' in NarrativeQA and Qasper, where there is more then one valid target). The dataset that contain `input_prefix` are: - SQuAD - the question - HotpotQA - the question - qmsum - the query - qasper - the question - narrative_qa - the question - quality - the question + the four choices - contract_nli - the hypothesis ## Controlled experiments To test multiple properties of SLED, we modify SQuAD 1.1 [Rajpurkar et al., 2016](https://arxiv.org/pdf/1606.05250.pdf) and HotpotQA [Yang et al., 2018](https://arxiv.org/pdf/1809.09600.pdf) to create a few controlled experiments settings. Those are accessible via the following configurations: - squad - Contains the original version of SQuAD 1.1 (question + passage) - squad_ordered_distractors - For each example, 9 random distrctor passages are concatenated (separated by '\n') - squad_shuffled_distractors - For each example, 9 random distrctor passages are added (separated by '\n'), and jointly the 10 passages are randomly shuffled - hotpotqa - A clean version of HotpotQA, where each input contains only the two gold passages (separated by '\n') - hotpotqa_second_only - In each example, the input contains only the second gold passage ## Citation If you use this dataset, **please make sure to cite all the original dataset papers as well SCROLLS.** [[bibtex](https://drive.google.com/uc?export=download&id=1IUYIzQD9DPsECw0JWkwk4Ildn8JOMtuU)] ``` @inproceedings{Ivgi2022EfficientLU, title={Efficient Long-Text Understanding with Short-Text Models}, author={Maor Ivgi and Uri Shaham and Jonathan Berant}, year={2022} } ```
Nexusflow/NexusRaven_API_evaluation
2023-09-29T05:19:42.000Z
[ "arxiv:2306.05301", "arxiv:2307.16789", "region:us" ]
Nexusflow
null
null
null
3
1,170
--- dataset_info: - config_name: outputs_in_toolllm_format features: - name: response list: - name: function_call dtype: string - name: query dtype: string - name: task_id dtype: int64 - name: timestamp dtype: float64 splits: - name: train num_bytes: 303376 num_examples: 348 download_size: 83053 dataset_size: 303376 - config_name: raw_api_list features: - name: dataset dtype: string - name: name dtype: string - name: description dtype: string - name: args_dicts list: - name: default dtype: 'null' - name: description dtype: string - name: name dtype: string - name: required dtype: bool - name: type dtype: string splits: - name: train num_bytes: 22276 num_examples: 2 download_size: 10949 dataset_size: 22276 - config_name: raw_queries features: - name: dataset dtype: string - name: query_dict dtype: string splits: - name: train num_bytes: 466227 num_examples: 339 download_size: 98527 dataset_size: 466227 - config_name: standardized_api_list features: - name: dataset dtype: string - name: name dtype: string - name: description dtype: string - name: args_dicts list: - name: default dtype: string - name: description dtype: string - name: name dtype: string - name: required dtype: bool - name: type dtype: string splits: - name: train num_bytes: 47776 num_examples: 65 download_size: 27751 dataset_size: 47776 - config_name: standardized_queries features: - name: dataset dtype: string - name: prompt dtype: string - name: python_function_name dtype: string - name: python_args_dict dtype: string - name: context_functions sequence: string splits: - name: train num_bytes: 153860 num_examples: 318 download_size: 36721 dataset_size: 153860 configs: - config_name: outputs_in_toolllm_format data_files: - split: train path: outputs_in_toolllm_format/train-* - config_name: raw_queries data_files: - split: train path: raw_queries/train-* - config_name: standardized_api_list data_files: - split: train path: standardized_api_list/train-* - config_name: standardized_queries data_files: - split: train path: standardized_queries/train-* --- # NexusRaven API Evaluation dataset Please see [blog post](http://nexusflow.ai/blog) or [NexusRaven Github repo](https://github.com/nexusflowai/NexusRaven) for more information. ## License The evaluation data in this repository consists primarily of our own curated evaluation data that only uses open source commercializable models. However, we include general domain data from the ToolLLM and ToolAlpaca papers. Since the data in the ToolLLM and ToolAlpaca works use OpenAI's GPT models for the generated content, the data is not commercially licensable, even if our own data is. As a result, the evaluation data used here is strictly non-commercial under [CC-BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/). Thank you for understanding! ## References We thank the following authors and entities for their evaluation data, which we leveraged to produce the results contained in this repository. Their citations can be found below 1. ToolAlpaca team 2. ToolLLM team ``` @misc{tang2023toolalpaca, title={ToolAlpaca: Generalized Tool Learning for Language Models with 3000 Simulated Cases}, author={Qiaoyu Tang and Ziliang Deng and Hongyu Lin and Xianpei Han and Qiao Liang and Boxi Cao and Le Sun}, year={2023}, eprint={2306.05301}, archivePrefix={arXiv}, primaryClass={cs.CL} } @misc{qin2023toolllm, title={ToolLLM: Facilitating Large Language Models to Master 16000+ Real-world APIs}, author={Yujia Qin and Shihao Liang and Yining Ye and Kunlun Zhu and Lan Yan and Yaxi Lu and Yankai Lin and Xin Cong and Xiangru Tang and Bill Qian and Sihan Zhao and Runchu Tian and Ruobing Xie and Jie Zhou and Mark Gerstein and Dahai Li and Zhiyuan Liu and Maosong Sun}, year={2023}, eprint={2307.16789}, archivePrefix={arXiv}, primaryClass={cs.AI} } ``` ## Citation ``` @misc{nexusraven, title={NexusRaven: Surpassing the state-of-the-art in open-source function calling LLMs}, author={Nexusflow.ai team}, year={2023}, url={http://nexusflow.ai/blog} } ``` ## Contact Please reach out to info@nexusflow.ai for any questions!
huggingface/semantic-segmentation-test-sample
2022-04-11T09:15:24.000Z
[ "region:us" ]
huggingface
null
null
null
0
1,154
This dataset contains 10 examples of the [segments/sidewalk-semantic](https://huggingface.co/datasets/segments/sidewalk-semantic) dataset (i.e. 10 images with corresponding ground-truth segmentation maps).
BeIR/fever
2022-10-23T06:04:31.000Z
[ "task_categories:text-retrieval", "task_ids:entity-linking-retrieval", "task_ids:fact-checking-retrieval", "multilinguality:monolingual", "language:en", "license:cc-by-sa-4.0", "region:us" ]
BeIR
null
null
null
2
1,150
--- annotations_creators: [] language_creators: [] language: - en license: - cc-by-sa-4.0 multilinguality: - monolingual paperswithcode_id: beir pretty_name: BEIR Benchmark size_categories: msmarco: - 1M<n<10M trec-covid: - 100k<n<1M nfcorpus: - 1K<n<10K nq: - 1M<n<10M hotpotqa: - 1M<n<10M fiqa: - 10K<n<100K arguana: - 1K<n<10K touche-2020: - 100K<n<1M cqadupstack: - 100K<n<1M quora: - 100K<n<1M dbpedia: - 1M<n<10M scidocs: - 10K<n<100K fever: - 1M<n<10M climate-fever: - 1M<n<10M scifact: - 1K<n<10K source_datasets: [] task_categories: - text-retrieval - zero-shot-retrieval - information-retrieval - zero-shot-information-retrieval task_ids: - passage-retrieval - entity-linking-retrieval - fact-checking-retrieval - tweet-retrieval - citation-prediction-retrieval - duplication-question-retrieval - argument-retrieval - news-retrieval - biomedical-information-retrieval - question-answering-retrieval --- # Dataset Card for BEIR Benchmark ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://github.com/UKPLab/beir - **Repository:** https://github.com/UKPLab/beir - **Paper:** https://openreview.net/forum?id=wCu6T5xFjeJ - **Leaderboard:** https://docs.google.com/spreadsheets/d/1L8aACyPaXrL8iEelJLGqlMqXKPX2oSP_R10pZoy77Ns - **Point of Contact:** nandan.thakur@uwaterloo.ca ### Dataset Summary BEIR is a heterogeneous benchmark that has been built from 18 diverse datasets representing 9 information retrieval tasks: - Fact-checking: [FEVER](http://fever.ai), [Climate-FEVER](http://climatefever.ai), [SciFact](https://github.com/allenai/scifact) - Question-Answering: [NQ](https://ai.google.com/research/NaturalQuestions), [HotpotQA](https://hotpotqa.github.io), [FiQA-2018](https://sites.google.com/view/fiqa/) - Bio-Medical IR: [TREC-COVID](https://ir.nist.gov/covidSubmit/index.html), [BioASQ](http://bioasq.org), [NFCorpus](https://www.cl.uni-heidelberg.de/statnlpgroup/nfcorpus/) - News Retrieval: [TREC-NEWS](https://trec.nist.gov/data/news2019.html), [Robust04](https://trec.nist.gov/data/robust/04.guidelines.html) - Argument Retrieval: [Touche-2020](https://webis.de/events/touche-20/shared-task-1.html), [ArguAna](tp://argumentation.bplaced.net/arguana/data) - Duplicate Question Retrieval: [Quora](https://www.quora.com/q/quoradata/First-Quora-Dataset-Release-Question-Pairs), [CqaDupstack](http://nlp.cis.unimelb.edu.au/resources/cqadupstack/) - Citation-Prediction: [SCIDOCS](https://allenai.org/data/scidocs) - Tweet Retrieval: [Signal-1M](https://research.signal-ai.com/datasets/signal1m-tweetir.html) - Entity Retrieval: [DBPedia](https://github.com/iai-group/DBpedia-Entity/) All these datasets have been preprocessed and can be used for your experiments. ```python ``` ### Supported Tasks and Leaderboards The dataset supports a leaderboard that evaluates models against task-specific metrics such as F1 or EM, as well as their ability to retrieve supporting information from Wikipedia. The current best performing models can be found [here](https://eval.ai/web/challenges/challenge-page/689/leaderboard/). ### Languages All tasks are in English (`en`). ## Dataset Structure All BEIR datasets must contain a corpus, queries and qrels (relevance judgments file). They must be in the following format: - `corpus` file: a `.jsonl` file (jsonlines) that contains a list of dictionaries, each with three fields `_id` with unique document identifier, `title` with document title (optional) and `text` with document paragraph or passage. For example: `{"_id": "doc1", "title": "Albert Einstein", "text": "Albert Einstein was a German-born...."}` - `queries` file: a `.jsonl` file (jsonlines) that contains a list of dictionaries, each with two fields `_id` with unique query identifier and `text` with query text. For example: `{"_id": "q1", "text": "Who developed the mass-energy equivalence formula?"}` - `qrels` file: a `.tsv` file (tab-seperated) that contains three columns, i.e. the `query-id`, `corpus-id` and `score` in this order. Keep 1st row as header. For example: `q1 doc1 1` ### Data Instances A high level example of any beir dataset: ```python corpus = { "doc1" : { "title": "Albert Einstein", "text": "Albert Einstein was a German-born theoretical physicist. who developed the theory of relativity, \ one of the two pillars of modern physics (alongside quantum mechanics). His work is also known for \ its influence on the philosophy of science. He is best known to the general public for his mass–energy \ equivalence formula E = mc2, which has been dubbed 'the world's most famous equation'. He received the 1921 \ Nobel Prize in Physics 'for his services to theoretical physics, and especially for his discovery of the law \ of the photoelectric effect', a pivotal step in the development of quantum theory." }, "doc2" : { "title": "", # Keep title an empty string if not present "text": "Wheat beer is a top-fermented beer which is brewed with a large proportion of wheat relative to the amount of \ malted barley. The two main varieties are German Weißbier and Belgian witbier; other types include Lambic (made\ with wild yeast), Berliner Weisse (a cloudy, sour beer), and Gose (a sour, salty beer)." }, } queries = { "q1" : "Who developed the mass-energy equivalence formula?", "q2" : "Which beer is brewed with a large proportion of wheat?" } qrels = { "q1" : {"doc1": 1}, "q2" : {"doc2": 1}, } ``` ### Data Fields Examples from all configurations have the following features: ### Corpus - `corpus`: a `dict` feature representing the document title and passage text, made up of: - `_id`: a `string` feature representing the unique document id - `title`: a `string` feature, denoting the title of the document. - `text`: a `string` feature, denoting the text of the document. ### Queries - `queries`: a `dict` feature representing the query, made up of: - `_id`: a `string` feature representing the unique query id - `text`: a `string` feature, denoting the text of the query. ### Qrels - `qrels`: a `dict` feature representing the query document relevance judgements, made up of: - `_id`: a `string` feature representing the query id - `_id`: a `string` feature, denoting the document id. - `score`: a `int32` feature, denoting the relevance judgement between query and document. ### Data Splits | Dataset | Website| BEIR-Name | Type | Queries | Corpus | Rel D/Q | Down-load | md5 | | -------- | -----| ---------| --------- | ----------- | ---------| ---------| :----------: | :------:| | MSMARCO | [Homepage](https://microsoft.github.io/msmarco/)| ``msmarco`` | ``train``<br>``dev``<br>``test``| 6,980 | 8.84M | 1.1 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/msmarco.zip) | ``444067daf65d982533ea17ebd59501e4`` | | TREC-COVID | [Homepage](https://ir.nist.gov/covidSubmit/index.html)| ``trec-covid``| ``test``| 50| 171K| 493.5 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/trec-covid.zip) | ``ce62140cb23feb9becf6270d0d1fe6d1`` | | NFCorpus | [Homepage](https://www.cl.uni-heidelberg.de/statnlpgroup/nfcorpus/) | ``nfcorpus`` | ``train``<br>``dev``<br>``test``| 323 | 3.6K | 38.2 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/nfcorpus.zip) | ``a89dba18a62ef92f7d323ec890a0d38d`` | | BioASQ | [Homepage](http://bioasq.org) | ``bioasq``| ``train``<br>``test`` | 500 | 14.91M | 8.05 | No | [How to Reproduce?](https://github.com/UKPLab/beir/blob/main/examples/dataset#2-bioasq) | | NQ | [Homepage](https://ai.google.com/research/NaturalQuestions) | ``nq``| ``train``<br>``test``| 3,452 | 2.68M | 1.2 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/nq.zip) | ``d4d3d2e48787a744b6f6e691ff534307`` | | HotpotQA | [Homepage](https://hotpotqa.github.io) | ``hotpotqa``| ``train``<br>``dev``<br>``test``| 7,405 | 5.23M | 2.0 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/hotpotqa.zip) | ``f412724f78b0d91183a0e86805e16114`` | | FiQA-2018 | [Homepage](https://sites.google.com/view/fiqa/) | ``fiqa`` | ``train``<br>``dev``<br>``test``| 648 | 57K | 2.6 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/fiqa.zip) | ``17918ed23cd04fb15047f73e6c3bd9d9`` | | Signal-1M(RT) | [Homepage](https://research.signal-ai.com/datasets/signal1m-tweetir.html)| ``signal1m`` | ``test``| 97 | 2.86M | 19.6 | No | [How to Reproduce?](https://github.com/UKPLab/beir/blob/main/examples/dataset#4-signal-1m) | | TREC-NEWS | [Homepage](https://trec.nist.gov/data/news2019.html) | ``trec-news`` | ``test``| 57 | 595K | 19.6 | No | [How to Reproduce?](https://github.com/UKPLab/beir/blob/main/examples/dataset#1-trec-news) | | ArguAna | [Homepage](http://argumentation.bplaced.net/arguana/data) | ``arguana``| ``test`` | 1,406 | 8.67K | 1.0 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/arguana.zip) | ``8ad3e3c2a5867cdced806d6503f29b99`` | | Touche-2020| [Homepage](https://webis.de/events/touche-20/shared-task-1.html) | ``webis-touche2020``| ``test``| 49 | 382K | 19.0 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/webis-touche2020.zip) | ``46f650ba5a527fc69e0a6521c5a23563`` | | CQADupstack| [Homepage](http://nlp.cis.unimelb.edu.au/resources/cqadupstack/) | ``cqadupstack``| ``test``| 13,145 | 457K | 1.4 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/cqadupstack.zip) | ``4e41456d7df8ee7760a7f866133bda78`` | | Quora| [Homepage](https://www.quora.com/q/quoradata/First-Quora-Dataset-Release-Question-Pairs) | ``quora``| ``dev``<br>``test``| 10,000 | 523K | 1.6 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/quora.zip) | ``18fb154900ba42a600f84b839c173167`` | | DBPedia | [Homepage](https://github.com/iai-group/DBpedia-Entity/) | ``dbpedia-entity``| ``dev``<br>``test``| 400 | 4.63M | 38.2 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/dbpedia-entity.zip) | ``c2a39eb420a3164af735795df012ac2c`` | | SCIDOCS| [Homepage](https://allenai.org/data/scidocs) | ``scidocs``| ``test``| 1,000 | 25K | 4.9 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/scidocs.zip) | ``38121350fc3a4d2f48850f6aff52e4a9`` | | FEVER | [Homepage](http://fever.ai) | ``fever``| ``train``<br>``dev``<br>``test``| 6,666 | 5.42M | 1.2| [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/fever.zip) | ``5a818580227bfb4b35bb6fa46d9b6c03`` | | Climate-FEVER| [Homepage](http://climatefever.ai) | ``climate-fever``|``test``| 1,535 | 5.42M | 3.0 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/climate-fever.zip) | ``8b66f0a9126c521bae2bde127b4dc99d`` | | SciFact| [Homepage](https://github.com/allenai/scifact) | ``scifact``| ``train``<br>``test``| 300 | 5K | 1.1 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/scifact.zip) | ``5f7d1de60b170fc8027bb7898e2efca1`` | | Robust04 | [Homepage](https://trec.nist.gov/data/robust/04.guidelines.html) | ``robust04``| ``test``| 249 | 528K | 69.9 | No | [How to Reproduce?](https://github.com/UKPLab/beir/blob/main/examples/dataset#3-robust04) | ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information [Needs More Information] ### Citation Information Cite as: ``` @inproceedings{ thakur2021beir, title={{BEIR}: A Heterogeneous Benchmark for Zero-shot Evaluation of Information Retrieval Models}, author={Nandan Thakur and Nils Reimers and Andreas R{\"u}ckl{\'e} and Abhishek Srivastava and Iryna Gurevych}, booktitle={Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2)}, year={2021}, url={https://openreview.net/forum?id=wCu6T5xFjeJ} } ``` ### Contributions Thanks to [@Nthakur20](https://github.com/Nthakur20) for adding this dataset.
mteb/sickr-sts
2022-09-27T19:13:22.000Z
[ "language:en", "region:us" ]
mteb
null
null
null
1
1,145
--- language: - en ---
nsmc
2023-01-25T14:41:49.000Z
[ "task_categories:text-classification", "task_ids:sentiment-classification", "annotations_creators:crowdsourced", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:ko", "license:cc-by-2.0", "region:us" ]
null
This is a movie review dataset in the Korean language. Reviews were scraped from Naver movies. The dataset construction is based on the method noted in Large movie review dataset from Maas et al., 2011.
@InProceedings{Park:2016, title = "Naver Sentiment Movie Corpus", author = "Lucy Park", year = "2016", howpublished = {\\url{https://github.com/e9t/nsmc}} }
null
3
1,139
--- annotations_creators: - crowdsourced language_creators: - found language: - ko license: - cc-by-2.0 multilinguality: - monolingual size_categories: - 100K<n<1M source_datasets: - original task_categories: - text-classification task_ids: - sentiment-classification paperswithcode_id: nsmc pretty_name: Naver Sentiment Movie Corpus dataset_info: features: - name: id dtype: string - name: document dtype: string - name: label dtype: class_label: names: '0': negative '1': positive splits: - name: train num_bytes: 16423803 num_examples: 150000 - name: test num_bytes: 5491417 num_examples: 50000 download_size: 19522142 dataset_size: 21915220 --- # Dataset Card for Naver sentiment movie corpus ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Github](https://github.com/e9t/nsmc/) - **Repository:** [Github](https://github.com/e9t/nsmc/) - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields Each instance is a movie review written by Korean internet users on Naver, the most commonly used search engine in Korea. Each row can be broken down into the following fields: - `id`: A unique review ID, provided by Naver - `document`: The actual movie review - `label`: Binary labels for sentiment analysis, where `0` denotes negative, and `1`, positive ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ``` @InProceedings{Park:2016, title = "Naver Sentiment Movie Corpus", author = "Lucy Park", year = "2016", howpublished = {\\url{https://github.com/e9t/nsmc}} } ``` ### Contributions Thanks to [@jaketae](https://github.com/jaketae) for adding this dataset.
bigbio/biored
2023-01-12T05:54:49.000Z
[ "multilinguality:monolingual", "language:en", "license:unknown", "arxiv:2204.04263", "region:us" ]
bigbio
Relation Extraction corpus with multiple entity types (e.g., gene/protein, disease, chemical) and relation pairs (e.g., gene-disease; chemical-chemical), on a set of 600 PubMed articles
@article{DBLP:journals/corr/abs-2204-04263, author = {Ling Luo and Po{-}Ting Lai and Chih{-}Hsuan Wei and Cecilia N. Arighi and Zhiyong Lu}, title = {BioRED: {A} Comprehensive Biomedical Relation Extraction Dataset}, journal = {CoRR}, volume = {abs/2204.04263}, year = {2022}, url = {https://doi.org/10.48550/arXiv.2204.04263}, doi = {10.48550/arXiv.2204.04263}, eprinttype = {arXiv}, eprint = {2204.04263}, timestamp = {Wed, 11 May 2022 15:24:37 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-2204-04263.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} }
null
0
1,139
--- language: - en bigbio_language: - English license: unknown multilinguality: monolingual bigbio_license_shortname: UNKNOWN pretty_name: BioRED homepage: https://ftp.ncbi.nlm.nih.gov/pub/lu/BioRED/ bigbio_pubmed: True bigbio_public: True bigbio_tasks: - NAMED_ENTITY_RECOGNITION - RELATION_EXTRACTION --- # Dataset Card for BioRED ## Dataset Description - **Homepage:** https://ftp.ncbi.nlm.nih.gov/pub/lu/BioRED/ - **Pubmed:** True - **Public:** True - **Tasks:** NER,RE Relation Extraction corpus with multiple entity types (e.g., gene/protein, disease, chemical) and relation pairs (e.g., gene-disease; chemical-chemical), on a set of 600 PubMed articles ## Citation Information ``` @article{DBLP:journals/corr/abs-2204-04263, author = {Ling Luo and Po{-}Ting Lai and Chih{-}Hsuan Wei and Cecilia N. Arighi and Zhiyong Lu}, title = {BioRED: {A} Comprehensive Biomedical Relation Extraction Dataset}, journal = {CoRR}, volume = {abs/2204.04263}, year = {2022}, url = {https://doi.org/10.48550/arXiv.2204.04263}, doi = {10.48550/arXiv.2204.04263}, eprinttype = {arXiv}, eprint = {2204.04263}, timestamp = {Wed, 11 May 2022 15:24:37 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-2204-04263.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ```
dlb/plue
2022-10-29T12:19:26.000Z
[ "task_categories:text-classification", "task_ids:acceptability-classification", "task_ids:natural-language-inference", "task_ids:semantic-similarity-scoring", "task_ids:sentiment-classification", "task_ids:text-scoring", "annotations_creators:found", "language_creators:machine-generated", "multilinguality:monolingual", "multilinguality:translation", "size_categories:10K<n<100K", "source_datasets:extended|glue", "language:pt", "license:lgpl-3.0", "paraphrase-identification", "qa-nli", "coreference-nli", "region:us" ]
dlb
PLUE: Portuguese Language Understanding Evaluationis a Portuguese translation of the GLUE benchmark and Scitail using OPUS-MT model and Google Cloud Translation.
@misc{Gomes2020, author = {GOMES, J. R. S.}, title = {Portuguese Language Understanding Evaluation}, year = {2020}, publisher = {GitHub}, journal = {GitHub repository}, howpublished = {\\url{https://github.com/jubs12/PLUE}}, commit = {CURRENT_COMMIT} } @inproceedings{wang2019glue, title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding}, author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.}, note={In the Proceedings of ICLR.}, year={2019} }
null
6
1,138
--- annotations_creators: - found language_creators: - machine-generated language: - pt license: - lgpl-3.0 multilinguality: - monolingual - translation size_categories: - 10K<n<100K source_datasets: - extended|glue task_categories: - text-classification task_ids: - acceptability-classification - natural-language-inference - semantic-similarity-scoring - sentiment-classification - text-scoring pretty_name: PLUE (Portuguese Language Understanding Evaluation benchmark) tags: - paraphrase-identification - qa-nli - coreference-nli --- # Dataset Card for PLUE ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Repository:** https://github.com/ju-resplande/PLUE - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Portuguese translation of the <a href="https://gluebenchmark.com/">GLUE benchmark</a>, <a href=https://nlp.stanford.edu/projects/snli/>SNLI</a>, and <a href=https://allenai.org/data/scitail> Scitail</a> using <a href=https://github.com/Helsinki-NLP/OPUS-MT>OPUS-MT model</a> and <a href=https://cloud.google.com/translate/docs>Google Cloud Translation</a>. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The language data in PLUE is Brazilian Portuguese (BCP-47 pt-BR) ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ```bibtex @misc{Gomes2020, author = {GOMES, J. R. S.}, title = {PLUE: Portuguese Language Understanding Evaluation}, year = {2020}, publisher = {GitHub}, journal = {GitHub repository}, howpublished = {\url{https://github.com/jubs12/PLUE}}, commit = {CURRENT_COMMIT} } ``` ### Contributions Thanks to [@ju-resplande](https://github.com/ju-resplande) for adding this dataset.
pharaouk/dharma-1
2023-09-14T23:50:58.000Z
[ "region:us" ]
pharaouk
null
null
null
19
1,135
--- configs: - config_name: default data_files: - split: 'dharma_1_full' path: dharma_1_full* - split: 'dharma_1_mini' path: dharma_1_mini* - split: 'dharma_1_micro' path: dharma_1_micro* - split: 'dharma_1_unshuffled' path: dharma_eval_unshuffled* --- # "Dharma-1" A new carefully curated benchmark set, designed for a new era where the true end user uses LLM's for zero-shot and one-shot tasks, for a vast majority of the time. Stop training your models on mindless targets (eval_loss, train_loss), start training your LLM on lightweight Dharma as an eval target. A mix of all the top benchmarks. Formed to have an equal distribution of some of the most trusted benchmarks used by those developing SOTA LLMs, comprised of only 3,000 examples for the largest size, as well as 450 and 90 for Dharma-mini and Dharma-micro respectively. The current version of Dharma is comprised of a curated sampling of the following benchmarks: - AGIEval - Bigbench - MMLU - Winogrande - Arc-C - Arc- E - OBQA - TruthfulQA - Bool-q Each of these original benchmark datasets have their own subsections, careful work has gone into also choosing an equal distribution of the important subsections of each these, to have the best representation of the original benchmark creators goals. Dharma-1 is now integrated into Axolotl as well!, so you can focus on optimizing the other aspects of your training pipeline, model architecture and/or dataset, as opposed to worrying about what is the best evaluation measurement or optimization target that will best represent capabilities for the end user. Benchmarking for top base model will be listed here when completed and verified. Special thanks to @LDJnr for their contributions. Check out their Puffin dataset here: https://huggingface.co/LDJnr [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
HumanCompatibleAI/ppo-seals-CartPole-v0
2023-05-29T09:52:49.000Z
[ "region:us" ]
HumanCompatibleAI
null
null
null
0
1,134
--- dataset_info: features: - name: obs sequence: sequence: float32 - name: acts sequence: int64 - name: infos sequence: string - name: terminal dtype: bool - name: rews sequence: float64 splits: - name: train num_bytes: 516313 num_examples: 24 download_size: 297546 dataset_size: 516313 --- # Dataset Card for "ppo-seals-CartPole-v0" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Ryan-sjtu/celebahq-caption
2023-05-26T15:54:04.000Z
[ "license:mit", "region:us" ]
Ryan-sjtu
null
null
null
0
1,129
--- license: mit dataset_info: features: - name: image dtype: image - name: text dtype: string splits: - name: train num_bytes: 2756863400.0 num_examples: 30000 download_size: 2762815442 dataset_size: 2756863400.0 ---
laion/laion400m
2023-04-04T06:35:23.000Z
[ "license:cc-by-4.0", "region:us" ]
laion
null
null
null
17
1,127
--- license: cc-by-4.0 --- # LAION-400m_new This datasets has two improvements compared to original LAION_400m dataset: 1. It uses a multilingual text filter to filter out malicious content 2. The better open_clip VitH model was used to detect potential harmful content in the images All in all, we filtered out around 6 million additional image-text pairs - probably with a high false positive rate - in order to improve dataset safety.
fujiki/japanese_hh-rlhf-49k
2023-05-28T06:08:04.000Z
[ "language:ja", "license:mit", "region:us" ]
fujiki
null
null
null
1
1,126
--- license: mit dataset_info: features: - name: instruction dtype: string - name: input dtype: string - name: output dtype: string - name: index dtype: string splits: - name: train num_bytes: 34168978 num_examples: 49332 download_size: 18427777 dataset_size: 34168978 language: - ja --- - This is a little bit different version of [`kunishou/hh-rlhf-49k-ja`](https://huggingface.co/datasets/kunishou/hh-rlhf-49k-ja) without `ng_translation == 1` examples. - Please also refer to the original dataset [`kunishou/hh-rlhf-49k-ja`](https://huggingface.co/datasets/kunishou/hh-rlhf-49k-ja).
deepset/germanquad
2023-04-06T13:58:35.000Z
[ "task_categories:question-answering", "task_categories:text-retrieval", "task_ids:extractive-qa", "task_ids:closed-domain-qa", "task_ids:open-domain-qa", "multilinguality:monolingual", "source_datasets:original", "language:de", "license:cc-by-4.0", "arxiv:2104.12741", "region:us" ]
deepset
In order to raise the bar for non-English QA, we are releasing a high-quality, human-labeled German QA dataset consisting of 13 722 questions, incl. a three-way annotated test set. The creation of GermanQuAD is inspired by insights from existing datasets as well as our labeling experience from several industry projects. We combine the strengths of SQuAD, such as high out-of-domain performance, with self-sufficient questions that contain all relevant information for open-domain QA as in the NaturalQuestions dataset. Our training and test datasets do not overlap like other popular datasets and include complex questions that cannot be answered with a single entity or only a few words.
@misc{möller2021germanquad, title={GermanQuAD and GermanDPR: Improving Non-English Question Answering and Passage Retrieval}, author={Timo Möller and Julian Risch and Malte Pietsch}, year={2021}, eprint={2104.12741}, archivePrefix={arXiv}, primaryClass={cs.CL} }
null
21
1,123
--- thumbnail: >- https://thumb.tildacdn.com/tild3433-3637-4830-a533-353833613061/-/resize/720x/-/format/webp/germanquad.jpg language: - de multilinguality: - monolingual source_datasets: - original task_categories: - question-answering - text-retrieval task_ids: - extractive-qa - closed-domain-qa - open-domain-qa train-eval-index: - config: plain_text task: question-answering task_id: extractive_question_answering splits: train_split: train eval_split: test col_mapping: context: context question: question answers.text: answers.text answers.answer_start: answers.answer_start license: cc-by-4.0 --- ![bert_image](https://thumb.tildacdn.com/tild3433-3637-4830-a533-353833613061/-/resize/720x/-/format/webp/germanquad.jpg) # Dataset Card for germanquad ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://deepset.ai/germanquad - **Repository:** https://github.com/deepset-ai/haystack - **Paper:** https://arxiv.org/abs/2104.12741 ### Dataset Summary In order to raise the bar for non-English QA, we are releasing a high-quality, human-labeled German QA dataset consisting of 13 722 questions, incl. a three-way annotated test set. The creation of GermanQuAD is inspired by insights from existing datasets as well as our labeling experience from several industry projects. We combine the strengths of SQuAD, such as high out-of-domain performance, with self-sufficient questions that contain all relevant information for open-domain QA as in the NaturalQuestions dataset. Our training and test datasets do not overlap like other popular datasets and include complex questions that cannot be answered with a single entity or only a few words. ### Supported Tasks and Leaderboards - `extractive-qa`, `closed-domain-qa`, `open-domain-qa`, `text-retrieval`: This dataset is intended to be used for `open-domain-qa`, but can also be used for information retrieval tasks. ### Languages The sentences in the dataset are in German (de). ## Dataset Structure ### Data Instances A sample from the training set is provided below: ``` { "paragraphs": [ { "qas": [ { "question": "Von welchem Gesetzt stammt das Amerikanische ab? ", "id": 51870, "answers": [ { "answer_id": 53778, "document_id": 43958, "question_id": 51870, "text": "britischen Common Laws", "answer_start": 146, "answer_category": "SHORT" } ], "is_impossible": false } ], "context": "Recht_der_Vereinigten_Staaten\ \ === Amerikanisches Common Law ===\ Obwohl die Vereinigten Staaten wie auch viele Staaten des Commonwealth Erben des britischen Common Laws sind, setzt sich das amerikanische Recht bedeutend davon ab. Dies rührt größtenteils von dem langen Zeitraum her, in dem sich das amerikanische Recht unabhängig vom Britischen entwickelt hat. Entsprechend schauen die Gerichte in den Vereinigten Staaten bei der Analyse von eventuell zutreffenden britischen Rechtsprinzipien im Common Law gewöhnlich nur bis ins frühe 19. Jahrhundert.\ Während es in den Commonwealth-Staaten üblich ist, dass Gerichte sich Entscheidungen und Prinzipien aus anderen Commonwealth-Staaten importieren, ist das in der amerikanischen Rechtsprechung selten. Ausnahmen bestehen hier nur, wenn sich überhaupt keine relevanten amerikanischen Fälle finden lassen, die Fakten nahezu identisch sind und die Begründung außerordentlich überzeugend ist. Frühe amerikanische Entscheidungen zitierten oft britische Fälle, solche Zitate verschwanden aber während des 19. Jahrhunderts, als die Gerichte eindeutig amerikanische Lösungen zu lokalen Konflikten fanden. In der aktuellen Rechtsprechung beziehen sich fast alle Zitate auf amerikanische Fälle.\ Einige Anhänger des Originalismus und der strikten Gesetzestextauslegung (''strict constructionism''), wie zum Beispiel der verstorbene Bundesrichter am Obersten Gerichtshof, Antonin Scalia, vertreten die Meinung, dass amerikanische Gerichte ''nie'' ausländische Fälle überprüfen sollten, die nach dem Unabhängigkeitskrieg entschieden wurden, unabhängig davon, ob die Argumentation überzeugend ist oder nicht. Die einzige Ausnahme wird hier in Fällen gesehen, die durch die Vereinigten Staaten ratifizierte völkerrechtliche Verträge betreffen. Andere Richter, wie zum Beispiel Anthony Kennedy und Stephen Breyer vertreten eine andere Ansicht und benutzen ausländische Rechtsprechung, sofern ihre Argumentation für sie überzeugend, nützlich oder hilfreich ist.", "document_id": 43958 } ] }, ``` ### Data Fields - `id`: a `string` feature. - `context`: a `string` feature. - `question`: a `string` feature. - `answers`: a dictionary feature containing: - `text`: a `string` feature. - `answer_start`: a `int32` feature. ### Data Splits The dataset is split into a one-way annotated training set and a three-way annotated test set of German Wikipedia passages (paragraphs). Each passage is from a different article. | |passages|questions|answers| |----------|----:|---------:|---------:| |train|2540| 11518|11518| |test|474| 2204|6536| ## Additional Information ### Dataset Curators The dataset was initially created by Timo Möller, Julian Risch, Malte Pietsch, Julian Gutsch, Tom Hersperger, Luise Köhler, Iuliia Mozhina, and Justus Peter, during work done at deepset.ai ### Citation Information ``` @misc{möller2021germanquad, title={GermanQuAD and GermanDPR: Improving Non-English Question Answering and Passage Retrieval}, author={Timo Möller and Julian Risch and Malte Pietsch}, year={2021}, eprint={2104.12741}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
mstz/arcene
2023-04-17T08:46:30.000Z
[ "task_categories:tabular-classification", "size_categories:n<1K", "language:en", "arcene", "tabular_classification", "binary_classification", "UCI", "region:us" ]
mstz
null
@misc{misc_arcene_167, author = {Guyon,Isabelle, Gunn,Steve, Ben-Hur,Asa & Dror,Gideon}, title = {{Arcene}}, year = {2008}, howpublished = {UCI Machine Learning Repository}, note = {{DOI}: \\url{10.24432/C58P55}} }
null
0
1,123
--- language: - en tags: - arcene - tabular_classification - binary_classification - UCI pretty_name: Arcene size_categories: - n<1K task_categories: # Full list at https://github.com/huggingface/hub-docs/blob/main/js/src/lib/interfaces/Types.ts - tabular-classification configs: - arcene --- # Arcene The [Arcene dataset](https://archive-beta.ics.uci.edu/dataset/167/arcene) from the [UCI repository](https://archive-beta.ics.uci.edu/).
nampdn-ai/tiny-textbooks
2023-10-04T03:56:50.000Z
[ "task_categories:text-generation", "size_categories:100K<n<1M", "language:en", "license:cc-by-nc-sa-4.0", "arxiv:2309.05463", "arxiv:2306.01116", "arxiv:2304.08442", "arxiv:2305.07759", "doi:10.57967/hf/1126", "region:us" ]
nampdn-ai
null
null
null
45
1,117
--- task_categories: - text-generation language: - en pretty_name: Tiny Textbooks size_categories: - 100K<n<1M license: cc-by-nc-sa-4.0 --- # Textbook-like Dataset: A High-Quality Resource for Small Language Models The idea is simply inspired by the [Textbooks Are All You Need II: phi-1.5 technical report](https://arxiv.org/abs/2309.05463) paper. The source texts in this dataset have been gathered and carefully select the best of the [falcon-refinedweb](https://arxiv.org/abs/2306.01116) and [minipile](https://arxiv.org/abs/2304.08442) datasets to ensure the diversity, quality while tiny in size. The dataset was synthesized using 4x3090 Ti cards over a period of 500 hours, thanks to [Nous-Hermes-Llama2-13b](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-13b) finetuned model. Why settle for low-quality text when you can train on a high-quality, textbook-like dataset? Training language models on subpar text can lead to several issues: 1. **Noise**: Such text often contains typos, grammatical errors, and poorly structured sentences, which can confuse models and degrade performance. 2. **Misinformation**: Low-quality web text may contain incorrect or misleading information, leading to models propagating these inaccuracies. 3. **Lack of Depth**: Subpar text often lacks the depth and detail found in high-quality content, limiting a model's understanding of complex topics. Conversely, training on my clean and high-quality dataset offers numerous advantages: 1. **Accuracy**: The theoretical concepts in my dataset provide near accurate and detailed information, akin to a well-written textbook. (Need more contribute for facts check) 2. **Context**: Practical examples demonstrate how these concepts apply in real-world situations, offering valuable context. 3. **Performance**: Models trained on high-quality data can generate more accurate, insightful, and human-like text. A standout feature of this dataset is its volume. It boasts a whopping **420,000 textbook documents**. This extensive collection ensures a wide coverage of topics and concepts, providing your models with a comprehensive and diverse learning resource. Moreover, this dataset is generated using an open-source language model, ensuring the data is open for every researcher to process. I love the openness and that's why I want to contribute this dataset for the community to push over the limit. Quality over quantity is a principle that holds true even in machine learning. Training on a large amount of low-quality tokens can lead to models learning and propagating the noise, inaccuracies, and poor structures present in the bad text. This can result in models that generate less accurate and less coherent outputs. On the other hand, training on a smaller amount of high-quality tokens, like those in this dataset, can yield significantly better results. High-quality tokens provide accurate, well-structured, and meaningful information from which models can learn effectively. This leads to models that can generate more accurate, insightful, and human-like text. In essence, it's about making every token count. Each high-quality token that a model learns from is a step towards better performance. So why waste computational resources and learning capacity on bad tokens when you can focus on high-quality ones? It's a more efficient and effective approach to training language models. Choosing high-quality dataset over low-quality web text is akin to opting for a reliable textbook over scattered internet articles. This choice can significantly enhance the performance and reliability of your causal language models. I'm excited to present this unique blend of theoretical concepts and practical examples designed to supercharge your causal language models. This isn't just another dataset; it's a high-quality resource that can help your models learn more effectively and with better common sense. I hope this dataset is an useful resource for ML researchers working with small causal language models. I eagerly await your feedback and suggestions as I continue to refine and expand the dataset. Together, let's push the boundaries of what's possible with a **tiny language models**! ## Visualization [Nomic Atlas](https://atlas.nomic.ai/map/0348f3f7-9280-404f-b6d3-d0b5993a6693/846bcd82-fcc5-474d-b24b-82d1b791f80b) 230k data points visualized thanks to Nomic AI platform. ### Disclaimer While every effort has been made to ensure the accuracy of the information contained within this dataset, please note that it is provided 'as is' and without any warranties. The use of the `textbook` field in this dataset is intended for research purposes only. You are advised to verify any information obtained from this dataset before acting upon it. ## Tiny Series Explore the possibilities and limitations of building Small Language Models with these tiny gems of data! - [TinyStories](https://arxiv.org/abs/2305.07759): The paper that sparked my interest in the journey of the tiny-* series. - [tiny-codes](https://huggingface.co/datasets/nampdn-ai/tiny-codes): Collection of 1.6M short and clear code snippets that can help LLM models learn how to reason. - [tiny-orca-textbooks](https://huggingface.co/datasets/nampdn-ai/tiny-orca-textbooks): Synthetic textbook to help model learn in-context on how it should perform task the right way. - [tiny-webtext](https://huggingface.co/datasets/nampdn-ai/tiny-webtext): A 6GB (4.5M records) variety of diverse webtext enriched with critical thinking methods to make unbiased English dataset. - [tiny-lessons](https://huggingface.co/datasets/nampdn-ai/tiny-lessons): Subset of this dataset, various lessons about "things of internet" augmented in a bite-sized textbook Markdown format. - [tiny-bridgedict](https://huggingface.co/datasets/nampdn-ai/tiny-bridgedict): A dataset that links and transfers knowledge between English, Vietnamese, Chinese in a tiny multilingual models. ### Others small HQ datasets with textbook-like quality - [devdocs.io](https://huggingface.co/datasets/nampdn-ai/devdocs.io): FreeCodeCamp has provided 189k comprehensive API documentation across a wide range of tech stacks and programming languages. - [sciphi-python-textbook](https://huggingface.co/datasets/emrgnt-cmplxty/sciphi-python-textbook) - [textbook_quality_programming](https://huggingface.co/datasets/vikp/textbook_quality_programming) - [sciphi-textbooks-are-all-you-need](https://huggingface.co/datasets/emrgnt-cmplxty/sciphi-textbooks-are-all-you-need)
bigbio/ddi_corpus
2022-12-22T15:44:31.000Z
[ "multilinguality:monolingual", "language:en", "license:cc-by-nc-4.0", "region:us" ]
bigbio
The DDI corpus has been manually annotated with drugs and pharmacokinetics and pharmacodynamics interactions. It contains 1025 documents from two different sources: DrugBank database and MedLine.
@article{HERREROZAZO2013914, title = { The DDI corpus: An annotated corpus with pharmacological substances and drug-drug interactions }, author = { María Herrero-Zazo and Isabel Segura-Bedmar and Paloma Martínez and Thierry Declerck }, year = 2013, journal = {Journal of Biomedical Informatics}, volume = 46, number = 5, pages = {914--920}, doi = {https://doi.org/10.1016/j.jbi.2013.07.011}, issn = {1532-0464}, url = {https://www.sciencedirect.com/science/article/pii/S1532046413001123}, keywords = {Biomedical corpora, Drug interaction, Information extraction} }
null
1
1,113
--- language: - en bigbio_language: - English license: cc-by-nc-4.0 multilinguality: monolingual bigbio_license_shortname: CC_BY_NC_4p0 pretty_name: DDI Corpus homepage: https://github.com/isegura/DDICorpus bigbio_pubmed: True bigbio_public: True bigbio_tasks: - NAMED_ENTITY_RECOGNITION - RELATION_EXTRACTION --- # Dataset Card for DDI Corpus ## Dataset Description - **Homepage:** https://github.com/isegura/DDICorpus - **Pubmed:** True - **Public:** True - **Tasks:** NER,RE The DDI corpus has been manually annotated with drugs and pharmacokinetics and pharmacodynamics interactions. It contains 1025 documents from two different sources: DrugBank database and MedLine. ## Citation Information ``` @article{HERREROZAZO2013914, title = { The DDI corpus: An annotated corpus with pharmacological substances and drug-drug interactions }, author = { María Herrero-Zazo and Isabel Segura-Bedmar and Paloma Martínez and Thierry Declerck }, year = 2013, journal = {Journal of Biomedical Informatics}, volume = 46, number = 5, pages = {914--920}, doi = {https://doi.org/10.1016/j.jbi.2013.07.011}, issn = {1532-0464}, url = {https://www.sciencedirect.com/science/article/pii/S1532046413001123}, keywords = {Biomedical corpora, Drug interaction, Information extraction} } ```
argilla/news-summary
2023-03-16T09:36:12.000Z
[ "task_categories:summarization", "task_ids:news-articles-summarization", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:cc-by-nc-4.0", "region:us" ]
argilla
null
null
null
28
1,111
--- language: - en license: - cc-by-nc-4.0 size_categories: - 10K<n<100K source_datasets: - original task_categories: - summarization task_ids: - news-articles-summarization dataset_info: features: - name: text dtype: string - name: prediction list: - name: score dtype: float64 - name: text dtype: string - name: prediction_agent dtype: string - name: annotation dtype: 'null' - name: annotation_agent dtype: 'null' - name: id dtype: string - name: metadata dtype: 'null' - name: status dtype: string - name: event_timestamp dtype: timestamp[us] - name: metrics struct: - name: text_length dtype: int64 splits: - name: train num_bytes: 2563132.0446374374 num_examples: 1000 - name: test num_bytes: 52331466.955362566 num_examples: 20417 download_size: 33207109 dataset_size: 54894599.0 --- # Dataset Card for "news-summary" ## Dataset Description - **Homepage:** Kaggle Challenge - **Repository:** https://www.kaggle.com/datasets/clmentbisaillon/fake-and-real-news-dataset?select=True.csv - **Paper:** N.A. - **Leaderboard:** N.A. - **Point of Contact:** N.A. ### Dataset Summary Officially it was supposed to be used for classification but, can you use this data set to summarize news articles? ### Languages english ### Citation Information Acknowledgements Ahmed H, Traore I, Saad S. “Detecting opinion spams and fake news using text classification”, Journal of Security and Privacy, Volume 1, Issue 1, Wiley, January/February 2018. Ahmed H, Traore I, Saad S. (2017) “Detection of Online Fake News Using N-Gram Analysis and Machine Learning Techniques. In: Traore I., Woungang I., Awad A. (eds) Intelligent, Secure, and Dependable Systems in Distributed and Cloud Environments. ISDDC 2017. Lecture Notes in Computer Science, vol 10618. Springer, Cham (pp. 127-138). ### Contributions Thanks to [@davidberenstein1957](https://github.com/davidberenstein1957) for adding this dataset.
alzoubi36/opp_115
2023-06-24T07:08:08.000Z
[ "region:us" ]
alzoubi36
null
null
null
0
1,110
--- dataset_info: features: - name: text dtype: string - name: label sequence: int64 splits: - name: train num_bytes: 1047118 num_examples: 2185 - name: validation num_bytes: 270827 num_examples: 550 - name: test num_bytes: 316635 num_examples: 697 download_size: 811600 dataset_size: 1634580 --- # Dataset for the OPP-115 task in the [PrivacyGLUE](https://github.com/infsys-lab/privacy-glue) dataset
knowrohit07/know_sql
2023-09-20T20:13:06.000Z
[ "license:openrail", "region:us" ]
knowrohit07
null
null
null
78
1,108
--- license: openrail --- please use the val ign file for training, its much cleaner. thanks :)
ncbi_disease
2023-01-25T14:41:18.000Z
[ "task_categories:token-classification", "task_ids:named-entity-recognition", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:unknown", "region:us" ]
null
This paper presents the disease name and concept annotations of the NCBI disease corpus, a collection of 793 PubMed abstracts fully annotated at the mention and concept level to serve as a research resource for the biomedical natural language processing community. Each PubMed abstract was manually annotated by two annotators with disease mentions and their corresponding concepts in Medical Subject Headings (MeSH®) or Online Mendelian Inheritance in Man (OMIM®). Manual curation was performed using PubTator, which allowed the use of pre-annotations as a pre-step to manual annotations. Fourteen annotators were randomly paired and differing annotations were discussed for reaching a consensus in two annotation phases. In this setting, a high inter-annotator agreement was observed. Finally, all results were checked against annotations of the rest of the corpus to assure corpus-wide consistency. For more details, see: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3951655/ The original dataset can be downloaded from: https://www.ncbi.nlm.nih.gov/CBBresearch/Dogan/DISEASE/NCBI_corpus.zip This dataset has been converted to CoNLL format for NER using the following tool: https://github.com/spyysalo/standoff2conll Note: there is a duplicate document (PMID 8528200) in the original data, and the duplicate is recreated in the converted data.
@article{dougan2014ncbi, title={NCBI disease corpus: a resource for disease name recognition and concept normalization}, author={Dogan, Rezarta Islamaj and Leaman, Robert and Lu, Zhiyong}, journal={Journal of biomedical informatics}, volume={47}, pages={1--10}, year={2014}, publisher={Elsevier} }
null
18
1,107
--- annotations_creators: - expert-generated language_creators: - expert-generated language: - en license: - unknown multilinguality: - monolingual size_categories: - 1K<n<10K source_datasets: - original task_categories: - token-classification task_ids: - named-entity-recognition paperswithcode_id: ncbi-disease-1 pretty_name: NCBI Disease dataset_info: features: - name: id dtype: string - name: tokens sequence: string - name: ner_tags sequence: class_label: names: '0': O '1': B-Disease '2': I-Disease config_name: ncbi_disease splits: - name: train num_bytes: 2355516 num_examples: 5433 - name: validation num_bytes: 413900 num_examples: 924 - name: test num_bytes: 422842 num_examples: 941 download_size: 1546492 dataset_size: 3192258 train-eval-index: - config: ncbi_disease task: token-classification task_id: multi_class_classification splits: train_split: train eval_split: test col_mapping: tokens: text ner_tags: target metrics: - type: accuracy name: Accuracy - type: f1 name: F1 macro args: average: macro - type: f1 name: F1 micro args: average: micro - type: f1 name: F1 weighted args: average: weighted - type: precision name: Precision macro args: average: macro - type: precision name: Precision micro args: average: micro - type: precision name: Precision weighted args: average: weighted - type: recall name: Recall macro args: average: macro - type: recall name: Recall micro args: average: micro - type: recall name: Recall weighted args: average: weighted --- # Dataset Card for NCBI Disease ## Table of Contents - [Dataset Card for NCBI Disease](#dataset-card-for-ncbi-disease) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Initial Data Collection and Normalization](#initial-data-collection-and-normalization) - [Who are the source language producers?](#who-are-the-source-language-producers) - [Annotations](#annotations) - [Annotation process](#annotation-process) - [Who are the annotators?](#who-are-the-annotators) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [NCBI](https://www.ncbi.nlm.nih.gov/research/bionlp/Data/disease) - **Repository:** [Github](https://github.com/spyysalo/ncbi-disease) - **Paper:** [NCBI disease corpus: A resource for disease name recognition and concept normalization](https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3951655) - **Leaderboard:** [Named Entity Recognition on NCBI-disease](https://paperswithcode.com/sota/named-entity-recognition-ner-on-ncbi-disease) - **Point of Contact:** [email](zhiyong.lu@nih.gov) ### Dataset Summary This dataset contains the disease name and concept annotations of the NCBI disease corpus, a collection of 793 PubMed abstracts fully annotated at the mention and concept level to serve as a research resource for the biomedical natural language processing community. ### Supported Tasks and Leaderboards Named Entity Recognition: [Leaderboard](https://paperswithcode.com/sota/named-entity-recognition-ner-on-ncbi-disease) ### Languages The text in the dataset is in English. The associated BCP-47 code is `en`. ## Dataset Structure ### Data Instances Instances of the dataset contain an array of `tokens`, `ner_tags` and an `id`. An example of an instance of the dataset: ``` { 'tokens': ['Identification', 'of', 'APC2', ',', 'a', 'homologue', 'of', 'the', 'adenomatous', 'polyposis', 'coli', 'tumour', 'suppressor', '.'], 'ner_tags': [0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 0, 0], 'id': '0' } ``` ### Data Fields - `id`: Sentence identifier. - `tokens`: Array of tokens composing a sentence. - `ner_tags`: Array of tags, where `0` indicates no disease mentioned, `1` signals the first token of a disease and `2` the subsequent disease tokens. ### Data Splits The data is split into a train (5433 instances), validation (924 instances) and test set (941 instances). ## Dataset Creation ### Curation Rationale The goal of the dataset consists on improving the state-of-the-art in disease name recognition and normalization research, by providing a high-quality gold standard thus enabling the development of machine-learning based approaches for such tasks. ### Source Data #### Initial Data Collection and Normalization The dataset consists on abstracts extracted from PubMed. #### Who are the source language producers? The source language producers are the authors of publication abstracts hosted in PubMed. ### Annotations #### Annotation process Each PubMed abstract was manually annotated by two annotators with disease mentions and their corresponding concepts in Medical Subject Headings (MeSH®) or Online Mendelian Inheritance in Man (OMIM®). Manual curation was performed using PubTator, which allowed the use of pre-annotations as a pre-step to manual annotations. Fourteen annotators were randomly paired and differing annotations were discussed for reaching a consensus in two annotation phases. Finally, all results were checked against annotations of the rest of the corpus to assure corpus-wide consistency. #### Who are the annotators? The annotator group consisted of 14 people with backgrounds in biomedical informatics research and experience in biomedical text corpus annotation. ### Personal and Sensitive Information [N/A] ## Considerations for Using the Data ### Social Impact of Dataset Information encoded in natural language in biomedical literature publications is only useful if efficient and reliable ways of accessing and analyzing that information are available. Natural language processing and text mining tools are therefore essential for extracting valuable information. This dataset provides an annotated corpora that can be used to develop highly effective tools to automatically detect central biomedical concepts such as diseases. ### Discussion of Biases To avoid annotator bias, pairs of annotators were chosen randomly for each set, so that each pair of annotators overlapped for at most two sets. ### Other Known Limitations A handful of disease concepts were discovered that were not included in MEDIC. For those, we decided to include the appropriate OMIM identifiers. In addition, certain disease mentions were found to not be easily represented using the standard categorizations. Also, each PMID document was pre-annotated using the Inference Method developed for disease name normalization, which properly handles abbreviation recognition, robust string matching, etc. As such, human annotators were given the pre-annotated documents as a starting point and allowed to see each pre-annotation with a computed confidence. ## Additional Information ### Dataset Curators Rezarta Islamaj Doğan, Robert Leaman, Zhiyong Lu ### Licensing Information ``` PUBLIC DOMAIN NOTICE This work is a "United States Government Work" under the terms of the United States Copyright Act. It was written as part of the authors' official duties as a United States Government employee and thus cannot be copyrighted within the United States. The data is freely available to the public for use. The National Library of Medicine and the U.S. Government have not placed any restriction on its use or reproduction. Although all reasonable efforts have been taken to ensure the accuracy and reliability of the data and its source code, the NLM and the U.S. Government do not and cannot warrant the performance or results that may be obtained by using it. The NLM and the U.S. Government disclaim all warranties, express or implied, including warranties of performance, merchantability or fitness for any particular purpose. Please cite the authors in any work or product based on this material: An improved corpus of disease mentions in PubMed citations http://aclweb.org/anthology-new/W/W12/W12-2411.pdf NCBI Disease Corpus: A Resource for Disease Name Recognition and Normalization http://www.ncbi.nlm.nih.gov/pubmed/24393765 Disease Name Normalization with Pairwise Learning to Rank http://www.ncbi.nlm.nih.gov/pubmed/23969135 ``` ### Citation Information ``` @article{dougan2014ncbi, title={NCBI disease corpus: a resource for disease name recognition and concept normalization}, author={Do{\u{g}}an, Rezarta Islamaj and Leaman, Robert and Lu, Zhiyong}, journal={Journal of biomedical informatics}, volume={47}, pages={1--10}, year={2014}, publisher={Elsevier} } ``` ### Contributions Thanks to [@edugp](https://github.com/edugp) for adding this dataset.
unicamp-dl/mmarco
2022-11-30T17:31:26.000Z
[ "arxiv:2108.13897", "arxiv:2105.06813", "region:us" ]
unicamp-dl
mMARCO translated datasets
@misc{bonifacio2021mmarco, title={mMARCO: A Multilingual Version of the MS MARCO Passage Ranking Dataset}, author={Luiz Henrique Bonifacio and Israel Campiotti and Vitor Jeronymo and Hugo Queiroz Abonizio and Roberto Lotufo and Rodrigo Nogueira}, year={2021}, eprint={2108.13897}, archivePrefix={arXiv}, primaryClass={cs.CL} }
null
35
1,104
# Dataset Summary **mMARCO** is a multilingual version of the [MS MARCO passage ranking dataset](https://microsoft.github.io/msmarco/). For more information, checkout our papers: * [**mMARCO: A Multilingual Version of the MS MARCO Passage Ranking Dataset**](https://arxiv.org/abs/2108.13897) * [**A cost-benefit analysis of cross-lingual transfer methods**](https://arxiv.org/abs/2105.06813) The first (deprecated) version comprises 8 languages: Chinese, French, German, Indonesian, Italian, Portuguese, Russian and Spanish. The current version included translations for Japanese, Dutch, Vietnamese, Hindi and Arabic. The current version is composed of 14 languages (including the original English version). ### Supported languages | Language name | Language code | |---------------|---------------| | English | english | | Chinese | chinese | | French | french | | German | german | | Indonesian | indonesian | | Italian | italian | | Portuguese | portuguese | | Russian | russian | | Spanish | spanish | | Arabic | arabic | | Dutch | dutch | | Hindi | hindi | | Japanese | japanese | | Vietnamese | vietnamese | # Dataset Structure You can load mMARCO dataset by choosing a specific language. We include training triples (query, positive and negative example), the translated collections of documents and queries. #### Training triples ```python >>> dataset = load_dataset('unicamp-dl/mmarco', 'english') >>> dataset['train'][1] {'query': 'what fruit is native to australia', 'positive': 'Passiflora herbertiana. A rare passion fruit native to Australia. Fruits are green-skinned, white fleshed, with an unknown edible rating. Some sources list the fruit as edible, sweet and tasty, while others list the fruits as being bitter and inedible.assiflora herbertiana. A rare passion fruit native to Australia. Fruits are green-skinned, white fleshed, with an unknown edible rating. Some sources list the fruit as edible, sweet and tasty, while others list the fruits as being bitter and inedible.', 'negative': 'The kola nut is the fruit of the kola tree, a genus (Cola) of trees that are native to the tropical rainforests of Africa.'} ``` #### Queries ```python >>> dataset = load_dataset('unicamp-dl/mmarco', 'queries-spanish') >>> dataset['train'][1] {'id': 634306, 'text': '¿Qué significa Chattel en el historial de crédito'} ``` #### Collection ```python >>> dataset = load_dataset('unicamp-dl/mmarco', 'collection-portuguese') >>> dataset['collection'][100] {'id': 100, 'text': 'Antonín Dvorák (1841-1904) Antonin Dvorak era filho de açougueiro, mas ele não seguiu o negócio de seu pai. Enquanto ajudava seu pai a meio tempo, estudou música e se formou na Escola de Órgãos de Praga em 1859.'} ``` # Citation Information ``` @misc{bonifacio2021mmarco, title={mMARCO: A Multilingual Version of MS MARCO Passage Ranking Dataset}, author={Luiz Henrique Bonifacio and Vitor Jeronymo and Hugo Queiroz Abonizio and Israel Campiotti and Marzieh Fadaee and and Roberto Lotufo and Rodrigo Nogueira}, year={2021}, eprint={2108.13897}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
augtoma/usmle_step_1
2023-08-11T21:25:08.000Z
[ "region:us" ]
augtoma
null
null
null
0
1,102
--- configs: - config_name: default data_files: - split: test path: data/test-* dataset_info: features: - name: question dtype: string - name: options struct: - name: A dtype: string - name: B dtype: string - name: C dtype: string - name: D dtype: string - name: E dtype: string - name: F dtype: string - name: G dtype: string - name: H dtype: string - name: I dtype: string - name: answer dtype: string - name: answer_idx dtype: string splits: - name: test num_bytes: 80576 num_examples: 94 download_size: 60551 dataset_size: 80576 --- # Dataset Card for "usmle_self_eval_step1" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
cats_vs_dogs
2023-01-25T14:27:39.000Z
[ "task_categories:image-classification", "task_ids:multi-class-image-classification", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:unknown", "region:us" ]
null
null
@Inproceedings (Conference){asirra-a-captcha-that-exploits-interest-aligned-manual-image-categorization, author = {Elson, Jeremy and Douceur, John (JD) and Howell, Jon and Saul, Jared}, title = {Asirra: A CAPTCHA that Exploits Interest-Aligned Manual Image Categorization}, booktitle = {Proceedings of 14th ACM Conference on Computer and Communications Security (CCS)}, year = {2007}, month = {October}, publisher = {Association for Computing Machinery, Inc.}, url = {https://www.microsoft.com/en-us/research/publication/asirra-a-captcha-that-exploits-interest-aligned-manual-image-categorization/}, edition = {Proceedings of 14th ACM Conference on Computer and Communications Security (CCS)}, }
null
13
1,098
--- annotations_creators: - crowdsourced language_creators: - crowdsourced language: - en license: - unknown multilinguality: - monolingual size_categories: - 10K<n<100K source_datasets: - original task_categories: - image-classification task_ids: - multi-class-image-classification paperswithcode_id: cats-vs-dogs pretty_name: Cats Vs. Dogs dataset_info: features: - name: image dtype: image - name: labels dtype: class_label: names: '0': cat '1': dog splits: - name: train num_bytes: 4219400 num_examples: 23410 download_size: 824887076 dataset_size: 4219400 --- # Dataset Card for Cats Vs. Dogs ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Cats vs Dogs Dataset](https://www.microsoft.com/en-us/download/details.aspx?id=54765) - **Repository:** - **Paper:** [Asirra: A CAPTCHA that Exploits Interest-Aligned Manual Image Categorization](https://www.microsoft.com/en-us/research/wp-content/uploads/2007/10/CCS2007.pdf) - **Leaderboard:** [Dogs vs. Cats](https://www.kaggle.com/competitions/dogs-vs-cats) - **Point of Contact:** ### Dataset Summary A large set of images of cats and dogs. There are 1738 corrupted images that are dropped. This dataset is part of a now-closed Kaggle competition and represents a subset of the so-called Asirra dataset. From the competition page: > The Asirra data set > > Web services are often protected with a challenge that's supposed to be easy for people to solve, but difficult for computers. Such a challenge is often called a [CAPTCHA](http://www.captcha.net/) (Completely Automated Public Turing test to tell Computers and Humans Apart) or HIP (Human Interactive Proof). HIPs are used for many purposes, such as to reduce email and blog spam and prevent brute-force attacks on web site passwords. > > Asirra (Animal Species Image Recognition for Restricting Access) is a HIP that works by asking users to identify photographs of cats and dogs. This task is difficult for computers, but studies have shown that people can accomplish it quickly and accurately. Many even think it's fun! Here is an example of the Asirra interface: > > Asirra is unique because of its partnership with [Petfinder.com](https://www.petfinder.com/), the world's largest site devoted to finding homes for homeless pets. They've provided Microsoft Research with over three million images of cats and dogs, manually classified by people at thousands of animal shelters across the United States. Kaggle is fortunate to offer a subset of this data for fun and research. ### Supported Tasks and Leaderboards - `image-classification`: The goal of this task is to classify a given image as either containing a cat or a dog. The leaderboard is available [here](https://paperswithcode.com/sota/image-classification-on-cats-vs-dogs). ### Languages English. ## Dataset Structure ### Data Instances A sample from the training set is provided below: ``` { 'image': <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=500x375 at 0x29CEAD71780>, 'labels': 0 } ``` ### Data Fields The data instances have the following fields: - `image`: A `PIL.Image.Image` object containing the image. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]`. - `labels`: an `int` classification label. Class Label Mappings: ``` { "cat": 0, "dog": 1, } ``` ### Data Splits | | train | |---------------|------:| | # of examples | 23410 | ## Dataset Creation ### Curation Rationale This subset was to built to test whether computer vision algorithms can beat the Asirra CAPTCHA: From the competition page: > Image recognition attacks > > While random guessing is the easiest form of attack, various forms of image recognition can allow an attacker to make guesses that are better than random. There is enormous diversity in the photo database (a wide variety of backgrounds, angles, poses, lighting, etc.), making accurate automatic classification difficult. In an informal poll conducted many years ago, computer vision experts posited that a classifier with better than 60% accuracy would be difficult without a major advance in the state of the art. For reference, a 60% classifier improves the guessing probability of a 12-image HIP from 1/4096 to 1/459. ### Source Data #### Initial Data Collection and Normalization This dataset is a subset of the Asirra dataset. From the competition page: > Asirra is unique because of its partnership with Petfinder.com, the world's largest site devoted to finding homes for homeless pets. They've provided Microsoft Research with over three million images of cats and dogs, manually classified by people at thousands of animal shelters across the United States. #### Who are the source language producers? The users of [Petfinder.com](https://www.petfinder.com/). ### Annotations #### Annotation process The images were annotated by selecting a pet category on [Petfinder.com](https://www.petfinder.com/). #### Who are the annotators? The users of [Petfinder.com](https://www.petfinder.com/). ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases From the paper: > Unlike many image-based CAPTCHAs which are abstract or subjective, Asirra’s challenges are concrete, inoffensive (cute, by some accounts), require no specialized or culturally biased knowledge, and have definite ground truth. This makes Asirra less frustrating for humans. Some beta-testers found it fun. The four-year-old child of one asked several times to “play the cat and dog game again.” ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ```bibtex @Inproceedings (Conference){asirra-a-captcha-that-exploits-interest-aligned-manual-image-categorization, author = {Elson, Jeremy and Douceur, John (JD) and Howell, Jon and Saul, Jared}, title = {Asirra: A CAPTCHA that Exploits Interest-Aligned Manual Image Categorization}, booktitle = {Proceedings of 14th ACM Conference on Computer and Communications Security (CCS)}, year = {2007}, month = {October}, publisher = {Association for Computing Machinery, Inc.}, url = {https://www.microsoft.com/en-us/research/publication/asirra-a-captcha-that-exploits-interest-aligned-manual-image-categorization/}, edition = {Proceedings of 14th ACM Conference on Computer and Communications Security (CCS)}, } ``` ### Contributions Thanks to [@nateraw](https://github.com/nateraw) for adding this dataset.
Graphcore/gqa
2022-10-25T08:59:27.000Z
[ "language:en", "license:cc-by-4.0", "region:us" ]
Graphcore
GQA is a new dataset for real-world visual reasoning and compositional question answering, seeking to address key shortcomings of previous visual question answering (VQA) datasets.
@inproceedings{hudson2019gqa, title={Gqa: A new dataset for real-world visual reasoning and compositional question answering}, author={Hudson, Drew A and Manning, Christopher D}, booktitle={Proceedings of the IEEE/CVF conference on computer vision and pattern recognition}, pages={6700--6709}, year={2019} }
null
0
1,095
--- language: - en license: - cc-by-4.0 ---
cyrilzhang/TinyStories2-ascii-bpe-32k
2023-09-08T06:00:38.000Z
[ "region:us" ]
cyrilzhang
null
null
null
0
1,095
--- configs: - config_name: default data_files: - split: train path: data/train-* - split: validation path: data/validation-* dataset_info: features: - name: input_ids sequence: int32 splits: - name: train num_bytes: 2116666000 num_examples: 516260 - name: validation num_bytes: 21369200 num_examples: 5212 download_size: 881246333 dataset_size: 2138035200 --- # Dataset Card for "TinyStories2-ascii-bpe-32k" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
dongyoung4091/hh-generated_flan_t5_large_flan_t5_zeroshot
2023-09-08T11:53:45.000Z
[ "region:us" ]
dongyoung4091
null
null
null
0
1,087
--- dataset_info: features: - name: prompt dtype: string - name: response dtype: string - name: zeroshot_helpfulness dtype: float64 - name: zeroshot_specificity dtype: float64 - name: zeroshot_intent dtype: int64 - name: zeroshot_factuality dtype: int64 - name: zeroshot_easy-to-understand dtype: int64 - name: zeroshot_relevance dtype: int64 - name: zeroshot_readability dtype: float64 - name: zeroshot_enough-detail dtype: float64 - name: 'zeroshot_biased:' dtype: int64 - name: zeroshot_fail-to-consider-individual-preferences dtype: int64 - name: zeroshot_repetetive dtype: int64 - name: zeroshot_fail-to-consider-context dtype: int64 - name: zeroshot_too-long dtype: float64 splits: - name: train num_bytes: 6352357 num_examples: 25600 download_size: 798475 dataset_size: 6352357 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "hh-generated_flan_t5_large_flan_t5_zeroshot" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
cfilt/iitb-english-hindi
2022-04-26T13:50:22.000Z
[ "region:us" ]
cfilt
null
null
null
11
1,082
<p align="center"><img src="https://huggingface.co/datasets/cfilt/HiNER-collapsed/raw/main/cfilt-dark-vec.png" alt="Computation for Indian Language Technology Logo" width="150" height="150"/></p> # IITB-English-Hindi Parallel Corpus [![License: CC BY-NC 4.0](https://img.shields.io/badge/License-CC%20BY--NC%204.0-lightgrey.svg)](https://creativecommons.org/licenses/by-nc/4.0/) [![Twitter Follow](https://img.shields.io/twitter/follow/cfiltnlp?color=1DA1F2&logo=twitter&style=flat-square)](https://twitter.com/cfiltnlp) [![Twitter Follow](https://img.shields.io/twitter/follow/PeopleCentredAI?color=1DA1F2&logo=twitter&style=flat-square)](https://twitter.com/PeopleCentredAI) ## About The IIT Bombay English-Hindi corpus contains parallel corpus for English-Hindi as well as monolingual Hindi corpus collected from a variety of existing sources and corpora developed at the Center for Indian Language Technology, IIT Bombay over the years. This page describes the corpus. This corpus has been used at the Workshop on Asian Language Translation Shared Task since 2016 the Hindi-to-English and English-to-Hindi languages pairs and as a pivot language pair for the Hindi-to-Japanese and Japanese-to-Hindi language pairs. The complete details of this corpus are available at [this URL](https://www.cfilt.iitb.ac.in/iitb_parallel/). We also provide this parallel corpus via browser download from the same URL. We also provide a monolingual Hindi corpus on the same URL. ### Recent Updates * Version 3.1 - December 2021 - Added 49,400 sentence pairs to the parallel corpus. * Version 3.0 - August 2020 - Added ~47,000 sentence pairs to the parallel corpus. ## Usage We provide a notebook that shows how to import the IITB English-Hindi Parallel Corpus from the HuggingFace datasets repository. The notebook also shows how to segment the corpus using BPE tokenization which can be used to train an English-Hindi MT System. [https://github.com/cfiltnlp/IITB-English-Hindi-PC](https://github.com/cfiltnlp/IITB-English-Hindi-PC) ## Other You can find a catalogue of other English-Hindi and other Indian language parallel corpora here: [Indic NLP Catalog](https://github.com/indicnlpweb/indicnlp_catalog) ## Maintainer(s) [Diptesh Kanojia](https://dipteshkanojia.github.io)<br/> Shivam Mhasker<br/> ## Citation If you use this corpus or its derivate resources for your research, kindly cite it as follows: Anoop Kunchukuttan, Pratik Mehta, Pushpak Bhattacharyya. The IIT Bombay English-Hindi Parallel Corpus. Language Resources and Evaluation Conference. 2018. ### BiBTeX Citation ```latex @inproceedings{kunchukuttan-etal-2018-iit, title = "The {IIT} {B}ombay {E}nglish-{H}indi Parallel Corpus", author = "Kunchukuttan, Anoop and Mehta, Pratik and Bhattacharyya, Pushpak", booktitle = "Proceedings of the Eleventh International Conference on Language Resources and Evaluation ({LREC} 2018)", month = may, year = "2018", address = "Miyazaki, Japan", publisher = "European Language Resources Association (ELRA)", url = "https://aclanthology.org/L18-1548", } ```
crows_pairs
2023-07-06T09:23:23.000Z
[ "task_categories:text-classification", "task_ids:text-scoring", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:cc-by-sa-4.0", "bias-evaluation", "region:us" ]
null
CrowS-Pairs, a challenge dataset for measuring the degree to which U.S. stereotypical biases present in the masked language models (MLMs).
@inproceedings{nangia2020crows, title = "{CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language Models}", author = "Nangia, Nikita and Vania, Clara and Bhalerao, Rasika and Bowman, Samuel R.", booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing", month = nov, year = "2020", address = "Online", publisher = "Association for Computational Linguistics" }
null
3
1,078
--- annotations_creators: - crowdsourced language_creators: - crowdsourced language: - en license: - cc-by-sa-4.0 multilinguality: - monolingual size_categories: - 1K<n<10K source_datasets: - original task_categories: - text-classification task_ids: - text-scoring paperswithcode_id: crows-pairs pretty_name: CrowS-Pairs tags: - bias-evaluation dataset_info: features: - name: id dtype: int32 - name: sent_more dtype: string - name: sent_less dtype: string - name: stereo_antistereo dtype: class_label: names: '0': stereo '1': antistereo - name: bias_type dtype: class_label: names: '0': race-color '1': socioeconomic '2': gender '3': disability '4': nationality '5': sexual-orientation '6': physical-appearance '7': religion '8': age - name: annotations sequence: sequence: class_label: names: '0': race-color '1': socioeconomic '2': gender '3': disability '4': nationality '5': sexual-orientation '6': physical-appearance '7': religion '8': age - name: anon_writer dtype: string - name: anon_annotators sequence: string config_name: crows_pairs splits: - name: test num_bytes: 419976 num_examples: 1508 download_size: 437764 dataset_size: 419976 --- # Dataset Card for CrowS-Pairs ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Add homepage URL here if available (unless it's a GitHub repository)]() - **Repository:** https://github.com/nyu-mll/crows-pairs - **Paper:** https://aclanthology.org/2020.emnlp-main.154 - **Leaderboard:** [If the dataset supports an active leaderboard, add link here]() - **Point of Contact:** [If known, name and email of at least one person the reader can contact for questions about the dataset.]() ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information CrowS-Pairs is licensed under a [Creative Commons Attribution-ShareAlike 4.0 International License](https://creativecommons.org/licenses/by-sa/4.0/). It is created using prompts taken from the [ROCStories corpora](https://cs.rochester.edu/nlp/rocstories/) and the fiction part of [MNLI](https://cims.nyu.edu/~sbowman/multinli/). Please refer to their papers for more details. ### Citation Information ``` @inproceedings{nangia-etal-2020-crows, title = "{C}row{S}-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language Models", author = "Nangia, Nikita and Vania, Clara and Bhalerao, Rasika and Bowman, Samuel R.", booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)", month = nov, year = "2020", address = "Online", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2020.emnlp-main.154", doi = "10.18653/v1/2020.emnlp-main.154", pages = "1953--1967", } ``` ### Contributions Thanks to [@patil-suraj](https://github.com/patil-suraj) for adding this dataset.
lucasmccabe-lmi/CodeAlpaca-20k
2023-05-19T00:10:02.000Z
[ "region:us" ]
lucasmccabe-lmi
null
null
null
4
1,072
--- dataset_info: features: - name: instruction dtype: string - name: input dtype: string - name: output dtype: string splits: - name: train num_bytes: 6576710.0 num_examples: 20022 download_size: 3450938 dataset_size: 6576710.0 --- # Dataset Card for "CodeAlpaca-20k" We provide a minor modification of the [CodeAlpaca-20k](https://huggingface.co/datasets/sahil2801/CodeAlpaca-20k) dataset. In particular, we add the phrase, "Write corresponding code in Python." if the intended language is not explicitly stated. ## Numbers: Prompts: 20022 Tokens: 1561716 using the EleutherAI/gpt-neox-20b tokenizer (counting instruction+input+output)
conceptofmind/cot_submix_original
2023-04-28T22:57:04.000Z
[ "region:us" ]
conceptofmind
null
null
null
52
1,071
--- dataset_info: features: - name: inputs dtype: string - name: targets dtype: string - name: task_source dtype: string - name: task_name dtype: string - name: template_type dtype: string splits: - name: train num_bytes: 209004809 num_examples: 183848 download_size: 100293073 dataset_size: 209004809 --- # Dataset Card for "cot_submix_original" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Critiquers/gsm8k_pairwise
2023-08-23T19:29:20.000Z
[ "region:us" ]
Critiquers
null
null
null
1
1,063
--- dataset_info: features: - name: prompt dtype: string - name: selected dtype: string - name: rejected dtype: string splits: - name: train num_bytes: 411013 num_examples: 512 download_size: 234406 dataset_size: 411013 --- # Dataset Card for "gsm8k_pairwise" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
NeelNanda/counterfact-tracing
2022-11-05T15:19:43.000Z
[ "arxiv:2211.00593", "region:us" ]
NeelNanda
null
null
null
5
1,062
--- dataset_info: features: - name: relation dtype: string - name: relation_prefix dtype: string - name: relation_suffix dtype: string - name: prompt dtype: string - name: relation_id dtype: string - name: target_false_id dtype: string - name: target_true_id dtype: string - name: target_true dtype: string - name: target_false dtype: string - name: subject dtype: string splits: - name: train num_bytes: 3400668 num_examples: 21919 download_size: 1109314 dataset_size: 3400668 --- # Dataset Card for "counterfact-tracing" This is adapted from the counterfact dataset from the excellent [ROME paper](https://rome.baulab.info/) from David Bau and Kevin Meng. This is a dataset of 21919 factual relations, formatted as `data["prompt"]==f"{data['relation_prefix']}{data['subject']}{data['relation_suffix']}"`. Each has two responses `data["target_true"]` and `data["target_false"]` which is intended to go immediately after the prompt. The dataset was originally designed for memory editing in models. I made this for a research project doing mechanistic interpretability of how models recall factual knowledge, building on their causal tracing technique, and so stripped their data down to the information relevant to causal tracing. I also prepended spaces where relevant so that the subject and targets can be properly tokenized as is (spaces are always prepended to targets, and are prepended to subjects unless the subject is at the start of a sentence). Each fact has both a true and false target. I recommend measuring the logit *difference* between the true and false target (at least, if it's a single token target!), so as to control for eg the parts of the model which identify that it's supposed to be giving a fact of this type at all. (Idea inspired by the excellent [Interpretability In the Wild](https://arxiv.org/abs/2211.00593) paper).
speechcolab/gigaspeech
2023-09-25T17:54:37.000Z
[ "task_categories:automatic-speech-recognition", "multilinguality:monolingual", "language:en", "license:apache-2.0", "arxiv:2106.06909", "region:us" ]
speechcolab
GigaSpeech is an evolving, multi-domain English speech recognition corpus with 10,000 hours of high quality labeled audio suitable for supervised training, and 40,000 hours of total audio suitable for semi-supervised and unsupervised training. Around 40,000 hours of transcribed audio is first collected from audiobooks, podcasts and YouTube, covering both read and spontaneous speaking styles, and a variety of topics, such as arts, science, sports, etc. A new forced alignment and segmentation pipeline is proposed to create sentence segments suitable for speech recognition training, and to filter out segments with low-quality transcription. For system training, GigaSpeech provides five subsets of different sizes, 10h, 250h, 1000h, 2500h, and 10000h. For our 10,000-hour XL training subset, we cap the word error rate at 4% during the filtering/validation stage, and for all our other smaller training subsets, we cap it at 0%. The DEV and TEST evaluation sets, on the other hand, are re-processed by professional human transcribers to ensure high transcription quality.
@article{DBLP:journals/corr/abs-2106-06909, author = {Guoguo Chen and Shuzhou Chai and Guanbo Wang and Jiayu Du and Wei{-}Qiang Zhang and Chao Weng and Dan Su and Daniel Povey and Jan Trmal and Junbo Zhang and Mingjie Jin and Sanjeev Khudanpur and Shinji Watanabe and Shuaijiang Zhao and Wei Zou and Xiangang Li and Xuchen Yao and Yongqing Wang and Yujun Wang and Zhao You and Zhiyong Yan}, title = {GigaSpeech: An Evolving, Multi-domain {ASR} Corpus with 10, 000 Hours of Transcribed Audio}, journal = {CoRR}, volume = {abs/2106.06909}, year = {2021}, url = {https://arxiv.org/abs/2106.06909}, eprinttype = {arXiv}, eprint = {2106.06909}, timestamp = {Wed, 29 Dec 2021 14:29:26 +0100}, biburl = {https://dblp.org/rec/journals/corr/abs-2106-06909.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} }
null
28
1,058
--- annotations_creators: [] language_creators: [] language: - en license: - apache-2.0 multilinguality: - monolingual pretty_name: Gigaspeech size_categories: [] source_datasets: [] task_categories: - automatic-speech-recognition extra_gated_prompt: |- SpeechColab does not own the copyright of the audio files. For researchers and educators who wish to use the audio files for non-commercial research and/or educational purposes, we can provide access through the Hub under certain conditions and terms. Terms of Access: The "Researcher" has requested permission to use the GigaSpeech database (the "Database") at Tsinghua University. In exchange for such permission, Researcher hereby agrees to the following terms and conditions: 1. Researcher shall use the Database only for non-commercial research and educational purposes. 2. The SpeechColab team and Tsinghua University make no representations or warranties regarding the Database, including but not limited to warranties of non-infringement or fitness for a particular purpose. 3. Researcher accepts full responsibility for his or her use of the Database and shall defend and indemnify the SpeechColab team and Tsinghua University, including their employees, Trustees, officers and agents, against any and all claims arising from Researcher's use of the Database, including but not limited to Researcher's use of any copies of copyrighted audio files that he or she may create from the Database. 4. Researcher may provide research associates and colleagues with access to the Database provided that they first agree to be bound by these terms and conditions. 5. The SpeechColab team and Tsinghua University reserve the right to terminate Researcher's access to the Database at any time. 6. If Researcher is employed by a for-profit, commercial entity, Researcher's employer shall also be bound by these terms and conditions, and Researcher hereby represents that he or she is fully authorized to enter into this agreement on behalf of such employer. !!! Please also fill out the Google Form https://forms.gle/UuGQAPyscGRrUMLq6 to request access to the Gigaspeech dataset. extra_gated_fields: Name: text Email: text Organization: text Address: text I hereby confirm that I have requested access via the Google Form provided above: checkbox I accept the terms of access: checkbox --- # Dataset Card for Gigaspeech ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) - [Terms of Access](#terms-of-access) ## Dataset Description - **Homepage:** https://github.com/SpeechColab/GigaSpeech - **Repository:** https://github.com/SpeechColab/GigaSpeech - **Paper:** https://arxiv.org/abs/2106.06909 - **Leaderboard:** https://github.com/SpeechColab/GigaSpeech#leaderboard - **Point of Contact:** [gigaspeech@speechcolab.org](mailto:gigaspeech@speechcolab.org) ## Dataset Description GigaSpeech is an evolving, multi-domain English speech recognition corpus with 10,000 hours of high quality labeled audio suitable for supervised training. The transcribed audio data is collected from audiobooks, podcasts and YouTube, covering both read and spontaneous speaking styles, and a variety of topics, such as arts, science, sports, etc. ### Example Usage The training split has several configurations of various size: XS, S, M, L, XL. See the Section on "Data Splits" for more information. To download the XS configuration: ```python from datasets import load_dataset gs = load_dataset("speechcolab/gigaspeech", "xs", use_auth_token=True) # see structure print(gs) # load audio sample on the fly audio_input = gs["train"][0]["audio"] # first decoded audio sample transcription = gs["train"][0]["text"] # first transcription ``` It is possible to download only the development or test data: ```python gs_dev = load_dataset("speechcolab/gigaspeech", "dev", use_auth_token=True) gs_test = load_dataset("speechcolab/gigaspeech", "test", use_auth_token=True) ``` ### Supported Tasks and Leaderboards - `automatic-speech-recognition`: The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). The task has an active leaderboard which can be found at https://github.com/SpeechColab/GigaSpeech#leaderboard and ranks models based on their WER. ### Languages Gigaspeech contains audio and transcription data in English. ## Dataset Structure ### Data Instances ```python { 'segment_id': 'YOU0000000315_S0000660', 'speaker': 'N/A', 'text': "AS THEY'RE LEAVING <COMMA> CAN KASH PULL ZAHRA ASIDE REALLY QUICKLY <QUESTIONMARK>", 'audio': { # in streaming mode 'path' will be 'xs_chunks_0000/YOU0000000315_S0000660.wav' 'path': '/home/user/.cache/huggingface/datasets/downloads/extracted/9d48cf31/xs_chunks_0000/YOU0000000315_S0000660.wav', 'array': array([0.0005188 , 0.00085449, 0.00012207, ..., 0.00125122, 0.00076294, 0.00036621], dtype=float32), 'sampling_rate': 16000 }, 'begin_time': 2941.889892578125, 'end_time': 2945.070068359375, 'audio_id': 'YOU0000000315', 'title': 'Return to Vasselheim | Critical Role: VOX MACHINA | Episode 43', 'url': 'https://www.youtube.com/watch?v=zr2n1fLVasU', 'source': 2, 'category': 24, 'original_full_path': 'audio/youtube/P0004/YOU0000000315.opus' } ``` ### Data Fields * segment_id (string) - string id of the segment. * speaker (string) - string id of the speaker (can be "N/A"). * text (string) - transcription of the segment. * begin_time (float) - start time of the segment in an original full audio. * end_time (float32) - end time of the segment in an original full audio. * audio (Audio feature) - a dictionary containing the path to the audio, the decoded audio array, and the sampling rate. In non-streaming mode (default), the path point to the locally extracted audio. In streaming mode, the path is the relative path of an audio. segment inside its archive (as files are not downloaded and extracted locally). * audio_id (string) - string idea of the original full audio. * title (string) - title of the original full audio. * url (string) - url of the original full audio. * source (ClassLabel) - id of the audio source. Sources are audiobook (0), podcast (1), and YouYube (2). * category (ClassLabel) - id of the audio category, categories are listed below. * original_full_path (string) - the relative path to the original full audio sample in the original data directory. Categories are assigned from the following labels: "People and Blogs", "Business", "Nonprofits and Activism", "Crime", "History", "Pets and Animals", "News and Politics", "Travel and Events", "Kids and Family", "Leisure", "N/A", "Comedy", "News and Politics", "Sports", "Arts", "Science and Technology", "Autos and Vehicles", "Science and Technology", "People and Blogs", "Music", "Society and Culture", "Education", "Howto and Style", "Film and Animation", "Gaming", "Entertainment", "Travel and Events", "Health and Fitness", "audiobook". ### Data Splits The dataset has three splits: train, evaluation (dev) and test. The train split has five configurations of various sizes: XS, S, M, L, XL. Larger subsets are supersets of smaller subsets, e.g., the L subset contains all the data from the M subset. #### Transcribed Training Subsets Size | Subset | Hours | Remarks | |:---------------:|:-------------:|:-------------| | XS | 10 | System building and debugging | | S | 250 | Quick research experiments | | M | 1,000 | Large-scale research experiments | | L | 2,500 | Medium-scale industrial experiments | | XL | 10,000 | Large-scale industrial experiments | #### Transcribed Evaluation Subsets | Subset | Hours | Remarks | |:------:|:-----:|:--------| | Dev | 12 | Randomly selected from the crawled Podcast and YouTube Data | | Test | 40 | Part of the subset was randomly selected from the crawled Podcast and YouTube data; part of it was manually collected through other channels to have better coverage. | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data | Audio Source | Transcribed Hours | Acoustic Condition | |:-------------|:----------------------:|:-------------------| | Audiobook | 2,655 | <li>Reading</li><li>Various ages and accents</li> | | Podcast | 3,498 | <li>Clean or background music</li><li>Indoor</li><li>Near-field</li><li>Spontaneous</li><li>Various ages and accents</li>| | YouTube | 3,845 | <li>Clean and noisy</li><li>Indoor and outdoor</li><li>Near- and far-field</li><li>Reading and spontaneous</li><li>Various ages and accents</li> | | ***Total*** | ***10,000*** || #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? Development and test subsets are annotated by professional human annotators. ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information SpeechColab does not own the copyright of the audio files. For researchers and educators who wish to use the audio files for non-commercial research and/or educational purposes, we can provide access through our site under certain conditions and terms. In general, when training a machine learning model on a given dataset, the license of the model is **independent** to that of the dataset. That is to say, speech recognition models trained on the GigaSpeech dataset may be eligible for commercial license, provided they abide to the 'Fair Use' terms of the underlying data and do not violate any explicit copyright restrictions. This is likely to be true in most use-cases. However, it is your responsiblity to verify the appropriate model license for your specific use-case by confirming that the dataset usage abides by the Fair Use terms. SpeechColab is not responsible for the license of any machine learning model trained on the GigaSpeech dataset. ### Citation Information Please cite this paper if you find this work useful: ```bibtext @inproceedings{GigaSpeech2021, title={GigaSpeech: An Evolving, Multi-domain ASR Corpus with 10,000 Hours of Transcribed Audio}, booktitle={Proc. Interspeech 2021}, year=2021, author={Guoguo Chen, Shuzhou Chai, Guanbo Wang, Jiayu Du, Wei-Qiang Zhang, Chao Weng, Dan Su, Daniel Povey, Jan Trmal, Junbo Zhang, Mingjie Jin, Sanjeev Khudanpur, Shinji Watanabe, Shuaijiang Zhao, Wei Zou, Xiangang Li, Xuchen Yao, Yongqing Wang, Yujun Wang, Zhao You, Zhiyong Yan} } ``` ### Contributions Thanks to [@polinaeterna](https://github.com/polinaeterna) and [@sanchit-gandhi](https://github.com/sanchit-gandhi) for adding this dataset. ## Terms of Access The "Researcher" has requested permission to use the GigaSpeech database (the "Database") at Tsinghua University. In exchange for such permission, Researcher hereby agrees to the following terms and conditions: 1. Researcher shall use the Database only for non-commercial research and educational purposes. 2. The SpeechColab team and Tsinghua University make no representations or warranties regarding the Database, including but not limited to warranties of non-infringement or fitness for a particular purpose. 3. Researcher accepts full responsibility for his or her use of the Database and shall defend and indemnify the SpeechColab team and Tsinghua University, including their employees, Trustees, officers and agents, against any and all claims arising from Researcher's use of the Database, including but not limited to Researcher's use of any copies of copyrighted audio files that he or she may create from the Database. 4. Researcher may provide research associates and colleagues with access to the Database provided that they first agree to be bound by these terms and conditions. 5. The SpeechColab team and Tsinghua University reserve the right to terminate Researcher's access to the Database at any time. 6. If Researcher is employed by a for-profit, commercial entity, Researcher's employer shall also be bound by these terms and conditions, and Researcher hereby represents that he or she is fully authorized to enter into this agreement on behalf of such employer.
dongyoung4091/shp-generated_flan_t5_large_flan_t5_zeroshot
2023-09-09T02:42:23.000Z
[ "region:us" ]
dongyoung4091
null
null
null
0
1,058
--- dataset_info: features: - name: prompt dtype: string - name: response dtype: string - name: zeroshot_helpfulness dtype: float64 - name: zeroshot_specificity dtype: float64 - name: zeroshot_intent dtype: float64 - name: zeroshot_factuality dtype: float64 - name: zeroshot_easy-to-understand dtype: int64 - name: zeroshot_relevance dtype: int64 - name: zeroshot_readability dtype: int64 - name: zeroshot_enough-detail dtype: int64 - name: 'zeroshot_biased:' dtype: float64 - name: zeroshot_fail-to-consider-individual-preferences dtype: float64 - name: zeroshot_repetetive dtype: float64 - name: zeroshot_fail-to-consider-context dtype: float64 - name: zeroshot_too-long dtype: int64 splits: - name: train num_bytes: 29519465 num_examples: 25600 download_size: 1900231 dataset_size: 29519465 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "shp-generated_flan_t5_large_flan_t5_zeroshot" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
assin2
2023-01-25T14:26:53.000Z
[ "task_categories:text-classification", "task_ids:text-scoring", "task_ids:natural-language-inference", "task_ids:semantic-similarity-scoring", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:pt", "license:unknown", "region:us" ]
null
The ASSIN 2 corpus is composed of rather simple sentences. Following the procedures of SemEval 2014 Task 1. The training and validation data are composed, respectively, of 6,500 and 500 sentence pairs in Brazilian Portuguese, annotated for entailment and semantic similarity. Semantic similarity values range from 1 to 5, and text entailment classes are either entailment or none. The test data are composed of approximately 3,000 sentence pairs with the same annotation. All data were manually annotated.
@inproceedings{real2020assin, title={The assin 2 shared task: a quick overview}, author={Real, Livy and Fonseca, Erick and Oliveira, Hugo Goncalo}, booktitle={International Conference on Computational Processing of the Portuguese Language}, pages={406--412}, year={2020}, organization={Springer} }
null
9
1,057
--- annotations_creators: - expert-generated language_creators: - found language: - pt license: - unknown multilinguality: - monolingual size_categories: - 1K<n<10K source_datasets: - original task_categories: - text-classification task_ids: - text-scoring - natural-language-inference - semantic-similarity-scoring paperswithcode_id: assin2 pretty_name: ASSIN 2 dataset_info: features: - name: sentence_pair_id dtype: int64 - name: premise dtype: string - name: hypothesis dtype: string - name: relatedness_score dtype: float32 - name: entailment_judgment dtype: class_label: names: '0': NONE '1': ENTAILMENT splits: - name: train num_bytes: 864816 num_examples: 6500 - name: test num_bytes: 339580 num_examples: 2448 - name: validation num_bytes: 66895 num_examples: 500 download_size: 2113646 dataset_size: 1271291 --- # Dataset Card for ASSIN 2 ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [ASSIN 2 homepage](https://sites.google.com/view/assin2) - **Repository:** [ASSIN 2 repository](https://sites.google.com/view/assin2) - **Paper:** [The ASSIN 2 shared task: a quick overview](https://drive.google.com/file/d/1ft1VU6xiVm-N58dfAp6FHWjQ4IvcXgqp/view) - **Point of Contact:** [Livy Real](mailto:livyreal@gmail.com) ### Dataset Summary The ASSIN 2 corpus is composed of rather simple sentences. Following the procedures of SemEval 2014 Task 1. The training and validation data are composed, respectively, of 6,500 and 500 sentence pairs in Brazilian Portuguese, annotated for entailment and semantic similarity. Semantic similarity values range from 1 to 5, and text entailment classes are either entailment or none. The test data are composed of approximately 3,000 sentence pairs with the same annotation. All data were manually annotated. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The language supported is Portuguese. ## Dataset Structure ### Data Instances An example from the ASSIN 2 dataset looks as follows: ``` { "entailment_judgment": 1, "hypothesis": "Uma criança está segurando uma pistola de água", "premise": "Uma criança risonha está segurando uma pistola de água e sendo espirrada com água", "relatedness_score": 4.5, "sentence_pair_id": 1 } ``` ### Data Fields - `sentence_pair_id`: a `int64` feature. - `premise`: a `string` feature. - `hypothesis`: a `string` feature. - `relatedness_score`: a `float32` feature. - `entailment_judgment`: a classification label, with possible values including `NONE`, `ENTAILMENT`. ### Data Splits The data is split into train, validation and test set. The split sizes are as follow: | Train | Val | Test | | ------ | ----- | ---- | | 6500 | 500 | 2448 | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ``` @inproceedings{real2020assin, title={The assin 2 shared task: a quick overview}, author={Real, Livy and Fonseca, Erick and Oliveira, Hugo Goncalo}, booktitle={International Conference on Computational Processing of the Portuguese Language}, pages={406--412}, year={2020}, organization={Springer} } ``` ### Contributions Thanks to [@jonatasgrosman](https://github.com/jonatasgrosman) for adding this dataset.
hippocrates/re_train
2023-10-09T16:55:29.000Z
[ "region:us" ]
hippocrates
null
null
null
0
1,056
--- configs: - config_name: default data_files: - split: train path: data/train-* - split: valid path: data/valid-* dataset_info: features: - name: id dtype: string - name: conversations list: - name: from dtype: string - name: value dtype: string - name: text dtype: string splits: - name: train num_bytes: 19219537 num_examples: 3572 - name: valid num_bytes: 1626844 num_examples: 305 download_size: 1753501 dataset_size: 20846381 --- # Dataset Card for "re_train" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
llm-book/wrime-sentiment
2023-10-06T00:56:38.000Z
[ "task_categories:text-classification", "size_categories:10K<n<100K", "language:ja", "region:us" ]
llm-book
null
null
null
1
1,053
--- task_categories: - text-classification language: - ja size_categories: - 10K<n<100K --- # Dataset Card for llm-book/wrime-sentiment 日本語の感情分析データセット WRIME を、ポジティブ/ネガティブの二値分類のタスクに加工したデータセットです。 GitHub リポジトリ [ids-cv/wrime](https://github.com/ids-cv/wrime) で公開されているデータセットを利用しています。 `Avg. Readers_Sentiment` の値が0より大きいものをポジティブ、0より小さいものをネガティブとラベル付をしています。 書籍『大規模言語モデル入門』のサンプルコードで利用することを想定しています。 詳しくは[書籍のGitHubリポジトリ](https://github.com/ghmagazine/llm-book)をご覧ください。 ## 使い方 以下のようにデータセットを読み込むことができます。 ```python from datasets import load_dataset dataset = load_dataset("hf_datasets/wrime-sentiment") print(dataset["train"].features["label"]) print(dataset) ``` ```python ClassLabel(names=['positive', 'negative'], id=None) DatasetDict({ train: Dataset({ features: ['sentence', 'label'], num_rows: 20149 }) validation: Dataset({ features: ['sentence', 'label'], num_rows: 1608 }) test: Dataset({ features: ['sentence', 'label'], num_rows: 1781 }) }) ``` デフォルトの設定では、元のデータセットから極性がニュートラルであるものを除いています。 `remove_netural=False`と指定することで、ニュートラルなデータも含めた三値分類のデータセットを読み込むことができます。 ```python from datasets import load_dataset dataset = load_dataset("hf_datasets/wrime-sentiment", remove_neutral=False) print(dataset["train"].features["label"]) print(dataset) ``` ```python ClassLabel(names=['positive', 'negative', 'neutral'], id=None) DatasetDict({ train: Dataset({ features: ['sentence', 'label'], num_rows: 30000 }) validation: Dataset({ features: ['sentence', 'label'], num_rows: 2500 }) test: Dataset({ features: ['sentence', 'label'], num_rows: 2500 }) }) ```
allenai/lila
2023-03-15T18:36:28.000Z
[ "license:cc-by-4.0", "region:us" ]
allenai
Līla is a comprehensive benchmark for mathematical reasoning with over 140K natural language questions annotated with Python programs and natural language instructions. The data set comes with multiple splits: Līla-IID (train, dev, test), Līla-OOD (train, dev, test), and Līla-Robust.
@INPROCEEDINGS{Mishra2022Lila, author = { Swaroop Mishra and Matthew Finlayson and Pan Lu and Leonard Tang and Sean Welleck and Chitta Baral and Tanmay Rajpurohit and Oyvind Tafjord and Ashish Sabharwal and Peter Clark and Ashwin Kalyan}, title = {Lila: A Unified Benchmark for Mathematical Reasoning}, booktitle = {Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP)}, year = {2022} }
null
15
1,050
--- license: cc-by-4.0 --- ## Dataset Description - **Repository:** [allenai/lila](https://github.com/allenai/lila) - **Paper:** [LILA: A Unified Benchmark for Mathematical Reasoning](https://aclanthology.org/2022.emnlp-main.392.pdf) - **Point of Contact:** [Matthew Finlayson](https://mattf1n.github.io/), [Sean Welleck](https://wellecks.com/) # Lila: A Unified Benchmark for Mathematical Reasoning - **Homepage: https://lila.apps.allenai.org/** - **Repository: https://github.com/allenai/lila** - **Paper: https://aclanthology.org/2022.emnlp-main.392.pdf** ### Licensing Information Creative Commons Attribution 4.0 International ### Citation Information Cite this dataset and the source datasets (see [sources.bib](https://github.com/allenai/Lila/blob/main/sources.bib)). ```bib @INPROCEEDINGS{Mishra2022Lila, author = { Swaroop Mishra and Matthew Finlayson and Pan Lu and Leonard Tang and Sean Welleck and Chitta Baral and Tanmay Rajpurohit and Oyvind Tafjord and Ashish Sabharwal and Peter Clark and Ashwin Kalyan}, title = {Lila: A Unified Benchmark for Mathematical Reasoning}, booktitle = {Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP)}, year = {2022} } ```
code_x_glue_cc_defect_detection
2022-11-18T19:31:11.000Z
[ "task_categories:text-classification", "task_ids:multi-class-classification", "annotations_creators:found", "language_creators:found", "multilinguality:other-programming-languages", "size_categories:10K<n<100K", "source_datasets:original", "language:code", "license:c-uda", "region:us" ]
null
Given a source code, the task is to identify whether it is an insecure code that may attack software systems, such as resource leaks, use-after-free vulnerabilities and DoS attack. We treat the task as binary classification (0/1), where 1 stands for insecure code and 0 for secure code. The dataset we use comes from the paper Devign: Effective Vulnerability Identification by Learning Comprehensive Program Semantics via Graph Neural Networks. We combine all projects and split 80%/10%/10% for training/dev/test.
@inproceedings{zhou2019devign, title={Devign: Effective vulnerability identification by learning comprehensive program semantics via graph neural networks}, author={Zhou, Yaqin and Liu, Shangqing and Siow, Jingkai and Du, Xiaoning and Liu, Yang}, booktitle={Advances in Neural Information Processing Systems}, pages={10197--10207}, year={2019}
null
5
1,046
--- annotations_creators: - found language_creators: - found language: - code license: - c-uda multilinguality: - other-programming-languages size_categories: - 10K<n<100K source_datasets: - original task_categories: - text-classification task_ids: - multi-class-classification pretty_name: CodeXGlueCcDefectDetection dataset_info: features: - name: id dtype: int32 - name: func dtype: string - name: target dtype: bool - name: project dtype: string - name: commit_id dtype: string splits: - name: train num_bytes: 45723487 num_examples: 21854 - name: validation num_bytes: 5582545 num_examples: 2732 - name: test num_bytes: 5646752 num_examples: 2732 download_size: 61685715 dataset_size: 56952784 --- # Dataset Card for "code_x_glue_cc_defect_detection" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits-sample-size) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://github.com/microsoft/CodeXGLUE/tree/main/Code-Code/Defect-detection ### Dataset Summary CodeXGLUE Defect-detection dataset, available at https://github.com/microsoft/CodeXGLUE/tree/main/Code-Code/Defect-detection Given a source code, the task is to identify whether it is an insecure code that may attack software systems, such as resource leaks, use-after-free vulnerabilities and DoS attack. We treat the task as binary classification (0/1), where 1 stands for insecure code and 0 for secure code. The dataset we use comes from the paper Devign: Effective Vulnerability Identification by Learning Comprehensive Program Semantics via Graph Neural Networks. We combine all projects and split 80%/10%/10% for training/dev/test. ### Supported Tasks and Leaderboards - `multi-class-classification`: The dataset can be used to train a model for detecting if code has a defect in it. ### Languages - C **programming** language ## Dataset Structure ### Data Instances An example of 'validation' looks as follows. ``` { "commit_id": "aa1530dec499f7525d2ccaa0e3a876dc8089ed1e", "func": "static void filter_mirror_setup(NetFilterState *nf, Error **errp)\n{\n MirrorState *s = FILTER_MIRROR(nf);\n Chardev *chr;\n chr = qemu_chr_find(s->outdev);\n if (chr == NULL) {\n error_set(errp, ERROR_CLASS_DEVICE_NOT_FOUND,\n \"Device '%s' not found\", s->outdev);\n qemu_chr_fe_init(&s->chr_out, chr, errp);", "id": 8, "project": "qemu", "target": true } ``` ### Data Fields In the following each data field in go is explained for each config. The data fields are the same among all splits. #### default |field name| type | description | |----------|------|------------------------------------------| |id |int32 | Index of the sample | |func |string| The source code | |target |bool | 0 or 1 (vulnerability or not) | |project |string| Original project that contains this code | |commit_id |string| Commit identifier in the original project| ### Data Splits | name |train|validation|test| |-------|----:|---------:|---:| |default|21854| 2732|2732| ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators https://github.com/microsoft, https://github.com/madlag ### Licensing Information Computational Use of Data Agreement (C-UDA) License. ### Citation Information ``` @inproceedings{zhou2019devign, title={Devign: Effective vulnerability identification by learning comprehensive program semantics via graph neural networks}, author={Zhou, Yaqin and Liu, Shangqing and Siow, Jingkai and Du, Xiaoning and Liu, Yang}, booktitle={Advances in Neural Information Processing Systems}, pages={10197--10207}, year={2019} ``` ### Contributions Thanks to @madlag (and partly also @ncoop57) for adding this dataset.
argilla/databricks-dolly-15k-curated-multilingual
2023-06-14T07:47:54.000Z
[ "task_categories:text-generation", "task_categories:text2text-generation", "size_categories:10K<n<100K", "language:es", "language:de", "language:fr", "license:cc-by-sa-3.0", "machine-translated", "instruction-following", "region:us" ]
argilla
null
null
null
33
1,046
--- dataset_info: features: - name: instruction dtype: string - name: context dtype: string - name: response dtype: string - name: category dtype: string - name: instruction_original_en dtype: string - name: context_original_en dtype: string - name: response_original_en dtype: string - name: id dtype: int64 splits: - name: de num_bytes: 25985140 num_examples: 15015 - name: en num_bytes: 24125109 num_examples: 15015 - name: es num_bytes: 25902709 num_examples: 15015 - name: fr num_bytes: 26704314 num_examples: 15015 download_size: 65586669 dataset_size: 102717272 license: cc-by-sa-3.0 task_categories: - text-generation - text2text-generation language: - es - de - fr tags: - machine-translated - instruction-following pretty_name: Databrick Dolly Instructions Multilingual size_categories: - 10K<n<100K --- # Dataset Card for "databricks-dolly-15k-curated-multilingual" A curated and multilingual version of the Databricks Dolly instructions dataset. It includes a programmatically and manually corrected version of the original `en` dataset. See below. **STATUS**: Currently, the original Dolly v2 English version has been curated combining automatic processing and collaborative human curation using Argilla (~400 records have been manually edited and fixed). The following graph shows a summary about the number of edited fields. ![Edited records](edited_records.png) ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage: https://huggingface.co/datasets/argilla/databricks-dolly-15k-multilingual/** - **Repository: https://huggingface.co/datasets/argilla/databricks-dolly-15k-multilingual/** - **Paper:** - **Leaderboard:** - **Point of Contact: contact@argilla.io, https://github.com/argilla-io/argilla** ### Dataset Summary This dataset collection is a curated and machine-translated version of the `databricks-dolly-15k` [dataset](https://github.com/databrickslabs/dolly/tree/master/data) originally created by Databricks, Inc. in 2023. The goal is to give practitioners a starting point for training open-source instruction-following models with better-quality English data and translated data beyond English. However, as the translation quality will not be perfect, we highly recommend dedicating time to curate and fix translation issues. Below we explain how to load the datasets into [Argilla for data curation and fixing](https://github.com/argilla-io/argilla). Additionally, we'll be improving the datasets made available here, with the help of different communities. Currently, the original English version has been curated combining automatic processing and collaborative human curation using Argilla (~400 records have been manually edited and fixed). The following graph shows a summary of the number of edited fields. The main issues (likely many issues still remaining) are the following: 1. Some labelers misunderstood the usage of the `context` field. This `context` field is used as part of the prompt for instruction-tuning and in other works it's called `input` (e.g., Alpaca). Likely, the name context, has led to some labelers using it to provide the full context of where they have extracted the response. This is problematic for some types of tasks (summarization, closed-qa or information-extraction) because sometimes the context is shorter than or unrelated to summaries, or the information cannot be extracted from the context (closed-qa, information-extraction). 2. Some labelers misunderstood the way to give instructions for summarization or closed-qa, for example, they ask: Who is Thomas Jefferson? then provide a very long context and a response equally long. We programmatically identified records with these potential issues and ran a campaign to fix it and as a result more than 400 records have been adapted. See below for statistics: ![Edited records](edited_records.png) As a result of this curation process the content of the fields has been reduced, counted in number of tokens, especially for the responses: ![Edited records](tokens_diff.png) If you want to browse and curate your dataset with Argilla, you can: 1. [Duplicate this Space](https://huggingface.co/spaces/argilla/dolly-multilingual-curation/settings?duplicate=true). IMPORTANT: The Space's Visibility need to be Public, but you can setup your own password and API KEYS [following this guide](https://docs.argilla.io/en/latest/getting_started/installation/deployments/huggingface-spaces.html#setting-up-secret-environment-variables). 2. Setup two secrets: `HF_TOKEN` and `LANG` for indicating the language split 3. Login with `admin`/`12345678` and start browsing and labelling. 4. Start labeling. Every 5 min the validations will be stored on a Hub dataset in your personal HF space. 5. Please get in touch to contribute fixes and improvements to the source datasets. There's one split per language: ```python from datasets import load_dataset # loads all splits load_dataset("argilla/databricks-dolly-15k-curate-multilingual") # loads Spanish splits load_dataset("argilla/databricks-dolly-15k-curated-multilingual", split="es") ``` ### Supported Tasks and Leaderboards As described in the README of the original dataset, this dataset can be used for: * Training LLMs * Synthetic Data Generation * Data Augmentation ### Languages Currently: `es`, `fr`, `de`, `en` Join Argilla [Slack community](https://join.slack.com/t/rubrixworkspace/shared_invite/zt-whigkyjn-a3IUJLD7gDbTZ0rKlvcJ5g) if you want to help us include other languages. ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits There's one split per language: ```python from datasets import load_dataset # loads all splits load_dataset("argilla/databricks-dolly-15k-multilingual") # loads Spanish splits load_dataset("argilla/databricks-dolly-15k-multilingual", split="es") ``` ## Dataset Creation These datasets have been translated using the DeepL API from the original English dataset between the 13th and 14th of April ### Curation Logbook * 28/04/23: Removed references from Wikipedia copy pastes for 8113 rows. Applied to context and response fields with the following regex: `r'\[[\w]+\]'` ### Source Data #### Initial Data Collection and Normalization Refer to the [original dataset](https://github.com/databrickslabs/dolly/tree/master/data) for more information. #### Who are the source language producers? [More Information Needed] ### Annotations Annotations are planned but not performed yet. #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information This dataset can be used for any purpose, whether academic or commercial, under the terms of the [Creative Commons Attribution-ShareAlike 3.0 Unported License](https://creativecommons.org/licenses/by-sa/3.0/legalcode). **Original dataset Owner: Databricks, Inc.** ### Citation Information [More Information Needed]
Yijia-Xiao/pii-wikidoc_patient_information
2023-09-12T22:24:25.000Z
[ "region:us" ]
Yijia-Xiao
null
null
null
2
1,046
--- dataset_info: features: - name: output dtype: string - name: instruction dtype: string - name: input dtype: string - name: cleaned_output dtype: string splits: - name: train num_bytes: 6017940 num_examples: 5942 download_size: 2993583 dataset_size: 6017940 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "pii-wikidoc_patient_information" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
climatebert/climate_sentiment
2023-04-18T14:37:00.000Z
[ "task_categories:text-classification", "task_ids:sentiment-classification", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:cc-by-nc-sa-4.0", "region:us" ]
climatebert
null
null
null
1
1,037
--- annotations_creators: - expert-generated language_creators: - found language: - en license: cc-by-nc-sa-4.0 multilinguality: - monolingual size_categories: - 1K<n<10K source_datasets: - original task_categories: - text-classification task_ids: - sentiment-classification pretty_name: ClimateSentiment dataset_info: features: - name: text dtype: string - name: label dtype: class_label: names: '0': risk '1': neutral '2': opportunity splits: - name: train num_bytes: 492077 num_examples: 1000 - name: test num_bytes: 174265 num_examples: 320 download_size: 373638 dataset_size: 666342 --- # Dataset Card for climate_sentiment ## Dataset Description - **Homepage:** [climatebert.ai](https://climatebert.ai) - **Repository:** - **Paper:** [papers.ssrn.com/sol3/papers.cfm?abstract_id=3998435](https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3998435) - **Leaderboard:** - **Point of Contact:** [Nicolas Webersinke](mailto:nicolas.webersinke@fau.de) ### Dataset Summary We introduce an expert-annotated dataset for classifying climate-related sentiment of climate-related paragraphs in corporate disclosures. ### Supported Tasks and Leaderboards The dataset supports a ternary sentiment classification task of whether a given climate-related paragraph has sentiment opportunity, neutral, or risk. ### Languages The text in the dataset is in English. ## Dataset Structure ### Data Instances ``` { 'text': '− Scope 3: Optional scope that includes indirect emissions associated with the goods and services supply chain produced outside the organization. Included are emissions from the transport of products from our logistics centres to stores (downstream) performed by external logistics operators (air, land and sea transport) as well as the emissions associated with electricity consumption in franchise stores.', 'label': 1 } ``` ### Data Fields - text: a climate-related paragraph extracted from corporate annual reports and sustainability reports - label: the label (0 -> risk, 1 -> neutral, 2 -> opportunity) ### Data Splits The dataset is split into: - train: 1,000 - test: 320 ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization Our dataset contains climate-related paragraphs extracted from financial disclosures by firms. We collect text from corporate annual reports and sustainability reports. For more information regarding our sample selection, please refer to the Appendix of our paper (see [citation](#citation-information)). #### Who are the source language producers? Mainly large listed companies. ### Annotations #### Annotation process For more information on our annotation process and annotation guidelines, please refer to the Appendix of our paper (see [citation](#citation-information)). #### Who are the annotators? The authors and students at Universität Zürich and Friedrich-Alexander-Universität Erlangen-Nürnberg with majors in finance and sustainable finance. ### Personal and Sensitive Information Since our text sources contain public information, no personal and sensitive information should be included. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators - Julia Anna Bingler - Mathias Kraus - Markus Leippold - Nicolas Webersinke ### Licensing Information This dataset is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International license (cc-by-nc-sa-4.0). To view a copy of this license, visit [creativecommons.org/licenses/by-nc-sa/4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/). If you are interested in commercial use of the dataset, please contact [markus.leippold@bf.uzh.ch](mailto:markus.leippold@bf.uzh.ch). ### Citation Information ```bibtex @techreport{bingler2023cheaptalk, title={How Cheap Talk in Climate Disclosures Relates to Climate Initiatives, Corporate Emissions, and Reputation Risk}, author={Bingler, Julia and Kraus, Mathias and Leippold, Markus and Webersinke, Nicolas}, type={Working paper}, institution={Available at SSRN 3998435}, year={2023} } ``` ### Contributions Thanks to [@webersni](https://github.com/webersni) for adding this dataset.
cbt
2023-06-01T14:59:53.000Z
[ "task_categories:other", "task_categories:question-answering", "task_ids:multiple-choice-qa", "annotations_creators:machine-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "size_categories:n<1K", "source_datasets:original", "language:en", "license:gfdl", "arxiv:1511.02301", "region:us" ]
null
The Children’s Book Test (CBT) is designed to measure directly how well language models can exploit wider linguistic context. The CBT is built from books that are freely available.
@misc{hill2016goldilocks, title={The Goldilocks Principle: Reading Children's Books with Explicit Memory Representations}, author={Felix Hill and Antoine Bordes and Sumit Chopra and Jason Weston}, year={2016}, eprint={1511.02301}, archivePrefix={arXiv}, primaryClass={cs.CL} }
null
9
1,033
--- pretty_name: Children’s Book Test (CBT) annotations_creators: - machine-generated language_creators: - found language: - en license: - gfdl multilinguality: - monolingual size_categories: - 100K<n<1M - n<1K source_datasets: - original task_categories: - other - question-answering task_ids: - multiple-choice-qa paperswithcode_id: cbt dataset_info: - config_name: raw features: - name: title dtype: string - name: content dtype: string splits: - name: train num_bytes: 25741580 num_examples: 98 - name: test num_bytes: 1528704 num_examples: 5 - name: validation num_bytes: 1182657 num_examples: 5 download_size: 120547669 dataset_size: 28452941 - config_name: V features: - name: sentences sequence: string - name: question dtype: string - name: answer dtype: string - name: options sequence: string splits: - name: train num_bytes: 252177649 num_examples: 105825 - name: test num_bytes: 5806625 num_examples: 2500 - name: validation num_bytes: 4556425 num_examples: 2000 download_size: 120547669 dataset_size: 262540699 - config_name: P features: - name: sentences sequence: string - name: question dtype: string - name: answer dtype: string - name: options sequence: string splits: - name: train num_bytes: 852852601 num_examples: 334030 - name: test num_bytes: 6078048 num_examples: 2500 - name: validation num_bytes: 4776981 num_examples: 2000 download_size: 120547669 dataset_size: 863707630 - config_name: NE features: - name: sentences sequence: string - name: question dtype: string - name: answer dtype: string - name: options sequence: string splits: - name: train num_bytes: 253551931 num_examples: 108719 - name: test num_bytes: 5707734 num_examples: 2500 - name: validation num_bytes: 4424316 num_examples: 2000 download_size: 120547669 dataset_size: 263683981 - config_name: CN features: - name: sentences sequence: string - name: question dtype: string - name: answer dtype: string - name: options sequence: string splits: - name: train num_bytes: 301730151 num_examples: 120769 - name: test num_bytes: 6138376 num_examples: 2500 - name: validation num_bytes: 4737257 num_examples: 2000 download_size: 120547669 dataset_size: 312605784 config_names: - CN - NE - P - V - raw --- # Dataset Card for CBT ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:**[The bAbI project](https://research.fb.com/downloads/babi/) - **Repository:** - **Paper:** [arXiv Paper](https://arxiv.org/pdf/1511.02301.pdf) - **Leaderboard:** - **Point of Contact:** [Felix Hill](mailto:felix.hill@cl.cam.ac.uk) or [Antoine Bordes](mailto:abordes@fb.com). ### Dataset Summary The Children’s Book Test (CBT) is designed to measure directly how well language models can exploit wider linguistic context. The CBT is built from books that are freely available. This dataset contains four different configurations: - `V`: where the answers to the questions are verbs. - `P`: where the answers to the questions are pronouns. - `NE`: where the answers to the questions are named entities. - `CN`: where the answers to the questions are common nouns. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The data is present in English language as written by authors Lucy Maud Montgomery, Charles Dickens,Andrew Lang, etc. in story books for children. ## Dataset Structure ### Data Instances An instance from the `V` config: ``` {'answer': 'said', 'options': ['christening', 'existed', 'hear', 'knows', 'read', 'remarked', 'said', 'sitting', 'talking', 'wearing'], 'question': "`` They are very kind old ladies in their way , '' XXXXX the king ; `` and were nice to me when I was a boy . ''", 'sentences': ['This vexed the king even more than the queen , who was very clever and learned , and who had hated dolls when she was a child .', 'However , she , too in spite of all the books she read and all the pictures she painted , would have been glad enough to be the mother of a little prince .', 'The king was anxious to consult the fairies , but the queen would not hear of such a thing .', 'She did not believe in fairies : she said that they had never existed ; and that she maintained , though The History of the Royal Family was full of chapters about nothing else .', 'Well , at long and at last they had a little boy , who was generally regarded as the finest baby that had ever been seen .', 'Even her majesty herself remarked that , though she could never believe all the courtiers told her , yet he certainly was a fine child -- a very fine child .', 'Now , the time drew near for the christening party , and the king and queen were sitting at breakfast in their summer parlour talking over it .', 'It was a splendid room , hung with portraits of the royal ancestors .', 'There was Cinderella , the grandmother of the reigning monarch , with her little foot in her glass slipper thrust out before her .', 'There was the Marquis de Carabas , who , as everyone knows , was raised to the throne as prince consort after his marriage with the daughter of the king of the period .', 'On the arm of the throne was seated his celebrated cat , wearing boots .', 'There , too , was a portrait of a beautiful lady , sound asleep : this was Madame La Belle au Bois-dormant , also an ancestress of the royal family .', 'Many other pictures of celebrated persons were hanging on the walls .', "`` You have asked all the right people , my dear ? ''", 'said the king .', "`` Everyone who should be asked , '' answered the queen .", "`` People are so touchy on these occasions , '' said his majesty .", "`` You have not forgotten any of our aunts ? ''", "`` No ; the old cats ! ''", "replied the queen ; for the king 's aunts were old-fashioned , and did not approve of her , and she knew it ."]} ``` ### Data Fields For the `raw` config, the data fields are: - `title`: a `string` feature containing the title of the book present in the dataset. - `content`: a `string` feature containing the content of the book present in the dataset. For all other configs, the data fields are: - `sentences`: a `list` of `string` features containing 20 sentences from a book. - `question`: a `string` feature containing a question with blank marked as `XXXX` which is to be filled with one of the options. - `answer`: a `string` feature containing the answer. - `options`: a `list` of `string` features containing the options for the question. ### Data Splits The splits and corresponding sizes are: | |train |test |validation| |:--|------:|----:|---------:| |raw|98 |5 |5 | |V |105825 |2500 |2000 | |P |334030 |2500 |2000 | |CN |120769 |2500 |2000 | |NE |108719 |2500 |2000 | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? Children's Book Authors ### Annotations #### Annotation process From the [homepage](https://research.fb.com/downloads/babi/): >After allocating books to either training, validation or test sets, we formed example ‘questions’ from chapters in the book by enumerating 21 consecutive sentences. In each question, the first 20 sentences form the context, and a word is removed from the 21st sentence, which becomes the query. Models must identify the answer word among a selection of 10 candidate answers appearing in the context sentences and the query. For finer-grained analyses, we evaluated four classes of question by removing distinct types of word: Named Entities, (Common) Nouns, Verbs and Prepositions. #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information ``` GNU Free Documentation License v1.3 ``` ### Citation Information ``` @misc{hill2016goldilocks, title={The Goldilocks Principle: Reading Children's Books with Explicit Memory Representations}, author={Felix Hill and Antoine Bordes and Sumit Chopra and Jason Weston}, year={2016}, eprint={1511.02301}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` ### Contributions Thanks to [@gchhablani](https://github.com/gchhablani) for adding this dataset.
dbrd
2023-01-25T14:29:14.000Z
[ "task_categories:text-generation", "task_categories:fill-mask", "task_categories:text-classification", "task_ids:language-modeling", "task_ids:masked-language-modeling", "task_ids:sentiment-classification", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:nl", "license:cc-by-nc-sa-4.0", "arxiv:1910.00896", "region:us" ]
null
The Dutch Book Review Dataset (DBRD) contains over 110k book reviews of which 22k have associated binary sentiment polarity labels. It is intended as a benchmark for sentiment classification in Dutch and created due to a lack of annotated datasets in Dutch that are suitable for this task.
@article{DBLP:journals/corr/abs-1910-00896, author = {Benjamin van der Burgh and Suzan Verberne}, title = {The merits of Universal Language Model Fine-tuning for Small Datasets - a case with Dutch book reviews}, journal = {CoRR}, volume = {abs/1910.00896}, year = {2019}, url = {http://arxiv.org/abs/1910.00896}, archivePrefix = {arXiv}, eprint = {1910.00896}, timestamp = {Fri, 04 Oct 2019 12:28:06 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-1910-00896.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} }
null
3
1,032
--- annotations_creators: - found language_creators: - found language: - nl license: - cc-by-nc-sa-4.0 multilinguality: - monolingual size_categories: - 100K<n<1M source_datasets: - original task_categories: - text-generation - fill-mask - text-classification task_ids: - language-modeling - masked-language-modeling - sentiment-classification paperswithcode_id: dbrd pretty_name: DBRD dataset_info: features: - name: text dtype: string - name: label dtype: class_label: names: '0': neg '1': pos config_name: plain_text splits: - name: train num_bytes: 29496333 num_examples: 20028 - name: test num_bytes: 3246243 num_examples: 2224 - name: unsupervised num_bytes: 152733031 num_examples: 96264 download_size: 79065872 dataset_size: 185475607 --- # Dataset Card for DBRD ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Dutch Book Review Dataset (DBRD) homepage](https://benjaminvdb.github.io/DBRD) - **Repository:** https://github.com/benjaminvdb/DBRD - **Paper:** [The merits of Universal Language Model Fine-tuning for Small Datasets - a case with Dutch book reviews](https://arxiv.org/abs/1910.00896) - **Leaderboard:** - **Point of Contact:** [Benjamin van der Burgh](mailto:benjaminvdb@gmail.com) ### Dataset Summary The DBRD (pronounced *dee-bird*) dataset contains over 110k book reviews of which 22k have associated binary sentiment polarity labels. It is intended as a benchmark for sentiment classification in Dutch and was created due to a lack of annotated datasets in Dutch that are suitable for this task. ### Supported Tasks and Leaderboards - `text-generation`: The dataset can be used to train a model for sequence modeling, more specifically language modeling. - `text-classification`: The dataset can be used to train a model for text classification, more specifically sentiment classification, using the provided positive/negative sentiment polarity labels. ### Languages Non-Dutch reviews were filtered out using [langdetect](https://github.com/Mimino666/langdetect), and all reviews should therefore be in Dutch (nl). They are written by reviewers on [Hebban](https://www.hebban.nl), a Dutch website for book reviews. ## Dataset Structure ### Data Instances The dataset contains three subsets: train, test, and unsupervised. The `train` and `test` sets contain labels, while the `unsupervised` set doesn't (the label value is -1 for each instance in `unsupervised`). Here's an example of a positive review, indicated with a label value of `1`. ``` { 'label': 1, 'text': 'Super om te lezen hoe haar leven is vergaan.\nBijzonder dat ze zo openhartig is geweest.' } ``` ### Data Fields - `label`: either 0 (negative) or 1 (positive) in the supervised sets `train` and `test`. These are always -1 for the unsupervised set. - `text`: book review as a utf-8 encoded string. ### Data Splits The `train` and `test` sets were constructed by extracting all non-neutral reviews because we want to assign either a positive or negative polarity label to each instance. Furthermore, the positive (pos) and negative (neg) labels were balanced in both train and test sets. The remainder was added to the unsupervised set. | | Train | Test | Unsupervised | | ----- | ------ | ----- | ----------- | | # No. texts | 20028 | 2224 | 96264 | | % of total | 16.9% | 1.9% | 81.2% | ## Dataset Creation ### Curation Rationale This dataset was created due to a lack of annotated Dutch text that is suitable for sentiment classification. Non-Dutch texts were therefore removed, but other than that, no curation was done. ### Source Data The book reviews were taken from [Hebban](https://www.hebban.nl), a Dutch platform for book reviews. #### Initial Data Collection and Normalization The source code of the scraper and preprocessing process can be found in the [DBRD GitHub repository](https://github.com/benjaminvdb/DBRD). #### Who are the source language producers? The reviews are written by users of [Hebban](https://www.hebban.nl) and are of varying quality. Some are short, others long, and many contain spelling mistakes and other errors. ### Annotations Each book review was accompanied by a 1 to 5-star rating. The annotations are produced by mapping the user-provided ratings to either a positive or negative label. 1 and 2-star ratings are given the negative label `0` and 4 and 5-star ratings the positive label `1`. Reviews with a rating of 3 stars are considered neutral and left out of the `train`/`test` sets and added to the unsupervised set. #### Annotation process Users of [Hebban](https://www.hebban.nl) were unaware that their reviews would be used in the creation of this dataset. #### Who are the annotators? The annotators are the [Hebban](https://www.hebban.nl) users who wrote the book reviews associated with the annotation. Anyone can register on [Hebban](https://www.hebban.nl) and it's impossible to know the demographics of this group. ### Personal and Sensitive Information The book reviews and ratings are publicly available on [Hebban](https://www.hebban.nl) and no personal or otherwise sensitive information is contained in this dataset. ## Considerations for Using the Data ### Social Impact of Dataset While predicting sentiment of book reviews in itself is not that interesting, the value of this dataset lies in its usage for benchmarking models. The dataset contains some challenges that are common to outings on the internet, such as spelling mistakes and other errors. It is therefore very useful for validating models for their real-world performance. These datasets are abundant for English but are harder to find for Dutch, making them a valuable resource for ML tasks in this language. ### Discussion of Biases [More Information Needed] ### Other Known Limitations Reviews on [Hebban](https://www.hebban.nl) are usually written in Dutch, but some have been written in English and possibly in other languages. While we've done our best to filter out non-Dutch texts, it's hard to do this without errors. For example, some reviews are in multiple languages, and these might slip through. Also be aware that some commercial outings can appear in the text, making them different from other reviews and influencing your models. While this doesn't pose a major issue in most cases, we just wanted to mention it briefly. ## Additional Information ### Dataset Curators This dataset was created by [Benjamin van der Burgh](mailto:benjaminvdb@gmail.com), who was working at [Leiden Institute of Advanced Computer Science (LIACS)](https://liacs.leidenuniv.nl/) at the time. ### Licensing Information The dataset is licensed under a [Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License](https://creativecommons.org/licenses/by-nc-sa/4.0/). ### Citation Information Please use the following citation when making use of this dataset in your work. ``` @article{DBLP:journals/corr/abs-1910-00896, author = {Benjamin van der Burgh and Suzan Verberne}, title = {The merits of Universal Language Model Fine-tuning for Small Datasets - a case with Dutch book reviews}, journal = {CoRR}, volume = {abs/1910.00896}, year = {2019}, url = {http://arxiv.org/abs/1910.00896}, archivePrefix = {arXiv}, eprint = {1910.00896}, timestamp = {Fri, 04 Oct 2019 12:28:06 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-1910-00896.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ``` ### Contributions Thanks to [@benjaminvdb](https://github.com/benjaminvdb) for adding this dataset.
zeroshot/twitter-financial-news-sentiment
2022-12-12T14:32:59.000Z
[ "task_categories:text-classification", "task_ids:multi-class-classification", "annotations_creators:other", "language_creators:other", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:mit", "twitter", "finance", "markets", "stocks", "wallstreet", "quant", "hedgefunds", "region:us" ]
zeroshot
null
null
null
28
1,017
--- annotations_creators: - other language: - en language_creators: - other license: - mit multilinguality: - monolingual pretty_name: twitter financial news size_categories: - 10K<n<100K source_datasets: - original tags: - twitter - finance - markets - stocks - wallstreet - quant - hedgefunds - markets task_categories: - text-classification task_ids: - multi-class-classification --- Read this [BLOG](https://neuralmagic.com/blog/classifying-finance-tweets-in-real-time-with-sparse-transformers/) to see how I fine-tuned a sparse transformer on this dataset. ### Dataset Description The Twitter Financial News dataset is an English-language dataset containing an annotated corpus of finance-related tweets. This dataset is used to classify finance-related tweets for their sentiment. 1. The dataset holds 11,932 documents annotated with 3 labels: ```python sentiments = { "LABEL_0": "Bearish", "LABEL_1": "Bullish", "LABEL_2": "Neutral" } ``` The data was collected using the Twitter API. The current dataset supports the multi-class classification task. ### Task: Sentiment Analysis # Data Splits There are 2 splits: train and validation. Below are the statistics: | Dataset Split | Number of Instances in Split | | ------------- | ------------------------------------------- | | Train | 9,938 | | Validation | 2,486 | # Licensing Information The Twitter Financial Dataset (sentiment) version 1.0.0 is released under the MIT License.
assin
2023-01-25T14:26:50.000Z
[ "task_categories:text-classification", "task_ids:text-scoring", "task_ids:natural-language-inference", "task_ids:semantic-similarity-scoring", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:pt", "license:unknown", "region:us" ]
null
The ASSIN (Avaliação de Similaridade Semântica e INferência textual) corpus is a corpus annotated with pairs of sentences written in Portuguese that is suitable for the exploration of textual entailment and paraphrasing classifiers. The corpus contains pairs of sentences extracted from news articles written in European Portuguese (EP) and Brazilian Portuguese (BP), obtained from Google News Portugal and Brazil, respectively. To create the corpus, the authors started by collecting a set of news articles describing the same event (one news article from Google News Portugal and another from Google News Brazil) from Google News. Then, they employed Latent Dirichlet Allocation (LDA) models to retrieve pairs of similar sentences between sets of news articles that were grouped together around the same topic. For that, two LDA models were trained (for EP and for BP) on external and large-scale collections of unannotated news articles from Portuguese and Brazilian news providers, respectively. Then, the authors defined a lower and upper threshold for the sentence similarity score of the retrieved pairs of sentences, taking into account that high similarity scores correspond to sentences that contain almost the same content (paraphrase candidates), and low similarity scores correspond to sentences that are very different in content from each other (no-relation candidates). From the collection of pairs of sentences obtained at this stage, the authors performed some manual grammatical corrections and discarded some of the pairs wrongly retrieved. Furthermore, from a preliminary analysis made to the retrieved sentence pairs the authors noticed that the number of contradictions retrieved during the previous stage was very low. Additionally, they also noticed that event though paraphrases are not very frequent, they occur with some frequency in news articles. Consequently, in contrast with the majority of the currently available corpora for other languages, which consider as labels “neutral”, “entailment” and “contradiction” for the task of RTE, the authors of the ASSIN corpus decided to use as labels “none”, “entailment” and “paraphrase”. Finally, the manual annotation of pairs of sentences was performed by human annotators. At least four annotators were randomly selected to annotate each pair of sentences, which is done in two steps: (i) assigning a semantic similarity label (a score between 1 and 5, from unrelated to very similar); and (ii) providing an entailment label (one sentence entails the other, sentences are paraphrases, or no relation). Sentence pairs where at least three annotators do not agree on the entailment label were considered controversial and thus discarded from the gold standard annotations. The full dataset has 10,000 sentence pairs, half of which in Brazilian Portuguese and half in European Portuguese. Either language variant has 2,500 pairs for training, 500 for validation and 2,000 for testing.
@inproceedings{fonseca2016assin, title={ASSIN: Avaliacao de similaridade semantica e inferencia textual}, author={Fonseca, E and Santos, L and Criscuolo, Marcelo and Aluisio, S}, booktitle={Computational Processing of the Portuguese Language-12th International Conference, Tomar, Portugal}, pages={13--15}, year={2016} }
null
8
1,015
--- annotations_creators: - expert-generated language_creators: - found language: - pt license: - unknown multilinguality: - monolingual size_categories: - 10K<n<100K source_datasets: - original task_categories: - text-classification task_ids: - text-scoring - natural-language-inference - semantic-similarity-scoring paperswithcode_id: assin pretty_name: ASSIN dataset_info: - config_name: full features: - name: sentence_pair_id dtype: int64 - name: premise dtype: string - name: hypothesis dtype: string - name: relatedness_score dtype: float32 - name: entailment_judgment dtype: class_label: names: '0': NONE '1': ENTAILMENT '2': PARAPHRASE splits: - name: train num_bytes: 986507 num_examples: 5000 - name: test num_bytes: 767312 num_examples: 4000 - name: validation num_bytes: 196829 num_examples: 1000 download_size: 749735 dataset_size: 1950648 - config_name: ptpt features: - name: sentence_pair_id dtype: int64 - name: premise dtype: string - name: hypothesis dtype: string - name: relatedness_score dtype: float32 - name: entailment_judgment dtype: class_label: names: '0': NONE '1': ENTAILMENT '2': PARAPHRASE splits: - name: train num_bytes: 523002 num_examples: 2500 - name: test num_bytes: 392888 num_examples: 2000 - name: validation num_bytes: 105626 num_examples: 500 download_size: 749735 dataset_size: 1021516 - config_name: ptbr features: - name: sentence_pair_id dtype: int64 - name: premise dtype: string - name: hypothesis dtype: string - name: relatedness_score dtype: float32 - name: entailment_judgment dtype: class_label: names: '0': NONE '1': ENTAILMENT '2': PARAPHRASE splits: - name: train num_bytes: 463513 num_examples: 2500 - name: test num_bytes: 374432 num_examples: 2000 - name: validation num_bytes: 91211 num_examples: 500 download_size: 749735 dataset_size: 929156 --- # Dataset Card for ASSIN ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [ASSIN homepage](http://nilc.icmc.usp.br/assin/) - **Repository:** [ASSIN repository](http://nilc.icmc.usp.br/assin/) - **Paper:** [ASSIN: Evaluation of Semantic Similarity and Textual Inference](http://propor2016.di.fc.ul.pt/wp-content/uploads/2015/10/assin-overview.pdf) - **Point of Contact:** [Erick Rocha Fonseca](mailto:erickrf@icmc.usp.br) ### Dataset Summary The ASSIN (Avaliação de Similaridade Semântica e INferência textual) corpus is a corpus annotated with pairs of sentences written in Portuguese that is suitable for the exploration of textual entailment and paraphrasing classifiers. The corpus contains pairs of sentences extracted from news articles written in European Portuguese (EP) and Brazilian Portuguese (BP), obtained from Google News Portugal and Brazil, respectively. To create the corpus, the authors started by collecting a set of news articles describing the same event (one news article from Google News Portugal and another from Google News Brazil) from Google News. Then, they employed Latent Dirichlet Allocation (LDA) models to retrieve pairs of similar sentences between sets of news articles that were grouped together around the same topic. For that, two LDA models were trained (for EP and for BP) on external and large-scale collections of unannotated news articles from Portuguese and Brazilian news providers, respectively. Then, the authors defined a lower and upper threshold for the sentence similarity score of the retrieved pairs of sentences, taking into account that high similarity scores correspond to sentences that contain almost the same content (paraphrase candidates), and low similarity scores correspond to sentences that are very different in content from each other (no-relation candidates). From the collection of pairs of sentences obtained at this stage, the authors performed some manual grammatical corrections and discarded some of the pairs wrongly retrieved. Furthermore, from a preliminary analysis made to the retrieved sentence pairs the authors noticed that the number of contradictions retrieved during the previous stage was very low. Additionally, they also noticed that event though paraphrases are not very frequent, they occur with some frequency in news articles. Consequently, in contrast with the majority of the currently available corpora for other languages, which consider as labels “neutral”, “entailment” and “contradiction” for the task of RTE, the authors of the ASSIN corpus decided to use as labels “none”, “entailment” and “paraphrase”. Finally, the manual annotation of pairs of sentences was performed by human annotators. At least four annotators were randomly selected to annotate each pair of sentences, which is done in two steps: (i) assigning a semantic similarity label (a score between 1 and 5, from unrelated to very similar); and (ii) providing an entailment label (one sentence entails the other, sentences are paraphrases, or no relation). Sentence pairs where at least three annotators do not agree on the entailment label were considered controversial and thus discarded from the gold standard annotations. The full dataset has 10,000 sentence pairs, half of which in Brazilian Portuguese (ptbr) and half in European Portuguese (ptpt). Either language variant has 2,500 pairs for training, 500 for validation and 2,000 for testing. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The language supported is Portuguese. ## Dataset Structure ### Data Instances An example from the ASSIN dataset looks as follows: ``` { "entailment_judgment": 0, "hypothesis": "André Gomes entra em campo quatro meses depois de uma lesão na perna esquerda o ter afastado dos relvados.", "premise": "Relembre-se que o atleta estava afastado dos relvados desde maio, altura em que contraiu uma lesão na perna esquerda.", "relatedness_score": 3.5, "sentence_pair_id": 1 } ``` ### Data Fields - `sentence_pair_id`: a `int64` feature. - `premise`: a `string` feature. - `hypothesis`: a `string` feature. - `relatedness_score`: a `float32` feature. - `entailment_judgment`: a classification label, with possible values including `NONE`, `ENTAILMENT`, `PARAPHRASE`. ### Data Splits The data is split into train, validation and test set. The split sizes are as follow: | | Train | Val | Test | | ----- | ------ | ----- | ---- | | full | 5000 | 1000 | 4000 | | ptbr | 2500 | 500 | 2000 | | ptpt | 2500 | 500 | 2000 | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ``` @inproceedings{fonseca2016assin, title={ASSIN: Avaliacao de similaridade semantica e inferencia textual}, author={Fonseca, E and Santos, L and Criscuolo, Marcelo and Aluisio, S}, booktitle={Computational Processing of the Portuguese Language-12th International Conference, Tomar, Portugal}, pages={13--15}, year={2016} } ``` ### Contributions Thanks to [@jonatasgrosman](https://github.com/jonatasgrosman) for adding this dataset.
mteb/amazon_reviews_multi
2022-09-27T19:10:01.000Z
[ "language:de", "language:en", "language:es", "language:fr", "language:ja", "language:zh", "region:us" ]
mteb
We provide an Amazon product reviews dataset for multilingual text classification. The dataset contains reviews in English, Japanese, German, French, Chinese and Spanish, collected between November 1, 2015 and November 1, 2019. Each record in the dataset contains the review text, the review title, the star rating, an anonymized reviewer ID, an anonymized product ID and the coarse-grained product category (e.g. ‘books’, ‘appliances’, etc.) The corpus is balanced across stars, so each star rating constitutes 20% of the reviews in each language. For each language, there are 200,000, 5,000 and 5,000 reviews in the training, development and test sets respectively. The maximum number of reviews per reviewer is 20 and the maximum number of reviews per product is 20. All reviews are truncated after 2,000 characters, and all reviews are at least 20 characters long. Note that the language of a review does not necessarily match the language of its marketplace (e.g. reviews from amazon.de are primarily written in German, but could also be written in English, etc.). For this reason, we applied a language detection algorithm based on the work in Bojanowski et al. (2017) to determine the language of the review text and we removed reviews that were not written in the expected language.
@inproceedings{marc_reviews, title={The Multilingual Amazon Reviews Corpus}, author={Keung, Phillip and Lu, Yichao and Szarvas, György and Smith, Noah A.}, booktitle={Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing}, year={2020} }
null
3
1,011
--- language: - de - en - es - fr - ja - zh ---