Datasets:
Tasks:
Question Answering
Formats:
parquet
Sub-tasks:
multiple-choice-qa
Languages:
English
Size:
10K - 100K
ArXiv:
License:
| annotations_creators: | |
| - crowdsourced | |
| language_creators: | |
| - crowdsourced | |
| language: | |
| - en | |
| license: odc-by | |
| multilinguality: | |
| - monolingual | |
| size_categories: | |
| - 1K<n<10K | |
| source_datasets: | |
| - original | |
| task_categories: | |
| - question-answering | |
| task_ids: | |
| - multiple-choice-qa | |
| paperswithcode_id: codah | |
| pretty_name: COmmonsense Dataset Adversarially-authored by Humans | |
| dataset_info: | |
| - config_name: codah | |
| features: | |
| - name: id | |
| dtype: int32 | |
| - name: question_category | |
| dtype: | |
| class_label: | |
| names: | |
| '0': Idioms | |
| '1': Reference | |
| '2': Polysemy | |
| '3': Negation | |
| '4': Quantitative | |
| '5': Others | |
| - name: question_propmt | |
| dtype: string | |
| - name: candidate_answers | |
| sequence: string | |
| - name: correct_answer_idx | |
| dtype: int32 | |
| splits: | |
| - name: train | |
| num_bytes: 571196 | |
| num_examples: 2776 | |
| download_size: 352902 | |
| dataset_size: 571196 | |
| - config_name: fold_0 | |
| features: | |
| - name: id | |
| dtype: int32 | |
| - name: question_category | |
| dtype: | |
| class_label: | |
| names: | |
| '0': Idioms | |
| '1': Reference | |
| '2': Polysemy | |
| '3': Negation | |
| '4': Quantitative | |
| '5': Others | |
| - name: question_propmt | |
| dtype: string | |
| - name: candidate_answers | |
| sequence: string | |
| - name: correct_answer_idx | |
| dtype: int32 | |
| splits: | |
| - name: train | |
| num_bytes: 344900 | |
| num_examples: 1665 | |
| - name: validation | |
| num_bytes: 114199 | |
| num_examples: 556 | |
| - name: test | |
| num_bytes: 112097 | |
| num_examples: 555 | |
| download_size: 379179 | |
| dataset_size: 571196 | |
| - config_name: fold_1 | |
| features: | |
| - name: id | |
| dtype: int32 | |
| - name: question_category | |
| dtype: | |
| class_label: | |
| names: | |
| '0': Idioms | |
| '1': Reference | |
| '2': Polysemy | |
| '3': Negation | |
| '4': Quantitative | |
| '5': Others | |
| - name: question_propmt | |
| dtype: string | |
| - name: candidate_answers | |
| sequence: string | |
| - name: correct_answer_idx | |
| dtype: int32 | |
| splits: | |
| - name: train | |
| num_bytes: 340978 | |
| num_examples: 1665 | |
| - name: validation | |
| num_bytes: 114199 | |
| num_examples: 556 | |
| - name: test | |
| num_bytes: 116019 | |
| num_examples: 555 | |
| download_size: 379728 | |
| dataset_size: 571196 | |
| - config_name: fold_2 | |
| features: | |
| - name: id | |
| dtype: int32 | |
| - name: question_category | |
| dtype: | |
| class_label: | |
| names: | |
| '0': Idioms | |
| '1': Reference | |
| '2': Polysemy | |
| '3': Negation | |
| '4': Quantitative | |
| '5': Others | |
| - name: question_propmt | |
| dtype: string | |
| - name: candidate_answers | |
| sequence: string | |
| - name: correct_answer_idx | |
| dtype: int32 | |
| splits: | |
| - name: train | |
| num_bytes: 342281 | |
| num_examples: 1665 | |
| - name: validation | |
| num_bytes: 114199 | |
| num_examples: 556 | |
| - name: test | |
| num_bytes: 114716 | |
| num_examples: 555 | |
| download_size: 379126 | |
| dataset_size: 571196 | |
| - config_name: fold_3 | |
| features: | |
| - name: id | |
| dtype: int32 | |
| - name: question_category | |
| dtype: | |
| class_label: | |
| names: | |
| '0': Idioms | |
| '1': Reference | |
| '2': Polysemy | |
| '3': Negation | |
| '4': Quantitative | |
| '5': Others | |
| - name: question_propmt | |
| dtype: string | |
| - name: candidate_answers | |
| sequence: string | |
| - name: correct_answer_idx | |
| dtype: int32 | |
| splits: | |
| - name: train | |
| num_bytes: 342832 | |
| num_examples: 1665 | |
| - name: validation | |
| num_bytes: 114199 | |
| num_examples: 556 | |
| - name: test | |
| num_bytes: 114165 | |
| num_examples: 555 | |
| download_size: 379178 | |
| dataset_size: 571196 | |
| - config_name: fold_4 | |
| features: | |
| - name: id | |
| dtype: int32 | |
| - name: question_category | |
| dtype: | |
| class_label: | |
| names: | |
| '0': Idioms | |
| '1': Reference | |
| '2': Polysemy | |
| '3': Negation | |
| '4': Quantitative | |
| '5': Others | |
| - name: question_propmt | |
| dtype: string | |
| - name: candidate_answers | |
| sequence: string | |
| - name: correct_answer_idx | |
| dtype: int32 | |
| splits: | |
| - name: train | |
| num_bytes: 342832 | |
| num_examples: 1665 | |
| - name: validation | |
| num_bytes: 114165 | |
| num_examples: 555 | |
| - name: test | |
| num_bytes: 114199 | |
| num_examples: 556 | |
| download_size: 379178 | |
| dataset_size: 571196 | |
| configs: | |
| - config_name: codah | |
| data_files: | |
| - split: train | |
| path: codah/train-* | |
| - config_name: fold_0 | |
| data_files: | |
| - split: train | |
| path: fold_0/train-* | |
| - split: validation | |
| path: fold_0/validation-* | |
| - split: test | |
| path: fold_0/test-* | |
| - config_name: fold_1 | |
| data_files: | |
| - split: train | |
| path: fold_1/train-* | |
| - split: validation | |
| path: fold_1/validation-* | |
| - split: test | |
| path: fold_1/test-* | |
| - config_name: fold_2 | |
| data_files: | |
| - split: train | |
| path: fold_2/train-* | |
| - split: validation | |
| path: fold_2/validation-* | |
| - split: test | |
| path: fold_2/test-* | |
| - config_name: fold_3 | |
| data_files: | |
| - split: train | |
| path: fold_3/train-* | |
| - split: validation | |
| path: fold_3/validation-* | |
| - split: test | |
| path: fold_3/test-* | |
| - config_name: fold_4 | |
| data_files: | |
| - split: train | |
| path: fold_4/train-* | |
| - split: validation | |
| path: fold_4/validation-* | |
| - split: test | |
| path: fold_4/test-* | |
| # Dataset Card for COmmonsense Dataset Adversarially-authored by Humans | |
| ## Table of Contents | |
| - [Dataset Description](#dataset-description) | |
| - [Dataset Summary](#dataset-summary) | |
| - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) | |
| - [Languages](#languages) | |
| - [Dataset Structure](#dataset-structure) | |
| - [Data Instances](#data-instances) | |
| - [Data Fields](#data-fields) | |
| - [Data Splits](#data-splits) | |
| - [Dataset Creation](#dataset-creation) | |
| - [Curation Rationale](#curation-rationale) | |
| - [Source Data](#source-data) | |
| - [Annotations](#annotations) | |
| - [Personal and Sensitive Information](#personal-and-sensitive-information) | |
| - [Considerations for Using the Data](#considerations-for-using-the-data) | |
| - [Social Impact of Dataset](#social-impact-of-dataset) | |
| - [Discussion of Biases](#discussion-of-biases) | |
| - [Other Known Limitations](#other-known-limitations) | |
| - [Additional Information](#additional-information) | |
| - [Dataset Curators](#dataset-curators) | |
| - [Licensing Information](#licensing-information) | |
| - [Citation Information](#citation-information) | |
| - [Contributions](#contributions) | |
| ## Dataset Description | |
| - **Homepage:** [Add homepage URL here if available (unless it's a GitHub repository)]() | |
| - **Repository:** https://github.com/Websail-NU/CODAH | |
| - **Paper:** https://aclanthology.org/W19-2008/ | |
| - **Paper:** https://arxiv.org/abs/1904.04365 | |
| ### Dataset Summary | |
| The COmmonsense Dataset Adversarially-authored by Humans (CODAH) is an evaluation set for commonsense | |
| question-answering in the sentence completion style of SWAG. As opposed to other automatically generated | |
| NLI datasets, CODAH is adversarially constructed by humans who can view feedback from a pre-trained model | |
| and use this information to design challenging commonsense questions. | |
| ### Supported Tasks and Leaderboards | |
| [More Information Needed] | |
| ### Languages | |
| [More Information Needed] | |
| ## Dataset Structure | |
| ### Data Instances | |
| [More Information Needed] | |
| ### Data Fields | |
| [More Information Needed] | |
| ### Data Splits | |
| [More Information Needed] | |
| ## Dataset Creation | |
| ### Curation Rationale | |
| [More Information Needed] | |
| ### Source Data | |
| [More Information Needed] | |
| #### Initial Data Collection and Normalization | |
| [More Information Needed] | |
| #### Who are the source language producers? | |
| [More Information Needed] | |
| ### Annotations | |
| [More Information Needed] | |
| #### Annotation process | |
| [More Information Needed] | |
| #### Who are the annotators? | |
| [More Information Needed] | |
| ### Personal and Sensitive Information | |
| [More Information Needed] | |
| ## Considerations for Using the Data | |
| ### Social Impact of Dataset | |
| [More Information Needed] | |
| ### Discussion of Biases | |
| [More Information Needed] | |
| ### Other Known Limitations | |
| [More Information Needed] | |
| ## Additional Information | |
| ### Dataset Curators | |
| [More Information Needed] | |
| ### Licensing Information | |
| The CODAH dataset is made available under the Open Data Commons Attribution License: http://opendatacommons.org/licenses/by/1.0/ | |
| ### Citation Information | |
| ``` | |
| @inproceedings{chen-etal-2019-codah, | |
| title = "{CODAH}: An Adversarially-Authored Question Answering Dataset for Common Sense", | |
| author = "Chen, Michael and | |
| D{'}Arcy, Mike and | |
| Liu, Alisa and | |
| Fernandez, Jared and | |
| Downey, Doug", | |
| editor = "Rogers, Anna and | |
| Drozd, Aleksandr and | |
| Rumshisky, Anna and | |
| Goldberg, Yoav", | |
| booktitle = "Proceedings of the 3rd Workshop on Evaluating Vector Space Representations for {NLP}", | |
| month = jun, | |
| year = "2019", | |
| address = "Minneapolis, USA", | |
| publisher = "Association for Computational Linguistics", | |
| url = "https://aclanthology.org/W19-2008", | |
| doi = "10.18653/v1/W19-2008", | |
| pages = "63--69", | |
| abstract = "Commonsense reasoning is a critical AI capability, but it is difficult to construct challenging datasets that test common sense. Recent neural question answering systems, based on large pre-trained models of language, have already achieved near-human-level performance on commonsense knowledge benchmarks. These systems do not possess human-level common sense, but are able to exploit limitations of the datasets to achieve human-level scores. We introduce the CODAH dataset, an adversarially-constructed evaluation dataset for testing common sense. CODAH forms a challenging extension to the recently-proposed SWAG dataset, which tests commonsense knowledge using sentence-completion questions that describe situations observed in video. To produce a more difficult dataset, we introduce a novel procedure for question acquisition in which workers author questions designed to target weaknesses of state-of-the-art neural question answering systems. Workers are rewarded for submissions that models fail to answer correctly both before and after fine-tuning (in cross-validation). We create 2.8k questions via this procedure and evaluate the performance of multiple state-of-the-art question answering systems on our dataset. We observe a significant gap between human performance, which is 95.3{\%}, and the performance of the best baseline accuracy of 65.3{\%} by the OpenAI GPT model.", | |
| } | |
| ``` | |
| ### Contributions | |
| Thanks to [@patil-suraj](https://github.com/patil-suraj) for adding this dataset. |