| | --- |
| | annotations_creators: |
| | - crowdsourced |
| | language_creators: |
| | - crowdsourced |
| | language: |
| | - en |
| | language_bcp47: |
| | - en-US |
| | license: |
| | - cc-by-4.0 |
| | multilinguality: |
| | - monolingual |
| | size_categories: |
| | - 10K<n<100K |
| | source_datasets: |
| | - original |
| | task_categories: |
| | - question-answering |
| | - text-generation |
| | - fill-mask |
| | task_ids: |
| | - open-domain-qa |
| | - dialogue-modeling |
| | pretty_name: ConvQuestions |
| | dataset_info: |
| | features: |
| | - name: domain |
| | dtype: string |
| | - name: seed_entity |
| | dtype: string |
| | - name: seed_entity_text |
| | dtype: string |
| | - name: questions |
| | sequence: string |
| | - name: answers |
| | sequence: |
| | sequence: string |
| | - name: answer_texts |
| | sequence: string |
| | splits: |
| | - name: train |
| | num_bytes: 3589880 |
| | num_examples: 6720 |
| | - name: validation |
| | num_bytes: 1241778 |
| | num_examples: 2240 |
| | - name: test |
| | num_bytes: 1175656 |
| | num_examples: 2240 |
| | download_size: 3276017 |
| | dataset_size: 6007314 |
| | --- |
| | |
| | # Dataset Card for ConvQuestions |
| |
|
| | ## Table of Contents |
| | - [Dataset Description](#dataset-description) |
| | - [Dataset Summary](#dataset-summary) |
| | - [Supported Tasks](#supported-tasks-and-leaderboards) |
| | - [Languages](#languages) |
| | - [Dataset Structure](#dataset-structure) |
| | - [Data Instances](#data-instances) |
| | - [Data Fields](#data-instances) |
| | - [Data Splits](#data-instances) |
| | - [Dataset Creation](#dataset-creation) |
| | - [Curation Rationale](#curation-rationale) |
| | - [Source Data](#source-data) |
| | - [Annotations](#annotations) |
| | - [Personal and Sensitive Information](#personal-and-sensitive-information) |
| | - [Considerations for Using the Data](#considerations-for-using-the-data) |
| | - [Social Impact of Dataset](#social-impact-of-dataset) |
| | - [Discussion of Biases](#discussion-of-biases) |
| | - [Other Known Limitations](#other-known-limitations) |
| | - [Additional Information](#additional-information) |
| | - [Dataset Curators](#dataset-curators) |
| | - [Licensing Information](#licensing-information) |
| | - [Citation Information](#citation-information) |
| | - [Contributions](#contributions) |
| |
|
| | ## Dataset Description |
| |
|
| | - **Homepage:** [ConvQuestions page](https://convex.mpi-inf.mpg.de) |
| | - **Repository:** [GitHub](https://github.com/PhilippChr/CONVEX) |
| | - **Paper:** [Look before you hop: Conversational question answering over knowledge graphs using judicious context expansion](https://arxiv.org/abs/1910.03262) |
| | - **Leaderboard:** [ConvQuestions leaderboard](https://convex.mpi-inf.mpg.de) |
| | - **Point of Contact:** [Philipp Christmann](mailto:pchristm@mpi-inf.mpg.de) |
| |
|
| | ### Dataset Summary |
| |
|
| | ConvQuestions is the first realistic benchmark for conversational question answering over |
| | knowledge graphs. It contains 11,200 conversations which can be evaluated over Wikidata. |
| | They are compiled from the inputs of 70 Master crowdworkers on Amazon Mechanical Turk, |
| | with conversations from five domains: Books, Movies, Soccer, Music, and TV Series. |
| | The questions feature a variety of complex question phenomena like comparisons, aggregations, |
| | compositionality, and temporal reasoning. Answers are grounded in Wikidata entities to enable |
| | fair comparison across diverse methods. The data gathering setup was kept as natural as |
| | possible, with the annotators selecting entities of their choice from each of the five domains, |
| | and formulating the entire conversation in one session. All questions in a conversation are |
| | from the same Turker, who also provided gold answers to the questions. For suitability to knowledge |
| | graphs, questions were constrained to be objective or factoid in nature, but no other restrictive |
| | guidelines were set. A notable property of ConvQuestions is that several questions are not |
| | answerable by Wikidata alone (as of September 2019), but the required facts can, for example, |
| | be found in the open Web or in Wikipedia. For details, please refer to the CIKM 2019 full paper |
| | (https://dl.acm.org/citation.cfm?id=3358016). |
| |
|
| | ### Supported Tasks and Leaderboards |
| |
|
| | [Needs More Information] |
| |
|
| | ### Languages |
| |
|
| | en |
| |
|
| | ## Dataset Structure |
| |
|
| | ### Data Instances |
| |
|
| | An example of 'train' looks as follows. |
| | ``` |
| | { |
| | 'domain': 'music', |
| | 'seed_entity': 'https://www.wikidata.org/wiki/Q223495', |
| | 'seed_entity_text': 'The Carpenters', |
| | 'questions': [ |
| | 'When did The Carpenters sign with A&M Records?', |
| | 'What song was their first hit?', |
| | 'When did Karen die?', |
| | 'Karen had what eating problem?', |
| | 'and how did she die?' |
| | ], |
| | 'answers': [ |
| | [ |
| | '1969' |
| | ], |
| | [ |
| | 'https://www.wikidata.org/wiki/Q928282' |
| | ], |
| | [ |
| | '1983' |
| | ], |
| | [ |
| | 'https://www.wikidata.org/wiki/Q131749' |
| | ], |
| | [ |
| | 'https://www.wikidata.org/wiki/Q181754' |
| | ] |
| | ], |
| | 'answer_texts': [ |
| | '1969', |
| | '(They Long to Be) Close to You', |
| | '1983', |
| | 'anorexia nervosa', |
| | 'heart failure' |
| | ] |
| | } |
| | ``` |
| |
|
| | ### Data Fields |
| |
|
| | - `domain`: a `string` feature. Any of: ['books', 'movies', 'music', 'soccer', 'tv_series'] |
| | - `seed_entity`: a `string` feature. Wikidata ID of the topic entity. |
| | - `seed_entity_text`: a `string` feature. Surface form of the topic entity. |
| | - `questions`: a `list` of `string` features. List of questions (initial question and follow-up questions). |
| | - `answers`: a `list` of `lists` of `string` features. List of answers, given as Wikidata IDs or literals (e.g. timestamps or names). |
| | - `answer_texts`: a `list` of `string` features. List of surface forms of the answers. |
| |
|
| | ### Data Splits |
| |
|
| | |train|validation|tests| |
| | |----:|---------:|----:| |
| | | 6720| 2240| 2240| |
| |
|
| | ## Dataset Creation |
| |
|
| | ### Curation Rationale |
| |
|
| | [Needs More Information] |
| |
|
| | ### Source Data |
| |
|
| | #### Initial Data Collection and Normalization |
| |
|
| | [Needs More Information] |
| |
|
| | #### Who are the source language producers? |
| |
|
| | [Needs More Information] |
| |
|
| | ### Annotations |
| |
|
| | #### Annotation process |
| |
|
| | With insights from a meticulous in-house pilot study with ten students over two weeks, the authors posed the conversation generation task on Amazon Mechanical Turk (AMT) in the most natural setup: Each crowdworker was asked to build a conversation by asking five sequential questions starting from any seed entity of his/her choice, as this is an intuitive mental model that humans may have when satisfying their real information needs via their search assistants. |
| |
|
| | #### Who are the annotators? |
| |
|
| | Local students (Saarland Informatics Campus) and AMT Master Workers. |
| |
|
| | ### Personal and Sensitive Information |
| |
|
| | [Needs More Information] |
| |
|
| | ## Considerations for Using the Data |
| |
|
| | ### Social Impact of Dataset |
| |
|
| | [Needs More Information] |
| |
|
| | ### Discussion of Biases |
| |
|
| | [Needs More Information] |
| |
|
| | ### Other Known Limitations |
| |
|
| | [Needs More Information] |
| |
|
| | ## Additional Information |
| |
|
| | ### Dataset Curators |
| |
|
| | [Needs More Information] |
| |
|
| | ### Licensing Information |
| |
|
| | The ConvQuestions benchmark is licensed under a Creative Commons Attribution 4.0 International License. |
| |
|
| | ### Citation Information |
| |
|
| | ``` |
| | @InProceedings{christmann2019look, |
| | title={Look before you hop: Conversational question answering over knowledge graphs using judicious context expansion}, |
| | author={Christmann, Philipp and Saha Roy, Rishiraj and Abujabal, Abdalghani and Singh, Jyotsna and Weikum, Gerhard}, |
| | booktitle={Proceedings of the 28th ACM International Conference on Information and Knowledge Management}, |
| | pages={729--738}, |
| | year={2019} |
| | } |
| | ``` |
| |
|
| | ### Contributions |
| |
|
| | Thanks to [@PhilippChr](https://github.com/PhilippChr) for adding this dataset. |