|
|
--- |
|
|
dataset_info: |
|
|
features: |
|
|
- name: id |
|
|
dtype: string |
|
|
- name: question |
|
|
dtype: string |
|
|
- name: answer |
|
|
dtype: string |
|
|
- name: type |
|
|
dtype: string |
|
|
- name: evidences |
|
|
sequence: |
|
|
sequence: string |
|
|
- name: supporting_facts |
|
|
struct: |
|
|
- name: title |
|
|
sequence: string |
|
|
- name: sent_id |
|
|
sequence: int64 |
|
|
- name: context |
|
|
struct: |
|
|
- name: title |
|
|
sequence: string |
|
|
- name: sentences |
|
|
sequence: |
|
|
sequence: string |
|
|
splits: |
|
|
- name: train |
|
|
num_bytes: 664062413 |
|
|
num_examples: 167454 |
|
|
- name: validation |
|
|
num_bytes: 54492966 |
|
|
num_examples: 12576 |
|
|
- name: test |
|
|
num_bytes: 51538723 |
|
|
num_examples: 12576 |
|
|
download_size: 388043174 |
|
|
dataset_size: 770094102 |
|
|
configs: |
|
|
- config_name: default |
|
|
data_files: |
|
|
- split: train |
|
|
path: data/train-* |
|
|
- split: validation |
|
|
path: data/validation-* |
|
|
- split: test |
|
|
path: data/test-* |
|
|
--- |
|
|
|
|
|
|
|
|
|
|
|
# 2WikiMultihopQA |
|
|
|
|
|
**This repository only repackages the original 2WikiMultihopQA data so that every example follows the field layout used by [HotpotQA](https://hotpotqa.github.io/).** The content of the underlying questions, answers and contexts is **unaltered**. |
|
|
All intellectual credit for creating 2WikiMultihopQA belongs to the authors of the paper *Constructing a Multi‑hop QA Dataset for Comprehensive Evaluation of Reasoning Steps* (COLING 2020) and the accompanying code/data in their GitHub project [https://github.com/Alab-NII/2wikimultihop](https://github.com/Alab-NII/2wikimultihop). |
|
|
|
|
|
## Dataset Summary |
|
|
|
|
|
* **Name:** 2WikiMultihopQA |
|
|
* **What’s different:** only the JSON schema. Each instance now has `id`, `question`, `answer`, `type`, `evidences`, `supporting_facts`, and `context` keys arranged exactly like HotpotQA so that existing HotpotQA data pipelines work out‑of‑the‑box. |
|
|
|
|
|
The dataset still contains **multi‑hop question‑answer pairs** with supporting evidence chains drawn from Wikipedia. There are three splits: |
|
|
|
|
|
| split | #examples | |
|
|
| ---------- | --------: | |
|
|
| train | 113,284 | |
|
|
| validation | 12,981 | |
|
|
| test | 12,995 | |
|
|
|
|
|
## Dataset Structure |
|
|
|
|
|
Each JSON Lines file contains records like: |
|
|
|
|
|
```json |
|
|
{ |
|
|
"id": "13f5ad2c088c11ebbd6fac1f6bf848b6", |
|
|
"question": "Are director of film Move (1970 Film) and director of film Méditerranée (1963 Film) from the same country?", |
|
|
"answer": "no", |
|
|
"type": "bridge_comparison", |
|
|
"level": "unknown", |
|
|
"supporting_facts": { |
|
|
"title": ["Move (1970 film)", "Méditerranée (1963 film)", "Stuart Rosenberg", "Jean-Daniel Pollet"], |
|
|
"sent_id": [0, 0, 0, 0] |
|
|
}, |
|
|
"context": { |
|
|
"title": ["Stuart Rosenberg", "Méditerranée (1963 film)", "Move (1970 film)", ...], |
|
|
"sentences": [["Stuart Rosenberg (August 11, 1927 – March 15, 2007) was an American film and television director ..."], |
|
|
["Méditerranée is a 1963 French experimental film directed by Jean-Daniel Pollet ..."], |
|
|
...] |
|
|
} |
|
|
} |
|
|
``` |
|
|
|
|
|
### Field definitions |
|
|
|
|
|
| Field | Type | Description | |
|
|
| -------------------------- | -------------------- | ------------------------------------------------------------------------ | |
|
|
| `id` | `string` | Unique identifier (original `_id`). | |
|
|
| `question` | `string` | Natural‑language question. | |
|
|
| `answer` | `string` | Short answer span (may be "yes"/"no" for binary questions). | |
|
|
| `type` | `string` | Original 2Wiki question type (e.g. `bridge_comparison`, `comparison`). | |
|
|
| `evidences` | `List[List[string]]` | Structured data (subject, property, object) obtained from Wikidata. | |
|
|
| `supporting_facts.title` | `List[string]` | Wikipedia page titles that contain evidence sentences. | |
|
|
| `supporting_facts.sent_id` | `List[int]` | Zero‑based sentence indices within each page that support the answer. | |
|
|
| `context.title` | `List[string]` | Titles for every paragraph provided to the model. | |
|
|
| `context.sentences` | `List[List[string]]` | Tokenised sentences for each corresponding title. | |
|
|
|
|
|
## Data Splits |
|
|
|
|
|
The conversion keeps the same train/validation/test division as the original dataset. No documents or examples were removed or added. |
|
|
|
|
|
## Source Data |
|
|
|
|
|
* **Original repository:** [https://github.com/Alab-NII/2wikimultihop](https://github.com/Alab-NII/2wikimultihop) |
|
|
Contains data generation scripts, the Apache‑2.0 license and citation information. |
|
|
* **Paper:** Ho et al., *Constructing A Multi-hop QA Dataset for Comprehensive Evaluation of Reasoning Steps*, COLING 2020. |
|
|
[\[ACL Anthology\]](https://aclanthology.org/2020.coling-main.580/) – [\[arXiv\]](https://arxiv.org/abs/2011.01060) |
|
|
|
|
|
No new text was crawled; every paragraph is already present in the original dataset. |
|
|
|
|
|
## License |
|
|
|
|
|
The original 2WikiMultihopQA is released under the **Apache License 2.0**. |
|
|
This redistribution keeps the same license. See [`LICENSE`](./LICENSE) copied verbatim from the upstream repo. |
|
|
|
|
|
## How to Use |
|
|
|
|
|
```python |
|
|
from datasets import load_dataset |
|
|
|
|
|
ds = load_dataset("framolfese/2WikiMultihopQA") |
|
|
print(ds["train"][0]["question"]) |
|
|
``` |