File size: 5,462 Bytes
73ef86d a5d42f3 fe713bf a5d42f3 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 |
---
dataset_info:
features:
- name: id
dtype: string
- name: question
dtype: string
- name: answer
dtype: string
- name: type
dtype: string
- name: evidences
sequence:
sequence: string
- name: supporting_facts
struct:
- name: title
sequence: string
- name: sent_id
sequence: int64
- name: context
struct:
- name: title
sequence: string
- name: sentences
sequence:
sequence: string
splits:
- name: train
num_bytes: 664062413
num_examples: 167454
- name: validation
num_bytes: 54492966
num_examples: 12576
- name: test
num_bytes: 51538723
num_examples: 12576
download_size: 388043174
dataset_size: 770094102
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
- split: test
path: data/test-*
---
# 2WikiMultihopQA
**This repository only repackages the original 2WikiMultihopQA data so that every example follows the field layout used by [HotpotQA](https://hotpotqa.github.io/).** The content of the underlying questions, answers and contexts is **unaltered**.
All intellectual credit for creating 2WikiMultihopQA belongs to the authors of the paper *Constructing a Multi‑hop QA Dataset for Comprehensive Evaluation of Reasoning Steps* (COLING 2020) and the accompanying code/data in their GitHub project [https://github.com/Alab-NII/2wikimultihop](https://github.com/Alab-NII/2wikimultihop).
## Dataset Summary
* **Name:** 2WikiMultihopQA
* **What’s different:** only the JSON schema. Each instance now has `id`, `question`, `answer`, `type`, `evidences`, `supporting_facts`, and `context` keys arranged exactly like HotpotQA so that existing HotpotQA data pipelines work out‑of‑the‑box.
The dataset still contains **multi‑hop question‑answer pairs** with supporting evidence chains drawn from Wikipedia. There are three splits:
| split | #examples |
| ---------- | --------: |
| train | 113,284 |
| validation | 12,981 |
| test | 12,995 |
## Dataset Structure
Each JSON Lines file contains records like:
```json
{
"id": "13f5ad2c088c11ebbd6fac1f6bf848b6",
"question": "Are director of film Move (1970 Film) and director of film Méditerranée (1963 Film) from the same country?",
"answer": "no",
"type": "bridge_comparison",
"level": "unknown",
"supporting_facts": {
"title": ["Move (1970 film)", "Méditerranée (1963 film)", "Stuart Rosenberg", "Jean-Daniel Pollet"],
"sent_id": [0, 0, 0, 0]
},
"context": {
"title": ["Stuart Rosenberg", "Méditerranée (1963 film)", "Move (1970 film)", ...],
"sentences": [["Stuart Rosenberg (August 11, 1927 – March 15, 2007) was an American film and television director ..."],
["Méditerranée is a 1963 French experimental film directed by Jean-Daniel Pollet ..."],
...]
}
}
```
### Field definitions
| Field | Type | Description |
| -------------------------- | -------------------- | ------------------------------------------------------------------------ |
| `id` | `string` | Unique identifier (original `_id`). |
| `question` | `string` | Natural‑language question. |
| `answer` | `string` | Short answer span (may be "yes"/"no" for binary questions). |
| `type` | `string` | Original 2Wiki question type (e.g. `bridge_comparison`, `comparison`). |
| `evidences` | `List[List[string]]` | Structured data (subject, property, object) obtained from Wikidata. |
| `supporting_facts.title` | `List[string]` | Wikipedia page titles that contain evidence sentences. |
| `supporting_facts.sent_id` | `List[int]` | Zero‑based sentence indices within each page that support the answer. |
| `context.title` | `List[string]` | Titles for every paragraph provided to the model. |
| `context.sentences` | `List[List[string]]` | Tokenised sentences for each corresponding title. |
## Data Splits
The conversion keeps the same train/validation/test division as the original dataset. No documents or examples were removed or added.
## Source Data
* **Original repository:** [https://github.com/Alab-NII/2wikimultihop](https://github.com/Alab-NII/2wikimultihop)
Contains data generation scripts, the Apache‑2.0 license and citation information.
* **Paper:** Ho et al., *Constructing A Multi-hop QA Dataset for Comprehensive Evaluation of Reasoning Steps*, COLING 2020.
[\[ACL Anthology\]](https://aclanthology.org/2020.coling-main.580/) – [\[arXiv\]](https://arxiv.org/abs/2011.01060)
No new text was crawled; every paragraph is already present in the original dataset.
## License
The original 2WikiMultihopQA is released under the **Apache License 2.0**.
This redistribution keeps the same license. See [`LICENSE`](./LICENSE) copied verbatim from the upstream repo.
## How to Use
```python
from datasets import load_dataset
ds = load_dataset("framolfese/2WikiMultihopQA")
print(ds["train"][0]["question"])
``` |