File size: 5,462 Bytes
73ef86d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a5d42f3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fe713bf
a5d42f3
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
---
dataset_info:
  features:
  - name: id
    dtype: string
  - name: question
    dtype: string
  - name: answer
    dtype: string
  - name: type
    dtype: string
  - name: evidences
    sequence:
      sequence: string
  - name: supporting_facts
    struct:
    - name: title
      sequence: string
    - name: sent_id
      sequence: int64
  - name: context
    struct:
    - name: title
      sequence: string
    - name: sentences
      sequence:
        sequence: string
  splits:
  - name: train
    num_bytes: 664062413
    num_examples: 167454
  - name: validation
    num_bytes: 54492966
    num_examples: 12576
  - name: test
    num_bytes: 51538723
    num_examples: 12576
  download_size: 388043174
  dataset_size: 770094102
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
  - split: validation
    path: data/validation-*
  - split: test
    path: data/test-*
---



# 2WikiMultihopQA

**This repository only repackages the original 2WikiMultihopQA data so that every example follows the field layout used by [HotpotQA](https://hotpotqa.github.io/).** The content of the underlying questions, answers and contexts is **unaltered**.
All intellectual credit for creating 2WikiMultihopQA belongs to the authors of the paper *Constructing a Multi‑hop QA Dataset for Comprehensive Evaluation of Reasoning Steps* (COLING 2020) and the accompanying code/data in their GitHub project [https://github.com/Alab-NII/2wikimultihop](https://github.com/Alab-NII/2wikimultihop).

## Dataset Summary

* **Name:** 2WikiMultihopQA
* **What’s different:** only the JSON schema. Each instance now has `id`, `question`, `answer`, `type`, `evidences`, `supporting_facts`, and `context` keys arranged exactly like HotpotQA so that existing HotpotQA data pipelines work out‑of‑the‑box.

The dataset still contains **multi‑hop question‑answer pairs** with supporting evidence chains drawn from Wikipedia. There are three splits:

| split      | #examples |
| ---------- | --------: |
| train      |   113,284 |
| validation |    12,981 |
| test       |    12,995 |

## Dataset Structure

Each JSON Lines file contains records like:

```json
{
  "id": "13f5ad2c088c11ebbd6fac1f6bf848b6",
  "question": "Are director of film Move (1970 Film) and director of film Méditerranée (1963 Film) from the same country?",
  "answer": "no",
  "type": "bridge_comparison",
  "level": "unknown",
  "supporting_facts": {
    "title": ["Move (1970 film)", "Méditerranée (1963 film)", "Stuart Rosenberg", "Jean-Daniel Pollet"],
    "sent_id": [0, 0, 0, 0]
  },
  "context": {
    "title": ["Stuart Rosenberg", "Méditerranée (1963 film)", "Move (1970 film)", ...],
    "sentences": [["Stuart Rosenberg (August 11, 1927 – March 15, 2007) was an American film and television director ..."],
                   ["Méditerranée is a 1963 French experimental film directed by Jean-Daniel Pollet ..."],
                   ...]
  }
}
```

### Field definitions

| Field                      | Type                 | Description                                                              |
| -------------------------- | -------------------- | ------------------------------------------------------------------------ |
| `id`                       | `string`             | Unique identifier (original `_id`).                                      |
| `question`                 | `string`             | Natural‑language question.                                               |
| `answer`                   | `string`             | Short answer span (may be "yes"/"no" for binary questions).              |
| `type`                     | `string`             | Original 2Wiki question type (e.g. `bridge_comparison`, `comparison`).   |
| `evidences`                | `List[List[string]]` | Structured data (subject, property, object) obtained from Wikidata.      |
| `supporting_facts.title`   | `List[string]`       | Wikipedia page titles that contain evidence sentences.                   |
| `supporting_facts.sent_id` | `List[int]`          | Zero‑based sentence indices within each page that support the answer.    |
| `context.title`            | `List[string]`       | Titles for every paragraph provided to the model.                        |
| `context.sentences`        | `List[List[string]]` | Tokenised sentences for each corresponding title.                        |

## Data Splits

The conversion keeps the same train/validation/test division as the original dataset. No documents or examples were removed or added.

## Source Data

* **Original repository:** [https://github.com/Alab-NII/2wikimultihop](https://github.com/Alab-NII/2wikimultihop)
  Contains data generation scripts, the Apache‑2.0 license and citation information.
* **Paper:** Ho et al., *Constructing A Multi-hop QA Dataset for Comprehensive Evaluation of Reasoning Steps*, COLING 2020.
  [\[ACL Anthology\]](https://aclanthology.org/2020.coling-main.580/) – [\[arXiv\]](https://arxiv.org/abs/2011.01060)

No new text was crawled; every paragraph is already present in the original dataset.

## License

The original 2WikiMultihopQA is released under the **Apache License 2.0**.
This redistribution keeps the same license. See [`LICENSE`](./LICENSE) copied verbatim from the upstream repo.

## How to Use

```python
from datasets import load_dataset

ds = load_dataset("framolfese/2WikiMultihopQA")
print(ds["train"][0]["question"])
```