christti commited on
Commit
81a69ae
·
1 Parent(s): 302b847

Delete README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -240
README.md DELETED
@@ -1,240 +0,0 @@
1
- ---
2
- pretty_name: SQuAD
3
- annotations_creators:
4
- - crowdsourced
5
- language_creators:
6
- - crowdsourced
7
- - found
8
- language:
9
- - en
10
- license:
11
- - cc-by-4.0
12
- multilinguality:
13
- - monolingual
14
- size_categories:
15
- - 10K<n<100K
16
- source_datasets:
17
- - extended|wikipedia
18
- task_categories:
19
- - question-answering
20
- task_ids:
21
- - extractive-qa
22
- paperswithcode_id: squad
23
- train-eval-index:
24
- - config: plain_text
25
- task: question-answering
26
- task_id: extractive_question_answering
27
- splits:
28
- train_split: train
29
- eval_split: validation
30
- col_mapping:
31
- question: question
32
- context: context
33
- answers:
34
- text: text
35
- answer_start: answer_start
36
- metrics:
37
- - type: squad
38
- name: SQuAD
39
- dataset_info:
40
- features:
41
- - name: id
42
- dtype: string
43
- - name: title
44
- dtype: string
45
- - name: context
46
- dtype: string
47
- - name: question
48
- dtype: string
49
- - name: answers
50
- sequence:
51
- - name: text
52
- dtype: string
53
- - name: answer_start
54
- dtype: int32
55
- config_name: plain_text
56
- splits:
57
- - name: train
58
- num_bytes: 156093316
59
- num_examples: 169211
60
- - name: validation
61
- num_bytes: 10472653
62
- num_examples: 10570
63
- download_size: 35142551
64
- dataset_size: 89789763
65
- ---
66
-
67
- # Dataset Card for "squad"
68
-
69
- ## Table of Contents
70
- - [Dataset Card for "squad"](#dataset-card-for-squad)
71
- - [Table of Contents](#table-of-contents)
72
- - [Dataset Description](#dataset-description)
73
- - [Dataset Summary](#dataset-summary)
74
- - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
75
- - [Languages](#languages)
76
- - [Dataset Structure](#dataset-structure)
77
- - [Data Instances](#data-instances)
78
- - [plain_text](#plain_text)
79
- - [Data Fields](#data-fields)
80
- - [plain_text](#plain_text-1)
81
- - [Data Splits](#data-splits)
82
- - [Dataset Creation](#dataset-creation)
83
- - [Curation Rationale](#curation-rationale)
84
- - [Source Data](#source-data)
85
- - [Initial Data Collection and Normalization](#initial-data-collection-and-normalization)
86
- - [Who are the source language producers?](#who-are-the-source-language-producers)
87
- - [Annotations](#annotations)
88
- - [Annotation process](#annotation-process)
89
- - [Who are the annotators?](#who-are-the-annotators)
90
- - [Personal and Sensitive Information](#personal-and-sensitive-information)
91
- - [Considerations for Using the Data](#considerations-for-using-the-data)
92
- - [Social Impact of Dataset](#social-impact-of-dataset)
93
- - [Discussion of Biases](#discussion-of-biases)
94
- - [Other Known Limitations](#other-known-limitations)
95
- - [Additional Information](#additional-information)
96
- - [Dataset Curators](#dataset-curators)
97
- - [Licensing Information](#licensing-information)
98
- - [Citation Information](#citation-information)
99
- - [Contributions](#contributions)
100
-
101
- ## Dataset Description
102
-
103
- - **Homepage:** [https://rajpurkar.github.io/SQuAD-explorer/](https://rajpurkar.github.io/SQuAD-explorer/)
104
- - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
105
- - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
106
- - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
107
- - **Size of downloaded dataset files:** 35.14 MB
108
- - **Size of the generated dataset:** 89.92 MB
109
- - **Total amount of disk used:** 125.06 MB
110
-
111
- ### Dataset Summary
112
-
113
- Stanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by crowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span, from the corresponding reading passage, or the question might be unanswerable.
114
-
115
- ### Supported Tasks and Leaderboards
116
-
117
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
118
-
119
- ### Languages
120
-
121
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
122
-
123
- ## Dataset Structure
124
-
125
- ### Data Instances
126
-
127
- #### plain_text
128
-
129
- - **Size of downloaded dataset files:** 35.14 MB
130
- - **Size of the generated dataset:** 89.92 MB
131
- - **Total amount of disk used:** 125.06 MB
132
-
133
- An example of 'train' looks as follows.
134
- ```
135
- {
136
- "answers": {
137
- "answer_start": [1],
138
- "text": ["This is a test text"]
139
- },
140
- "context": "This is a test context.",
141
- "id": "1",
142
- "question": "Is this a test?",
143
- "title": "train test"
144
- }
145
- ```
146
-
147
- ### Data Fields
148
-
149
- The data fields are the same among all splits.
150
-
151
- #### plain_text
152
- - `id`: a `string` feature.
153
- - `title`: a `string` feature.
154
- - `context`: a `string` feature.
155
- - `question`: a `string` feature.
156
- - `answers`: a dictionary feature containing:
157
- - `text`: a `string` feature.
158
- - `answer_start`: a `int32` feature.
159
-
160
- ### Data Splits
161
-
162
- | name |train|validation|
163
- |----------|----:|---------:|
164
- |plain_text|87599| 10570|
165
-
166
- ## Dataset Creation
167
-
168
- ### Curation Rationale
169
-
170
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
171
-
172
- ### Source Data
173
-
174
- #### Initial Data Collection and Normalization
175
-
176
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
177
-
178
- #### Who are the source language producers?
179
-
180
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
181
-
182
- ### Annotations
183
-
184
- #### Annotation process
185
-
186
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
187
-
188
- #### Who are the annotators?
189
-
190
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
191
-
192
- ### Personal and Sensitive Information
193
-
194
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
195
-
196
- ## Considerations for Using the Data
197
-
198
- ### Social Impact of Dataset
199
-
200
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
201
-
202
- ### Discussion of Biases
203
-
204
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
205
-
206
- ### Other Known Limitations
207
-
208
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
209
-
210
- ## Additional Information
211
-
212
- ### Dataset Curators
213
-
214
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
215
-
216
- ### Licensing Information
217
-
218
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
219
-
220
- ### Citation Information
221
-
222
- ```
223
- @article{2016arXiv160605250R,
224
- author = {{Rajpurkar}, Pranav and {Zhang}, Jian and {Lopyrev},
225
- Konstantin and {Liang}, Percy},
226
- title = "{SQuAD: 100,000+ Questions for Machine Comprehension of Text}",
227
- journal = {arXiv e-prints},
228
- year = 2016,
229
- eid = {arXiv:1606.05250},
230
- pages = {arXiv:1606.05250},
231
- archivePrefix = {arXiv},
232
- eprint = {1606.05250},
233
- }
234
-
235
- ```
236
-
237
-
238
- ### Contributions
239
-
240
- Thanks to [@lewtun](https://github.com/lewtun), [@albertvillanova](https://github.com/albertvillanova), [@patrickvonplaten](https://github.com/patrickvonplaten), [@thomwolf](https://github.com/thomwolf) for adding this dataset.