parquet-converter commited on
Commit
d5045ee
·
1 Parent(s): e326197

Update parquet files

Browse files
.gitattributes DELETED
@@ -1,27 +0,0 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bin.* filter=lfs diff=lfs merge=lfs -text
5
- *.bz2 filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.model filter=lfs diff=lfs merge=lfs -text
12
- *.msgpack filter=lfs diff=lfs merge=lfs -text
13
- *.onnx filter=lfs diff=lfs merge=lfs -text
14
- *.ot filter=lfs diff=lfs merge=lfs -text
15
- *.parquet filter=lfs diff=lfs merge=lfs -text
16
- *.pb filter=lfs diff=lfs merge=lfs -text
17
- *.pt filter=lfs diff=lfs merge=lfs -text
18
- *.pth filter=lfs diff=lfs merge=lfs -text
19
- *.rar filter=lfs diff=lfs merge=lfs -text
20
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
21
- *.tar.* filter=lfs diff=lfs merge=lfs -text
22
- *.tflite filter=lfs diff=lfs merge=lfs -text
23
- *.tgz filter=lfs diff=lfs merge=lfs -text
24
- *.xz filter=lfs diff=lfs merge=lfs -text
25
- *.zip filter=lfs diff=lfs merge=lfs -text
26
- *.zstandard filter=lfs diff=lfs merge=lfs -text
27
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
README.md DELETED
@@ -1,334 +0,0 @@
1
- ---
2
- annotations_creators:
3
- - crowdsourced
4
- language_creators:
5
- - crowdsourced
6
- language:
7
- - en
8
- license:
9
- - unknown
10
- multilinguality:
11
- - monolingual
12
- size_categories:
13
- - 1K<n<10K
14
- source_datasets:
15
- - original
16
- task_categories:
17
- - question-answering
18
- task_ids:
19
- - multiple-choice-qa
20
- paperswithcode_id: codah
21
- pretty_name: COmmonsense Dataset Adversarially-authored by Humans
22
- dataset_info:
23
- - config_name: codah
24
- features:
25
- - name: id
26
- dtype: int32
27
- - name: question_category
28
- dtype:
29
- class_label:
30
- names:
31
- 0: Idioms
32
- 1: Reference
33
- 2: Polysemy
34
- 3: Negation
35
- 4: Quantitative
36
- 5: Others
37
- - name: question_propmt
38
- dtype: string
39
- - name: candidate_answers
40
- sequence: string
41
- - name: correct_answer_idx
42
- dtype: int32
43
- splits:
44
- - name: train
45
- num_bytes: 571208
46
- num_examples: 2776
47
- download_size: 485130
48
- dataset_size: 571208
49
- - config_name: fold_0
50
- features:
51
- - name: id
52
- dtype: int32
53
- - name: question_category
54
- dtype:
55
- class_label:
56
- names:
57
- 0: Idioms
58
- 1: Reference
59
- 2: Polysemy
60
- 3: Negation
61
- 4: Quantitative
62
- 5: Others
63
- - name: question_propmt
64
- dtype: string
65
- - name: candidate_answers
66
- sequence: string
67
- - name: correct_answer_idx
68
- dtype: int32
69
- splits:
70
- - name: train
71
- num_bytes: 344912
72
- num_examples: 1665
73
- - name: validation
74
- num_bytes: 114211
75
- num_examples: 556
76
- - name: test
77
- num_bytes: 112109
78
- num_examples: 555
79
- download_size: 485130
80
- dataset_size: 571232
81
- - config_name: fold_1
82
- features:
83
- - name: id
84
- dtype: int32
85
- - name: question_category
86
- dtype:
87
- class_label:
88
- names:
89
- 0: Idioms
90
- 1: Reference
91
- 2: Polysemy
92
- 3: Negation
93
- 4: Quantitative
94
- 5: Others
95
- - name: question_propmt
96
- dtype: string
97
- - name: candidate_answers
98
- sequence: string
99
- - name: correct_answer_idx
100
- dtype: int32
101
- splits:
102
- - name: train
103
- num_bytes: 340990
104
- num_examples: 1665
105
- - name: validation
106
- num_bytes: 114211
107
- num_examples: 556
108
- - name: test
109
- num_bytes: 116031
110
- num_examples: 555
111
- download_size: 485130
112
- dataset_size: 571232
113
- - config_name: fold_2
114
- features:
115
- - name: id
116
- dtype: int32
117
- - name: question_category
118
- dtype:
119
- class_label:
120
- names:
121
- 0: Idioms
122
- 1: Reference
123
- 2: Polysemy
124
- 3: Negation
125
- 4: Quantitative
126
- 5: Others
127
- - name: question_propmt
128
- dtype: string
129
- - name: candidate_answers
130
- sequence: string
131
- - name: correct_answer_idx
132
- dtype: int32
133
- splits:
134
- - name: train
135
- num_bytes: 342293
136
- num_examples: 1665
137
- - name: validation
138
- num_bytes: 114211
139
- num_examples: 556
140
- - name: test
141
- num_bytes: 114728
142
- num_examples: 555
143
- download_size: 485130
144
- dataset_size: 571232
145
- - config_name: fold_3
146
- features:
147
- - name: id
148
- dtype: int32
149
- - name: question_category
150
- dtype:
151
- class_label:
152
- names:
153
- 0: Idioms
154
- 1: Reference
155
- 2: Polysemy
156
- 3: Negation
157
- 4: Quantitative
158
- 5: Others
159
- - name: question_propmt
160
- dtype: string
161
- - name: candidate_answers
162
- sequence: string
163
- - name: correct_answer_idx
164
- dtype: int32
165
- splits:
166
- - name: train
167
- num_bytes: 342844
168
- num_examples: 1665
169
- - name: validation
170
- num_bytes: 114211
171
- num_examples: 556
172
- - name: test
173
- num_bytes: 114177
174
- num_examples: 555
175
- download_size: 485130
176
- dataset_size: 571232
177
- - config_name: fold_4
178
- features:
179
- - name: id
180
- dtype: int32
181
- - name: question_category
182
- dtype:
183
- class_label:
184
- names:
185
- 0: Idioms
186
- 1: Reference
187
- 2: Polysemy
188
- 3: Negation
189
- 4: Quantitative
190
- 5: Others
191
- - name: question_propmt
192
- dtype: string
193
- - name: candidate_answers
194
- sequence: string
195
- - name: correct_answer_idx
196
- dtype: int32
197
- splits:
198
- - name: train
199
- num_bytes: 342844
200
- num_examples: 1665
201
- - name: validation
202
- num_bytes: 114177
203
- num_examples: 555
204
- - name: test
205
- num_bytes: 114211
206
- num_examples: 556
207
- download_size: 485130
208
- dataset_size: 571232
209
- ---
210
-
211
- # Dataset Card for COmmonsense Dataset Adversarially-authored by Humans
212
-
213
- ## Table of Contents
214
- - [Dataset Description](#dataset-description)
215
- - [Dataset Summary](#dataset-summary)
216
- - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
217
- - [Languages](#languages)
218
- - [Dataset Structure](#dataset-structure)
219
- - [Data Instances](#data-instances)
220
- - [Data Fields](#data-fields)
221
- - [Data Splits](#data-splits)
222
- - [Dataset Creation](#dataset-creation)
223
- - [Curation Rationale](#curation-rationale)
224
- - [Source Data](#source-data)
225
- - [Annotations](#annotations)
226
- - [Personal and Sensitive Information](#personal-and-sensitive-information)
227
- - [Considerations for Using the Data](#considerations-for-using-the-data)
228
- - [Social Impact of Dataset](#social-impact-of-dataset)
229
- - [Discussion of Biases](#discussion-of-biases)
230
- - [Other Known Limitations](#other-known-limitations)
231
- - [Additional Information](#additional-information)
232
- - [Dataset Curators](#dataset-curators)
233
- - [Licensing Information](#licensing-information)
234
- - [Citation Information](#citation-information)
235
- - [Contributions](#contributions)
236
-
237
- ## Dataset Description
238
-
239
- - **Homepage:** [Add homepage URL here if available (unless it's a GitHub repository)]()
240
- - **Repository:** [If the dataset is hosted on github or has a github homepage, add URL here]()
241
- - **Paper:** [If the dataset was introduced by a paper or there was a paper written describing the dataset, add URL here (landing page for Arxiv paper preferred)]()
242
- - **Leaderboard:** [If the dataset supports an active leaderboard, add link here]()
243
- - **Point of Contact:** [If known, name and email of at least one person the reader can contact for questions about the dataset.]()
244
-
245
- ### Dataset Summary
246
-
247
- [More Information Needed]
248
-
249
- ### Supported Tasks and Leaderboards
250
-
251
- [More Information Needed]
252
-
253
- ### Languages
254
-
255
- [More Information Needed]
256
-
257
- ## Dataset Structure
258
-
259
- ### Data Instances
260
-
261
- [More Information Needed]
262
-
263
- ### Data Fields
264
-
265
- [More Information Needed]
266
-
267
- ### Data Splits
268
-
269
- [More Information Needed]
270
- ## Dataset Creation
271
-
272
- ### Curation Rationale
273
-
274
- [More Information Needed]
275
-
276
- ### Source Data
277
-
278
- [More Information Needed]
279
-
280
- #### Initial Data Collection and Normalization
281
-
282
- [More Information Needed]
283
-
284
- #### Who are the source language producers?
285
-
286
- [More Information Needed]
287
-
288
- ### Annotations
289
-
290
- [More Information Needed]
291
-
292
- #### Annotation process
293
-
294
- [More Information Needed]
295
-
296
- #### Who are the annotators?
297
-
298
- [More Information Needed]
299
-
300
- ### Personal and Sensitive Information
301
-
302
- [More Information Needed]
303
-
304
- ## Considerations for Using the Data
305
-
306
- ### Social Impact of Dataset
307
-
308
- [More Information Needed]
309
-
310
- ### Discussion of Biases
311
-
312
- [More Information Needed]
313
-
314
- ### Other Known Limitations
315
-
316
- [More Information Needed]
317
-
318
- ## Additional Information
319
-
320
- ### Dataset Curators
321
-
322
- [More Information Needed]
323
-
324
- ### Licensing Information
325
-
326
- [More Information Needed]
327
-
328
- ### Citation Information
329
-
330
- [More Information Needed]
331
-
332
- ### Contributions
333
-
334
- Thanks to [@patil-suraj](https://github.com/patil-suraj) for adding this dataset.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
codah.py DELETED
@@ -1,141 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
- """The COmmonsense Dataset Adversarially-authored by Humans (CODAH)"""
16
-
17
-
18
- import csv
19
-
20
- import datasets
21
-
22
-
23
- _CITATION = """\
24
- @inproceedings{chen2019codah,
25
- title={CODAH: An Adversarially-Authored Question Answering Dataset for Common Sense},
26
- author={Chen, Michael and D'Arcy, Mike and Liu, Alisa and Fernandez, Jared and Downey, Doug},
27
- booktitle={Proceedings of the 3rd Workshop on Evaluating Vector Space Representations for NLP},
28
- pages={63--69},
29
- year={2019}
30
- }
31
- """
32
-
33
- _DESCRIPTION = """\
34
- The COmmonsense Dataset Adversarially-authored by Humans (CODAH) is an evaluation set for commonsense \
35
- question-answering in the sentence completion style of SWAG. As opposed to other automatically \
36
- generated NLI datasets, CODAH is adversarially constructed by humans who can view feedback \
37
- from a pre-trained model and use this information to design challenging commonsense questions. \
38
- Our experimental results show that CODAH questions present a complementary extension to the SWAG dataset, testing additional modes of common sense.
39
- """
40
-
41
- _URL = "https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/"
42
- _FULL_DATA_URL = _URL + "full_data.tsv"
43
-
44
- QUESTION_CATEGORIES_MAPPING = {
45
- "i": "Idioms",
46
- "r": "Reference",
47
- "p": "Polysemy",
48
- "n": "Negation",
49
- "q": "Quantitative",
50
- "o": "Others",
51
- }
52
-
53
-
54
- class CodahConfig(datasets.BuilderConfig):
55
- """BuilderConfig for CODAH."""
56
-
57
- def __init__(self, fold=None, **kwargs):
58
- """BuilderConfig for CODAH.
59
-
60
- Args:
61
- fold: `string`, official cross validation fold.
62
- **kwargs: keyword arguments forwarded to super.
63
- """
64
- super(CodahConfig, self).__init__(**kwargs)
65
- self.fold = fold
66
-
67
-
68
- class Codah(datasets.GeneratorBasedBuilder):
69
- """The COmmonsense Dataset Adversarially-authored by Humans (CODAH)"""
70
-
71
- VERSION = datasets.Version("1.0.0")
72
- BUILDER_CONFIGS = [
73
- CodahConfig(name="codah", version=datasets.Version("1.0.0"), description="Full CODAH dataset", fold=None),
74
- CodahConfig(
75
- name="fold_0", version=datasets.Version("1.0.0"), description="Official CV split (fold_0)", fold="fold_0"
76
- ),
77
- CodahConfig(
78
- name="fold_1", version=datasets.Version("1.0.0"), description="Official CV split (fold_1)", fold="fold_1"
79
- ),
80
- CodahConfig(
81
- name="fold_2", version=datasets.Version("1.0.0"), description="Official CV split (fold_2)", fold="fold_2"
82
- ),
83
- CodahConfig(
84
- name="fold_3", version=datasets.Version("1.0.0"), description="Official CV split (fold_3)", fold="fold_3"
85
- ),
86
- CodahConfig(
87
- name="fold_4", version=datasets.Version("1.0.0"), description="Official CV split (fold_4)", fold="fold_4"
88
- ),
89
- ]
90
-
91
- def _info(self):
92
- return datasets.DatasetInfo(
93
- description=_DESCRIPTION,
94
- features=datasets.Features(
95
- {
96
- "id": datasets.Value("int32"),
97
- "question_category": datasets.features.ClassLabel(
98
- names=["Idioms", "Reference", "Polysemy", "Negation", "Quantitative", "Others"]
99
- ),
100
- "question_propmt": datasets.Value("string"),
101
- "candidate_answers": datasets.features.Sequence(datasets.Value("string")),
102
- "correct_answer_idx": datasets.Value("int32"),
103
- }
104
- ),
105
- supervised_keys=None,
106
- homepage="https://github.com/Websail-NU/CODAH",
107
- citation=_CITATION,
108
- )
109
-
110
- def _split_generators(self, dl_manager):
111
- if self.config.name == "codah":
112
- data_file = dl_manager.download(_FULL_DATA_URL)
113
- return [datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"data_file": data_file})]
114
-
115
- base_url = f"{_URL}cv_split/{self.config.fold}/"
116
- _urls = {
117
- "train": base_url + "train.tsv",
118
- "dev": base_url + "dev.tsv",
119
- "test": base_url + "test.tsv",
120
- }
121
- downloaded_files = dl_manager.download_and_extract(_urls)
122
-
123
- return [
124
- datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"data_file": downloaded_files["train"]}),
125
- datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={"data_file": downloaded_files["dev"]}),
126
- datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"data_file": downloaded_files["test"]}),
127
- ]
128
-
129
- def _generate_examples(self, data_file):
130
- with open(data_file, encoding="utf-8") as f:
131
- rows = csv.reader(f, delimiter="\t")
132
- for i, row in enumerate(rows):
133
- question_category = QUESTION_CATEGORIES_MAPPING[row[0]] if row[0] != "" else -1
134
- example = {
135
- "id": i,
136
- "question_category": question_category,
137
- "question_propmt": row[1],
138
- "candidate_answers": row[2:-1],
139
- "correct_answer_idx": int(row[-1]),
140
- }
141
- yield i, example
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
codah/codah-train.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:485013bd93fc717d6223c2e51e176c785af20b094a053a34205d415f9c575430
3
+ size 352901
dataset_infos.json DELETED
@@ -1 +0,0 @@
1
- {"codah": {"description": "The COmmonsense Dataset Adversarially-authored by Humans (CODAH) is an evaluation set for commonsense question-answering in the sentence completion style of SWAG. As opposed to other automatically generated NLI datasets, CODAH is adversarially constructed by humans who can view feedback from a pre-trained model and use this information to design challenging commonsense questions. Our experimental results show that CODAH questions present a complementary extension to the SWAG dataset, testing additional modes of common sense.\n", "citation": "@inproceedings{chen2019codah,\n title={CODAH: An Adversarially-Authored Question Answering Dataset for Common Sense},\n author={Chen, Michael and D'Arcy, Mike and Liu, Alisa and Fernandez, Jared and Downey, Doug},\n booktitle={Proceedings of the 3rd Workshop on Evaluating Vector Space Representations for NLP},\n pages={63--69},\n year={2019}\n}\n", "homepage": "https://github.com/Websail-NU/CODAH", "license": "", "features": {"id": {"dtype": "int32", "id": null, "_type": "Value"}, "question_category": {"num_classes": 6, "names": ["Idioms", "Reference", "Polysemy", "Negation", "Quantitative", "Others"], "names_file": null, "id": null, "_type": "ClassLabel"}, "question_propmt": {"dtype": "string", "id": null, "_type": "Value"}, "candidate_answers": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_idx": {"dtype": "int32", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "codah", "config_name": "codah", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 571208, "num_examples": 2776, "dataset_name": "codah"}}, "download_checksums": {"https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/full_data.tsv": {"num_bytes": 485130, "checksum": "96689f2abf2f09eb91af50e2d8f92bb553e157601908a1cee83fe68670b5c3ea"}}, "download_size": 485130, "post_processing_size": null, "dataset_size": 571208, "size_in_bytes": 1056338}, "fold_0": {"description": "The COmmonsense Dataset Adversarially-authored by Humans (CODAH) is an evaluation set for commonsense question-answering in the sentence completion style of SWAG. As opposed to other automatically generated NLI datasets, CODAH is adversarially constructed by humans who can view feedback from a pre-trained model and use this information to design challenging commonsense questions. Our experimental results show that CODAH questions present a complementary extension to the SWAG dataset, testing additional modes of common sense.\n", "citation": "@inproceedings{chen2019codah,\n title={CODAH: An Adversarially-Authored Question Answering Dataset for Common Sense},\n author={Chen, Michael and D'Arcy, Mike and Liu, Alisa and Fernandez, Jared and Downey, Doug},\n booktitle={Proceedings of the 3rd Workshop on Evaluating Vector Space Representations for NLP},\n pages={63--69},\n year={2019}\n}\n", "homepage": "https://github.com/Websail-NU/CODAH", "license": "", "features": {"id": {"dtype": "int32", "id": null, "_type": "Value"}, "question_category": {"num_classes": 6, "names": ["Idioms", "Reference", "Polysemy", "Negation", "Quantitative", "Others"], "names_file": null, "id": null, "_type": "ClassLabel"}, "question_propmt": {"dtype": "string", "id": null, "_type": "Value"}, "candidate_answers": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_idx": {"dtype": "int32", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "codah", "config_name": "fold_0", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 344912, "num_examples": 1665, "dataset_name": "codah"}, "validation": {"name": "validation", "num_bytes": 114211, "num_examples": 556, "dataset_name": "codah"}, "test": {"name": "test", "num_bytes": 112109, "num_examples": 555, "dataset_name": "codah"}}, "download_checksums": {"https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/cv_split/fold_0/train.tsv": {"num_bytes": 293278, "checksum": "790edcdfcba7d07b66ae2f3074c9ed17239b1f0865962f1bea0e0768207f0d43"}, "https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/cv_split/fold_0/dev.tsv": {"num_bytes": 96963, "checksum": "96c69a4240e6b0ef1f2364aeaa9d549e1b36188814f63d52bbc678a57b855923"}, "https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/cv_split/fold_0/test.tsv": {"num_bytes": 94889, "checksum": "2065089015df59d7cf0328bdd32a9e2088ff66ab9639d926acf26a3b0203454e"}}, "download_size": 485130, "post_processing_size": null, "dataset_size": 571232, "size_in_bytes": 1056362}, "fold_1": {"description": "The COmmonsense Dataset Adversarially-authored by Humans (CODAH) is an evaluation set for commonsense question-answering in the sentence completion style of SWAG. As opposed to other automatically generated NLI datasets, CODAH is adversarially constructed by humans who can view feedback from a pre-trained model and use this information to design challenging commonsense questions. Our experimental results show that CODAH questions present a complementary extension to the SWAG dataset, testing additional modes of common sense.\n", "citation": "@inproceedings{chen2019codah,\n title={CODAH: An Adversarially-Authored Question Answering Dataset for Common Sense},\n author={Chen, Michael and D'Arcy, Mike and Liu, Alisa and Fernandez, Jared and Downey, Doug},\n booktitle={Proceedings of the 3rd Workshop on Evaluating Vector Space Representations for NLP},\n pages={63--69},\n year={2019}\n}\n", "homepage": "https://github.com/Websail-NU/CODAH", "license": "", "features": {"id": {"dtype": "int32", "id": null, "_type": "Value"}, "question_category": {"num_classes": 6, "names": ["Idioms", "Reference", "Polysemy", "Negation", "Quantitative", "Others"], "names_file": null, "id": null, "_type": "ClassLabel"}, "question_propmt": {"dtype": "string", "id": null, "_type": "Value"}, "candidate_answers": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_idx": {"dtype": "int32", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "codah", "config_name": "fold_1", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 340990, "num_examples": 1665, "dataset_name": "codah"}, "validation": {"name": "validation", "num_bytes": 114211, "num_examples": 556, "dataset_name": "codah"}, "test": {"name": "test", "num_bytes": 116031, "num_examples": 555, "dataset_name": "codah"}}, "download_checksums": {"https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/cv_split/fold_1/train.tsv": {"num_bytes": 289356, "checksum": "99841c58cbef1afb8aff8bf4ee8827d0824c3e787ff55d9a365a39284edde7ec"}, "https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/cv_split/fold_1/dev.tsv": {"num_bytes": 96963, "checksum": "96c69a4240e6b0ef1f2364aeaa9d549e1b36188814f63d52bbc678a57b855923"}, "https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/cv_split/fold_1/test.tsv": {"num_bytes": 98811, "checksum": "0f63cf4b30a7bd01f44f466cfa072bd99156e9f9a9042ac659f87b11e9bdbebb"}}, "download_size": 485130, "post_processing_size": null, "dataset_size": 571232, "size_in_bytes": 1056362}, "fold_2": {"description": "The COmmonsense Dataset Adversarially-authored by Humans (CODAH) is an evaluation set for commonsense question-answering in the sentence completion style of SWAG. As opposed to other automatically generated NLI datasets, CODAH is adversarially constructed by humans who can view feedback from a pre-trained model and use this information to design challenging commonsense questions. Our experimental results show that CODAH questions present a complementary extension to the SWAG dataset, testing additional modes of common sense.\n", "citation": "@inproceedings{chen2019codah,\n title={CODAH: An Adversarially-Authored Question Answering Dataset for Common Sense},\n author={Chen, Michael and D'Arcy, Mike and Liu, Alisa and Fernandez, Jared and Downey, Doug},\n booktitle={Proceedings of the 3rd Workshop on Evaluating Vector Space Representations for NLP},\n pages={63--69},\n year={2019}\n}\n", "homepage": "https://github.com/Websail-NU/CODAH", "license": "", "features": {"id": {"dtype": "int32", "id": null, "_type": "Value"}, "question_category": {"num_classes": 6, "names": ["Idioms", "Reference", "Polysemy", "Negation", "Quantitative", "Others"], "names_file": null, "id": null, "_type": "ClassLabel"}, "question_propmt": {"dtype": "string", "id": null, "_type": "Value"}, "candidate_answers": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_idx": {"dtype": "int32", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "codah", "config_name": "fold_2", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 342293, "num_examples": 1665, "dataset_name": "codah"}, "validation": {"name": "validation", "num_bytes": 114211, "num_examples": 556, "dataset_name": "codah"}, "test": {"name": "test", "num_bytes": 114728, "num_examples": 555, "dataset_name": "codah"}}, "download_checksums": {"https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/cv_split/fold_2/train.tsv": {"num_bytes": 290657, "checksum": "79ea621ba442cb8ebb64bf06179a038cfbccc875947a49fca1a487ce79bdc652"}, "https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/cv_split/fold_2/dev.tsv": {"num_bytes": 96963, "checksum": "96c69a4240e6b0ef1f2364aeaa9d549e1b36188814f63d52bbc678a57b855923"}, "https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/cv_split/fold_2/test.tsv": {"num_bytes": 97510, "checksum": "285996f708245858789a9a99740de07cb20c33b8fb071e772a047907727ed0ac"}}, "download_size": 485130, "post_processing_size": null, "dataset_size": 571232, "size_in_bytes": 1056362}, "fold_3": {"description": "The COmmonsense Dataset Adversarially-authored by Humans (CODAH) is an evaluation set for commonsense question-answering in the sentence completion style of SWAG. As opposed to other automatically generated NLI datasets, CODAH is adversarially constructed by humans who can view feedback from a pre-trained model and use this information to design challenging commonsense questions. Our experimental results show that CODAH questions present a complementary extension to the SWAG dataset, testing additional modes of common sense.\n", "citation": "@inproceedings{chen2019codah,\n title={CODAH: An Adversarially-Authored Question Answering Dataset for Common Sense},\n author={Chen, Michael and D'Arcy, Mike and Liu, Alisa and Fernandez, Jared and Downey, Doug},\n booktitle={Proceedings of the 3rd Workshop on Evaluating Vector Space Representations for NLP},\n pages={63--69},\n year={2019}\n}\n", "homepage": "https://github.com/Websail-NU/CODAH", "license": "", "features": {"id": {"dtype": "int32", "id": null, "_type": "Value"}, "question_category": {"num_classes": 6, "names": ["Idioms", "Reference", "Polysemy", "Negation", "Quantitative", "Others"], "names_file": null, "id": null, "_type": "ClassLabel"}, "question_propmt": {"dtype": "string", "id": null, "_type": "Value"}, "candidate_answers": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_idx": {"dtype": "int32", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "codah", "config_name": "fold_3", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 342844, "num_examples": 1665, "dataset_name": "codah"}, "validation": {"name": "validation", "num_bytes": 114211, "num_examples": 556, "dataset_name": "codah"}, "test": {"name": "test", "num_bytes": 114177, "num_examples": 555, "dataset_name": "codah"}}, "download_checksums": {"https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/cv_split/fold_3/train.tsv": {"num_bytes": 291210, "checksum": "27e6a6b55770bde68cedb97f8046af2fa469ca6082eb3ffc803ab61b39555bae"}, "https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/cv_split/fold_3/dev.tsv": {"num_bytes": 96963, "checksum": "96c69a4240e6b0ef1f2364aeaa9d549e1b36188814f63d52bbc678a57b855923"}, "https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/cv_split/fold_3/test.tsv": {"num_bytes": 96957, "checksum": "eb91713dd8aaca7bf7997168bd072ea0aa55c947c91e6f0869f48a2904a3c8fd"}}, "download_size": 485130, "post_processing_size": null, "dataset_size": 571232, "size_in_bytes": 1056362}, "fold_4": {"description": "The COmmonsense Dataset Adversarially-authored by Humans (CODAH) is an evaluation set for commonsense question-answering in the sentence completion style of SWAG. As opposed to other automatically generated NLI datasets, CODAH is adversarially constructed by humans who can view feedback from a pre-trained model and use this information to design challenging commonsense questions. Our experimental results show that CODAH questions present a complementary extension to the SWAG dataset, testing additional modes of common sense.\n", "citation": "@inproceedings{chen2019codah,\n title={CODAH: An Adversarially-Authored Question Answering Dataset for Common Sense},\n author={Chen, Michael and D'Arcy, Mike and Liu, Alisa and Fernandez, Jared and Downey, Doug},\n booktitle={Proceedings of the 3rd Workshop on Evaluating Vector Space Representations for NLP},\n pages={63--69},\n year={2019}\n}\n", "homepage": "https://github.com/Websail-NU/CODAH", "license": "", "features": {"id": {"dtype": "int32", "id": null, "_type": "Value"}, "question_category": {"num_classes": 6, "names": ["Idioms", "Reference", "Polysemy", "Negation", "Quantitative", "Others"], "names_file": null, "id": null, "_type": "ClassLabel"}, "question_propmt": {"dtype": "string", "id": null, "_type": "Value"}, "candidate_answers": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_idx": {"dtype": "int32", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "codah", "config_name": "fold_4", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 342844, "num_examples": 1665, "dataset_name": "codah"}, "validation": {"name": "validation", "num_bytes": 114177, "num_examples": 555, "dataset_name": "codah"}, "test": {"name": "test", "num_bytes": 114211, "num_examples": 556, "dataset_name": "codah"}}, "download_checksums": {"https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/cv_split/fold_4/train.tsv": {"num_bytes": 291210, "checksum": "27e6a6b55770bde68cedb97f8046af2fa469ca6082eb3ffc803ab61b39555bae"}, "https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/cv_split/fold_4/dev.tsv": {"num_bytes": 96957, "checksum": "eb91713dd8aaca7bf7997168bd072ea0aa55c947c91e6f0869f48a2904a3c8fd"}, "https://raw.githubusercontent.com/Websail-NU/CODAH/master/data/cv_split/fold_4/test.tsv": {"num_bytes": 96963, "checksum": "96c69a4240e6b0ef1f2364aeaa9d549e1b36188814f63d52bbc678a57b855923"}}, "download_size": 485130, "post_processing_size": null, "dataset_size": 571232, "size_in_bytes": 1056362}}
 
 
fold_0/codah-test.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6864a0a63cf1f9d4f04257288f0d19a3b6e8b334d18fbe0090148b50b699a766
3
+ size 76392
fold_0/codah-train.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85ce9063b0cce27705d4ada6f697c7bab7a187b9864b4951ba9e513188a7a7e4
3
+ size 225465
fold_0/codah-validation.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58e1fd703a5d6d86e3b13c3d2e97ea6e5336053e2d0287f2388ae20de0accb14
3
+ size 77319
fold_1/codah-test.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63e5dd70554d17506a52d1429eb70e9a409f1e7e72bcf18c09e11b7d8677ce93
3
+ size 78823
fold_1/codah-train.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b824c205cb4e82bb4f546091c219c0063c69362e2215fcbc31ebe6bb49843f32
3
+ size 223583
fold_1/codah-validation.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58e1fd703a5d6d86e3b13c3d2e97ea6e5336053e2d0287f2388ae20de0accb14
3
+ size 77319
fold_2/codah-test.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c45755b3fe0379d72e6419112fe2f2ab6f293549955082d537a7b42004fb6f3
3
+ size 78335
fold_2/codah-train.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a52f12c2087cd3b71da74ee849e57d9bed1478be7b7281f47134a84e39d87860
3
+ size 223469
fold_2/codah-validation.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58e1fd703a5d6d86e3b13c3d2e97ea6e5336053e2d0287f2388ae20de0accb14
3
+ size 77319
fold_3/codah-test.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8907f25a7fe9d1e2788681510e352327d68aedd283bcf0ca31c6180a45b53543
3
+ size 77543
fold_3/codah-train.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:153c7be79d582b0397121623dc464c1e5971a0f9280c4a89d3e19373cfbde5a9
3
+ size 224313
fold_3/codah-validation.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58e1fd703a5d6d86e3b13c3d2e97ea6e5336053e2d0287f2388ae20de0accb14
3
+ size 77319
fold_4/codah-test.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58e1fd703a5d6d86e3b13c3d2e97ea6e5336053e2d0287f2388ae20de0accb14
3
+ size 77319
fold_4/codah-train.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:153c7be79d582b0397121623dc464c1e5971a0f9280c4a89d3e19373cfbde5a9
3
+ size 224313
fold_4/codah-validation.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8907f25a7fe9d1e2788681510e352327d68aedd283bcf0ca31c6180a45b53543
3
+ size 77543