polinaeterna commited on
Commit
bdcd4e1
·
1 Parent(s): 916e3ff

Update glue.py

Browse files
Files changed (1) hide show
  1. glue.py +290 -290
glue.py CHANGED
@@ -135,296 +135,296 @@ class Glue(datasets.GeneratorBasedBuilder):
135
 
136
  BUILDER_CONFIGS = [
137
  GlueConfig(
138
- name="cola",
139
- description=textwrap.dedent(
140
- """\
141
- The Corpus of Linguistic Acceptability consists of English
142
- acceptability judgments drawn from books and journal articles on
143
- linguistic theory. Each example is a sequence of words annotated
144
- with whether it is a grammatical English sentence."""
145
- ),
146
- text_features={"sentence": "sentence"},
147
- label_classes=["unacceptable", "acceptable"],
148
- label_column="is_acceptable",
149
- data_url="https://dl.fbaipublicfiles.com/glue/data/CoLA.zip",
150
- data_dir="CoLA",
151
- citation=textwrap.dedent(
152
- """\
153
- @article{warstadt2018neural,
154
- title={Neural Network Acceptability Judgments},
155
- author={Warstadt, Alex and Singh, Amanpreet and Bowman, Samuel R},
156
- journal={arXiv preprint arXiv:1805.12471},
157
- year={2018}
158
- }"""
159
- ),
160
- url="https://nyu-mll.github.io/CoLA/",
161
- ),
162
- GlueConfig(
163
- name="sst2",
164
- description=textwrap.dedent(
165
- """\
166
- The Stanford Sentiment Treebank consists of sentences from movie reviews and
167
- human annotations of their sentiment. The task is to predict the sentiment of a
168
- given sentence. We use the two-way (positive/negative) class split, and use only
169
- sentence-level labels."""
170
- ),
171
- text_features={"sentence": "sentence"},
172
- label_classes=["negative", "positive"],
173
- label_column="label",
174
- data_url="https://dl.fbaipublicfiles.com/glue/data/SST-2.zip",
175
- data_dir="SST-2",
176
- citation=textwrap.dedent(
177
- """\
178
- @inproceedings{socher2013recursive,
179
- title={Recursive deep models for semantic compositionality over a sentiment treebank},
180
- author={Socher, Richard and Perelygin, Alex and Wu, Jean and Chuang, Jason and Manning, Christopher D and Ng, Andrew and Potts, Christopher},
181
- booktitle={Proceedings of the 2013 conference on empirical methods in natural language processing},
182
- pages={1631--1642},
183
- year={2013}
184
- }"""
185
- ),
186
- url="https://datasets.stanford.edu/sentiment/index.html",
187
- ),
188
- GlueConfig(
189
- name="mrpc",
190
- description=textwrap.dedent(
191
- """\
192
- The Microsoft Research Paraphrase Corpus (Dolan & Brockett, 2005) is a corpus of
193
- sentence pairs automatically extracted from online news sources, with human annotations
194
- for whether the sentences in the pair are semantically equivalent."""
195
- ), # pylint: disable=line-too-long
196
- text_features={"sentence1": "", "sentence2": ""},
197
- label_classes=["not_equivalent", "equivalent"],
198
- label_column="Quality",
199
- data_url="", # MRPC isn't hosted by GLUE.
200
- data_dir="MRPC",
201
- citation=textwrap.dedent(
202
- """\
203
- @inproceedings{dolan2005automatically,
204
- title={Automatically constructing a corpus of sentential paraphrases},
205
- author={Dolan, William B and Brockett, Chris},
206
- booktitle={Proceedings of the Third International Workshop on Paraphrasing (IWP2005)},
207
- year={2005}
208
- }"""
209
- ),
210
- url="https://www.microsoft.com/en-us/download/details.aspx?id=52398",
211
- ),
212
- GlueConfig(
213
- name="qqp",
214
- description=textwrap.dedent(
215
- """\
216
- The Quora Question Pairs2 dataset is a collection of question pairs from the
217
- community question-answering website Quora. The task is to determine whether a
218
- pair of questions are semantically equivalent."""
219
- ),
220
- text_features={
221
- "question1": "question1",
222
- "question2": "question2",
223
- },
224
- label_classes=["not_duplicate", "duplicate"],
225
- label_column="is_duplicate",
226
- data_url="https://dl.fbaipublicfiles.com/glue/data/QQP-clean.zip",
227
- data_dir="QQP",
228
- citation=textwrap.dedent(
229
- """\
230
- @online{WinNT,
231
- author = {Iyer, Shankar and Dandekar, Nikhil and Csernai, Kornel},
232
- title = {First Quora Dataset Release: Question Pairs},
233
- year = {2017},
234
- url = {https://data.quora.com/First-Quora-Dataset-Release-Question-Pairs},
235
- urldate = {2019-04-03}
236
- }"""
237
- ),
238
- url="https://data.quora.com/First-Quora-Dataset-Release-Question-Pairs",
239
- ),
240
- GlueConfig(
241
- name="stsb",
242
- description=textwrap.dedent(
243
- """\
244
- The Semantic Textual Similarity Benchmark (Cer et al., 2017) is a collection of
245
- sentence pairs drawn from news headlines, video and image captions, and natural
246
- language inference data. Each pair is human-annotated with a similarity score
247
- from 1 to 5."""
248
- ),
249
- text_features={
250
- "sentence1": "sentence1",
251
- "sentence2": "sentence2",
252
- },
253
- label_column="score",
254
- data_url="https://dl.fbaipublicfiles.com/glue/data/STS-B.zip",
255
- data_dir="STS-B",
256
- citation=textwrap.dedent(
257
- """\
258
- @article{cer2017semeval,
259
- title={Semeval-2017 task 1: Semantic textual similarity-multilingual and cross-lingual focused evaluation},
260
- author={Cer, Daniel and Diab, Mona and Agirre, Eneko and Lopez-Gazpio, Inigo and Specia, Lucia},
261
- journal={arXiv preprint arXiv:1708.00055},
262
- year={2017}
263
- }"""
264
- ),
265
- url="http://ixa2.si.ehu.es/stswiki/index.php/STSbenchmark",
266
- process_label=np.float32,
267
- ),
268
- GlueConfig(
269
- name="mnli",
270
- description=textwrap.dedent(
271
- """\
272
- The Multi-Genre Natural Language Inference Corpus is a crowdsourced
273
- collection of sentence pairs with textual entailment annotations. Given a premise sentence
274
- and a hypothesis sentence, the task is to predict whether the premise entails the hypothesis
275
- (entailment), contradicts the hypothesis (contradiction), or neither (neutral). The premise sentences are
276
- gathered from ten different sources, including transcribed speech, fiction, and government reports.
277
- We use the standard test set, for which we obtained private labels from the authors, and evaluate
278
- on both the matched (in-domain) and mismatched (cross-domain) section. We also use and recommend
279
- the SNLI corpus as 550k examples of auxiliary training data."""
280
- ),
281
- **_MNLI_BASE_KWARGS,
282
- ),
283
- GlueConfig(
284
- name="mnli_mismatched",
285
- description=textwrap.dedent(
286
- """\
287
- The mismatched validation and test splits from MNLI.
288
- See the "mnli" BuilderConfig for additional information."""
289
- ),
290
- **_MNLI_BASE_KWARGS,
291
- ),
292
- GlueConfig(
293
- name="mnli_matched",
294
- description=textwrap.dedent(
295
- """\
296
- The matched validation and test splits from MNLI.
297
- See the "mnli" BuilderConfig for additional information."""
298
- ),
299
- **_MNLI_BASE_KWARGS,
300
- ),
301
- GlueConfig(
302
- name="qnli",
303
- description=textwrap.dedent(
304
- """\
305
- The Stanford Question Answering Dataset is a question-answering
306
- dataset consisting of question-paragraph pairs, where one of the sentences in the paragraph (drawn
307
- from Wikipedia) contains the answer to the corresponding question (written by an annotator). We
308
- convert the task into sentence pair classification by forming a pair between each question and each
309
- sentence in the corresponding context, and filtering out pairs with low lexical overlap between the
310
- question and the context sentence. The task is to determine whether the context sentence contains
311
- the answer to the question. This modified version of the original task removes the requirement that
312
- the model select the exact answer, but also removes the simplifying assumptions that the answer
313
- is always present in the input and that lexical overlap is a reliable cue."""
314
- ), # pylint: disable=line-too-long
315
- text_features={
316
- "question": "question",
317
- "sentence": "sentence",
318
- },
319
- label_classes=["entailment", "not_entailment"],
320
- label_column="label",
321
- data_url="https://dl.fbaipublicfiles.com/glue/data/QNLIv2.zip",
322
- data_dir="QNLI",
323
- citation=textwrap.dedent(
324
- """\
325
- @article{rajpurkar2016squad,
326
- title={Squad: 100,000+ questions for machine comprehension of text},
327
- author={Rajpurkar, Pranav and Zhang, Jian and Lopyrev, Konstantin and Liang, Percy},
328
- journal={arXiv preprint arXiv:1606.05250},
329
- year={2016}
330
- }"""
331
- ),
332
- url="https://rajpurkar.github.io/SQuAD-explorer/",
333
- ),
334
- GlueConfig(
335
- name="rte",
336
- description=textwrap.dedent(
337
- """\
338
- The Recognizing Textual Entailment (RTE) datasets come from a series of annual textual
339
- entailment challenges. We combine the data from RTE1 (Dagan et al., 2006), RTE2 (Bar Haim
340
- et al., 2006), RTE3 (Giampiccolo et al., 2007), and RTE5 (Bentivogli et al., 2009).4 Examples are
341
- constructed based on news and Wikipedia text. We convert all datasets to a two-class split, where
342
- for three-class datasets we collapse neutral and contradiction into not entailment, for consistency."""
343
- ), # pylint: disable=line-too-long
344
- text_features={
345
- "sentence1": "sentence1",
346
- "sentence2": "sentence2",
347
- },
348
- label_classes=["entailment", "not_entailment"],
349
- label_column="label",
350
- data_url="https://dl.fbaipublicfiles.com/glue/data/RTE.zip",
351
- data_dir="RTE",
352
- citation=textwrap.dedent(
353
- """\
354
- @inproceedings{dagan2005pascal,
355
- title={The PASCAL recognising textual entailment challenge},
356
- author={Dagan, Ido and Glickman, Oren and Magnini, Bernardo},
357
- booktitle={Machine Learning Challenges Workshop},
358
- pages={177--190},
359
- year={2005},
360
- organization={Springer}
361
- }
362
- @inproceedings{bar2006second,
363
- title={The second pascal recognising textual entailment challenge},
364
- author={Bar-Haim, Roy and Dagan, Ido and Dolan, Bill and Ferro, Lisa and Giampiccolo, Danilo and Magnini, Bernardo and Szpektor, Idan},
365
- booktitle={Proceedings of the second PASCAL challenges workshop on recognising textual entailment},
366
- volume={6},
367
- number={1},
368
- pages={6--4},
369
- year={2006},
370
- organization={Venice}
371
- }
372
- @inproceedings{giampiccolo2007third,
373
- title={The third pascal recognizing textual entailment challenge},
374
- author={Giampiccolo, Danilo and Magnini, Bernardo and Dagan, Ido and Dolan, Bill},
375
- booktitle={Proceedings of the ACL-PASCAL workshop on textual entailment and paraphrasing},
376
- pages={1--9},
377
- year={2007},
378
- organization={Association for Computational Linguistics}
379
- }
380
- @inproceedings{bentivogli2009fifth,
381
- title={The Fifth PASCAL Recognizing Textual Entailment Challenge.},
382
- author={Bentivogli, Luisa and Clark, Peter and Dagan, Ido and Giampiccolo, Danilo},
383
- booktitle={TAC},
384
- year={2009}
385
- }"""
386
- ),
387
- url="https://aclweb.org/aclwiki/Recognizing_Textual_Entailment",
388
- ),
389
- GlueConfig(
390
- name="wnli",
391
- description=textwrap.dedent(
392
- """\
393
- The Winograd Schema Challenge (Levesque et al., 2011) is a reading comprehension task
394
- in which a system must read a sentence with a pronoun and select the referent of that pronoun from
395
- a list of choices. The examples are manually constructed to foil simple statistical methods: Each
396
- one is contingent on contextual information provided by a single word or phrase in the sentence.
397
- To convert the problem into sentence pair classification, we construct sentence pairs by replacing
398
- the ambiguous pronoun with each possible referent. The task is to predict if the sentence with the
399
- pronoun substituted is entailed by the original sentence. We use a small evaluation set consisting of
400
- new examples derived from fiction books that was shared privately by the authors of the original
401
- corpus. While the included training set is balanced between two classes, the test set is imbalanced
402
- between them (65% not entailment). Also, due to a data quirk, the development set is adversarial:
403
- hypotheses are sometimes shared between training and development examples, so if a model memorizes the
404
- training examples, they will predict the wrong label on corresponding development set
405
- example. As with QNLI, each example is evaluated separately, so there is not a systematic correspondence
406
- between a model's score on this task and its score on the unconverted original task. We
407
- call converted dataset WNLI (Winograd NLI)."""
408
- ),
409
- text_features={
410
- "sentence1": "sentence1",
411
- "sentence2": "sentence2",
412
- },
413
- label_classes=["not_entailment", "entailment"],
414
- label_column="label",
415
- data_url="https://dl.fbaipublicfiles.com/glue/data/WNLI.zip",
416
- data_dir="WNLI",
417
- citation=textwrap.dedent(
418
- """\
419
- @inproceedings{levesque2012winograd,
420
- title={The winograd schema challenge},
421
- author={Levesque, Hector and Davis, Ernest and Morgenstern, Leora},
422
- booktitle={Thirteenth International Conference on the Principles of Knowledge Representation and Reasoning},
423
- year={2012}
424
- }"""
425
- ),
426
- url="https://cs.nyu.edu/faculty/davise/papers/WinogradSchemas/WS.html",
427
- ),
428
  GlueConfig(
429
  name="ax",
430
  description=textwrap.dedent(
 
135
 
136
  BUILDER_CONFIGS = [
137
  GlueConfig(
138
+ # name="cola",
139
+ # description=textwrap.dedent(
140
+ # """\
141
+ # The Corpus of Linguistic Acceptability consists of English
142
+ # acceptability judgments drawn from books and journal articles on
143
+ # linguistic theory. Each example is a sequence of words annotated
144
+ # with whether it is a grammatical English sentence."""
145
+ # ),
146
+ # text_features={"sentence": "sentence"},
147
+ # label_classes=["unacceptable", "acceptable"],
148
+ # label_column="is_acceptable",
149
+ # data_url="https://dl.fbaipublicfiles.com/glue/data/CoLA.zip",
150
+ # data_dir="CoLA",
151
+ # citation=textwrap.dedent(
152
+ # """\
153
+ # @article{warstadt2018neural,
154
+ # title={Neural Network Acceptability Judgments},
155
+ # author={Warstadt, Alex and Singh, Amanpreet and Bowman, Samuel R},
156
+ # journal={arXiv preprint arXiv:1805.12471},
157
+ # year={2018}
158
+ # }"""
159
+ # ),
160
+ # url="https://nyu-mll.github.io/CoLA/",
161
+ # ),
162
+ # GlueConfig(
163
+ # name="sst2",
164
+ # description=textwrap.dedent(
165
+ # """\
166
+ # The Stanford Sentiment Treebank consists of sentences from movie reviews and
167
+ # human annotations of their sentiment. The task is to predict the sentiment of a
168
+ # given sentence. We use the two-way (positive/negative) class split, and use only
169
+ # sentence-level labels."""
170
+ # ),
171
+ # text_features={"sentence": "sentence"},
172
+ # label_classes=["negative", "positive"],
173
+ # label_column="label",
174
+ # data_url="https://dl.fbaipublicfiles.com/glue/data/SST-2.zip",
175
+ # data_dir="SST-2",
176
+ # citation=textwrap.dedent(
177
+ # """\
178
+ # @inproceedings{socher2013recursive,
179
+ # title={Recursive deep models for semantic compositionality over a sentiment treebank},
180
+ # author={Socher, Richard and Perelygin, Alex and Wu, Jean and Chuang, Jason and Manning, Christopher D and Ng, Andrew and Potts, Christopher},
181
+ # booktitle={Proceedings of the 2013 conference on empirical methods in natural language processing},
182
+ # pages={1631--1642},
183
+ # year={2013}
184
+ # }"""
185
+ # ),
186
+ # url="https://datasets.stanford.edu/sentiment/index.html",
187
+ # ),
188
+ # GlueConfig(
189
+ # name="mrpc",
190
+ # description=textwrap.dedent(
191
+ # """\
192
+ # The Microsoft Research Paraphrase Corpus (Dolan & Brockett, 2005) is a corpus of
193
+ # sentence pairs automatically extracted from online news sources, with human annotations
194
+ # for whether the sentences in the pair are semantically equivalent."""
195
+ # ), # pylint: disable=line-too-long
196
+ # text_features={"sentence1": "", "sentence2": ""},
197
+ # label_classes=["not_equivalent", "equivalent"],
198
+ # label_column="Quality",
199
+ # data_url="", # MRPC isn't hosted by GLUE.
200
+ # data_dir="MRPC",
201
+ # citation=textwrap.dedent(
202
+ # """\
203
+ # @inproceedings{dolan2005automatically,
204
+ # title={Automatically constructing a corpus of sentential paraphrases},
205
+ # author={Dolan, William B and Brockett, Chris},
206
+ # booktitle={Proceedings of the Third International Workshop on Paraphrasing (IWP2005)},
207
+ # year={2005}
208
+ # }"""
209
+ # ),
210
+ # url="https://www.microsoft.com/en-us/download/details.aspx?id=52398",
211
+ # ),
212
+ # GlueConfig(
213
+ # name="qqp",
214
+ # description=textwrap.dedent(
215
+ # """\
216
+ # The Quora Question Pairs2 dataset is a collection of question pairs from the
217
+ # community question-answering website Quora. The task is to determine whether a
218
+ # pair of questions are semantically equivalent."""
219
+ # ),
220
+ # text_features={
221
+ # "question1": "question1",
222
+ # "question2": "question2",
223
+ # },
224
+ # label_classes=["not_duplicate", "duplicate"],
225
+ # label_column="is_duplicate",
226
+ # data_url="https://dl.fbaipublicfiles.com/glue/data/QQP-clean.zip",
227
+ # data_dir="QQP",
228
+ # citation=textwrap.dedent(
229
+ # """\
230
+ # @online{WinNT,
231
+ # author = {Iyer, Shankar and Dandekar, Nikhil and Csernai, Kornel},
232
+ # title = {First Quora Dataset Release: Question Pairs},
233
+ # year = {2017},
234
+ # url = {https://data.quora.com/First-Quora-Dataset-Release-Question-Pairs},
235
+ # urldate = {2019-04-03}
236
+ # }"""
237
+ # ),
238
+ # url="https://data.quora.com/First-Quora-Dataset-Release-Question-Pairs",
239
+ # ),
240
+ # GlueConfig(
241
+ # name="stsb",
242
+ # description=textwrap.dedent(
243
+ # """\
244
+ # The Semantic Textual Similarity Benchmark (Cer et al., 2017) is a collection of
245
+ # sentence pairs drawn from news headlines, video and image captions, and natural
246
+ # language inference data. Each pair is human-annotated with a similarity score
247
+ # from 1 to 5."""
248
+ # ),
249
+ # text_features={
250
+ # "sentence1": "sentence1",
251
+ # "sentence2": "sentence2",
252
+ # },
253
+ # label_column="score",
254
+ # data_url="https://dl.fbaipublicfiles.com/glue/data/STS-B.zip",
255
+ # data_dir="STS-B",
256
+ # citation=textwrap.dedent(
257
+ # """\
258
+ # @article{cer2017semeval,
259
+ # title={Semeval-2017 task 1: Semantic textual similarity-multilingual and cross-lingual focused evaluation},
260
+ # author={Cer, Daniel and Diab, Mona and Agirre, Eneko and Lopez-Gazpio, Inigo and Specia, Lucia},
261
+ # journal={arXiv preprint arXiv:1708.00055},
262
+ # year={2017}
263
+ # }"""
264
+ # ),
265
+ # url="http://ixa2.si.ehu.es/stswiki/index.php/STSbenchmark",
266
+ # process_label=np.float32,
267
+ # ),
268
+ # GlueConfig(
269
+ # name="mnli",
270
+ # description=textwrap.dedent(
271
+ # """\
272
+ # The Multi-Genre Natural Language Inference Corpus is a crowdsourced
273
+ # collection of sentence pairs with textual entailment annotations. Given a premise sentence
274
+ # and a hypothesis sentence, the task is to predict whether the premise entails the hypothesis
275
+ # (entailment), contradicts the hypothesis (contradiction), or neither (neutral). The premise sentences are
276
+ # gathered from ten different sources, including transcribed speech, fiction, and government reports.
277
+ # We use the standard test set, for which we obtained private labels from the authors, and evaluate
278
+ # on both the matched (in-domain) and mismatched (cross-domain) section. We also use and recommend
279
+ # the SNLI corpus as 550k examples of auxiliary training data."""
280
+ # ),
281
+ # **_MNLI_BASE_KWARGS,
282
+ # ),
283
+ # GlueConfig(
284
+ # name="mnli_mismatched",
285
+ # description=textwrap.dedent(
286
+ # """\
287
+ # The mismatched validation and test splits from MNLI.
288
+ # See the "mnli" BuilderConfig for additional information."""
289
+ # ),
290
+ # **_MNLI_BASE_KWARGS,
291
+ # ),
292
+ # GlueConfig(
293
+ # name="mnli_matched",
294
+ # description=textwrap.dedent(
295
+ # """\
296
+ # The matched validation and test splits from MNLI.
297
+ # See the "mnli" BuilderConfig for additional information."""
298
+ # ),
299
+ # **_MNLI_BASE_KWARGS,
300
+ # ),
301
+ # GlueConfig(
302
+ # name="qnli",
303
+ # description=textwrap.dedent(
304
+ # """\
305
+ # The Stanford Question Answering Dataset is a question-answering
306
+ # dataset consisting of question-paragraph pairs, where one of the sentences in the paragraph (drawn
307
+ # from Wikipedia) contains the answer to the corresponding question (written by an annotator). We
308
+ # convert the task into sentence pair classification by forming a pair between each question and each
309
+ # sentence in the corresponding context, and filtering out pairs with low lexical overlap between the
310
+ # question and the context sentence. The task is to determine whether the context sentence contains
311
+ # the answer to the question. This modified version of the original task removes the requirement that
312
+ # the model select the exact answer, but also removes the simplifying assumptions that the answer
313
+ # is always present in the input and that lexical overlap is a reliable cue."""
314
+ # ), # pylint: disable=line-too-long
315
+ # text_features={
316
+ # "question": "question",
317
+ # "sentence": "sentence",
318
+ # },
319
+ # label_classes=["entailment", "not_entailment"],
320
+ # label_column="label",
321
+ # data_url="https://dl.fbaipublicfiles.com/glue/data/QNLIv2.zip",
322
+ # data_dir="QNLI",
323
+ # citation=textwrap.dedent(
324
+ # """\
325
+ # @article{rajpurkar2016squad,
326
+ # title={Squad: 100,000+ questions for machine comprehension of text},
327
+ # author={Rajpurkar, Pranav and Zhang, Jian and Lopyrev, Konstantin and Liang, Percy},
328
+ # journal={arXiv preprint arXiv:1606.05250},
329
+ # year={2016}
330
+ # }"""
331
+ # ),
332
+ # url="https://rajpurkar.github.io/SQuAD-explorer/",
333
+ # ),
334
+ # GlueConfig(
335
+ # name="rte",
336
+ # description=textwrap.dedent(
337
+ # """\
338
+ # The Recognizing Textual Entailment (RTE) datasets come from a series of annual textual
339
+ # entailment challenges. We combine the data from RTE1 (Dagan et al., 2006), RTE2 (Bar Haim
340
+ # et al., 2006), RTE3 (Giampiccolo et al., 2007), and RTE5 (Bentivogli et al., 2009).4 Examples are
341
+ # constructed based on news and Wikipedia text. We convert all datasets to a two-class split, where
342
+ # for three-class datasets we collapse neutral and contradiction into not entailment, for consistency."""
343
+ # ), # pylint: disable=line-too-long
344
+ # text_features={
345
+ # "sentence1": "sentence1",
346
+ # "sentence2": "sentence2",
347
+ # },
348
+ # label_classes=["entailment", "not_entailment"],
349
+ # label_column="label",
350
+ # data_url="https://dl.fbaipublicfiles.com/glue/data/RTE.zip",
351
+ # data_dir="RTE",
352
+ # citation=textwrap.dedent(
353
+ # """\
354
+ # @inproceedings{dagan2005pascal,
355
+ # title={The PASCAL recognising textual entailment challenge},
356
+ # author={Dagan, Ido and Glickman, Oren and Magnini, Bernardo},
357
+ # booktitle={Machine Learning Challenges Workshop},
358
+ # pages={177--190},
359
+ # year={2005},
360
+ # organization={Springer}
361
+ # }
362
+ # @inproceedings{bar2006second,
363
+ # title={The second pascal recognising textual entailment challenge},
364
+ # author={Bar-Haim, Roy and Dagan, Ido and Dolan, Bill and Ferro, Lisa and Giampiccolo, Danilo and Magnini, Bernardo and Szpektor, Idan},
365
+ # booktitle={Proceedings of the second PASCAL challenges workshop on recognising textual entailment},
366
+ # volume={6},
367
+ # number={1},
368
+ # pages={6--4},
369
+ # year={2006},
370
+ # organization={Venice}
371
+ # }
372
+ # @inproceedings{giampiccolo2007third,
373
+ # title={The third pascal recognizing textual entailment challenge},
374
+ # author={Giampiccolo, Danilo and Magnini, Bernardo and Dagan, Ido and Dolan, Bill},
375
+ # booktitle={Proceedings of the ACL-PASCAL workshop on textual entailment and paraphrasing},
376
+ # pages={1--9},
377
+ # year={2007},
378
+ # organization={Association for Computational Linguistics}
379
+ # }
380
+ # @inproceedings{bentivogli2009fifth,
381
+ # title={The Fifth PASCAL Recognizing Textual Entailment Challenge.},
382
+ # author={Bentivogli, Luisa and Clark, Peter and Dagan, Ido and Giampiccolo, Danilo},
383
+ # booktitle={TAC},
384
+ # year={2009}
385
+ # }"""
386
+ # ),
387
+ # url="https://aclweb.org/aclwiki/Recognizing_Textual_Entailment",
388
+ # ),
389
+ # GlueConfig(
390
+ # name="wnli",
391
+ # description=textwrap.dedent(
392
+ # """\
393
+ # The Winograd Schema Challenge (Levesque et al., 2011) is a reading comprehension task
394
+ # in which a system must read a sentence with a pronoun and select the referent of that pronoun from
395
+ # a list of choices. The examples are manually constructed to foil simple statistical methods: Each
396
+ # one is contingent on contextual information provided by a single word or phrase in the sentence.
397
+ # To convert the problem into sentence pair classification, we construct sentence pairs by replacing
398
+ # the ambiguous pronoun with each possible referent. The task is to predict if the sentence with the
399
+ # pronoun substituted is entailed by the original sentence. We use a small evaluation set consisting of
400
+ # new examples derived from fiction books that was shared privately by the authors of the original
401
+ # corpus. While the included training set is balanced between two classes, the test set is imbalanced
402
+ # between them (65% not entailment). Also, due to a data quirk, the development set is adversarial:
403
+ # hypotheses are sometimes shared between training and development examples, so if a model memorizes the
404
+ # training examples, they will predict the wrong label on corresponding development set
405
+ # example. As with QNLI, each example is evaluated separately, so there is not a systematic correspondence
406
+ # between a model's score on this task and its score on the unconverted original task. We
407
+ # call converted dataset WNLI (Winograd NLI)."""
408
+ # ),
409
+ # text_features={
410
+ # "sentence1": "sentence1",
411
+ # "sentence2": "sentence2",
412
+ # },
413
+ # label_classes=["not_entailment", "entailment"],
414
+ # label_column="label",
415
+ # data_url="https://dl.fbaipublicfiles.com/glue/data/WNLI.zip",
416
+ # data_dir="WNLI",
417
+ # citation=textwrap.dedent(
418
+ # """\
419
+ # @inproceedings{levesque2012winograd,
420
+ # title={The winograd schema challenge},
421
+ # author={Levesque, Hector and Davis, Ernest and Morgenstern, Leora},
422
+ # booktitle={Thirteenth International Conference on the Principles of Knowledge Representation and Reasoning},
423
+ # year={2012}
424
+ # }"""
425
+ # ),
426
+ # url="https://cs.nyu.edu/faculty/davise/papers/WinogradSchemas/WS.html",
427
+ # ),
428
  GlueConfig(
429
  name="ax",
430
  description=textwrap.dedent(