mathemakitten commited on
Commit
bfc67c0
·
1 Parent(s): 8035aad
Files changed (1) hide show
  1. dataset_infos.json +0 -791
dataset_infos.json CHANGED
@@ -1,142 +1,4 @@
1
  {
2
- "cola": {
3
- "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
4
- "citation": "@article{warstadt2018neural,\n title={Neural Network Acceptability Judgments},\n author={Warstadt, Alex and Singh, Amanpreet and Bowman, Samuel R},\n journal={arXiv preprint arXiv:1805.12471},\n year={2018}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n\nNote that each GLUE dataset has its own citation. Please see the source to see\nthe correct citation for each contained dataset.",
5
- "homepage": "https://nyu-mll.github.io/CoLA/",
6
- "license": "",
7
- "features": {
8
- "sentence": {
9
- "dtype": "string",
10
- "id": null,
11
- "_type": "Value"
12
- },
13
- "label": {
14
- "num_classes": 2,
15
- "names": [
16
- "unacceptable",
17
- "acceptable"
18
- ],
19
- "names_file": null,
20
- "id": null,
21
- "_type": "ClassLabel"
22
- },
23
- "idx": {
24
- "dtype": "int32",
25
- "id": null,
26
- "_type": "Value"
27
- }
28
- },
29
- "post_processed": null,
30
- "supervised_keys": null,
31
- "builder_name": "glue",
32
- "config_name": "cola",
33
- "version": {
34
- "version_str": "1.0.0",
35
- "description": "",
36
- "major": 1,
37
- "minor": 0,
38
- "patch": 0
39
- },
40
- "splits": {
41
- "test": {
42
- "name": "test",
43
- "num_bytes": 61049,
44
- "num_examples": 1063,
45
- "dataset_name": "glue"
46
- },
47
- "train": {
48
- "name": "train",
49
- "num_bytes": 489149,
50
- "num_examples": 8551,
51
- "dataset_name": "glue"
52
- },
53
- "validation": {
54
- "name": "validation",
55
- "num_bytes": 60850,
56
- "num_examples": 1043,
57
- "dataset_name": "glue"
58
- }
59
- },
60
- "download_checksums": {
61
- "https://dl.fbaipublicfiles.com/glue/data/CoLA.zip": {
62
- "num_bytes": 376971,
63
- "checksum": "f212fcd832b8f7b435fb991f101abf89f96b933ab400603bf198960dfc32cbff"
64
- }
65
- },
66
- "download_size": 376971,
67
- "post_processing_size": null,
68
- "dataset_size": 611048,
69
- "size_in_bytes": 988019
70
- },
71
- "sst2": {
72
- "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
73
- "citation": "@inproceedings{socher2013recursive,\n title={Recursive deep models for semantic compositionality over a sentiment treebank},\n author={Socher, Richard and Perelygin, Alex and Wu, Jean and Chuang, Jason and Manning, Christopher D and Ng, Andrew and Potts, Christopher},\n booktitle={Proceedings of the 2013 conference on empirical methods in natural language processing},\n pages={1631--1642},\n year={2013}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n\nNote that each GLUE dataset has its own citation. Please see the source to see\nthe correct citation for each contained dataset.",
74
- "homepage": "https://nlp.stanford.edu/sentiment/index.html",
75
- "license": "",
76
- "features": {
77
- "sentence": {
78
- "dtype": "string",
79
- "id": null,
80
- "_type": "Value"
81
- },
82
- "label": {
83
- "num_classes": 2,
84
- "names": [
85
- "negative",
86
- "positive"
87
- ],
88
- "names_file": null,
89
- "id": null,
90
- "_type": "ClassLabel"
91
- },
92
- "idx": {
93
- "dtype": "int32",
94
- "id": null,
95
- "_type": "Value"
96
- }
97
- },
98
- "post_processed": null,
99
- "supervised_keys": null,
100
- "builder_name": "glue",
101
- "config_name": "sst2",
102
- "version": {
103
- "version_str": "1.0.0",
104
- "description": "",
105
- "major": 1,
106
- "minor": 0,
107
- "patch": 0
108
- },
109
- "splits": {
110
- "test": {
111
- "name": "test",
112
- "num_bytes": 217556,
113
- "num_examples": 1821,
114
- "dataset_name": "glue"
115
- },
116
- "train": {
117
- "name": "train",
118
- "num_bytes": 4715283,
119
- "num_examples": 67349,
120
- "dataset_name": "glue"
121
- },
122
- "validation": {
123
- "name": "validation",
124
- "num_bytes": 106692,
125
- "num_examples": 872,
126
- "dataset_name": "glue"
127
- }
128
- },
129
- "download_checksums": {
130
- "https://dl.fbaipublicfiles.com/glue/data/SST-2.zip": {
131
- "num_bytes": 7439277,
132
- "checksum": "d67e16fb55739c1b32cdce9877596db1c127dc322d93c082281f64057c16deaa"
133
- }
134
- },
135
- "download_size": 7439277,
136
- "post_processing_size": null,
137
- "dataset_size": 5039531,
138
- "size_in_bytes": 12478808
139
- },
140
  "mrpc": {
141
  "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
142
  "citation": "@inproceedings{dolan2005automatically,\n title={Automatically constructing a corpus of sentential paraphrases},\n author={Dolan, William B and Brockett, Chris},\n booktitle={Proceedings of the Third International Workshop on Paraphrasing (IWP2005)},\n year={2005}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n\nNote that each GLUE dataset has its own citation. Please see the source to see\nthe correct citation for each contained dataset.",
@@ -218,658 +80,5 @@
218
  "post_processing_size": null,
219
  "dataset_size": 1495786,
220
  "size_in_bytes": 2990327
221
- },
222
- "qqp": {
223
- "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
224
- "citation": "@online{WinNT,\n author = {Iyer, Shankar and Dandekar, Nikhil and Csernai, Kornel},\n title = {First Quora Dataset Release: Question Pairs},\n year = {2017},\n url = {https://data.quora.com/First-Quora-Dataset-Release-Question-Pairs},\n urldate = {2019-04-03}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n",
225
- "homepage": "https://data.quora.com/First-Quora-Dataset-Release-Question-Pairs",
226
- "license": "",
227
- "features": {
228
- "question1": {
229
- "dtype": "string",
230
- "id": null,
231
- "_type": "Value"
232
- },
233
- "question2": {
234
- "dtype": "string",
235
- "id": null,
236
- "_type": "Value"
237
- },
238
- "label": {
239
- "num_classes": 2,
240
- "names": [
241
- "not_duplicate",
242
- "duplicate"
243
- ],
244
- "names_file": null,
245
- "id": null,
246
- "_type": "ClassLabel"
247
- },
248
- "idx": {
249
- "dtype": "int32",
250
- "id": null,
251
- "_type": "Value"
252
- }
253
- },
254
- "post_processed": null,
255
- "supervised_keys": null,
256
- "builder_name": "glue",
257
- "config_name": "qqp",
258
- "version": {
259
- "version_str": "1.0.0",
260
- "description": "",
261
- "major": 1,
262
- "minor": 0,
263
- "patch": 0
264
- },
265
- "splits": {
266
- "train": {
267
- "name": "train",
268
- "num_bytes": 50901116,
269
- "num_examples": 363846,
270
- "dataset_name": "glue"
271
- },
272
- "validation": {
273
- "name": "validation",
274
- "num_bytes": 5653794,
275
- "num_examples": 40430,
276
- "dataset_name": "glue"
277
- },
278
- "test": {
279
- "name": "test",
280
- "num_bytes": 55171431,
281
- "num_examples": 390965,
282
- "dataset_name": "glue"
283
- }
284
- },
285
- "download_checksums": {
286
- "https://dl.fbaipublicfiles.com/glue/data/QQP-clean.zip": {
287
- "num_bytes": 41696084,
288
- "checksum": "40e7c862c04eb26ee04b67fd900e76c45c6ba8e6d8fab4f8f1f8072a1a3fbae0"
289
- }
290
- },
291
- "download_size": 41696084,
292
- "post_processing_size": null,
293
- "dataset_size": 111726341,
294
- "size_in_bytes": 153422425
295
- },
296
- "stsb": {
297
- "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
298
- "citation": "@article{cer2017semeval,\n title={Semeval-2017 task 1: Semantic textual similarity-multilingual and cross-lingual focused evaluation},\n author={Cer, Daniel and Diab, Mona and Agirre, Eneko and Lopez-Gazpio, Inigo and Specia, Lucia},\n journal={arXiv preprint arXiv:1708.00055},\n year={2017}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n\nNote that each GLUE dataset has its own citation. Please see the source to see\nthe correct citation for each contained dataset.",
299
- "homepage": "http://ixa2.si.ehu.es/stswiki/index.php/STSbenchmark",
300
- "license": "",
301
- "features": {
302
- "sentence1": {
303
- "dtype": "string",
304
- "id": null,
305
- "_type": "Value"
306
- },
307
- "sentence2": {
308
- "dtype": "string",
309
- "id": null,
310
- "_type": "Value"
311
- },
312
- "label": {
313
- "dtype": "float32",
314
- "id": null,
315
- "_type": "Value"
316
- },
317
- "idx": {
318
- "dtype": "int32",
319
- "id": null,
320
- "_type": "Value"
321
- }
322
- },
323
- "post_processed": null,
324
- "supervised_keys": null,
325
- "builder_name": "glue",
326
- "config_name": "stsb",
327
- "version": {
328
- "version_str": "1.0.0",
329
- "description": "",
330
- "major": 1,
331
- "minor": 0,
332
- "patch": 0
333
- },
334
- "splits": {
335
- "test": {
336
- "name": "test",
337
- "num_bytes": 170847,
338
- "num_examples": 1379,
339
- "dataset_name": "glue"
340
- },
341
- "train": {
342
- "name": "train",
343
- "num_bytes": 758394,
344
- "num_examples": 5749,
345
- "dataset_name": "glue"
346
- },
347
- "validation": {
348
- "name": "validation",
349
- "num_bytes": 217012,
350
- "num_examples": 1500,
351
- "dataset_name": "glue"
352
- }
353
- },
354
- "download_checksums": {
355
- "https://dl.fbaipublicfiles.com/glue/data/STS-B.zip": {
356
- "num_bytes": 802872,
357
- "checksum": "e60a6393de5a8b5b9bac5020a1554b54e3691f9d600b775bd131e613ac179c85"
358
- }
359
- },
360
- "download_size": 802872,
361
- "post_processing_size": null,
362
- "dataset_size": 1146253,
363
- "size_in_bytes": 1949125
364
- },
365
- "mnli": {
366
- "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
367
- "citation": "@InProceedings{N18-1101,\n author = \"Williams, Adina\n and Nangia, Nikita\n and Bowman, Samuel\",\n title = \"A Broad-Coverage Challenge Corpus for\n Sentence Understanding through Inference\",\n booktitle = \"Proceedings of the 2018 Conference of\n the North American Chapter of the\n Association for Computational Linguistics:\n Human Language Technologies, Volume 1 (Long\n Papers)\",\n year = \"2018\",\n publisher = \"Association for Computational Linguistics\",\n pages = \"1112--1122\",\n location = \"New Orleans, Louisiana\",\n url = \"http://aclweb.org/anthology/N18-1101\"\n}\n@article{bowman2015large,\n title={A large annotated corpus for learning natural language inference},\n author={Bowman, Samuel R and Angeli, Gabor and Potts, Christopher and Manning, Christopher D},\n journal={arXiv preprint arXiv:1508.05326},\n year={2015}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n\nNote that each GLUE dataset has its own citation. Please see the source to see\nthe correct citation for each contained dataset.",
368
- "homepage": "http://www.nyu.edu/projects/bowman/multinli/",
369
- "license": "",
370
- "features": {
371
- "premise": {
372
- "dtype": "string",
373
- "id": null,
374
- "_type": "Value"
375
- },
376
- "hypothesis": {
377
- "dtype": "string",
378
- "id": null,
379
- "_type": "Value"
380
- },
381
- "label": {
382
- "num_classes": 3,
383
- "names": [
384
- "entailment",
385
- "neutral",
386
- "contradiction"
387
- ],
388
- "names_file": null,
389
- "id": null,
390
- "_type": "ClassLabel"
391
- },
392
- "idx": {
393
- "dtype": "int32",
394
- "id": null,
395
- "_type": "Value"
396
- }
397
- },
398
- "post_processed": null,
399
- "supervised_keys": null,
400
- "builder_name": "glue",
401
- "config_name": "mnli",
402
- "version": {
403
- "version_str": "1.0.0",
404
- "description": "",
405
- "major": 1,
406
- "minor": 0,
407
- "patch": 0
408
- },
409
- "splits": {
410
- "test_matched": {
411
- "name": "test_matched",
412
- "num_bytes": 1854787,
413
- "num_examples": 9796,
414
- "dataset_name": "glue"
415
- },
416
- "test_mismatched": {
417
- "name": "test_mismatched",
418
- "num_bytes": 1956866,
419
- "num_examples": 9847,
420
- "dataset_name": "glue"
421
- },
422
- "train": {
423
- "name": "train",
424
- "num_bytes": 74865118,
425
- "num_examples": 392702,
426
- "dataset_name": "glue"
427
- },
428
- "validation_matched": {
429
- "name": "validation_matched",
430
- "num_bytes": 1839926,
431
- "num_examples": 9815,
432
- "dataset_name": "glue"
433
- },
434
- "validation_mismatched": {
435
- "name": "validation_mismatched",
436
- "num_bytes": 1955384,
437
- "num_examples": 9832,
438
- "dataset_name": "glue"
439
- }
440
- },
441
- "download_checksums": {
442
- "https://dl.fbaipublicfiles.com/glue/data/MNLI.zip": {
443
- "num_bytes": 312783507,
444
- "checksum": "e7c1d896d26ed6caf700110645df426cc2d8ebf02a5ab743d5a5c68ac1c83633"
445
- }
446
- },
447
- "download_size": 312783507,
448
- "post_processing_size": null,
449
- "dataset_size": 82472081,
450
- "size_in_bytes": 395255588
451
- },
452
- "mnli_mismatched": {
453
- "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
454
- "citation": "@InProceedings{N18-1101,\n author = \"Williams, Adina\n and Nangia, Nikita\n and Bowman, Samuel\",\n title = \"A Broad-Coverage Challenge Corpus for\n Sentence Understanding through Inference\",\n booktitle = \"Proceedings of the 2018 Conference of\n the North American Chapter of the\n Association for Computational Linguistics:\n Human Language Technologies, Volume 1 (Long\n Papers)\",\n year = \"2018\",\n publisher = \"Association for Computational Linguistics\",\n pages = \"1112--1122\",\n location = \"New Orleans, Louisiana\",\n url = \"http://aclweb.org/anthology/N18-1101\"\n}\n@article{bowman2015large,\n title={A large annotated corpus for learning natural language inference},\n author={Bowman, Samuel R and Angeli, Gabor and Potts, Christopher and Manning, Christopher D},\n journal={arXiv preprint arXiv:1508.05326},\n year={2015}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n\nNote that each GLUE dataset has its own citation. Please see the source to see\nthe correct citation for each contained dataset.",
455
- "homepage": "http://www.nyu.edu/projects/bowman/multinli/",
456
- "license": "",
457
- "features": {
458
- "premise": {
459
- "dtype": "string",
460
- "id": null,
461
- "_type": "Value"
462
- },
463
- "hypothesis": {
464
- "dtype": "string",
465
- "id": null,
466
- "_type": "Value"
467
- },
468
- "label": {
469
- "num_classes": 3,
470
- "names": [
471
- "entailment",
472
- "neutral",
473
- "contradiction"
474
- ],
475
- "names_file": null,
476
- "id": null,
477
- "_type": "ClassLabel"
478
- },
479
- "idx": {
480
- "dtype": "int32",
481
- "id": null,
482
- "_type": "Value"
483
- }
484
- },
485
- "post_processed": null,
486
- "supervised_keys": null,
487
- "builder_name": "glue",
488
- "config_name": "mnli_mismatched",
489
- "version": {
490
- "version_str": "1.0.0",
491
- "description": "",
492
- "major": 1,
493
- "minor": 0,
494
- "patch": 0
495
- },
496
- "splits": {
497
- "test": {
498
- "name": "test",
499
- "num_bytes": 1956866,
500
- "num_examples": 9847,
501
- "dataset_name": "glue"
502
- },
503
- "validation": {
504
- "name": "validation",
505
- "num_bytes": 1955384,
506
- "num_examples": 9832,
507
- "dataset_name": "glue"
508
- }
509
- },
510
- "download_checksums": {
511
- "https://dl.fbaipublicfiles.com/glue/data/MNLI.zip": {
512
- "num_bytes": 312783507,
513
- "checksum": "e7c1d896d26ed6caf700110645df426cc2d8ebf02a5ab743d5a5c68ac1c83633"
514
- }
515
- },
516
- "download_size": 312783507,
517
- "post_processing_size": null,
518
- "dataset_size": 3912250,
519
- "size_in_bytes": 316695757
520
- },
521
- "mnli_matched": {
522
- "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
523
- "citation": "@InProceedings{N18-1101,\n author = \"Williams, Adina\n and Nangia, Nikita\n and Bowman, Samuel\",\n title = \"A Broad-Coverage Challenge Corpus for\n Sentence Understanding through Inference\",\n booktitle = \"Proceedings of the 2018 Conference of\n the North American Chapter of the\n Association for Computational Linguistics:\n Human Language Technologies, Volume 1 (Long\n Papers)\",\n year = \"2018\",\n publisher = \"Association for Computational Linguistics\",\n pages = \"1112--1122\",\n location = \"New Orleans, Louisiana\",\n url = \"http://aclweb.org/anthology/N18-1101\"\n}\n@article{bowman2015large,\n title={A large annotated corpus for learning natural language inference},\n author={Bowman, Samuel R and Angeli, Gabor and Potts, Christopher and Manning, Christopher D},\n journal={arXiv preprint arXiv:1508.05326},\n year={2015}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n\nNote that each GLUE dataset has its own citation. Please see the source to see\nthe correct citation for each contained dataset.",
524
- "homepage": "http://www.nyu.edu/projects/bowman/multinli/",
525
- "license": "",
526
- "features": {
527
- "premise": {
528
- "dtype": "string",
529
- "id": null,
530
- "_type": "Value"
531
- },
532
- "hypothesis": {
533
- "dtype": "string",
534
- "id": null,
535
- "_type": "Value"
536
- },
537
- "label": {
538
- "num_classes": 3,
539
- "names": [
540
- "entailment",
541
- "neutral",
542
- "contradiction"
543
- ],
544
- "names_file": null,
545
- "id": null,
546
- "_type": "ClassLabel"
547
- },
548
- "idx": {
549
- "dtype": "int32",
550
- "id": null,
551
- "_type": "Value"
552
- }
553
- },
554
- "post_processed": null,
555
- "supervised_keys": null,
556
- "builder_name": "glue",
557
- "config_name": "mnli_matched",
558
- "version": {
559
- "version_str": "1.0.0",
560
- "description": "",
561
- "major": 1,
562
- "minor": 0,
563
- "patch": 0
564
- },
565
- "splits": {
566
- "test": {
567
- "name": "test",
568
- "num_bytes": 1854787,
569
- "num_examples": 9796,
570
- "dataset_name": "glue"
571
- },
572
- "validation": {
573
- "name": "validation",
574
- "num_bytes": 1839926,
575
- "num_examples": 9815,
576
- "dataset_name": "glue"
577
- }
578
- },
579
- "download_checksums": {
580
- "https://dl.fbaipublicfiles.com/glue/data/MNLI.zip": {
581
- "num_bytes": 312783507,
582
- "checksum": "e7c1d896d26ed6caf700110645df426cc2d8ebf02a5ab743d5a5c68ac1c83633"
583
- }
584
- },
585
- "download_size": 312783507,
586
- "post_processing_size": null,
587
- "dataset_size": 3694713,
588
- "size_in_bytes": 316478220
589
- },
590
- "qnli": {
591
- "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
592
- "citation": "@article{rajpurkar2016squad,\n title={Squad: 100,000+ questions for machine comprehension of text},\n author={Rajpurkar, Pranav and Zhang, Jian and Lopyrev, Konstantin and Liang, Percy},\n journal={arXiv preprint arXiv:1606.05250},\n year={2016}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n\nNote that each GLUE dataset has its own citation. Please see the source to see\nthe correct citation for each contained dataset.",
593
- "homepage": "https://rajpurkar.github.io/SQuAD-explorer/",
594
- "license": "",
595
- "features": {
596
- "question": {
597
- "dtype": "string",
598
- "id": null,
599
- "_type": "Value"
600
- },
601
- "sentence": {
602
- "dtype": "string",
603
- "id": null,
604
- "_type": "Value"
605
- },
606
- "label": {
607
- "num_classes": 2,
608
- "names": [
609
- "entailment",
610
- "not_entailment"
611
- ],
612
- "names_file": null,
613
- "id": null,
614
- "_type": "ClassLabel"
615
- },
616
- "idx": {
617
- "dtype": "int32",
618
- "id": null,
619
- "_type": "Value"
620
- }
621
- },
622
- "post_processed": null,
623
- "supervised_keys": null,
624
- "builder_name": "glue",
625
- "config_name": "qnli",
626
- "version": {
627
- "version_str": "1.0.0",
628
- "description": "",
629
- "major": 1,
630
- "minor": 0,
631
- "patch": 0
632
- },
633
- "splits": {
634
- "test": {
635
- "name": "test",
636
- "num_bytes": 1376516,
637
- "num_examples": 5463,
638
- "dataset_name": "glue"
639
- },
640
- "train": {
641
- "name": "train",
642
- "num_bytes": 25677924,
643
- "num_examples": 104743,
644
- "dataset_name": "glue"
645
- },
646
- "validation": {
647
- "name": "validation",
648
- "num_bytes": 1371727,
649
- "num_examples": 5463,
650
- "dataset_name": "glue"
651
- }
652
- },
653
- "download_checksums": {
654
- "https://dl.fbaipublicfiles.com/glue/data/QNLIv2.zip": {
655
- "num_bytes": 10627589,
656
- "checksum": "e634e78627a29adaecd4f955359b22bf5e70f2cbd93b493f2d624138a0c0e5f5"
657
- }
658
- },
659
- "download_size": 10627589,
660
- "post_processing_size": null,
661
- "dataset_size": 28426167,
662
- "size_in_bytes": 39053756
663
- },
664
- "rte": {
665
- "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
666
- "citation": "@inproceedings{dagan2005pascal,\n title={The PASCAL recognising textual entailment challenge},\n author={Dagan, Ido and Glickman, Oren and Magnini, Bernardo},\n booktitle={Machine Learning Challenges Workshop},\n pages={177--190},\n year={2005},\n organization={Springer}\n}\n@inproceedings{bar2006second,\n title={The second pascal recognising textual entailment challenge},\n author={Bar-Haim, Roy and Dagan, Ido and Dolan, Bill and Ferro, Lisa and Giampiccolo, Danilo and Magnini, Bernardo and Szpektor, Idan},\n booktitle={Proceedings of the second PASCAL challenges workshop on recognising textual entailment},\n volume={6},\n number={1},\n pages={6--4},\n year={2006},\n organization={Venice}\n}\n@inproceedings{giampiccolo2007third,\n title={The third pascal recognizing textual entailment challenge},\n author={Giampiccolo, Danilo and Magnini, Bernardo and Dagan, Ido and Dolan, Bill},\n booktitle={Proceedings of the ACL-PASCAL workshop on textual entailment and paraphrasing},\n pages={1--9},\n year={2007},\n organization={Association for Computational Linguistics}\n}\n@inproceedings{bentivogli2009fifth,\n title={The Fifth PASCAL Recognizing Textual Entailment Challenge.},\n author={Bentivogli, Luisa and Clark, Peter and Dagan, Ido and Giampiccolo, Danilo},\n booktitle={TAC},\n year={2009}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n\nNote that each GLUE dataset has its own citation. Please see the source to see\nthe correct citation for each contained dataset.",
667
- "homepage": "https://aclweb.org/aclwiki/Recognizing_Textual_Entailment",
668
- "license": "",
669
- "features": {
670
- "sentence1": {
671
- "dtype": "string",
672
- "id": null,
673
- "_type": "Value"
674
- },
675
- "sentence2": {
676
- "dtype": "string",
677
- "id": null,
678
- "_type": "Value"
679
- },
680
- "label": {
681
- "num_classes": 2,
682
- "names": [
683
- "entailment",
684
- "not_entailment"
685
- ],
686
- "names_file": null,
687
- "id": null,
688
- "_type": "ClassLabel"
689
- },
690
- "idx": {
691
- "dtype": "int32",
692
- "id": null,
693
- "_type": "Value"
694
- }
695
- },
696
- "post_processed": null,
697
- "supervised_keys": null,
698
- "builder_name": "glue",
699
- "config_name": "rte",
700
- "version": {
701
- "version_str": "1.0.0",
702
- "description": "",
703
- "major": 1,
704
- "minor": 0,
705
- "patch": 0
706
- },
707
- "splits": {
708
- "test": {
709
- "name": "test",
710
- "num_bytes": 975936,
711
- "num_examples": 3000,
712
- "dataset_name": "glue"
713
- },
714
- "train": {
715
- "name": "train",
716
- "num_bytes": 848888,
717
- "num_examples": 2490,
718
- "dataset_name": "glue"
719
- },
720
- "validation": {
721
- "name": "validation",
722
- "num_bytes": 90911,
723
- "num_examples": 277,
724
- "dataset_name": "glue"
725
- }
726
- },
727
- "download_checksums": {
728
- "https://dl.fbaipublicfiles.com/glue/data/RTE.zip": {
729
- "num_bytes": 697150,
730
- "checksum": "6bf86de103ecd335f3441bd43574d23fef87ecc695977a63b82d5efb206556ee"
731
- }
732
- },
733
- "download_size": 697150,
734
- "post_processing_size": null,
735
- "dataset_size": 1915735,
736
- "size_in_bytes": 2612885
737
- },
738
- "wnli": {
739
- "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
740
- "citation": "@inproceedings{levesque2012winograd,\n title={The winograd schema challenge},\n author={Levesque, Hector and Davis, Ernest and Morgenstern, Leora},\n booktitle={Thirteenth International Conference on the Principles of Knowledge Representation and Reasoning},\n year={2012}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n\nNote that each GLUE dataset has its own citation. Please see the source to see\nthe correct citation for each contained dataset.",
741
- "homepage": "https://cs.nyu.edu/faculty/davise/papers/WinogradSchemas/WS.html",
742
- "license": "",
743
- "features": {
744
- "sentence1": {
745
- "dtype": "string",
746
- "id": null,
747
- "_type": "Value"
748
- },
749
- "sentence2": {
750
- "dtype": "string",
751
- "id": null,
752
- "_type": "Value"
753
- },
754
- "label": {
755
- "num_classes": 2,
756
- "names": [
757
- "not_entailment",
758
- "entailment"
759
- ],
760
- "names_file": null,
761
- "id": null,
762
- "_type": "ClassLabel"
763
- },
764
- "idx": {
765
- "dtype": "int32",
766
- "id": null,
767
- "_type": "Value"
768
- }
769
- },
770
- "post_processed": null,
771
- "supervised_keys": null,
772
- "builder_name": "glue",
773
- "config_name": "wnli",
774
- "version": {
775
- "version_str": "1.0.0",
776
- "description": "",
777
- "major": 1,
778
- "minor": 0,
779
- "patch": 0
780
- },
781
- "splits": {
782
- "test": {
783
- "name": "test",
784
- "num_bytes": 37992,
785
- "num_examples": 146,
786
- "dataset_name": "glue"
787
- },
788
- "train": {
789
- "name": "train",
790
- "num_bytes": 107517,
791
- "num_examples": 635,
792
- "dataset_name": "glue"
793
- },
794
- "validation": {
795
- "name": "validation",
796
- "num_bytes": 12215,
797
- "num_examples": 71,
798
- "dataset_name": "glue"
799
- }
800
- },
801
- "download_checksums": {
802
- "https://dl.fbaipublicfiles.com/glue/data/WNLI.zip": {
803
- "num_bytes": 28999,
804
- "checksum": "ae0e8e4d16f4d46d4a0a566ec7ecceccfd3fbfaa4a7a4b4e02848c0f2561ac46"
805
- }
806
- },
807
- "download_size": 28999,
808
- "post_processing_size": null,
809
- "dataset_size": 157724,
810
- "size_in_bytes": 186723
811
- },
812
- "ax": {
813
- "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
814
- "citation": "\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n\nNote that each GLUE dataset has its own citation. Please see the source to see\nthe correct citation for each contained dataset.",
815
- "homepage": "https://gluebenchmark.com/diagnostics",
816
- "license": "",
817
- "features": {
818
- "premise": {
819
- "dtype": "string",
820
- "id": null,
821
- "_type": "Value"
822
- },
823
- "hypothesis": {
824
- "dtype": "string",
825
- "id": null,
826
- "_type": "Value"
827
- },
828
- "label": {
829
- "num_classes": 3,
830
- "names": [
831
- "entailment",
832
- "neutral",
833
- "contradiction"
834
- ],
835
- "names_file": null,
836
- "id": null,
837
- "_type": "ClassLabel"
838
- },
839
- "idx": {
840
- "dtype": "int32",
841
- "id": null,
842
- "_type": "Value"
843
- }
844
- },
845
- "post_processed": null,
846
- "supervised_keys": null,
847
- "builder_name": "glue",
848
- "config_name": "ax",
849
- "version": {
850
- "version_str": "1.0.0",
851
- "description": "",
852
- "major": 1,
853
- "minor": 0,
854
- "patch": 0
855
- },
856
- "splits": {
857
- "test": {
858
- "name": "test",
859
- "num_bytes": 238392,
860
- "num_examples": 1104,
861
- "dataset_name": "glue"
862
- }
863
- },
864
- "download_checksums": {
865
- "https://dl.fbaipublicfiles.com/glue/data/AX.tsv": {
866
- "num_bytes": 222257,
867
- "checksum": "0e13510b1bb14436ff7e2ee82338f0efb0133ecf2e73507a697dc210db3f05fd"
868
- }
869
- },
870
- "download_size": 222257,
871
- "post_processing_size": null,
872
- "dataset_size": 238392,
873
- "size_in_bytes": 460649
874
  }
875
  }
 
1
  {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  "mrpc": {
3
  "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
4
  "citation": "@inproceedings{dolan2005automatically,\n title={Automatically constructing a corpus of sentential paraphrases},\n author={Dolan, William B and Brockett, Chris},\n booktitle={Proceedings of the Third International Workshop on Paraphrasing (IWP2005)},\n year={2005}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n\nNote that each GLUE dataset has its own citation. Please see the source to see\nthe correct citation for each contained dataset.",
 
80
  "post_processing_size": null,
81
  "dataset_size": 1495786,
82
  "size_in_bytes": 2990327
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
83
  }
84
  }