Datasets:
Commit
·
c234b36
1
Parent(s):
b04c7fe
Add mnli_matched data files
Browse files- README.md +12 -6
- dataset_infos.json +13 -27
- mnli_matched/test-00000-of-00001.parquet +3 -0
- mnli_matched/validation-00000-of-00001.parquet +3 -0
README.md
CHANGED
|
@@ -135,14 +135,14 @@ dataset_info:
|
|
| 135 |
- name: idx
|
| 136 |
dtype: int32
|
| 137 |
splits:
|
| 138 |
-
- name: test
|
| 139 |
-
num_bytes: 1854787
|
| 140 |
-
num_examples: 9796
|
| 141 |
- name: validation
|
| 142 |
-
num_bytes:
|
| 143 |
num_examples: 9815
|
| 144 |
-
|
| 145 |
-
|
|
|
|
|
|
|
|
|
|
| 146 |
- config_name: mnli_mismatched
|
| 147 |
features:
|
| 148 |
- name: premise
|
|
@@ -364,6 +364,12 @@ configs:
|
|
| 364 |
path: mnli/test_matched-*
|
| 365 |
- split: test_mismatched
|
| 366 |
path: mnli/test_mismatched-*
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 367 |
- config_name: mnli_mismatched
|
| 368 |
data_files:
|
| 369 |
- split: validation
|
|
|
|
| 135 |
- name: idx
|
| 136 |
dtype: int32
|
| 137 |
splits:
|
|
|
|
|
|
|
|
|
|
| 138 |
- name: validation
|
| 139 |
+
num_bytes: 1833783
|
| 140 |
num_examples: 9815
|
| 141 |
+
- name: test
|
| 142 |
+
num_bytes: 1848654
|
| 143 |
+
num_examples: 9796
|
| 144 |
+
download_size: 2435055
|
| 145 |
+
dataset_size: 3682437
|
| 146 |
- config_name: mnli_mismatched
|
| 147 |
features:
|
| 148 |
- name: premise
|
|
|
|
| 364 |
path: mnli/test_matched-*
|
| 365 |
- split: test_mismatched
|
| 366 |
path: mnli/test_mismatched-*
|
| 367 |
+
- config_name: mnli_matched
|
| 368 |
+
data_files:
|
| 369 |
+
- split: validation
|
| 370 |
+
path: mnli_matched/validation-*
|
| 371 |
+
- split: test
|
| 372 |
+
path: mnli_matched/test-*
|
| 373 |
- config_name: mnli_mismatched
|
| 374 |
data_files:
|
| 375 |
- split: validation
|
dataset_infos.json
CHANGED
|
@@ -418,40 +418,33 @@
|
|
| 418 |
},
|
| 419 |
"mnli_matched": {
|
| 420 |
"description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
|
| 421 |
-
"citation": "@InProceedings{N18-1101,\n author = \"Williams, Adina\n and Nangia, Nikita\n and Bowman, Samuel\",\n title = \"A Broad-Coverage Challenge Corpus for\n Sentence Understanding through Inference\",\n booktitle = \"Proceedings of the 2018 Conference of\n the North American Chapter of the\n Association for Computational Linguistics:\n Human Language Technologies, Volume 1 (Long\n Papers)\",\n year = \"2018\",\n publisher = \"Association for Computational Linguistics\",\n pages = \"1112--1122\",\n location = \"New Orleans, Louisiana\",\n url = \"http://aclweb.org/anthology/N18-1101\"\n}\n@article{bowman2015large,\n title={A large annotated corpus for learning natural language inference},\n author={Bowman, Samuel R and Angeli, Gabor and Potts, Christopher and Manning, Christopher D},\n journal={arXiv preprint arXiv:1508.05326},\n year={2015}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n
|
| 422 |
"homepage": "http://www.nyu.edu/projects/bowman/multinli/",
|
| 423 |
"license": "",
|
| 424 |
"features": {
|
| 425 |
"premise": {
|
| 426 |
"dtype": "string",
|
| 427 |
-
"id": null,
|
| 428 |
"_type": "Value"
|
| 429 |
},
|
| 430 |
"hypothesis": {
|
| 431 |
"dtype": "string",
|
| 432 |
-
"id": null,
|
| 433 |
"_type": "Value"
|
| 434 |
},
|
| 435 |
"label": {
|
| 436 |
-
"num_classes": 3,
|
| 437 |
"names": [
|
| 438 |
"entailment",
|
| 439 |
"neutral",
|
| 440 |
"contradiction"
|
| 441 |
],
|
| 442 |
-
"names_file": null,
|
| 443 |
-
"id": null,
|
| 444 |
"_type": "ClassLabel"
|
| 445 |
},
|
| 446 |
"idx": {
|
| 447 |
"dtype": "int32",
|
| 448 |
-
"id": null,
|
| 449 |
"_type": "Value"
|
| 450 |
}
|
| 451 |
},
|
| 452 |
-
"post_processed": null,
|
| 453 |
-
"supervised_keys": null,
|
| 454 |
"builder_name": "glue",
|
|
|
|
| 455 |
"config_name": "mnli_matched",
|
| 456 |
"version": {
|
| 457 |
"version_str": "1.0.0",
|
|
@@ -461,29 +454,22 @@
|
|
| 461 |
"patch": 0
|
| 462 |
},
|
| 463 |
"splits": {
|
| 464 |
-
"test": {
|
| 465 |
-
"name": "test",
|
| 466 |
-
"num_bytes": 1854787,
|
| 467 |
-
"num_examples": 9796,
|
| 468 |
-
"dataset_name": "glue"
|
| 469 |
-
},
|
| 470 |
"validation": {
|
| 471 |
"name": "validation",
|
| 472 |
-
"num_bytes":
|
| 473 |
"num_examples": 9815,
|
| 474 |
-
"dataset_name":
|
| 475 |
-
}
|
| 476 |
-
|
| 477 |
-
|
| 478 |
-
|
| 479 |
-
"
|
| 480 |
-
"
|
| 481 |
}
|
| 482 |
},
|
| 483 |
-
"download_size":
|
| 484 |
-
"
|
| 485 |
-
"
|
| 486 |
-
"size_in_bytes": 316478220
|
| 487 |
},
|
| 488 |
"qnli": {
|
| 489 |
"description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
|
|
|
|
| 418 |
},
|
| 419 |
"mnli_matched": {
|
| 420 |
"description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
|
| 421 |
+
"citation": "@InProceedings{N18-1101,\n author = \"Williams, Adina\n and Nangia, Nikita\n and Bowman, Samuel\",\n title = \"A Broad-Coverage Challenge Corpus for\n Sentence Understanding through Inference\",\n booktitle = \"Proceedings of the 2018 Conference of\n the North American Chapter of the\n Association for Computational Linguistics:\n Human Language Technologies, Volume 1 (Long\n Papers)\",\n year = \"2018\",\n publisher = \"Association for Computational Linguistics\",\n pages = \"1112--1122\",\n location = \"New Orleans, Louisiana\",\n url = \"http://aclweb.org/anthology/N18-1101\"\n}\n@article{bowman2015large,\n title={A large annotated corpus for learning natural language inference},\n author={Bowman, Samuel R and Angeli, Gabor and Potts, Christopher and Manning, Christopher D},\n journal={arXiv preprint arXiv:1508.05326},\n year={2015}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n",
|
| 422 |
"homepage": "http://www.nyu.edu/projects/bowman/multinli/",
|
| 423 |
"license": "",
|
| 424 |
"features": {
|
| 425 |
"premise": {
|
| 426 |
"dtype": "string",
|
|
|
|
| 427 |
"_type": "Value"
|
| 428 |
},
|
| 429 |
"hypothesis": {
|
| 430 |
"dtype": "string",
|
|
|
|
| 431 |
"_type": "Value"
|
| 432 |
},
|
| 433 |
"label": {
|
|
|
|
| 434 |
"names": [
|
| 435 |
"entailment",
|
| 436 |
"neutral",
|
| 437 |
"contradiction"
|
| 438 |
],
|
|
|
|
|
|
|
| 439 |
"_type": "ClassLabel"
|
| 440 |
},
|
| 441 |
"idx": {
|
| 442 |
"dtype": "int32",
|
|
|
|
| 443 |
"_type": "Value"
|
| 444 |
}
|
| 445 |
},
|
|
|
|
|
|
|
| 446 |
"builder_name": "glue",
|
| 447 |
+
"dataset_name": "glue",
|
| 448 |
"config_name": "mnli_matched",
|
| 449 |
"version": {
|
| 450 |
"version_str": "1.0.0",
|
|
|
|
| 454 |
"patch": 0
|
| 455 |
},
|
| 456 |
"splits": {
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 457 |
"validation": {
|
| 458 |
"name": "validation",
|
| 459 |
+
"num_bytes": 1833783,
|
| 460 |
"num_examples": 9815,
|
| 461 |
+
"dataset_name": null
|
| 462 |
+
},
|
| 463 |
+
"test": {
|
| 464 |
+
"name": "test",
|
| 465 |
+
"num_bytes": 1848654,
|
| 466 |
+
"num_examples": 9796,
|
| 467 |
+
"dataset_name": null
|
| 468 |
}
|
| 469 |
},
|
| 470 |
+
"download_size": 2435055,
|
| 471 |
+
"dataset_size": 3682437,
|
| 472 |
+
"size_in_bytes": 6117492
|
|
|
|
| 473 |
},
|
| 474 |
"qnli": {
|
| 475 |
"description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
|
mnli_matched/test-00000-of-00001.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a330c4f2aeb0bc92f1b4b133fbbaf51bf9c7d0f5cac3d06f49ef63af47dbb822
|
| 3 |
+
size 1220119
|
mnli_matched/validation-00000-of-00001.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7f918c09d9c35446b8e8f06a5672f8ab704e2897fecbf52e2e154141f3d7c421
|
| 3 |
+
size 1214936
|