Datasets:
Tasks:
Question Answering
Modalities:
Text
Formats:
parquet
Sub-tasks:
multiple-choice-qa
Languages:
English
Size:
1M - 10M
ArXiv:
License:
| { | |
| "abstract_algebra": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "abstract_algebra", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 21316, | |
| "num_examples": 100, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 2232, | |
| "num_examples": 11, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 918, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47179638, | |
| "dataset_size": 161025091, | |
| "size_in_bytes": 208204729 | |
| }, | |
| "anatomy": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "anatomy", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 34594, | |
| "num_examples": 135, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 3282, | |
| "num_examples": 14, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1010, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47191229, | |
| "dataset_size": 161039511, | |
| "size_in_bytes": 208230740 | |
| }, | |
| "astronomy": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "astronomy", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 48735, | |
| "num_examples": 152, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 5223, | |
| "num_examples": 16, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 2129, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47201551, | |
| "dataset_size": 161056712, | |
| "size_in_bytes": 208258263 | |
| }, | |
| "business_ethics": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "business_ethics", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 35140, | |
| "num_examples": 100, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 3235, | |
| "num_examples": 11, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 2273, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47193421, | |
| "dataset_size": 161041273, | |
| "size_in_bytes": 208234694 | |
| }, | |
| "clinical_knowledge": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "clinical_knowledge", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 68572, | |
| "num_examples": 265, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 7290, | |
| "num_examples": 29, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1308, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47213955, | |
| "dataset_size": 161077795, | |
| "size_in_bytes": 208291750 | |
| }, | |
| "college_biology": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "college_biology", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 51521, | |
| "num_examples": 144, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 5111, | |
| "num_examples": 16, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1615, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47205152, | |
| "dataset_size": 161058872, | |
| "size_in_bytes": 208264024 | |
| }, | |
| "college_chemistry": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "college_chemistry", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 26796, | |
| "num_examples": 100, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 2484, | |
| "num_examples": 8, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1424, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47188958, | |
| "dataset_size": 161031329, | |
| "size_in_bytes": 208220287 | |
| }, | |
| "college_computer_science": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "college_computer_science", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 45429, | |
| "num_examples": 100, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 4959, | |
| "num_examples": 11, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 2893, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47201966, | |
| "dataset_size": 161053906, | |
| "size_in_bytes": 208255872 | |
| }, | |
| "college_mathematics": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "college_mathematics", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 26999, | |
| "num_examples": 100, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 2909, | |
| "num_examples": 11, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1596, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47188597, | |
| "dataset_size": 161032129, | |
| "size_in_bytes": 208220726 | |
| }, | |
| "college_medicine": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "college_medicine", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 85845, | |
| "num_examples": 173, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 8337, | |
| "num_examples": 22, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1758, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47218201, | |
| "dataset_size": 161096565, | |
| "size_in_bytes": 208314766 | |
| }, | |
| "college_physics": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "college_physics", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 32107, | |
| "num_examples": 102, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 3687, | |
| "num_examples": 11, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1495, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47190901, | |
| "dataset_size": 161037914, | |
| "size_in_bytes": 208228815 | |
| }, | |
| "computer_security": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "computer_security", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 29212, | |
| "num_examples": 100, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 4768, | |
| "num_examples": 11, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1194, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47192155, | |
| "dataset_size": 161035799, | |
| "size_in_bytes": 208227954 | |
| }, | |
| "conceptual_physics": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "conceptual_physics", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 45867, | |
| "num_examples": 235, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 5034, | |
| "num_examples": 26, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1032, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47197231, | |
| "dataset_size": 161052558, | |
| "size_in_bytes": 208249789 | |
| }, | |
| "econometrics": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "econometrics", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 48359, | |
| "num_examples": 114, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 5147, | |
| "num_examples": 12, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1712, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47197846, | |
| "dataset_size": 161055843, | |
| "size_in_bytes": 208253689 | |
| }, | |
| "electrical_engineering": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "electrical_engineering", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 28900, | |
| "num_examples": 145, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 3307, | |
| "num_examples": 16, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1090, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47189021, | |
| "dataset_size": 161033922, | |
| "size_in_bytes": 208222943 | |
| }, | |
| "elementary_mathematics": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "elementary_mathematics", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 79924, | |
| "num_examples": 378, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 10042, | |
| "num_examples": 41, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1558, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47216972, | |
| "dataset_size": 161092149, | |
| "size_in_bytes": 208309121 | |
| }, | |
| "formal_logic": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "formal_logic", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 51789, | |
| "num_examples": 126, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 6464, | |
| "num_examples": 14, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1825, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47194349, | |
| "dataset_size": 161060703, | |
| "size_in_bytes": 208255052 | |
| }, | |
| "global_facts": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "global_facts", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 19991, | |
| "num_examples": 100, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 2013, | |
| "num_examples": 10, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1297, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47181634, | |
| "dataset_size": 161023926, | |
| "size_in_bytes": 208205560 | |
| }, | |
| "high_school_biology": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "high_school_biology", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 116850, | |
| "num_examples": 310, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 11746, | |
| "num_examples": 32, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1776, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47239946, | |
| "dataset_size": 161130997, | |
| "size_in_bytes": 208370943 | |
| }, | |
| "high_school_chemistry": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "high_school_chemistry", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 63527, | |
| "num_examples": 203, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 7630, | |
| "num_examples": 22, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1333, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47207769, | |
| "dataset_size": 161073115, | |
| "size_in_bytes": 208280884 | |
| }, | |
| "high_school_computer_science": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "high_school_computer_science", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 47664, | |
| "num_examples": 100, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 3619, | |
| "num_examples": 9, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 3066, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47200669, | |
| "dataset_size": 161054974, | |
| "size_in_bytes": 208255643 | |
| }, | |
| "high_school_european_history": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "high_school_european_history", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 275568, | |
| "num_examples": 165, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 30196, | |
| "num_examples": 18, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 11712, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47349494, | |
| "dataset_size": 161318101, | |
| "size_in_bytes": 208667595 | |
| }, | |
| "high_school_geography": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "high_school_geography", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 46972, | |
| "num_examples": 198, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 4870, | |
| "num_examples": 22, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1516, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47200648, | |
| "dataset_size": 161053983, | |
| "size_in_bytes": 208254631 | |
| }, | |
| "high_school_government_and_politics": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "high_school_government_and_politics", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 73589, | |
| "num_examples": 193, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 7870, | |
| "num_examples": 21, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1962, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47214961, | |
| "dataset_size": 161084046, | |
| "size_in_bytes": 208299007 | |
| }, | |
| "high_school_macroeconomics": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "high_school_macroeconomics", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 129375, | |
| "num_examples": 390, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 14298, | |
| "num_examples": 43, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1466, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47231078, | |
| "dataset_size": 161145764, | |
| "size_in_bytes": 208376842 | |
| }, | |
| "high_school_mathematics": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "high_school_mathematics", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 62132, | |
| "num_examples": 270, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 6536, | |
| "num_examples": 29, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1420, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47207019, | |
| "dataset_size": 161070713, | |
| "size_in_bytes": 208277732 | |
| }, | |
| "high_school_microeconomics": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "high_school_microeconomics", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 82831, | |
| "num_examples": 238, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 8321, | |
| "num_examples": 26, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1436, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47212056, | |
| "dataset_size": 161093213, | |
| "size_in_bytes": 208305269 | |
| }, | |
| "high_school_physics": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "high_school_physics", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 62999, | |
| "num_examples": 151, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 7150, | |
| "num_examples": 17, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1592, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47207223, | |
| "dataset_size": 161072366, | |
| "size_in_bytes": 208279589 | |
| }, | |
| "high_school_psychology": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "high_school_psychology", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 173565, | |
| "num_examples": 545, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 18817, | |
| "num_examples": 60, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 2023, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47274794, | |
| "dataset_size": 161195030, | |
| "size_in_bytes": 208469824 | |
| }, | |
| "high_school_statistics": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "high_school_statistics", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 116306, | |
| "num_examples": 216, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 10583, | |
| "num_examples": 23, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 2646, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47236110, | |
| "dataset_size": 161130160, | |
| "size_in_bytes": 208366270 | |
| }, | |
| "high_school_us_history": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "high_school_us_history", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 302026, | |
| "num_examples": 204, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 32266, | |
| "num_examples": 22, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 8982, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47355390, | |
| "dataset_size": 161343899, | |
| "size_in_bytes": 208699289 | |
| }, | |
| "high_school_world_history": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "high_school_world_history", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 385478, | |
| "num_examples": 237, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 46243, | |
| "num_examples": 26, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 5015, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47405363, | |
| "dataset_size": 161437361, | |
| "size_in_bytes": 208842724 | |
| }, | |
| "human_aging": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "human_aging", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 49431, | |
| "num_examples": 223, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 5040, | |
| "num_examples": 23, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1071, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47203557, | |
| "dataset_size": 161056167, | |
| "size_in_bytes": 208259724 | |
| }, | |
| "human_sexuality": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "human_sexuality", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 34587, | |
| "num_examples": 131, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 2637, | |
| "num_examples": 12, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1160, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47194557, | |
| "dataset_size": 161039009, | |
| "size_in_bytes": 208233566 | |
| }, | |
| "international_law": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "international_law", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 56060, | |
| "num_examples": 121, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 6734, | |
| "num_examples": 13, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 2511, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47203538, | |
| "dataset_size": 161065930, | |
| "size_in_bytes": 208269468 | |
| }, | |
| "jurisprudence": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "jurisprudence", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 35810, | |
| "num_examples": 108, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 3904, | |
| "num_examples": 11, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1376, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47195740, | |
| "dataset_size": 161041715, | |
| "size_in_bytes": 208237455 | |
| }, | |
| "logical_fallacies": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "logical_fallacies", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 53528, | |
| "num_examples": 163, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 5469, | |
| "num_examples": 18, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1666, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47196007, | |
| "dataset_size": 161061288, | |
| "size_in_bytes": 208257295 | |
| }, | |
| "machine_learning": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "machine_learning", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 36108, | |
| "num_examples": 112, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 3440, | |
| "num_examples": 11, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 2411, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47192772, | |
| "dataset_size": 161042584, | |
| "size_in_bytes": 208235356 | |
| }, | |
| "management": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "management", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 21432, | |
| "num_examples": 103, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 1962, | |
| "num_examples": 11, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 956, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47185208, | |
| "dataset_size": 161024975, | |
| "size_in_bytes": 208210183 | |
| }, | |
| "marketing": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "marketing", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 66055, | |
| "num_examples": 234, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 7707, | |
| "num_examples": 25, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1534, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47211905, | |
| "dataset_size": 161075921, | |
| "size_in_bytes": 208287826 | |
| }, | |
| "medical_genetics": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "medical_genetics", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 22852, | |
| "num_examples": 100, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 3213, | |
| "num_examples": 11, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1177, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47187825, | |
| "dataset_size": 161027867, | |
| "size_in_bytes": 208215692 | |
| }, | |
| "miscellaneous": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "miscellaneous", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 161003, | |
| "num_examples": 783, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 15780, | |
| "num_examples": 86, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 772, | |
| "num_examples": 5, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 47277261, | |
| "dataset_size": 161178180, | |
| "size_in_bytes": 208455441 | |
| }, | |
| "moral_disputes": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "length": -1, | |
| "id": null, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "num_classes": 4, | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "id": null, | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "post_processed": null, | |
| "supervised_keys": null, | |
| "task_templates": null, | |
| "builder_name": "mmlu", | |
| "config_name": "moral_disputes", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "description": null, | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 160601257, | |
| "num_examples": 99842, | |
| "dataset_name": "mmlu" | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 107806, | |
| "num_examples": 346, | |
| "dataset_name": "mmlu" | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 12408, | |
| "num_examples": 38, | |
| "dataset_name": "mmlu" | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1743, | |
| "num_examples": 5, | |
| "dataset_name": "mmlu" | |
| } | |
| }, | |
| "download_checksums": { | |
| "data.tar": { | |
| "num_bytes": 166184960, | |
| "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b" | |
| } | |
| }, | |
| "download_size": 166184960, | |
| "post_processing_size": null, | |
| "dataset_size": 160723214, | |
| "size_in_bytes": 326908174 | |
| }, | |
| "moral_scenarios": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "length": -1, | |
| "id": null, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "num_classes": 4, | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "id": null, | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "post_processed": null, | |
| "supervised_keys": null, | |
| "task_templates": null, | |
| "builder_name": "mmlu", | |
| "config_name": "moral_scenarios", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "description": null, | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 160601257, | |
| "num_examples": 99842, | |
| "dataset_name": "mmlu" | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 374014, | |
| "num_examples": 895, | |
| "dataset_name": "mmlu" | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 42326, | |
| "num_examples": 100, | |
| "dataset_name": "mmlu" | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 2046, | |
| "num_examples": 5, | |
| "dataset_name": "mmlu" | |
| } | |
| }, | |
| "download_checksums": { | |
| "data.tar": { | |
| "num_bytes": 166184960, | |
| "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b" | |
| } | |
| }, | |
| "download_size": 166184960, | |
| "post_processing_size": null, | |
| "dataset_size": 161019643, | |
| "size_in_bytes": 327204603 | |
| }, | |
| "nutrition": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "length": -1, | |
| "id": null, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "num_classes": 4, | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "id": null, | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "post_processed": null, | |
| "supervised_keys": null, | |
| "task_templates": null, | |
| "builder_name": "mmlu", | |
| "config_name": "nutrition", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "description": null, | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 160601257, | |
| "num_examples": 99842, | |
| "dataset_name": "mmlu" | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 92398, | |
| "num_examples": 306, | |
| "dataset_name": "mmlu" | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 8424, | |
| "num_examples": 33, | |
| "dataset_name": "mmlu" | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 2073, | |
| "num_examples": 5, | |
| "dataset_name": "mmlu" | |
| } | |
| }, | |
| "download_checksums": { | |
| "data.tar": { | |
| "num_bytes": 166184960, | |
| "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b" | |
| } | |
| }, | |
| "download_size": 166184960, | |
| "post_processing_size": null, | |
| "dataset_size": 160704152, | |
| "size_in_bytes": 326889112 | |
| }, | |
| "philosophy": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "length": -1, | |
| "id": null, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "num_classes": 4, | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "id": null, | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "post_processed": null, | |
| "supervised_keys": null, | |
| "task_templates": null, | |
| "builder_name": "mmlu", | |
| "config_name": "philosophy", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "description": null, | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 160601257, | |
| "num_examples": 99842, | |
| "dataset_name": "mmlu" | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 80061, | |
| "num_examples": 311, | |
| "dataset_name": "mmlu" | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 9172, | |
| "num_examples": 34, | |
| "dataset_name": "mmlu" | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 976, | |
| "num_examples": 5, | |
| "dataset_name": "mmlu" | |
| } | |
| }, | |
| "download_checksums": { | |
| "data.tar": { | |
| "num_bytes": 166184960, | |
| "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b" | |
| } | |
| }, | |
| "download_size": 166184960, | |
| "post_processing_size": null, | |
| "dataset_size": 160691466, | |
| "size_in_bytes": 326876426 | |
| }, | |
| "prehistory": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "length": -1, | |
| "id": null, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "num_classes": 4, | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "id": null, | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "post_processed": null, | |
| "supervised_keys": null, | |
| "task_templates": null, | |
| "builder_name": "mmlu", | |
| "config_name": "prehistory", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "description": null, | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 160601257, | |
| "num_examples": 99842, | |
| "dataset_name": "mmlu" | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 89582, | |
| "num_examples": 324, | |
| "dataset_name": "mmlu" | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 10273, | |
| "num_examples": 35, | |
| "dataset_name": "mmlu" | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1866, | |
| "num_examples": 5, | |
| "dataset_name": "mmlu" | |
| } | |
| }, | |
| "download_checksums": { | |
| "data.tar": { | |
| "num_bytes": 166184960, | |
| "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b" | |
| } | |
| }, | |
| "download_size": 166184960, | |
| "post_processing_size": null, | |
| "dataset_size": 160702978, | |
| "size_in_bytes": 326887938 | |
| }, | |
| "professional_accounting": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "length": -1, | |
| "id": null, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "num_classes": 4, | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "id": null, | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "post_processed": null, | |
| "supervised_keys": null, | |
| "task_templates": null, | |
| "builder_name": "mmlu", | |
| "config_name": "professional_accounting", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "description": null, | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 160601257, | |
| "num_examples": 99842, | |
| "dataset_name": "mmlu" | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 124538, | |
| "num_examples": 282, | |
| "dataset_name": "mmlu" | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 14360, | |
| "num_examples": 31, | |
| "dataset_name": "mmlu" | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 2136, | |
| "num_examples": 5, | |
| "dataset_name": "mmlu" | |
| } | |
| }, | |
| "download_checksums": { | |
| "data.tar": { | |
| "num_bytes": 166184960, | |
| "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b" | |
| } | |
| }, | |
| "download_size": 166184960, | |
| "post_processing_size": null, | |
| "dataset_size": 160742291, | |
| "size_in_bytes": 326927251 | |
| }, | |
| "professional_law": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "length": -1, | |
| "id": null, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "num_classes": 4, | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "id": null, | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "post_processed": null, | |
| "supervised_keys": null, | |
| "task_templates": null, | |
| "builder_name": "mmlu", | |
| "config_name": "professional_law", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "description": null, | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 160601257, | |
| "num_examples": 99842, | |
| "dataset_name": "mmlu" | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 1891750, | |
| "num_examples": 1534, | |
| "dataset_name": "mmlu" | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 203507, | |
| "num_examples": 170, | |
| "dataset_name": "mmlu" | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 6598, | |
| "num_examples": 5, | |
| "dataset_name": "mmlu" | |
| } | |
| }, | |
| "download_checksums": { | |
| "data.tar": { | |
| "num_bytes": 166184960, | |
| "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b" | |
| } | |
| }, | |
| "download_size": 166184960, | |
| "post_processing_size": null, | |
| "dataset_size": 162703112, | |
| "size_in_bytes": 328888072 | |
| }, | |
| "professional_medicine": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "length": -1, | |
| "id": null, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "num_classes": 4, | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "id": null, | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "post_processed": null, | |
| "supervised_keys": null, | |
| "task_templates": null, | |
| "builder_name": "mmlu", | |
| "config_name": "professional_medicine", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "description": null, | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 160601257, | |
| "num_examples": 99842, | |
| "dataset_name": "mmlu" | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 217549, | |
| "num_examples": 272, | |
| "dataset_name": "mmlu" | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 23835, | |
| "num_examples": 31, | |
| "dataset_name": "mmlu" | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 3795, | |
| "num_examples": 5, | |
| "dataset_name": "mmlu" | |
| } | |
| }, | |
| "download_checksums": { | |
| "data.tar": { | |
| "num_bytes": 166184960, | |
| "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b" | |
| } | |
| }, | |
| "download_size": 166184960, | |
| "post_processing_size": null, | |
| "dataset_size": 160846436, | |
| "size_in_bytes": 327031396 | |
| }, | |
| "professional_psychology": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "length": -1, | |
| "id": null, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "num_classes": 4, | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "id": null, | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "post_processed": null, | |
| "supervised_keys": null, | |
| "task_templates": null, | |
| "builder_name": "mmlu", | |
| "config_name": "professional_psychology", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "description": null, | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 160601257, | |
| "num_examples": 99842, | |
| "dataset_name": "mmlu" | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 225887, | |
| "num_examples": 612, | |
| "dataset_name": "mmlu" | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 29089, | |
| "num_examples": 69, | |
| "dataset_name": "mmlu" | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 2255, | |
| "num_examples": 5, | |
| "dataset_name": "mmlu" | |
| } | |
| }, | |
| "download_checksums": { | |
| "data.tar": { | |
| "num_bytes": 166184960, | |
| "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b" | |
| } | |
| }, | |
| "download_size": 166184960, | |
| "post_processing_size": null, | |
| "dataset_size": 160858488, | |
| "size_in_bytes": 327043448 | |
| }, | |
| "public_relations": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "length": -1, | |
| "id": null, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "num_classes": 4, | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "id": null, | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "post_processed": null, | |
| "supervised_keys": null, | |
| "task_templates": null, | |
| "builder_name": "mmlu", | |
| "config_name": "public_relations", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "description": null, | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 160601257, | |
| "num_examples": 99842, | |
| "dataset_name": "mmlu" | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 28748, | |
| "num_examples": 110, | |
| "dataset_name": "mmlu" | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 4554, | |
| "num_examples": 12, | |
| "dataset_name": "mmlu" | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1484, | |
| "num_examples": 5, | |
| "dataset_name": "mmlu" | |
| } | |
| }, | |
| "download_checksums": { | |
| "data.tar": { | |
| "num_bytes": 166184960, | |
| "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b" | |
| } | |
| }, | |
| "download_size": 166184960, | |
| "post_processing_size": null, | |
| "dataset_size": 160636043, | |
| "size_in_bytes": 326821003 | |
| }, | |
| "security_studies": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "length": -1, | |
| "id": null, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "num_classes": 4, | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "id": null, | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "post_processed": null, | |
| "supervised_keys": null, | |
| "task_templates": null, | |
| "builder_name": "mmlu", | |
| "config_name": "security_studies", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "description": null, | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 160601257, | |
| "num_examples": 99842, | |
| "dataset_name": "mmlu" | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 204832, | |
| "num_examples": 245, | |
| "dataset_name": "mmlu" | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 22625, | |
| "num_examples": 27, | |
| "dataset_name": "mmlu" | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 5323, | |
| "num_examples": 5, | |
| "dataset_name": "mmlu" | |
| } | |
| }, | |
| "download_checksums": { | |
| "data.tar": { | |
| "num_bytes": 166184960, | |
| "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b" | |
| } | |
| }, | |
| "download_size": 166184960, | |
| "post_processing_size": null, | |
| "dataset_size": 160834037, | |
| "size_in_bytes": 327018997 | |
| }, | |
| "sociology": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "length": -1, | |
| "id": null, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "num_classes": 4, | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "id": null, | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "post_processed": null, | |
| "supervised_keys": null, | |
| "task_templates": null, | |
| "builder_name": "mmlu", | |
| "config_name": "sociology", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "description": null, | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 160601257, | |
| "num_examples": 99842, | |
| "dataset_name": "mmlu" | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 66231, | |
| "num_examples": 201, | |
| "dataset_name": "mmlu" | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 7172, | |
| "num_examples": 22, | |
| "dataset_name": "mmlu" | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1601, | |
| "num_examples": 5, | |
| "dataset_name": "mmlu" | |
| } | |
| }, | |
| "download_checksums": { | |
| "data.tar": { | |
| "num_bytes": 166184960, | |
| "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b" | |
| } | |
| }, | |
| "download_size": 166184960, | |
| "post_processing_size": null, | |
| "dataset_size": 160676261, | |
| "size_in_bytes": 326861221 | |
| }, | |
| "us_foreign_policy": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "length": -1, | |
| "id": null, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "num_classes": 4, | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "id": null, | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "post_processed": null, | |
| "supervised_keys": null, | |
| "task_templates": null, | |
| "builder_name": "mmlu", | |
| "config_name": "us_foreign_policy", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "description": null, | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 160601257, | |
| "num_examples": 99842, | |
| "dataset_name": "mmlu" | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 28431, | |
| "num_examples": 100, | |
| "dataset_name": "mmlu" | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 3252, | |
| "num_examples": 11, | |
| "dataset_name": "mmlu" | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1599, | |
| "num_examples": 5, | |
| "dataset_name": "mmlu" | |
| } | |
| }, | |
| "download_checksums": { | |
| "data.tar": { | |
| "num_bytes": 166184960, | |
| "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b" | |
| } | |
| }, | |
| "download_size": 166184960, | |
| "post_processing_size": null, | |
| "dataset_size": 160634539, | |
| "size_in_bytes": 326819499 | |
| }, | |
| "virology": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "length": -1, | |
| "id": null, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "num_classes": 4, | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "id": null, | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "post_processed": null, | |
| "supervised_keys": null, | |
| "task_templates": null, | |
| "builder_name": "mmlu", | |
| "config_name": "virology", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "description": null, | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 160601257, | |
| "num_examples": 99842, | |
| "dataset_name": "mmlu" | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 38747, | |
| "num_examples": 166, | |
| "dataset_name": "mmlu" | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 5451, | |
| "num_examples": 18, | |
| "dataset_name": "mmlu" | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 1084, | |
| "num_examples": 5, | |
| "dataset_name": "mmlu" | |
| } | |
| }, | |
| "download_checksums": { | |
| "data.tar": { | |
| "num_bytes": 166184960, | |
| "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b" | |
| } | |
| }, | |
| "download_size": 166184960, | |
| "post_processing_size": null, | |
| "dataset_size": 160646539, | |
| "size_in_bytes": 326831499 | |
| }, | |
| "world_religions": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "id": null, | |
| "_type": "Value" | |
| }, | |
| "length": -1, | |
| "id": null, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "num_classes": 4, | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "id": null, | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "post_processed": null, | |
| "supervised_keys": null, | |
| "task_templates": null, | |
| "builder_name": "mmlu", | |
| "config_name": "world_religions", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "description": null, | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 160601257, | |
| "num_examples": 99842, | |
| "dataset_name": "mmlu" | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 25262, | |
| "num_examples": 171, | |
| "dataset_name": "mmlu" | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 2753, | |
| "num_examples": 19, | |
| "dataset_name": "mmlu" | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 658, | |
| "num_examples": 5, | |
| "dataset_name": "mmlu" | |
| } | |
| }, | |
| "download_checksums": { | |
| "data.tar": { | |
| "num_bytes": 166184960, | |
| "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b" | |
| } | |
| }, | |
| "download_size": 166184960, | |
| "post_processing_size": null, | |
| "dataset_size": 160629930, | |
| "size_in_bytes": 326814890 | |
| }, | |
| "all": { | |
| "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", | |
| "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", | |
| "homepage": "https://github.com/hendrycks/test", | |
| "license": "", | |
| "features": { | |
| "question": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "subject": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "choices": { | |
| "feature": { | |
| "dtype": "string", | |
| "_type": "Value" | |
| }, | |
| "_type": "Sequence" | |
| }, | |
| "answer": { | |
| "names": [ | |
| "A", | |
| "B", | |
| "C", | |
| "D" | |
| ], | |
| "_type": "ClassLabel" | |
| } | |
| }, | |
| "builder_name": "parquet", | |
| "dataset_name": "mmlu", | |
| "config_name": "all", | |
| "version": { | |
| "version_str": "1.0.0", | |
| "major": 1, | |
| "minor": 0, | |
| "patch": 0 | |
| }, | |
| "splits": { | |
| "auxiliary_train": { | |
| "name": "auxiliary_train", | |
| "num_bytes": 161000625, | |
| "num_examples": 99842, | |
| "dataset_name": null | |
| }, | |
| "test": { | |
| "name": "test", | |
| "num_bytes": 6967453, | |
| "num_examples": 14042, | |
| "dataset_name": null | |
| }, | |
| "validation": { | |
| "name": "validation", | |
| "num_bytes": 763484, | |
| "num_examples": 1531, | |
| "dataset_name": null | |
| }, | |
| "dev": { | |
| "name": "dev", | |
| "num_bytes": 125353, | |
| "num_examples": 285, | |
| "dataset_name": null | |
| } | |
| }, | |
| "download_size": 51132212, | |
| "dataset_size": 168856915, | |
| "size_in_bytes": 219989127 | |
| } | |
| } |