diff --git "a/dataset_infos.json" "b/dataset_infos.json" deleted file mode 100644--- "a/dataset_infos.json" +++ /dev/null @@ -1,4062 +0,0 @@ -{ - "abstract_algebra": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "abstract_algebra", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 21316, - "num_examples": 100, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 2232, - "num_examples": 11, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 918, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47179638, - "dataset_size": 161025091, - "size_in_bytes": 208204729 - }, - "anatomy": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "anatomy", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 34594, - "num_examples": 135, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 3282, - "num_examples": 14, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1010, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47191229, - "dataset_size": 161039511, - "size_in_bytes": 208230740 - }, - "astronomy": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "astronomy", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 48735, - "num_examples": 152, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 5223, - "num_examples": 16, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 2129, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47201551, - "dataset_size": 161056712, - "size_in_bytes": 208258263 - }, - "business_ethics": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "business_ethics", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 35140, - "num_examples": 100, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 3235, - "num_examples": 11, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 2273, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47193421, - "dataset_size": 161041273, - "size_in_bytes": 208234694 - }, - "clinical_knowledge": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "clinical_knowledge", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 68572, - "num_examples": 265, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 7290, - "num_examples": 29, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1308, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47213955, - "dataset_size": 161077795, - "size_in_bytes": 208291750 - }, - "college_biology": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "college_biology", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 51521, - "num_examples": 144, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 5111, - "num_examples": 16, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1615, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47205152, - "dataset_size": 161058872, - "size_in_bytes": 208264024 - }, - "college_chemistry": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "college_chemistry", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 26796, - "num_examples": 100, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 2484, - "num_examples": 8, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1424, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47188958, - "dataset_size": 161031329, - "size_in_bytes": 208220287 - }, - "college_computer_science": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "college_computer_science", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 45429, - "num_examples": 100, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 4959, - "num_examples": 11, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 2893, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47201966, - "dataset_size": 161053906, - "size_in_bytes": 208255872 - }, - "college_mathematics": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "college_mathematics", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 26999, - "num_examples": 100, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 2909, - "num_examples": 11, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1596, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47188597, - "dataset_size": 161032129, - "size_in_bytes": 208220726 - }, - "college_medicine": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "college_medicine", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 85845, - "num_examples": 173, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 8337, - "num_examples": 22, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1758, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47218201, - "dataset_size": 161096565, - "size_in_bytes": 208314766 - }, - "college_physics": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "college_physics", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 32107, - "num_examples": 102, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 3687, - "num_examples": 11, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1495, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47190901, - "dataset_size": 161037914, - "size_in_bytes": 208228815 - }, - "computer_security": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "computer_security", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 29212, - "num_examples": 100, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 4768, - "num_examples": 11, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1194, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47192155, - "dataset_size": 161035799, - "size_in_bytes": 208227954 - }, - "conceptual_physics": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "conceptual_physics", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 45867, - "num_examples": 235, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 5034, - "num_examples": 26, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1032, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47197231, - "dataset_size": 161052558, - "size_in_bytes": 208249789 - }, - "econometrics": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "econometrics", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 48359, - "num_examples": 114, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 5147, - "num_examples": 12, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1712, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47197846, - "dataset_size": 161055843, - "size_in_bytes": 208253689 - }, - "electrical_engineering": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "electrical_engineering", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 28900, - "num_examples": 145, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 3307, - "num_examples": 16, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1090, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47189021, - "dataset_size": 161033922, - "size_in_bytes": 208222943 - }, - "elementary_mathematics": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "elementary_mathematics", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 79924, - "num_examples": 378, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 10042, - "num_examples": 41, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1558, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47216972, - "dataset_size": 161092149, - "size_in_bytes": 208309121 - }, - "formal_logic": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "formal_logic", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 51789, - "num_examples": 126, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 6464, - "num_examples": 14, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1825, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47194349, - "dataset_size": 161060703, - "size_in_bytes": 208255052 - }, - "global_facts": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "global_facts", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 19991, - "num_examples": 100, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 2013, - "num_examples": 10, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1297, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47181634, - "dataset_size": 161023926, - "size_in_bytes": 208205560 - }, - "high_school_biology": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "high_school_biology", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 116850, - "num_examples": 310, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 11746, - "num_examples": 32, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1776, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47239946, - "dataset_size": 161130997, - "size_in_bytes": 208370943 - }, - "high_school_chemistry": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "high_school_chemistry", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 63527, - "num_examples": 203, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 7630, - "num_examples": 22, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1333, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47207769, - "dataset_size": 161073115, - "size_in_bytes": 208280884 - }, - "high_school_computer_science": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "high_school_computer_science", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 47664, - "num_examples": 100, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 3619, - "num_examples": 9, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 3066, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47200669, - "dataset_size": 161054974, - "size_in_bytes": 208255643 - }, - "high_school_european_history": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "high_school_european_history", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 275568, - "num_examples": 165, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 30196, - "num_examples": 18, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 11712, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47349494, - "dataset_size": 161318101, - "size_in_bytes": 208667595 - }, - "high_school_geography": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "high_school_geography", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 46972, - "num_examples": 198, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 4870, - "num_examples": 22, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1516, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47200648, - "dataset_size": 161053983, - "size_in_bytes": 208254631 - }, - "high_school_government_and_politics": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "high_school_government_and_politics", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 73589, - "num_examples": 193, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 7870, - "num_examples": 21, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1962, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47214961, - "dataset_size": 161084046, - "size_in_bytes": 208299007 - }, - "high_school_macroeconomics": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "high_school_macroeconomics", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 129375, - "num_examples": 390, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 14298, - "num_examples": 43, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1466, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47231078, - "dataset_size": 161145764, - "size_in_bytes": 208376842 - }, - "high_school_mathematics": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "high_school_mathematics", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 62132, - "num_examples": 270, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 6536, - "num_examples": 29, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1420, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47207019, - "dataset_size": 161070713, - "size_in_bytes": 208277732 - }, - "high_school_microeconomics": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "high_school_microeconomics", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 82831, - "num_examples": 238, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 8321, - "num_examples": 26, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1436, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47212056, - "dataset_size": 161093213, - "size_in_bytes": 208305269 - }, - "high_school_physics": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "high_school_physics", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 62999, - "num_examples": 151, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 7150, - "num_examples": 17, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1592, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47207223, - "dataset_size": 161072366, - "size_in_bytes": 208279589 - }, - "high_school_psychology": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "high_school_psychology", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 173565, - "num_examples": 545, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 18817, - "num_examples": 60, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 2023, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47274794, - "dataset_size": 161195030, - "size_in_bytes": 208469824 - }, - "high_school_statistics": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "high_school_statistics", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 116306, - "num_examples": 216, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 10583, - "num_examples": 23, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 2646, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47236110, - "dataset_size": 161130160, - "size_in_bytes": 208366270 - }, - "high_school_us_history": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "high_school_us_history", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 302026, - "num_examples": 204, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 32266, - "num_examples": 22, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 8982, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47355390, - "dataset_size": 161343899, - "size_in_bytes": 208699289 - }, - "high_school_world_history": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "high_school_world_history", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 385478, - "num_examples": 237, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 46243, - "num_examples": 26, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 5015, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47405363, - "dataset_size": 161437361, - "size_in_bytes": 208842724 - }, - "human_aging": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "human_aging", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 49431, - "num_examples": 223, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 5040, - "num_examples": 23, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1071, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47203557, - "dataset_size": 161056167, - "size_in_bytes": 208259724 - }, - "human_sexuality": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "human_sexuality", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 34587, - "num_examples": 131, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 2637, - "num_examples": 12, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1160, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47194557, - "dataset_size": 161039009, - "size_in_bytes": 208233566 - }, - "international_law": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "international_law", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 56060, - "num_examples": 121, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 6734, - "num_examples": 13, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 2511, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47203538, - "dataset_size": 161065930, - "size_in_bytes": 208269468 - }, - "jurisprudence": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "jurisprudence", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 35810, - "num_examples": 108, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 3904, - "num_examples": 11, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1376, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47195740, - "dataset_size": 161041715, - "size_in_bytes": 208237455 - }, - "logical_fallacies": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "logical_fallacies", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 53528, - "num_examples": 163, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 5469, - "num_examples": 18, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1666, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47196007, - "dataset_size": 161061288, - "size_in_bytes": 208257295 - }, - "machine_learning": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "machine_learning", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 36108, - "num_examples": 112, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 3440, - "num_examples": 11, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 2411, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47192772, - "dataset_size": 161042584, - "size_in_bytes": 208235356 - }, - "management": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "management", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 21432, - "num_examples": 103, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 1962, - "num_examples": 11, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 956, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47185208, - "dataset_size": 161024975, - "size_in_bytes": 208210183 - }, - "marketing": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "marketing", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 66055, - "num_examples": 234, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 7707, - "num_examples": 25, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1534, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47211905, - "dataset_size": 161075921, - "size_in_bytes": 208287826 - }, - "medical_genetics": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "medical_genetics", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 22852, - "num_examples": 100, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 3213, - "num_examples": 11, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1177, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47187825, - "dataset_size": 161027867, - "size_in_bytes": 208215692 - }, - "miscellaneous": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "miscellaneous", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 161003, - "num_examples": 783, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 15780, - "num_examples": 86, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 772, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47277261, - "dataset_size": 161178180, - "size_in_bytes": 208455441 - }, - "moral_disputes": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "moral_disputes", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 114034, - "num_examples": 346, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 13092, - "num_examples": 38, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1833, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47238228, - "dataset_size": 161129584, - "size_in_bytes": 208367812 - }, - "moral_scenarios": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "moral_scenarios", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 391019, - "num_examples": 895, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 44226, - "num_examples": 100, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 2141, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47270775, - "dataset_size": 161438011, - "size_in_bytes": 208708786 - }, - "nutrition": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "nutrition", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 96376, - "num_examples": 306, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 8853, - "num_examples": 33, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 2138, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47230622, - "dataset_size": 161107992, - "size_in_bytes": 208338614 - }, - "philosophy": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "philosophy", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 84415, - "num_examples": 311, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 9648, - "num_examples": 34, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1046, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47223948, - "dataset_size": 161095734, - "size_in_bytes": 208319682 - }, - "prehistory": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "prehistory", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 94118, - "num_examples": 324, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 10763, - "num_examples": 35, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1936, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47231112, - "dataset_size": 161107442, - "size_in_bytes": 208338554 - }, - "professional_accounting": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "professional_accounting", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 132152, - "num_examples": 282, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 15197, - "num_examples": 31, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 2271, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47249006, - "dataset_size": 161150245, - "size_in_bytes": 208399251 - }, - "professional_law": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "professional_law", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 1922430, - "num_examples": 1534, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 206907, - "num_examples": 170, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 6698, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 48319302, - "dataset_size": 163136660, - "size_in_bytes": 211455962 - }, - "professional_medicine": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "professional_medicine", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 224349, - "num_examples": 272, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 24610, - "num_examples": 31, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 3920, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47312228, - "dataset_size": 161253504, - "size_in_bytes": 208565732 - }, - "professional_psychology": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "professional_psychology", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 242411, - "num_examples": 612, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 30952, - "num_examples": 69, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 2390, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47321118, - "dataset_size": 161276378, - "size_in_bytes": 208597496 - }, - "public_relations": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "public_relations", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 30948, - "num_examples": 110, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 4794, - "num_examples": 12, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1584, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47193575, - "dataset_size": 161037951, - "size_in_bytes": 208231526 - }, - "security_studies": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "security_studies", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 209732, - "num_examples": 245, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 23165, - "num_examples": 27, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 5423, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47300997, - "dataset_size": 161238945, - "size_in_bytes": 208539942 - }, - "sociology": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "sociology", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 68844, - "num_examples": 201, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 7458, - "num_examples": 22, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1666, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47218696, - "dataset_size": 161078593, - "size_in_bytes": 208297289 - }, - "us_foreign_policy": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "us_foreign_policy", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 30531, - "num_examples": 100, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 3483, - "num_examples": 11, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1704, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47191346, - "dataset_size": 161036343, - "size_in_bytes": 208227689 - }, - "virology": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "virology", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 40739, - "num_examples": 166, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 5667, - "num_examples": 18, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 1144, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47200276, - "dataset_size": 161048175, - "size_in_bytes": 208248451 - }, - "world_religions": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "world_religions", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 28511, - "num_examples": 171, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 3114, - "num_examples": 19, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 753, - "num_examples": 5, - "dataset_name": null - } - }, - "download_size": 47189681, - "dataset_size": 161033003, - "size_in_bytes": 208222684 - }, - "all": { - "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", - "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", - "homepage": "https://github.com/hendrycks/test", - "license": "", - "features": { - "question": { - "dtype": "string", - "_type": "Value" - }, - "subject": { - "dtype": "string", - "_type": "Value" - }, - "choices": { - "feature": { - "dtype": "string", - "_type": "Value" - }, - "_type": "Sequence" - }, - "answer": { - "names": [ - "A", - "B", - "C", - "D" - ], - "_type": "ClassLabel" - } - }, - "builder_name": "parquet", - "dataset_name": "mmlu", - "config_name": "all", - "version": { - "version_str": "1.0.0", - "major": 1, - "minor": 0, - "patch": 0 - }, - "splits": { - "auxiliary_train": { - "name": "auxiliary_train", - "num_bytes": 161000625, - "num_examples": 99842, - "dataset_name": null - }, - "test": { - "name": "test", - "num_bytes": 6967453, - "num_examples": 14042, - "dataset_name": null - }, - "validation": { - "name": "validation", - "num_bytes": 763484, - "num_examples": 1531, - "dataset_name": null - }, - "dev": { - "name": "dev", - "num_bytes": 125353, - "num_examples": 285, - "dataset_name": null - } - }, - "download_size": 51132212, - "dataset_size": 168856915, - "size_in_bytes": 219989127 - } -} \ No newline at end of file