Datasets:
Tasks:
Text Generation
Modalities:
Text
Formats:
json
Sub-tasks:
language-modeling
Languages:
English
Size:
10K - 100K
ArXiv:
License:
| license: cc-by-nc-sa-4.0 | |
| language: | |
| - en | |
| annotations_creators: | |
| - no-annotation | |
| task_categories: | |
| - text-generation | |
| task_ids: | |
| - language-modeling | |
| size_categories: | |
| - 10K<n<100K | |
| configs: | |
| - config_name: python | |
| data_files: | |
| - split: test | |
| path: | |
| - data/python.jsonl | |
| - config_name: cc | |
| data_files: | |
| - split: test | |
| path: | |
| - data/cc.jsonl | |
| - config_name: arxiv_math | |
| data_files: | |
| - split: test | |
| path: | |
| - data/arxiv_math.jsonl | |
| This is the compression corpora dataset used in the paper "Compression Represents Intelligence Linearly". | |
| We find that LLMs’ intelligence – reflected by benchmark scores – almost **linearly** correlates with their ability to compress external text corpora. We measure intelligence along three key abilities: knowledge and commonsense, coding, and mathematical reasoning, and provide the corresponding compression corpora here respectively named cc, python, and arxiv_math. | |
| ### Load the data | |
| ```python | |
| from datasets import load_dataset | |
| dataset=load_dataset(r"hkust-nlp/llm-compression",name="python") | |
| print(dataset['test'][0]) | |
| ``` | |
| More details on compression evaluation are at our [github page](https://github.com/hkust-nlp/llm-compression-intelligence). | |
| ### Citation | |
| ``` | |
| @misc{huang2024compression, | |
| title={Compression Represents Intelligence Linearly}, | |
| author={Yuzhen Huang and Jinghan Zhang and Zifei Shan and Junxian He}, | |
| year={2024}, | |
| eprint={2404.09937}, | |
| archivePrefix={arXiv}, | |
| primaryClass={cs.CL} | |
| } | |
| ``` |