| | --- |
| | language: |
| | - en |
| | pretty_name: SlimPajama_300B |
| | --- |
| | |
| | The SlimPajama_300B is a 300B token sample of de-duplicated Slim Pajama dataset tokenized using the [EleutherAI/gpt-neox-20b](https://huggingface.co/EleutherAI/gpt-neox-20b) tokenizer |
| | |
| | Due to file size constraints, C4 and CommonCrawl has been uploaded in multiple chunks, you can use the following commands to merge them back into a single file: |
| | ```bash |
| | cat C4_part_* > C4.bin |
| | cat CommonCrawl_part_* > CommonCrawl.bin |
| | ``` |
| | |
| | #### Data Distribution |
| | |
| | | Data source | Composition | |
| | | ------------- | ------------------------------- | |
| | | Commoncrawl | 0.5208 | |
| | | C4 | 0.2668 | |
| | | GitHub | 0.0522 | |
| | | Books | 0.0420 | |
| | | ArXiv | 0.0442 | |
| | | Wikpedia | 0.0399 | |
| | | StackExchange | 0.0337 | |
| | |