| This is a preprocessed version of the realnewslike subdirectory of C4, with 4 different vocabulary sizes: 32k, 64k, 128k and 256k | |
| C4 from: https://huggingface.co/datasets/allenai/c4 | |
| Files generated by using SentencePieceTrainer https://github.com/google/sentencepiece and Megatron-LM https://github.com/NVIDIA/Megatron-LM/ | |
| ```py | |
| import sentencepiece as spm | |
| spm.SentencePieceTrainer.Train('--input=c4_dataset.txt --model_prefix=vp_sample_dataset --vocab_size=256000') | |
| ``` | |
| ```bash | |
| python tools/preprocess_data.py \ | |
| --input 'c4/realnewslike/c4-train.0000[0-2]-of-00512.json' \ | |
| --partitions 8 \ | |
| --output-prefix c4 \ | |
| --vocab-file vp_sample_dataset.vocab \ | |
| --tokenizer-type GPTSentencePieceTokenizer \ | |
| --tokenizer-model vp_sample_dataset.model \ | |
| --workers 8 | |
| ``` | |
| --- | |
| license: odc-by | |
| --- | |