--- license: mit --- # Stage 1 Packed Pretraining Dataset This dataset contains preprocessed and token-packed `.bin` files intended for use in pretraining a decoder-only Transformer language model. ## Dataset Contents - Each `.bin` file contains a fixed number of samples, where each sample is exactly 8192 tokens long. - Samples are grouped into batches of 125 samples, totaling **1.024 million tokens per batch**. - Each file (called a "block") contains 62500 samples (approximately 512 million tokens). - All samples are tokenized using the `GPT2TokenizerFast` from Hugging Face Transformers. ## Structure - Format: Binary files (`int32`) containing token IDs. - File naming: `stage1_block_0000.bin`, `stage1_block_0001.bin`, etc. - Tokenizer: `GPT2TokenizerFast` with `eos_token` used as a separator and padding token. - Context length: 8192 tokens per sample. ## Source Datasets Tokens were drawn from a diverse mix of high-quality open datasets: - `C4 (en)` - `Wikipedia (2023/11 dump)` - `OpenWebText` - `CCNews` - `Gutenberg` - `arXiv` - `BookCorpus Open` - `S2ORC` - `TriviaQA` - `PAQ` - `Natural Questions` Each dataset was assigned a token quota to ensure a balanced mix. ## Preprocessing & Packing Strategy - Samples were **streamed** using Hugging Face Datasets with shuffling. - Texts were **tokenized**, filtered using a garbage filter, and concatenated with separator tokens. - Samples were packed into fixed-length chunks of 8192 tokens. - Leftover tokens from one batch are carried forward to the next to ensure no token duplication or loss. ### Garbage Filtering Heuristics: - Removed texts with: - Too few words or characters. - High symbol-to-alphanumeric ratio. - Excessive character repetition. - Very low word diversity. ## Usage Example You can load and decode tokens using PyTorch: ```python import torch from transformers import GPT2TokenizerFast tokenizer = GPT2TokenizerFast.from_pretrained("gpt2") with open("stage1_block_0000.bin", "rb") as f: tokens = torch.frombuffer(f.read(), dtype=torch.int32) sample = tokens[:8192].tolist() text = tokenizer.decode(sample) print(text)