nano-start-data-bin / README.md
fs90's picture
Upload README.md with huggingface_hub
390f63f verified
metadata
annotations_creators:
  - machine-generated
language_creators:
  - expert-generated
language:
  - en
license:
  - mit
multilinguality:
  - monolingual
size_categories:
  - n<1K
source_datasets:
  - fs90/nano-start-data
task_categories:
  - text-generation
pretty_name: Nano-Start Tokenized Dataset
tags:
  - educational
  - llm-training
  - tokenized
  - binary
  - oxidizr

Nano-Start Tokenized Dataset

Pre-tokenized binary files ready for training with oxidizr. This is the tokenized version of fs90/nano-start-data.

What is Tokenization?

Language models don't process text directly - they work with numbers called tokens. Tokenization converts text into token IDs:

"Hello world" → [9906, 1917]

This dataset is pre-tokenized for simplicity - download and start training immediately. To learn how tokenization works and create your own datasets, see the splintr project.

Quick Start

Option A: Using hf

pip install huggingface_hub
hf download fs90/nano-start-data-bin --local-dir data/nano-start/tokenized --repo-type dataset

Option B: Direct download

Download combined.bin from the Files tab and place it in your project.

Train with oxidizr:

cargo run --release -- \
    --config models/nano-start.yaml \
    --data data/nano-start/tokenized/combined.bin

Files

Download combined.bin for training - it contains all data merged together:

File Size Tokens Description
combined.bin 25,516 bytes 6,379 All data merged (recommended)

Individual Files (Optional)

You can also train on individual subsets. Training on different data produces different model behavior:

File Size Tokens Description
completions.bin 8,788 bytes 2,197 Factual statements only
qa.bin 11,036 bytes 2,759 Q&A pairs only
chat.bin 5,692 bytes 1,423 Multi-turn conversations only

Experiment with different files to see how the training data affects model behavior!

Binary Format

Each .bin file contains raw token IDs:

  • Encoding: u32 (32-bit unsigned integer)
  • Byte order: Little-endian
  • Headers: None (raw token stream)
  • Tokenizer: cl100k_base (OpenAI, vocab size: 100,331)

Reading the Data

import struct

def read_tokens(path):
    with open(path, "rb") as f:
        data = f.read()
    return list(struct.unpack(f"<{len(data)//4}I", data))

tokens = read_tokens("combined.bin")
print(f"Total tokens: {len(tokens)}")

Tokenizer Details

Property Value
Tokenizer cl100k_base (OpenAI GPT-4/GPT-3.5)
Vocab size 100,331
EOS token <|endoftext|> (ID: 100257)

Special Tokens

Token ID Purpose
<|endoftext|> 100257 Separates examples
<|system|> 100277 System instructions
<|user|> 100278 User input
<|assistant|> 100279 Model response

Source Data

To see the human-readable text before tokenization: fs90/nano-start-data

Related Resources

License

MIT License

Citation

@dataset{nano_start_bin_2024,
  title={Nano-Start Tokenized Dataset},
  author={fs90},
  year={2024},
  publisher={Hugging Face},
  url={https://huggingface.co/datasets/fs90/nano-start-data-bin}
}