| --- |
| license: apache-2.0 |
| language: |
| - en |
| pretty_name: FineWeb 10B Bytes |
| task_categories: |
| - text-generation |
| --- |
| |
| # FineWeb 10B Bytes |
|
|
| This repository contains training shards for byte-level language model pretraining. |
|
|
| The dataset format is the same format used by [openai/parameter-golf](https://github.com/openai/parameter-golf), the OpenAI Model Craft Challenge repository for training compact language models and evaluating them on FineWeb in bits per byte. In that repository, evaluation is described as tokenizer-agnostic and based on compression performance on the FineWeb validation set. [oai_citation:0‡GitHub](https://github.com/openai/parameter-golf) |
|
|
| ## Origin |
|
|
| These byte shards were generated using the data conversion approach from [openai/parameter-golf Pull Request #705](https://github.com/openai/parameter-golf/pull/705), authored by GitHub user [`seanward`](https://github.com/seanward). That PR is titled **“Byte-Level Tokenizer-Free Transformer”** and explicitly includes a conversion script named `convert_to_bytes.py`, described there as **“Data conversion (sp1024 → raw bytes)”**. [oai_citation:1‡GitHub](https://github.com/openai/parameter-golf/pull/705) |
|
|
| ## Contents |
|
|
| This repository stores shard files such as: |
|
|
| - `fineweb_train_000000.bin` |
| - `fineweb_train_000001.bin` |
| - `fineweb_train_000002.bin` |
|
|
| and so on. |
|
|
| ## Dataset format |
|
|
| The `.bin` shards follow the same binary training-data convention used for byte-level experiments in `parameter-golf`. |
|
|
| At a high level: |
|
|
| - data is represented as raw UTF-8 bytes |
| - the byte vocabulary size is 256 |
| - shards are intended for training tokenizer-free / byte-level models |
| - the data layout is meant for efficient streaming during pretraining |
|
|
| The associated PR #705 describes the model as operating directly on raw UTF-8 bytes with `vocab=256`, and states that it uses raw byte input without BPE or SentencePiece. [oai_citation:2‡GitHub](https://github.com/openai/parameter-golf/pull/705) |
|
|
| ## Provenance |
|
|
| Source data is derived from FineWeb preprocessing workflows associated with byte-level training experiments for `parameter-golf`. |
|
|
| This repository republishes the resulting training shards only. It does **not** bundle the training code itself; for the original training setup, conversion logic, and experiment context, see: |
|
|
| - `openai/parameter-golf` |
| - PR #705 by `seanward` (“Byte-Level Tokenizer-Free Transformer”) [oai_citation:3‡GitHub](https://github.com/openai/parameter-golf) |
|
|
| ## Intended use |
|
|
| This dataset is intended for: |
|
|
| - byte-level language model pretraining |
| - tokenizer-free training experiments |
| - reproducing or adapting `parameter-golf`-style data pipelines |
| - benchmarking compact models on byte-level objectives |
|
|