| --- |
| license: apache-2.0 |
| task_categories: |
| - text-generation |
| tags: |
| - marin |
| - token-counts |
| - pretraining |
| pretty_name: Marin Token Counts |
| --- |
| |
| # Marin Token Counts |
|
|
| Token counts for all datasets used in [Marin](https://github.com/marin-community/marin) pretraining runs. |
|
|
| ## Schema |
|
|
| | Column | Type | Description | |
| |--------|------|-------------| |
| | `dataset` | string | Dataset identifier | |
| | `marin_tokens` | int | Number of tokens after tokenization | |
| | `category` | string | Content domain (web, code, math, academic, books, etc.) | |
| | `synthetic` | bool | Whether the data is LLM-generated or LLM-translated | |
|
|
| ## Categories |
|
|
| - **web** — Quality-classified Common Crawl text (Nemotron-CC) |
| - **code** — Source code and code-related documents |
| - **math** — Math-focused extractions and competition problems |
| - **academic** — Peer-reviewed papers and abstracts |
| - **reasoning** — Cross-domain reasoning and formal logic |
| - **books** — Digitized public domain and open access books |
| - **legal** — Court decisions, regulations, patents |
| - **government** — Parliamentary proceedings and publications |
| - **education** — Open educational resources and textbooks |
| - **encyclopedic** — Wiki-style reference content |
| - **forum** — Q&A sites and chat logs |
| - **documents** — PDF-extracted document text |
| - **translation** — Parallel translation corpora |
| - **news** — CC-licensed news articles |
| - **media** — Transcribed audio/video |
| - **supervised** — Curated task datasets |
| - **reference** — Niche reference sites |
| - **general** — General-domain content |
|
|
| ## Updates |
|
|
| This dataset is updated by running `experiments/count_tokens.py` from the Marin repo, |
| which reads tokenized dataset stats from GCS and pushes the results here. |
|
|