| # Comma v0.1 dataset | |
| This repository contains the dataset used to train [Comma v0.1-1T](https://huggingface.co/common-pile/comma-v0.1-1t) and [Comma v0.1-2T](https://huggingface.co/common-pile/comma-v0.1-2t). | |
| It is a slightly modified and consolidated version of the [Common Pile v0.1 "filtered" data](https://huggingface.co/collections/common-pile/common-pile-v01-filtered-data-68300bb0a946d10dda697663). | |
| If you are looknig for the raw Common Pile v0.1 data, please see [this collection](https://huggingface.co/collections/common-pile/common-pile-v01-raw-data-6826b454a5a6a445d0b51b37). | |
| You can learn more about Common Pile in [our paper](https://huggingface.co/papers/2506.05209). | |
| ## Mixing rates and token counts | |
| The Comma v0.1 models were trained in two stages, a "main" stage and a "cooldown" stage. | |
| During each stage, we heuristically set mixing rates to up or downweight different sources. | |
| In the two tables below, we provide per-source token count, repeat rate, and effective token count (after up/downweighting) for the main and cooldown stage of the Comma v0.1-1T training run. | |
| For the Comma v0.1-2T training run, all sources are repeated 2x as many times in both stages. | |
| Token counts are as provided by the Comma v0.1 tokenizer; using a different tokenizer may change these counts significantly. | |
| | Main stage | Tokens (B) | Repeats | Effective tokens (B) | | |
| |-------------------------------|------------|---------|----------------------| | |
| | arxiv_abstracts | 0.57 | 6 | 3.4 | | |
| | arxiv_papers | 6.0 | 6 | 35.8 | | |
| | biodiversity_heritage_library | 9.8 | 0.25 | 2.5 | | |
| | caselaw_access_project | 19.7 | 1 | 19.7 | | |
| | cccc | 15.2 | 6 | 91.4 | | |
| | data_provenance_initiative | 0.92 | 6 | 5.5 | | |
| | doab | 3.0 | 6 | 18.2 | | |
| | foodista | 0.025 | 6 | 0.15 | | |
| | github_archive | 11.0 | 6 | 66.1 | | |
| | library_of_congress | 9.5 | 0.25 | 2.4 | | |
| | libretexts | 0.093 | 6 | 0.56 | | |
| | news | 0.064 | 6 | 0.38 | | |
| | oercommons | 0.012 | 6 | 0.07 | | |
| | peS2o | 43.3 | 6 | 260.0 | | |
| | pre_1929_books | 12.4 | 1 | 12.4 | | |
| | pressbooks | 0.14 | 6 | 0.86 | | |
| | project_gutenberg | 5.7 | 1 | 5.7 | | |
| | public_domain_review | 0.0017 | 6 | 0.010 | | |
| | pubmed | 36.6 | 1 | 36.6 | | |
| | python_enhancement_proposals | 0.0027 | 6 | 0.016 | | |
| | regulations | 1.4 | 6 | 8.2 | | |
| | stackexchange | 23.9 | 6 | 143.2 | | |
| | stackv2_edu | 67.8 | 2 | 135.5 | | |
| | stackv2_html | 1.2 | 2 | 2.5 | | |
| | ubuntu_irc | 1.9 | 6 | 11.1 | | |
| | uk_hansard | 2.3 | 6 | 14.0 | | |
| | usgpo | 8.8 | 0.25 | 2.2 | | |
| | uspto | 157.4 | 0.25 | 39.4 | | |
| | wikimedia | 15.8 | 6 | 94.7 | | |
| | wikiteam | 4.3 | 4 | 17.2 | | |
| | youtube | 4.7 | 1 | 4.7 | | |
| | Total | 463.6 | | 1034.4 | | |
| | Cooldown stage | Tokens (B) | Repeats | Effective tokens (B) | | |
| |------------------------------|------------|---------|----------------------| | |
| | arxiv_papers | 6.0 | 0.5 | 3.0 | | |
| | cccc | 15.2 | 0.3 | 4.6 | | |
| | data_provenance_initiative | 0.92 | 2 | 1.8 | | |
| | doab | 3.0 | 2 | 6.1 | | |
| | foodista | 0.025 | 2 | 0.05 | | |
| | libretexts | 0.093 | 2 | 0.19 | | |
| | news | 0.064 | 2 | 0.13 | | |
| | oercommons | 0.012 | 2 | 0.02 | | |
| | peS2o | 43.3 | 0.1 | 4.3 | | |
| | pressbooks | 0.14 | 2 | 0.29 | | |
| | public_domain_review | 0.0017 | 2 | 0.003 | | |
| | python_enhancement_proposals | 0.0027 | 2 | 0.005 | | |
| | stackexchange | 23.9 | 0.25 | 6.0 | | |
| | stackv2_edu | 67.8 | 0.1 | 6.8 | | |
| | wikimedia | 15.8 | 0.4 | 6.3 | | |
| | Total | 176.2 | | 39.5 | |