# Comma v0.1 dataset This repository contains the dataset used to train [Comma v0.1-1T](https://huggingface.co/common-pile/comma-v0.1-1t) and [Comma v0.1-2T](https://huggingface.co/common-pile/comma-v0.1-2t). It is a slightly modified and consolidated version of the [Common Pile v0.1 "filtered" data](https://huggingface.co/collections/common-pile/common-pile-v01-filtered-data-68300bb0a946d10dda697663). If you are looknig for the raw Common Pile v0.1 data, please see [this collection](https://huggingface.co/collections/common-pile/common-pile-v01-raw-data-6826b454a5a6a445d0b51b37). You can learn more about Common Pile in [our paper](https://huggingface.co/papers/2506.05209). ## Mixing rates and token counts The Comma v0.1 models were trained in two stages, a "main" stage and a "cooldown" stage. During each stage, we heuristically set mixing rates to up or downweight different sources. In the two tables below, we provide per-source token count, repeat rate, and effective token count (after up/downweighting) for the main and cooldown stage of the Comma v0.1-1T training run. For the Comma v0.1-2T training run, all sources are repeated 2x as many times in both stages. Token counts are as provided by the Comma v0.1 tokenizer; using a different tokenizer may change these counts significantly. | Main stage | Tokens (B) | Repeats | Effective tokens (B) | |-------------------------------|------------|---------|----------------------| | arxiv_abstracts | 0.57 | 6 | 3.4 | | arxiv_papers | 6.0 | 6 | 35.8 | | biodiversity_heritage_library | 9.8 | 0.25 | 2.5 | | caselaw_access_project | 19.7 | 1 | 19.7 | | cccc | 15.2 | 6 | 91.4 | | data_provenance_initiative | 0.92 | 6 | 5.5 | | doab | 3.0 | 6 | 18.2 | | foodista | 0.025 | 6 | 0.15 | | github_archive | 11.0 | 6 | 66.1 | | library_of_congress | 9.5 | 0.25 | 2.4 | | libretexts | 0.093 | 6 | 0.56 | | news | 0.064 | 6 | 0.38 | | oercommons | 0.012 | 6 | 0.07 | | peS2o | 43.3 | 6 | 260.0 | | pre_1929_books | 12.4 | 1 | 12.4 | | pressbooks | 0.14 | 6 | 0.86 | | project_gutenberg | 5.7 | 1 | 5.7 | | public_domain_review | 0.0017 | 6 | 0.010 | | pubmed | 36.6 | 1 | 36.6 | | python_enhancement_proposals | 0.0027 | 6 | 0.016 | | regulations | 1.4 | 6 | 8.2 | | stackexchange | 23.9 | 6 | 143.2 | | stackv2_edu | 67.8 | 2 | 135.5 | | stackv2_html | 1.2 | 2 | 2.5 | | ubuntu_irc | 1.9 | 6 | 11.1 | | uk_hansard | 2.3 | 6 | 14.0 | | usgpo | 8.8 | 0.25 | 2.2 | | uspto | 157.4 | 0.25 | 39.4 | | wikimedia | 15.8 | 6 | 94.7 | | wikiteam | 4.3 | 4 | 17.2 | | youtube | 4.7 | 1 | 4.7 | | Total | 463.6 | | 1034.4 | | Cooldown stage | Tokens (B) | Repeats | Effective tokens (B) | |------------------------------|------------|---------|----------------------| | arxiv_papers | 6.0 | 0.5 | 3.0 | | cccc | 15.2 | 0.3 | 4.6 | | data_provenance_initiative | 0.92 | 2 | 1.8 | | doab | 3.0 | 2 | 6.1 | | foodista | 0.025 | 2 | 0.05 | | libretexts | 0.093 | 2 | 0.19 | | news | 0.064 | 2 | 0.13 | | oercommons | 0.012 | 2 | 0.02 | | peS2o | 43.3 | 0.1 | 4.3 | | pressbooks | 0.14 | 2 | 0.29 | | public_domain_review | 0.0017 | 2 | 0.003 | | python_enhancement_proposals | 0.0027 | 2 | 0.005 | | stackexchange | 23.9 | 0.25 | 6.0 | | stackv2_edu | 67.8 | 0.1 | 6.8 | | wikimedia | 15.8 | 0.4 | 6.3 | | Total | 176.2 | | 39.5 |