JayJayThrowThrow's picture
Add README
a332e38 verified

commonwealth-wiki-mix

FineWeb-style Parquet shards (LLM-training friendly) created by merging multiple Wikipedia language datasets into a single dataset to reduce looping during training.

What’s inside

  • Format: nanochat-parquet-v1
  • Layout: shard_*.parquet + metadata.json
  • Text column: text
  • Parquet settings: zstd (level 3), row_group_size=1024, use_dictionary=False, write_statistics=False

Sources included

This mix was built from the already-exported wiki datasets in ~/.cache/nanochat/datasets/monolingual/:

  • South Asia: Urdu, Hindi, Bengali, Tamil, Telugu, Malayalam, Marathi, Kannada, Gujarati, Punjabi, Nepali, Sinhala
  • Southeast Asia: Malay
  • Africa: Swahili, Yoruba, Afrikaans, Zulu

Loading

In Python, you can load shards directly with PyArrow / Pandas. For HuggingFace datasets, Parquet loading works out of the box:

from datasets import load_dataset

ds = load_dataset("JayJayThrowThrow/commonwealth-wiki-mix", split="train")
print(ds.column_names)  # ["text"]