JayJayThrowThrow's picture
Add README
a332e38 verified
# commonwealth-wiki-mix
FineWeb-style Parquet shards (LLM-training friendly) created by merging multiple Wikipedia language datasets into a single dataset to reduce looping during training.
## What’s inside
- **Format**: `nanochat-parquet-v1`
- **Layout**: `shard_*.parquet` + `metadata.json`
- **Text column**: `text`
- **Parquet settings**: zstd (level 3), `row_group_size=1024`, `use_dictionary=False`, `write_statistics=False`
## Sources included
This mix was built from the already-exported wiki datasets in `~/.cache/nanochat/datasets/monolingual/`:
- South Asia: Urdu, Hindi, Bengali, Tamil, Telugu, Malayalam, Marathi, Kannada, Gujarati, Punjabi, Nepali, Sinhala
- Southeast Asia: Malay
- Africa: Swahili, Yoruba, Afrikaans, Zulu
## Loading
In Python, you can load shards directly with PyArrow / Pandas. For HuggingFace `datasets`, Parquet loading works out of the box:
```python
from datasets import load_dataset
ds = load_dataset("JayJayThrowThrow/commonwealth-wiki-mix", split="train")
print(ds.column_names) # ["text"]
```