jspaulsen's picture
Upload dataset
604530d verified
---
license: cc-by-sa-4.0
dataset_info:
features:
- name: title
dtype: string
- name: text
dtype: string
splits:
- name: train
num_bytes: 18070709909
num_examples: 121082466
download_size: 9046108313
dataset_size: 18070709909
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Wikipedia Utterances
Text segments extracted from English Wikipedia, segmented into utterance-length chunks suitable for text-to-speech synthesis.
## Dataset Description
This dataset contains ~41M text utterances derived from the [wikimedia/wikipedia](https://huggingface.co/datasets/wikimedia/wikipedia) dataset (20231101.en snapshot). Each row contains:
| Field | Type | Description |
|-------|------|-------------|
| `title` | string | The Wikipedia article title |
| `text` | string | A text segment (10-4,880 characters) |
| `duration` | float | Estimated speech duration in seconds (at 150 WPM) |
## Processing
The transformation pipeline:
1. Tokenizes Wikipedia articles into paragraphs and sentences using NLTK
2. Combines consecutive sentences targeting 15-30 second utterances
3. Strips bracketed content (parentheses, braces, square brackets)
4. Filters for valid utterances ending in sentence-final punctuation
See `transform_wikipedia.py` in this repository for the full implementation.
## Usage
```python
from datasets import load_dataset
dataset = load_dataset("jspaulsen/wikipedia-utterances", split="train")
```
## License
This dataset is released under [CC BY-SA 4.0](https://creativecommons.org/licenses/by-sa/4.0/), consistent with the original Wikipedia content license.
## Citation
If you use this dataset, please cite the original Wikimedia source:
```bibtex
@ONLINE{wikidump,
author = "Wikimedia Foundation",
title = "Wikimedia Downloads",
url = "https://dumps.wikimedia.org"
}
```