File size: 16,254 Bytes
dbeb0df
 
 
 
 
 
04575a1
 
 
 
 
 
 
 
dbeb0df
 
8a6b508
dbeb0df
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8a6b508
dbeb0df
 
8fba08d
 
dbeb0df
 
 
 
 
 
 
 
 
 
5bf5452
157eb25
5bf5452
 
8a6b508
dbeb0df
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8a6b508
 
dbeb0df
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8a6b508
dbeb0df
8a6b508
dbeb0df
8a6b508
dbeb0df
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
935e3a6
dbeb0df
 
 
 
 
8a6b508
dbeb0df
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8a6b508
dbeb0df
 
 
 
8a6b508
dbeb0df
 
 
 
 
 
 
 
 
8a6b508
dbeb0df
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8a6b508
dbeb0df
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8a6b508
dbeb0df
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8a6b508
dbeb0df
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
04575a1
 
 
 
 
dbeb0df
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
---
license: other
license_name: mixed-per-dataset
license_link: LICENSE
language:
  - en
tags:
  - text-to-speech
  - tts
  - speech
  - audio
  - open-data
  - training-data
  - english
task_categories:
  - text-to-speech
pretty_name: Raon-OpenTTS-Pool
size_categories:
  - 100M<n<1B
configs:
  - config_name: all
    data_files:
      - split: pool
        path: "*/metadata_pool.parquet"
      - split: core
        path: "*/metadata_core.parquet"
  - config_name: Raon-YouTube-Commons
    data_files:
      - split: pool
        path: Raon-YouTube-Commons/metadata_pool.parquet
      - split: core
        path: Raon-YouTube-Commons/metadata_core.parquet
  - config_name: Emilia-YODAS2
    data_files:
      - split: pool
        path: Emilia-YODAS2/metadata_pool.parquet
      - split: core
        path: Emilia-YODAS2/metadata_core.parquet
  - config_name: Emilia
    data_files:
      - split: pool
        path: Emilia/metadata_pool.parquet
      - split: core
        path: Emilia/metadata_core.parquet
  - config_name: LibriHeavy
    data_files:
      - split: pool
        path: LibriHeavy/metadata_pool.parquet
      - split: core
        path: LibriHeavy/metadata_core.parquet
  - config_name: HiFiTTS
    data_files:
      - split: pool
        path: HiFiTTS/metadata_pool.parquet
      - split: core
        path: HiFiTTS/metadata_core.parquet
  - config_name: VoxPopuli
    data_files:
      - split: pool
        path: VoxPopuli/metadata_pool.parquet
      - split: core
        path: VoxPopuli/metadata_core.parquet
  - config_name: PeoplesSpeech-Clean
    data_files:
      - split: pool
        path: PeoplesSpeech-Clean/metadata_pool.parquet
      - split: core
        path: PeoplesSpeech-Clean/metadata_core.parquet
  - config_name: PeoplesSpeech-Dirty
    data_files:
      - split: pool
        path: PeoplesSpeech-Dirty/metadata_pool.parquet
      - split: core
        path: PeoplesSpeech-Dirty/metadata_core.parquet
  - config_name: LibriTTS-R
    data_files:
      - split: pool
        path: LibriTTS-R/metadata_pool.parquet
      - split: core
        path: LibriTTS-R/metadata_core.parquet
  - config_name: SPGISpeech2-Cut
    data_files:
      - split: pool
        path: SPGISpeech2-Cut/metadata_pool.parquet
      - split: core
        path: SPGISpeech2-Cut/metadata_core.parquet
---

# Raon-OpenTTS-Pool

<div align="center">
  <img class="block dark:hidden" src="assets/Raon-OpenTTS-Gradient-Black.png" alt="RAON-OpenTTS" width="600">
  <img class="hidden dark:block" src="assets/Raon-OpenTTS-Gradient-White.png" alt="RAON-OpenTTS" width="600">
</div>

<p align="center">
  <a href="https://www.krafton.ai/ko/"><img src="https://img.shields.io/badge/Homepage-KRAFTON%20AI-blue?style=flat&logo=google-chrome&logoColor=white" alt="Homepage"></a>
  <a href="https://github.com/krafton-ai/RAON-OpenTTS"><img src="https://img.shields.io/badge/GitHub-RAON--OpenTTS-white?style=flat&logo=github&logoColor=black" alt="GitHub"></a>
  <a href="https://huggingface.co/KRAFTON"><img src="https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-KRAFTON-yellow?style=flat" alt="Hugging Face"></a>
  <a href="https://x.com/Krafton_AI"><img src="https://img.shields.io/badge/X-KRAFTON%20AI-white?style=flat&logo=x&logoColor=black" alt="X"></a>
  <a href="#license"><img src="https://img.shields.io/badge/License-Mixed%20(see%20below)-lightgrey?style=flat" alt="License"></a>
</p>

<p align="center">
  Technical Report (Coming soon)
</p>

**Raon-OpenTTS-Pool** is a large-scale open English speech corpus for text-to-speech (TTS) training,
constructed from 8 publicly available speech corpora and a set of web-sourced recordings.
It is the training data behind [RAON-OpenTTS](https://github.com/krafton-ai/RAON-OpenTTS),
an open TTS model that performs on par with state-of-the-art closed-data systems.

- **615K hours** of speech audio
- **239.7M** speech segments
- **11 source datasets** aggregated into a unified format
- All audio stored as **16 kHz mono Opus (64 kbps)** in [WebDataset](https://github.com/webdataset/webdataset) tar shards

We restrict data sources to publicly available English speech datasets with more than 500 hours of audio.
All speech segments are limited to **30 seconds or shorter** to reduce alignment errors, multi-speaker content, and non-speech artifacts.
Existing public datasets (LibriHeavy, Emilia, VoxPopuli, etc.) are included as-is without modification,
with audio standardized to 16 kHz mono Opus 64 kbps for storage efficiency.
The Raon-YouTube-Commons portion is reconstructed from [YouTube-Commons](https://huggingface.co/datasets/PleIAs/YouTube-Commons)
through a dedicated preprocessing pipeline (see [below](#raon-youtube-commons)).

With a model-based filtering pipeline applied to Raon-OpenTTS-Pool, we derive
**Raon-OpenTTS-Core**, a curated high-quality subset of **510.1K hours** and **194.5M** segments.

For more details, see our paper: [Raon-OpenTTS: Open Models and Data for Robust Text-to-Speech](https://github.com/krafton-ai/RAON-OpenTTS)

## Format

Each WebDataset tar shard contains pairs of files per sample:

```
{sample_key}.opus   # 16 kHz mono Opus 64 kbps audio
{sample_key}.json   # {"text": "...", "duration": 8.42, "source": "..."}
```

> **Note:** The dataset viewer shows metadata only (sample_key, text, duration, shard_name).
> Audio is stored in WebDataset tar files — see [Usage](#usage) below to download and load audio.

## Splits

Each dataset config has two metadata splits:

- **pool** — all samples (sample_key, text, duration, shard_name)
- **core** — quality-filtered subset (**Raon-OpenTTS-Core**), retaining ~85% of the data

### Raon-OpenTTS-Core Filtering

Raon-OpenTTS-Core is constructed by applying three model-based quality filters and removing the bottom 15% of samples by combined score:

1. **WER-based**: Transcribe each segment with Whisper-small ASR and compute WER against the existing text annotation. Samples with excessively high WER (> 0.35) indicate severe transcription mismatches.
2. **DNSMOS-based**: Estimate perceptual speech quality using DNSMOS. Samples below 2.24 indicate strong background noise or distortion.
3. **VAD-based**: Estimate speech activity ratio (SAR) using Silero VAD. Samples with SAR below 0.79 are dominated by silence, music, or non-speech audio.
4. **Combined**: Compute an absolute rank for each segment along each criterion (DNSMOS, WER, SAR) and average the ranks into a single combined score. Segments falling below the 15th percentile are discarded.

This combined filtering achieves the best overall TTS performance across diverse evaluation benchmarks (see paper, Figure 3).

## Available Datasets

| Dataset | Source | Size (h) | Avg. Dur. (s) | Segments (M) | Tars | License | DNSMOS | WER | SAR |
|---|---|---|---|---|---|---|---|---|---|
| **Raon-YouTube-Commons** | [YouTube-Commons](https://huggingface.co/datasets/PleIAs/YouTube-Commons) | 335k | 8.5 | 141.70 | 1,017 | CC BY 4.0 | 2.74 | 0.30 | 0.90 |
| **Emilia-YODAS2** | [Emilia](https://huggingface.co/datasets/amphion/Emilia-Dataset) | 92k | 9.2 | 35.97 | 287 | CC BY-NC 4.0 | 2.82 | 0.19 | 0.90 |
| **Emilia** | [Emilia](https://huggingface.co/datasets/amphion/Emilia-Dataset) | 47k | 9.3 | 18.14 | 145 | CC BY 4.0 | 3.02 | 0.18 | 0.89 |
| **LibriHeavy** | [LibriHeavy](https://github.com/k2-fsa/libriheavy) | 42k | 14.2 | 10.77 | 127 | Public Domain | 3.22 | 0.11 | 0.83 |
| **HiFiTTS** | [HiFiTTS2](https://www.openslr.org/hifitts/) | 37k | 10.1 | 13.09 | 109 | CC BY 4.0 | 3.20 | 0.11 | 0.84 |
| **PeoplesSpeech-Dirty** | [People's Speech](https://huggingface.co/datasets/MLCommons/peoples_speech) | 28k | 14.2 | 5.48 | 63 | CC BY 4.0 | 2.63 | 0.25 | 0.86 |
| **VoxPopuli** | [VoxPopuli](https://github.com/facebookresearch/voxpopuli) | 17k | 27.8 | 2.24 | 50 | CC-0 | 2.82 | 0.36 | 0.83 |
| **PeoplesSpeech-Clean** | [People's Speech](https://huggingface.co/datasets/MLCommons/peoples_speech) | 10k | — | 1.50 | 18 | CC BY 4.0 | — | — | — |
| **LibriTTS-R** | [LibriTTS-R](https://www.openslr.org/141/) | 552 | 5.6 | 0.35 | 2 | CC BY 4.0 | 2.96 | 0.06 | 0.91 |
| **SPGISpeech2-Cut** | SPGISpeech 2.0 | 889 | 14.4 | 0.22 | 3 | Kensho UA | 2.72 | 0.08 | 0.90 |
| | | | | | | | | | |
| **Total** | | **615k** | **9.2** | **239.7** | **1,821** | — | 2.83 | 0.24 | 0.89 |

### Raon-YouTube-Commons

A substantial portion of Raon-OpenTTS-Pool (335K hours) is derived from [YouTube-Commons](https://huggingface.co/datasets/PleIAs/YouTube-Commons).
Since the original release provides only YouTube URLs with noisy or unreliable transcriptions,
we reconstructed it into a high-quality speech-text dataset through the following pipeline:

1. **Audio collection**: Download audio from YouTube URLs in the original dataset
2. **Source separation** (UVR-MDX): Suppress background music and non-vocal components
3. **Speaker diarization** (PyAnnote 3.1): Estimate speaker boundaries to ensure single-speaker segments
4. **Voice activity detection** (Silero VAD): Segment continuous speech regions into clips of 3--30 seconds
5. **Automatic transcription** (Whisper-large-v3): Transcribe each segment to obtain aligned speech-text pairs
6. **Standardization**: Resample to 16 kHz mono, encode as 64 kbps Opus

The resulting dataset is released as **Raon-YouTube-Commons** in this repository.

### Non-redistributable Datasets

Two additional datasets used in training cannot be included due to license restrictions.
Users who have agreed to the license on HuggingFace can automatically download and convert them
using `prepare_nonredist_datasets.py`:

| Dataset | Size (h) | License | Source |
|---|---|---|---|
| GigaSpeech | 10k | License agreement required | [speechcolab/gigaspeech](https://huggingface.co/datasets/speechcolab/gigaspeech) |
| SPGISpeech | 5k | Non-commercial (Kensho) | [kensho/spgispeech](https://huggingface.co/datasets/kensho/spgispeech) |

See [Preparing Non-redistributable Datasets](#preparing-non-redistributable-datasets) for instructions.

---

## Usage

### 1. Metadata (pool / core split)

```python
from datasets import load_dataset

# Core metadata for a single dataset
meta = load_dataset("KRAFTON/Raon-OpenTTS-Pool", "Raon-YouTube-Commons", split="core")
# Columns: sample_key, text, duration, shard_name
print(meta[0])

# All datasets combined
all_core = load_dataset("KRAFTON/Raon-OpenTTS-Pool", "all", split="core")
```

### 2. Audio (WebDataset, local tars)

Download tars first:

```python
from huggingface_hub import snapshot_download

local_dir = snapshot_download("KRAFTON/Raon-OpenTTS-Pool", repo_type="dataset",
                               ignore_patterns=["*.parquet"])
```

Then load with WebDataset:

```python
import webdataset as wds
import json, io, soundfile as sf

dataset = (
    wds.WebDataset(f"{local_dir}/LibriTTS-R/lr-{{000000..000001}}.tar")
    .to_tuple("opus", "json")
)
for opus_bytes, json_bytes in dataset:
    meta = json.loads(json_bytes)
    audio, sr = sf.read(io.BytesIO(opus_bytes))
    text = meta["text"]
```

### 3. Core-only training

The audio tars contain pool and core samples mixed. To train on core only, filter by sample_key:

```python
import webdataset as wds
from datasets import load_dataset
import json, io, soundfile as sf

# Step 1: load core sample keys from metadata
core_keys = set(
    load_dataset("KRAFTON/Raon-OpenTTS-Pool", "LibriTTS-R", split="core")["sample_key"]
)

# Step 2: stream tars, skip non-core samples
dataset = (
    wds.WebDataset(f"{local_dir}/LibriTTS-R/lr-{{000000..000001}}.tar")
    .select(lambda s: s["__key__"] in core_keys)
    .to_tuple("opus", "json")
)
for opus_bytes, json_bytes in dataset:
    meta = json.loads(json_bytes)
    audio, sr = sf.read(io.BytesIO(opus_bytes))
    text = meta["text"]
    duration = meta["duration"]
```

---

## Preparing Non-redistributable Datasets

The script `prepare_nonredist_datasets.py` automatically downloads and converts GigaSpeech
and SPGISpeech into the same WebDataset tar + parquet format used by Raon-OpenTTS-Pool.

### Prerequisites

1. **Accept the dataset license** on each HuggingFace dataset page:
   - GigaSpeech: https://huggingface.co/datasets/speechcolab/gigaspeech
   - SPGISpeech: https://huggingface.co/datasets/kensho/spgispeech

2. **Set your HuggingFace token** (from an account that has accepted the licenses):
   ```bash
   export HF_TOKEN=hf_your_token_here
   ```

3. **Install dependencies:**
   ```bash
   pip install "datasets<4.0" soundfile pyarrow numpy tqdm
   ```
   > **Note:** `datasets>=4.0` dropped `soundfile` audio decoding and requires `torchcodec`
   > with system FFmpeg libraries. Use `datasets<4.0` (e.g. `datasets==3.5.0`) to avoid this.

4. **ffmpeg** must be in PATH.

### GigaSpeech

```bash
# Download and convert xl subset from HuggingFace Hub
python prepare_nonredist_datasets.py gigaspeech \
    --output_dir ./GigaSpeech \
    --gigaspeech_subset xl \
    --num_workers 16

# Or from a local HF snapshot (no HF_TOKEN needed)
python prepare_nonredist_datasets.py gigaspeech \
    --source_dir /path/to/gigaspeech_local \
    --output_dir ./GigaSpeech \
    --gigaspeech_subset xl
```

Available subsets: `xs` (10h), `s` (250h), `m` (1000h), `l` (2500h), `xl` (10000h)

### SPGISpeech

```bash
# Download and convert L subset from HuggingFace Hub
python prepare_nonredist_datasets.py spgispeech \
    --output_dir ./SPGISpeech \
    --spgispeech_subset L \
    --num_workers 16

# Or from a local HF snapshot (no HF_TOKEN needed)
python prepare_nonredist_datasets.py spgispeech \
    --source_dir /path/to/spgispeech_local \
    --output_dir ./SPGISpeech \
    --num_workers 16
```

Available subsets: `L` (full ~5000h), `M` (~1000h), `S` (~200h), `dev`, `test`

### Output

```
<output_dir>/
  {prefix}-000000.tar     # WebDataset shard (~10 GB)
  {prefix}-000001.tar
  ...
  metadata_pool.parquet   # all samples
  metadata_core.parquet   # = pool (no quality filtering without --core_json)
```

By default `metadata_core.parquet` equals `metadata_pool.parquet` since quality filtering
requires an internal index file. If you have `pool_indices_filter_remove_15pct_combined.json`
from the Raon-OpenTTS maintainers, pass it with `--core_json` to generate a filtered core split.

### Using with RAON-OpenTTS training

Once prepared, pass the output directory as a `nonredist_dirs` entry in the training config:

```yaml
datasets:
  nonredist_dirs:
    - /path/to/GigaSpeech
    - /path/to/SPGISpeech
```

---

## License

**This repository contains data from multiple sources, each with its own license.**
Users must comply with the license of each individual sub-dataset they use.

| Dataset | License | Commercial Use |
|---|---|---|
| Raon-YouTube-Commons | [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) | Yes |
| Emilia | [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) | Yes |
| **Emilia-YODAS2** | **[CC BY-NC 4.0](https://creativecommons.org/licenses/by-nc/4.0/)** | **No** |
| LibriHeavy | Public Domain (LibriVox) | Yes |
| HiFiTTS | [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) | Yes |
| PeoplesSpeech-Clean / Dirty | [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) | Yes |
| VoxPopuli | [CC-0](https://creativecommons.org/publicdomain/zero/1.0/) | Yes |
| LibriTTS-R | [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) | Yes |
| SPGISpeech2-Cut | [Kensho User Agreement](https://huggingface.co/datasets/kensho/spgispeech) | Non-commercial |
| GigaSpeech (non-redist) | [License agreement required](https://huggingface.co/datasets/speechcolab/gigaspeech) | See terms |
| SPGISpeech (non-redist) | [Kensho User Agreement](https://huggingface.co/datasets/kensho/spgispeech) | Non-commercial |
| Metadata and dataset structure | [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) | Yes |

> **Note:** Emilia-YODAS2 and SPGISpeech2-Cut are licensed under non-commercial terms.
> If you require fully commercial-use data, exclude these sub-datasets via the `configs` parameter.

## Citation

```bibtex
@article{raon2026opentts,
  title     = {Raon-OpenTTS: Open Models and Data for Robust Text-to-Speech},
  author    = {TBD},
  year      = {2026},
  url       = {https://github.com/krafton-ai/Raon-OpenTTS}
}
```

&copy; 2026 KRAFTON