NMikka's picture
Update dataset card with train/eval/test split info
965fee3 verified
---
language:
- ka
license: cc0-1.0
task_categories:
- text-to-speech
- audio-to-audio
tags:
- georgian
- tts
- common-voice
- speech-synthesis
size_categories:
- 10K<n<100K
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: eval
path: data/eval-*
- split: test
path: data/test-*
dataset_info:
features:
- name: id
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 24000
- name: text
dtype: string
- name: speaker_id
dtype: string
- name: duration
dtype: float64
splits:
- name: train
num_bytes: 6137099626.8
num_examples: 20300
- name: eval
num_bytes: 282215307.562
num_examples: 1001
- name: test
num_bytes: 31410778.0
num_examples: 120
download_size: 5452817424
dataset_size: 6450725712.362
---
# Common Voice Georgian — Cleaned for TTS/STT
A high-quality subset of [Mozilla Common Voice Georgian](https://commonvoice.mozilla.org/en/datasets) cleaned and filtered specifically for text-to-speech fine-tuning.
## Dataset Summary
| | |
|---|---|
| **Total samples** | 21,421 |
| **Total duration** | 35.0 hours |
| **Speakers** | 12 |
| **Sample rate** | 24 kHz mono WAV |
| **Language** | Georgian (kat) |
| **Source** | Mozilla Common Voice 19.0 |
| **License** | CC-0 (public domain) |
## Splits
| Split | Samples | Description |
|-------|---------|-------------|
| `train` | 20,300 | Training data |
| `eval` | 1,001 | Validation data |
| `test` | 120 | Best quality speaker references (top NISQA scores) |
## Quality Pipeline
The dataset was cleaned from ~71K raw Common Voice recordings through a 6-stage pipeline:
1. **Standardize** — Resample to 24 kHz mono, normalize loudness to −23 LUFS, filter duration to [0.5s, 30s]
2. **Enhance** — VoiceFixer audio restoration + Sox spectral noise subtraction
3. **NISQA Filter** — NISQA MOS ≥ 3.0 (neural speech quality assessment)
4. **Duration Outlier** — IQR-based character duration filter (removes misaligned/rushed/slow speech)
5. **Transcript Verify** — Round-trip ASR (Meta Omnilingual 7B, 1.9% CER on Georgian) with CER ≤ 0.20 threshold
6. **Speaker Select** — Keep speakers with ≥ 1800 seconds total audio
## Fields
| Field | Type | Description |
|-------|------|-------------|
| `id` | string | Common Voice clip ID |
| `audio` | Audio | 24 kHz mono WAV |
| `text` | string | Georgian transcript |
| `speaker_id` | string | Anonymized speaker ID (0–11) |
| `duration` | float | Duration in seconds |
## Speaker Distribution
| Speaker | Samples | Duration |
|---------|---------|----------|
| 0 | 5,683 | 8.8h |
| 1 | 1,164 | 1.8h |
| 2 | 2,970 | 5.3h |
| 3 | 3,240 | 5.3h |
| 4 | 2,595 | 3.6h |
| 5 | 1,556 | 2.8h |
| 6 | 1,131 | 1.8h |
| 7 | 1,130 | 2.1h |
| 8 | 470 | 0.8h |
| 9 | 544 | 1.0h |
| 10 | 607 | 1.0h |
| 11 | 331 | 0.7h |
## Statistics
- **Duration**: min 2.4s, mean 5.9s, max 10.6s
## Usage
```python
from datasets import load_dataset
ds = load_dataset("NMikka/Common-Voice-Geo-Cleaned")
# Training
for sample in ds["train"]:
print(sample["text"], sample["duration"])
# Validation
for sample in ds["eval"]:
print(sample["text"])
# Best speaker references (for TTS inference/voice cloning)
for sample in ds["test"]:
print(sample["text"], sample["speaker_id"])
```
## Citation
If you use this dataset, please cite Mozilla Common Voice:
```bibtex
@inproceedings{ardila2020common,
title={Common Voice: A Massively-Multilingual Speech Corpus},
author={Ardila, Rosana and others},
booktitle={LREC},
year={2020}
}
```