| # Dataset Setup |
| We open-source our preprocessed datasets in the S3 link here. Download it at put them into the `data` directory. |
|
|
| Below are the details of how we preprocessed the dataset. |
|
|
| ## Preprocessing |
| ### VCTK (English multi-speaker, multi-accent) |
|
|
| We use a curated subset of the **CSTR VCTK Corpus: English Multi-speaker Corpus for CSTR Voice Cloning Toolkit (version 0.92)** [Yamagishi et al., 2019](https://doi.org/10.7488/ds/2645). |
|
|
| The original VCTK corpus contains ~44 hours of English read speech from 110 speakers with diverse accents (American, British, Irish, Scottish, Indian, etc.), each reading about 400 phonetically rich sentences recorded with high-quality microphones in a hemi-anechoic chamber. |
|
|
| In our benchmark, we select **40 speakers** to balance **gender** and **accent** while keeping the dataset compact: |
|
|
| - **Speakers:** 40 in total (22 female, 18 male) |
| - **Accents:** 12 accent categories |
| - Major accents (American, English, Canadian, Irish, Scottish, South African) have 4–6 speakers each |
| - Minor accents (e.g., Australian, New Zealand, Welsh, Northern Irish, Indian, British RP) have 1–3 speakers each |
|
|
| All audio is used at **48 kHz, 16-bit PCM, mono**, following the original VCTK speaker IDs (e.g., `p294`, `p303`, `s5`). |
|
|
| ### LibriTTS (English multi-speaker TTS) |
|
|
| We use a small, speaker-balanced subset of the **LibriTTS** corpus. |
| LibriTTS is a large-scale, multi-speaker English corpus of read audiobooks, derived from LibriSpeech and released at 24 kHz for TTS research: it contains approximately **585 hours** of speech from **2,456 speakers**, with sentence-level segmentation, normalized text, and noisy utterances removed. |
|
|
| Official LibriTTS page: [https://www.openslr.org/60/](https://www.openslr.org/60/) |
|
|
| In our benchmark subset: |
|
|
| - **Speakers:** 40 total, all from LibriTTS `dev` and `test` splits |
| - **Gender balance (overall):** |
| - 20 female speakers |
| - 20 male speakers |
|
|
| - **Split design (speaker-disjoint):** |
| - **dev:** 19 speakers (7 F / 12 M) |
| - **test:** 21 speakers (13 F / 8 M) |
|
|
| All audio is used at **24 kHz, mono, 16-bit PCM**, following the original LibriTTS speaker IDs |
|
|
| ### AISHELL-1 |
|
|
| We use a subset of the **AISHELL-1** Mandarin corpus for development and evaluation. |
| AISHELL-1 is an open-source Mandarin speech corpus containing about **178 hours** of transcribed speech from **400 speakers**, recorded in quiet indoor environments and released at **16 kHz**. The texts cover 11 application-oriented domains (finance, science & technology, sports, entertainment, news, etc.). |
|
|
| Official AISHELL-1 page: https://www.openslr.org/33/ |
|
|
| In our benchmark subset: |
|
|
| - **Speakers:** 40 total (IDs `S0724`–`S0763`) |
| - **Gender balance:** 28 female, 12 male |
| - **Per-speaker content:** 50 paired utterances (`gt`, `ori`) per speaker |
| - **Sampling rate:** 16 kHz, mono, 16-bit PCM |
|
|
| ### UEDIN_bilingual_96kHz (Mandarin/English bilingual) |
|
|
| We use the **Mandarin talkers** portion of the 96 kHz release of the **EMIME Bilingual English–Mandarin Database**. |
| The original EMIME Mandarin/English database contains studio recordings of 14 native Mandarin speakers (7 female, 7 male) who read parallel material in both Mandarin and English for research on personalized speech-to-speech translation and cross-lingual voice conversion. |
| Official EMIME bilingual database page: https://www.emime.org/participate/emime-bilingual-database.html |
|
|
| In our benchmark subset: |
|
|
| - **Speakers:** 13 total from `Mandarin_talkers/` |
| - Female: 6 (`MF1`, `MF3`, `MF4`, `MF5`, `MF6`, `MF7`) |
| - Male: 7 (`MM1`–`MM7`) |
| - We exclude **MF2** because some of her Mandarin recordings exhibit abnormal behaviour. |
| - **Gender balance:** 6 female, 7 male |
| - **Bilingual content per speaker:** |
| - 25 **English → Mandarin** sentence pairs |
| - 25 **Mandarin → English** sentence pairs |
| → 50 bilingual pairs per speaker |
|
|
| - **Sampling rate:** 96 kHz, mono, 16-bit PCM |
|
|
| ### CommonVoiceFR_dev (French crowd-sourced speech) |
| |
| We use a subset of **Mozilla Common Voice French v23.0**, taking clips from the `validated.tsv` split. |
| |
| Official Common Voice site: [https://commonvoice.mozilla.org/](https://datacollective.mozillafoundation.org/datasets/cmflnuzw5ahjms0zbrcl0vg4e) |
| |
| In our benchmark subset: |
| |
| - **Source split:** `validated.tsv` from Common Voice French v23.0 |
| - **Speakers:** 40 unique French speakers |
| - **Per-speaker content:** 50 utterance pairs per speaker. |
| - **Audio format:** mono, 16-bit PCM, 48 kHz (downsampled in code if needed) |
| |
| ### Long_context (LibriSpeech-Long subset) |
|
|
| We use a long-form English subset derived from **LibriSpeech-Long**, a benchmark dataset for long-form speech processing released with *“Long-Form Speech Generation with Spoken Language Models”* (Park et al., 2024). |
|
|
| - Original LibriSpeech-Long card: https://huggingface.co/datasets/ilyakam/librispeech-long |
|
|
| In our benchmark subset (“Long_context”): |
| |
| - **Language:** English |
| - **Speakers:** 10 |
| - Speaker IDs: `1272, 1673, 1919, 1993, 2078, 3576, 3853, 422, 6241, 8842` |
| - **Pairs per speaker:** 2 long-form utterance pairs (e.g., original vs processed) |
| - **Total pairs:** 20 |
| - **Audio duration:** long-form segments, typically **0.4–4 minutes** per file |
| - **Sampling rate:** 16 kHz, mono, 16-bit PCM (inherited from LibriSpeech) |
| |
| |
| ### VCTK natural_noise (VoiceBank+DEMAND subset) |
|
|
| We use a noisy VCTK subset derived from the **“Noisy speech database for training speech enhancement algorithms and TTS models”** (also known as the VoiceBank+DEMAND dataset). |
| The original database provides parallel **clean/noisy speech at 48 kHz** based on the VCTK multi-speaker corpus, where clean VCTK utterances are mixed with environmental noises (mainly from the DEMAND corpus) and additional speech-shaped / babble noise, and is widely used for speech enhancement and noise-robust TTS research. |
|
|
| Official dataset page: https://datashare.ed.ac.uk/handle/10283/2791 |
|
|
| In our benchmark subse: |
|
|
| - **Language:** English (multi-accent, inherited from VCTK) |
| - **Speakers:** 20 VCTK speakers (`p226`–`p273`) |
| - **Background noise:** 10 natural noise environments, all mixed at **10 dB SNR** |
| - `babble`, `cafeteria`, `car`, `kitchen`, `meeting`, `metro`, `restaurant`, |
| `ssn` (speech-shaped noise), `station`, `traffic` |
| - **Per-speaker structure:** |
| - 20 **noisy–clean pairs** (each noisy utterance has a clean VCTK reference) |
| - 10 additional **clean-only** utterances |
| - → 30 sentence-level items per speaker |
| - **Overall size:** 20 speakers × 30 items ≈ **600 sentence items** (clean/noisy together give about **1,200 audio files**) |
| - **Audio format:** 48 kHz, mono, 16-bit PCM |
| |
| ### Multispeaker_libri (English multi-speaker interference) |
| |
| We use a synthetic multi-speaker interference set built from our **LibriTTS** English audiobook subset. |
| Clean target utterances from 10 LibriTTS speakers are mixed with interfering speech from 2 additional speakers at controlled SNR levels, producing parallel **clean / mixture** pairs for studying multi-speaker robustness. |
| |
| In our benchmark subset: |
| |
| - **Target speakers:** 10 (5 male, 5 female) |
| - Male: `61, 908, 2300, 2830, 7729` |
| - Female: `237, 1221, 1284, 4970, 6829` |
| - **Interferer speakers:** 2 |
| - Female interferer: `121` |
| - Male interferer: `672` |
| - **Clean groundtruth segments:** 100 (10 per target speaker) |
| - **Mixtures per groundtruth:** 8 |
| - 4 SNR levels: `-5 dB`, `0 dB`, `+5 dB`, `+10 dB` |
| - 2 interferers: `121`, `672` |
| - **Total entries in `manifest.json`:** 800 mixture entries (each with a linked clean groundtruth) |
| - **Sampling rate:** 16 kHz, mono, 16-bit PCM |
| |
| ### iemocap (TO BE EXPLORED — emotional English speech) |
| |
| > Status: **to be explored** — this dataset is prepared but not yet integrated into the main benchmark pipeline. |
| |
| We construct an emotional subset from the **IEMOCAP (Interactive Emotional Dyadic Motion Capture)** corpus. |
| |
| In our prepared subset: |
| |
| - **Language:** English |
| - **Speakers:** 10 total (5 male, 5 female) |
| - Female: `Ses01_F, Ses02_F, Ses03_F, Ses04_F, Ses05_F` |
| - Male: `Ses01_M, Ses02_M, Ses03_M, Ses04_M, Ses05_M` |
| - **Total pairs:** 184 (`ori`, `gt`) audio pairs |
| - **Emotion categories (6):** `ang`, `exc`, `fru`, `hap`, `neu`, `sad` |
|
|
| **Pair construction rule** |
|
|
| For each speaker, we select utterances from the six emotion types and build pairs such that: |
|
|
| - `ori` and `gt` always come from the **same speaker**, but |
| - **their emotions are deliberately different** (e.g., `ori = ang`, `gt = neu`). |
|
|
| ### robocall_ftc (TO BE DONE — real-world scam calls) |
| |
| > Status: **to be done** — planned for trustworthy deepfake / telecom-fraud experiments. |
| |
| We plan to integrate the **Robocall Audio Dataset** released by Prasad & Reaves (NCSU) based on the FTC’s *Project Point of No Entry* (PPoNE). |
| This dataset contains real-world audio recordings of automated or semi-automated phone calls (“robocalls”), most of which are suspected to be **illegal scam or spam calls**. |
| |
| Official repo: https://github.com/wspr-ncsu/robocall-audio-dataset |
| |
| Planned usage in our benchmark: |
| |
| |
| |