--- license: apache-2.0 --- # SpeechFake Dataset Please use the download scripts from https://github.com/XIAOYixuan/AUDDT/tree/yixuan-dev to download and process the dataset. ``` chmod +x download/get_speechfake.sh ./download/get_speechfake.sh ``` ## Label Distribution The dataset is organized into four experiment types: ### baseline - **train_all**: 704,862 samples (spoof: 629,154, bonafide: 75,708) - **train_en**: 428,266 samples (spoof: 389,866, bonafide: 38,400) - **train_zh**: 276,596 samples (spoof: 239,288, bonafide: 37,308) - **test_all**: 346,313 samples (spoof: 309,065, bonafide: 37,248) - **test_en**: 208,655 samples (spoof: 189,455, bonafide: 19,200) - **test_zh**: 137,658 samples (spoof: 119,610, bonafide: 18,048) - **dev_all**: 117,463 samples (spoof: 104,845, bonafide: 12,618) - **dev_en**: 71,370 samples (spoof: 64,970, bonafide: 6,400) - **dev_zh**: 46,093 samples (spoof: 39,875, bonafide: 6,218) ### cross_generator - **train_TTS**: 276,422 samples (all spoof) - **train_VC**: 192,210 samples (all spoof) - **train_NV**: 160,522 samples (all spoof) - **test_TTS**: 132,733 samples (all spoof) - **test_VC**: 96,059 samples (all spoof) - **test_NV**: 80,273 samples (all spoof) - **dev_TTS**: 46,065 samples (all spoof) - **dev_VC**: 32,031 samples (all spoof) - **dev_NV**: 26,749 samples (all spoof) ### cross_lingual - **train**: 263,399 samples (spoof: 203,399, bonafide: 60,000) - **test**: 133,707 samples (spoof: 103,707, bonafide: 30,000) - **test_ko**: 85,228 samples (spoof: 82,728, bonafide: 2,500) - **test_zh**: 63,283 samples (spoof: 48,283, bonafide: 15,000) - **test_en**: 70,424 samples (spoof: 55,424, bonafide: 15,000) - **test_it**: 44,989 samples (spoof: 39,989, bonafide: 5,000) - **test_hu**: 44,981 samples (spoof: 39,981, bonafide: 5,000) - **test_id**: 44,904 samples (spoof: 39,936, bonafide: 4,968) - **test_es**: 48,868 samples (spoof: 43,868, bonafide: 5,000) - **test_gl**: 44,838 samples (spoof: 39,838, bonafide: 5,000) - **test_lv**: 37,784 samples (spoof: 32,845, bonafide: 4,939) - **test_fi**: 36,578 samples (spoof: 31,619, bonafide: 4,959) - **test_et**: 33,350 samples (spoof: 28,392, bonafide: 4,958) - **test_he**: 21,210 samples (spoof: 20,605, bonafide: 605) - **test_is**: 13,388 samples (spoof: 13,373, bonafide: 15) - **dev**: 44,563 samples (spoof: 34,563, bonafide: 10,000) ### cross_speaker - **train**: 34,305 samples (spoof: 27,734, bonafide: 6,571) - **test_overlap2**: 18,976 samples (spoof: 12,377, bonafide: 6,599) - **test_same_spk**: 20,470 samples (spoof: 13,871, bonafide: 6,599) - **test_overlap1**: 19,428 samples (spoof: 13,871, bonafide: 5,557) - **test_diff_spk**: 17,934 samples (spoof: 12,377, bonafide: 5,557) ## Data Attributes ``` ID path label dataset_name 0 0 Real/LibriTTS/train-clean-100/1841/150351/1841_150351_000026_000002.wav bonafide SpeechFake 1 1 Real/LibriTTS/train-clean-100/1116/132851/1116_132851_000040_000002.wav bonafide SpeechFake 2 2 Real/LibriTTS/train-clean-100/83/9960/83_9960_000022_000000.wav bonafide SpeechFake ``` ## How to Import ```python import pandas as pd # Example: Load baseline train_all split df = pd.read_parquet("baseline/train_all.parquet") print(df.head()) ```