cdminix commited on
Commit
ce8d818
·
verified ·
1 Parent(s): 30db0ca

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .DS_Store +0 -0
  2. .env +1 -0
  3. 01_eddacc_exploration.ipynb +554 -0
  4. dataset/.DS_Store +0 -0
  5. dataset/chinese/.DS_Store +0 -0
  6. dataset/chinese/female/hesitation/40-chinese-female-16385-hesitation.wav +3 -0
  7. dataset/chinese/female/hesitation/41-chinese-female-16385-hesitation.wav +3 -0
  8. dataset/chinese/female/hesitation/42-chinese-female-16385-hesitation.wav +3 -0
  9. dataset/chinese/female/hesitation/43-chinese-female-16385-hesitation.wav +3 -0
  10. dataset/chinese/female/hesitation/44-chinese-female-16385-hesitation.wav +3 -0
  11. dataset/chinese/female/hesitation/45-chinese-female-16385-hesitation.wav +3 -0
  12. dataset/chinese/female/hesitation/46-chinese-female-16385-hesitation.wav +3 -0
  13. dataset/chinese/female/hesitation/47-chinese-female-16385-hesitation.wav +3 -0
  14. dataset/chinese/female/hesitation/48-chinese-female-16385-hesitation.wav +3 -0
  15. dataset/chinese/female/hesitation/49-chinese-female-16385-hesitation.wav +3 -0
  16. dataset/chinese/female/hesitation/50-chinese-female-16385-hesitation.wav +3 -0
  17. dataset/chinese/female/hesitation/51-chinese-female-16385-hesitation.wav +3 -0
  18. dataset/chinese/female/hesitation/52-chinese-female-16385-hesitation.wav +3 -0
  19. dataset/chinese/female/hesitation/53-chinese-female-16385-hesitation.wav +3 -0
  20. dataset/chinese/female/hesitation/54-chinese-female-16385-hesitation.wav +3 -0
  21. dataset/chinese/female/hesitation/55-chinese-female-16385-hesitation.wav +3 -0
  22. dataset/chinese/female/hesitation/56-chinese-female-16385-hesitation.wav +3 -0
  23. dataset/chinese/female/hesitation/57-chinese-female-16385-hesitation.wav +3 -0
  24. dataset/chinese/female/hesitation/58-chinese-female-16385-hesitation.wav +3 -0
  25. dataset/chinese/female/hesitation/59-chinese-female-16385-hesitation.wav +3 -0
  26. dataset/chinese/female/hesitation/60-chinese-female-16385-hesitation.wav +3 -0
  27. dataset/chinese/female/hesitation/61-chinese-female-16385-hesitation.wav +3 -0
  28. dataset/chinese/female/hesitation/62-chinese-female-16385-hesitation.wav +3 -0
  29. dataset/chinese/female/hesitation/63-chinese-female-16385-hesitation.wav +3 -0
  30. dataset/chinese/female/hesitation/64-chinese-female-16385-hesitation.wav +3 -0
  31. dataset/chinese/female/hesitation/65-chinese-female-16385-hesitation.wav +3 -0
  32. dataset/chinese/female/hesitation/66-chinese-female-16385-hesitation.wav +3 -0
  33. dataset/chinese/female/hesitation/67-chinese-female-16385-hesitation.wav +3 -0
  34. dataset/chinese/female/hesitation/68-chinese-female-16385-hesitation.wav +3 -0
  35. dataset/chinese/female/hesitation/69-chinese-female-16385-hesitation.wav +3 -0
  36. dataset/chinese/female/hesitation/70-chinese-female-16385-hesitation.wav +3 -0
  37. dataset/chinese/female/hesitation/71-chinese-female-16385-hesitation.wav +3 -0
  38. dataset/chinese/female/hesitation/72-chinese-female-16385-hesitation.wav +3 -0
  39. dataset/chinese/female/hesitation/73-chinese-female-16385-hesitation.wav +3 -0
  40. dataset/chinese/female/hesitation/74-chinese-female-16385-hesitation.wav +3 -0
  41. dataset/chinese/female/hesitation/75-chinese-female-16385-hesitation.wav +3 -0
  42. dataset/chinese/female/hesitation/76-chinese-female-16385-hesitation.wav +3 -0
  43. dataset/chinese/female/hesitation/77-chinese-female-16385-hesitation.wav +3 -0
  44. dataset/chinese/female/hesitation/78-chinese-female-16385-hesitation.wav +3 -0
  45. dataset/chinese/female/hesitation/79-chinese-female-16385-hesitation.wav +3 -0
  46. dataset/chinese/female/no_hesitation/0-chinese-female-16385-no_hesitation.wav +3 -0
  47. dataset/chinese/female/no_hesitation/1-chinese-female-16385-no_hesitation.wav +3 -0
  48. dataset/chinese/female/no_hesitation/10-chinese-female-16385-no_hesitation.wav +3 -0
  49. dataset/chinese/female/no_hesitation/11-chinese-female-16385-no_hesitation.wav +3 -0
  50. dataset/chinese/female/no_hesitation/12-chinese-female-16385-no_hesitation.wav +3 -0
.DS_Store ADDED
Binary file (6.15 kB). View file
 
.env ADDED
@@ -0,0 +1 @@
 
 
1
+ REPLICATE_API_TOKEN=r8_DUSwbx0Ce5hM8rsZmDyizDX4vRrM3W00j2TjA
01_eddacc_exploration.ipynb ADDED
@@ -0,0 +1,554 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cells": [
3
+ {
4
+ "cell_type": "markdown",
5
+ "id": "d54feda5",
6
+ "metadata": {},
7
+ "source": [
8
+ "## EddAcc Exploration\n",
9
+ "We visualise the speaker identities/accents in the EddAcc test and dev splits (those are the splits used to avoid splits potentially seen by SLMs)."
10
+ ]
11
+ },
12
+ {
13
+ "cell_type": "code",
14
+ "execution_count": 10,
15
+ "id": "58b9fbe9",
16
+ "metadata": {
17
+ "vscode": {
18
+ "languageId": "bat"
19
+ }
20
+ },
21
+ "outputs": [
22
+ {
23
+ "name": "stdout",
24
+ "output_type": "stream",
25
+ "text": [
26
+ "ERROR: Python with virtualenvwrapper module not found!\n",
27
+ "Either, install virtualenvwrapper module for the default python3 interpreter\n",
28
+ "or set VIRTUALENVWRAPPER_PYTHON to the interpreter to use.\n",
29
+ "Requirement already satisfied: requests in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from -r requirements.txt (line 1)) (2.32.5)\n",
30
+ "Requirement already satisfied: datasets in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from -r requirements.txt (line 2)) (3.6.0)\n",
31
+ "Requirement already satisfied: tqdm in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from -r requirements.txt (line 3)) (4.67.1)\n",
32
+ "Requirement already satisfied: ipywidgets in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from -r requirements.txt (line 4)) (8.1.8)\n",
33
+ "Requirement already satisfied: torchaudio in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from -r requirements.txt (line 5)) (2.2.2)\n",
34
+ "Requirement already satisfied: torch in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from -r requirements.txt (line 6)) (2.2.2)\n",
35
+ "Requirement already satisfied: torchcodec in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from -r requirements.txt (line 8)) (0.8.1)\n",
36
+ "Requirement already satisfied: deepmultilingualpunctuation in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from -r requirements.txt (line 9)) (1.0.1)\n",
37
+ "Requirement already satisfied: idna<4,>=2.5 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from requests->-r requirements.txt (line 1)) (3.10)\n",
38
+ "Requirement already satisfied: urllib3<3,>=1.21.1 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from requests->-r requirements.txt (line 1)) (2.5.0)\n",
39
+ "Requirement already satisfied: certifi>=2017.4.17 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from requests->-r requirements.txt (line 1)) (2025.8.3)\n",
40
+ "Requirement already satisfied: charset_normalizer<4,>=2 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from requests->-r requirements.txt (line 1)) (3.4.3)\n",
41
+ "Requirement already satisfied: huggingface-hub>=0.24.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from datasets->-r requirements.txt (line 2)) (0.35.1)\n",
42
+ "Requirement already satisfied: pandas in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from datasets->-r requirements.txt (line 2)) (2.3.2)\n",
43
+ "Requirement already satisfied: xxhash in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from datasets->-r requirements.txt (line 2)) (3.6.0)\n",
44
+ "Requirement already satisfied: pyarrow>=15.0.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from datasets->-r requirements.txt (line 2)) (21.0.0)\n",
45
+ "Requirement already satisfied: filelock in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from datasets->-r requirements.txt (line 2)) (3.19.1)\n",
46
+ "Requirement already satisfied: numpy>=1.17 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from datasets->-r requirements.txt (line 2)) (1.26.4)\n",
47
+ "Requirement already satisfied: pyyaml>=5.1 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from datasets->-r requirements.txt (line 2)) (6.0.3)\n",
48
+ "Requirement already satisfied: dill<0.3.9,>=0.3.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from datasets->-r requirements.txt (line 2)) (0.3.8)\n",
49
+ "Requirement already satisfied: packaging in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from datasets->-r requirements.txt (line 2)) (25.0)\n",
50
+ "Requirement already satisfied: fsspec[http]<=2025.3.0,>=2023.1.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from datasets->-r requirements.txt (line 2)) (2025.3.0)\n",
51
+ "Requirement already satisfied: multiprocess<0.70.17 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from datasets->-r requirements.txt (line 2)) (0.70.16)\n",
52
+ "Requirement already satisfied: comm>=0.1.3 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from ipywidgets->-r requirements.txt (line 4)) (0.2.3)\n",
53
+ "Requirement already satisfied: traitlets>=4.3.1 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from ipywidgets->-r requirements.txt (line 4)) (5.14.3)\n",
54
+ "Requirement already satisfied: ipython>=6.1.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from ipywidgets->-r requirements.txt (line 4)) (8.37.0)\n",
55
+ "Requirement already satisfied: jupyterlab_widgets~=3.0.15 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from ipywidgets->-r requirements.txt (line 4)) (3.0.16)\n",
56
+ "Requirement already satisfied: widgetsnbextension~=4.0.14 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from ipywidgets->-r requirements.txt (line 4)) (4.0.15)\n",
57
+ "Requirement already satisfied: networkx in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from torch->-r requirements.txt (line 6)) (3.4.2)\n",
58
+ "Requirement already satisfied: jinja2 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from torch->-r requirements.txt (line 6)) (3.1.6)\n",
59
+ "Requirement already satisfied: sympy in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from torch->-r requirements.txt (line 6)) (1.14.0)\n",
60
+ "Requirement already satisfied: typing-extensions>=4.8.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from torch->-r requirements.txt (line 6)) (4.15.0)\n",
61
+ "Requirement already satisfied: librosa in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from datasets->-r requirements.txt (line 2)) (0.11.0)\n",
62
+ "Requirement already satisfied: soundfile>=0.12.1 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from datasets->-r requirements.txt (line 2)) (0.13.1)\n",
63
+ "Requirement already satisfied: soxr>=0.4.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from datasets->-r requirements.txt (line 2)) (1.0.0)\n",
64
+ "Requirement already satisfied: transformers in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from deepmultilingualpunctuation->-r requirements.txt (line 9)) (4.56.2)\n",
65
+ "Requirement already satisfied: aiohttp!=4.0.0a0,!=4.0.0a1 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from fsspec[http]<=2025.3.0,>=2023.1.0->datasets->-r requirements.txt (line 2)) (3.12.15)\n",
66
+ "Requirement already satisfied: hf-xet<2.0.0,>=1.1.3 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from huggingface-hub>=0.24.0->datasets->-r requirements.txt (line 2)) (1.1.10)\n",
67
+ "Requirement already satisfied: pygments>=2.4.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from ipython>=6.1.0->ipywidgets->-r requirements.txt (line 4)) (2.19.2)\n",
68
+ "Requirement already satisfied: pexpect>4.3 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from ipython>=6.1.0->ipywidgets->-r requirements.txt (line 4)) (4.9.0)\n",
69
+ "Requirement already satisfied: jedi>=0.16 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from ipython>=6.1.0->ipywidgets->-r requirements.txt (line 4)) (0.19.2)\n",
70
+ "Requirement already satisfied: prompt_toolkit<3.1.0,>=3.0.41 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from ipython>=6.1.0->ipywidgets->-r requirements.txt (line 4)) (3.0.52)\n",
71
+ "Requirement already satisfied: matplotlib-inline in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from ipython>=6.1.0->ipywidgets->-r requirements.txt (line 4)) (0.2.1)\n",
72
+ "Requirement already satisfied: decorator in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from ipython>=6.1.0->ipywidgets->-r requirements.txt (line 4)) (5.2.1)\n",
73
+ "Requirement already satisfied: stack_data in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from ipython>=6.1.0->ipywidgets->-r requirements.txt (line 4)) (0.6.3)\n",
74
+ "Requirement already satisfied: exceptiongroup in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from ipython>=6.1.0->ipywidgets->-r requirements.txt (line 4)) (1.3.0)\n",
75
+ "Requirement already satisfied: cffi>=1.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from soundfile>=0.12.1->datasets->-r requirements.txt (line 2)) (2.0.0)\n",
76
+ "Requirement already satisfied: MarkupSafe>=2.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from jinja2->torch->-r requirements.txt (line 6)) (2.1.5)\n",
77
+ "Requirement already satisfied: msgpack>=1.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from librosa->datasets->-r requirements.txt (line 2)) (1.1.1)\n",
78
+ "Requirement already satisfied: lazy_loader>=0.1 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from librosa->datasets->-r requirements.txt (line 2)) (0.4)\n",
79
+ "Requirement already satisfied: joblib>=1.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from librosa->datasets->-r requirements.txt (line 2)) (1.5.2)\n",
80
+ "Requirement already satisfied: audioread>=2.1.9 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from librosa->datasets->-r requirements.txt (line 2)) (3.0.1)\n",
81
+ "Requirement already satisfied: scipy>=1.6.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from librosa->datasets->-r requirements.txt (line 2)) (1.15.3)\n",
82
+ "Requirement already satisfied: scikit-learn>=1.1.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from librosa->datasets->-r requirements.txt (line 2)) (1.7.2)\n",
83
+ "Requirement already satisfied: numba>=0.51.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from librosa->datasets->-r requirements.txt (line 2)) (0.62.0)\n",
84
+ "Requirement already satisfied: pooch>=1.1 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from librosa->datasets->-r requirements.txt (line 2)) (1.8.2)\n",
85
+ "Requirement already satisfied: pytz>=2020.1 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from pandas->datasets->-r requirements.txt (line 2)) (2025.2)\n",
86
+ "Requirement already satisfied: tzdata>=2022.7 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from pandas->datasets->-r requirements.txt (line 2)) (2025.2)\n",
87
+ "Requirement already satisfied: python-dateutil>=2.8.2 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from pandas->datasets->-r requirements.txt (line 2)) (2.9.0.post0)\n",
88
+ "Requirement already satisfied: mpmath<1.4,>=1.1.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from sympy->torch->-r requirements.txt (line 6)) (1.3.0)\n",
89
+ "Requirement already satisfied: regex!=2019.12.17 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from transformers->deepmultilingualpunctuation->-r requirements.txt (line 9)) (2025.9.18)\n",
90
+ "Requirement already satisfied: tokenizers<=0.23.0,>=0.22.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from transformers->deepmultilingualpunctuation->-r requirements.txt (line 9)) (0.22.1)\n",
91
+ "Requirement already satisfied: safetensors>=0.4.3 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from transformers->deepmultilingualpunctuation->-r requirements.txt (line 9)) (0.6.2)\n",
92
+ "Requirement already satisfied: attrs>=17.3.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->-r requirements.txt (line 2)) (25.3.0)\n",
93
+ "Requirement already satisfied: multidict<7.0,>=4.5 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->-r requirements.txt (line 2)) (6.6.4)\n",
94
+ "Requirement already satisfied: async-timeout<6.0,>=4.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->-r requirements.txt (line 2)) (5.0.1)\n",
95
+ "Requirement already satisfied: aiosignal>=1.4.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->-r requirements.txt (line 2)) (1.4.0)\n",
96
+ "Requirement already satisfied: frozenlist>=1.1.1 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->-r requirements.txt (line 2)) (1.7.0)\n",
97
+ "Requirement already satisfied: propcache>=0.2.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->-r requirements.txt (line 2)) (0.3.2)\n",
98
+ "Requirement already satisfied: yarl<2.0,>=1.17.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->-r requirements.txt (line 2)) (1.20.1)\n",
99
+ "Requirement already satisfied: aiohappyeyeballs>=2.5.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->-r requirements.txt (line 2)) (2.6.1)\n",
100
+ "Requirement already satisfied: pycparser in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from cffi>=1.0->soundfile>=0.12.1->datasets->-r requirements.txt (line 2)) (2.23)\n",
101
+ "Requirement already satisfied: parso<0.9.0,>=0.8.4 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from jedi>=0.16->ipython>=6.1.0->ipywidgets->-r requirements.txt (line 4)) (0.8.5)\n",
102
+ "Requirement already satisfied: llvmlite<0.46,>=0.45.0dev0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from numba>=0.51.0->librosa->datasets->-r requirements.txt (line 2)) (0.45.0)\n",
103
+ "Requirement already satisfied: ptyprocess>=0.5 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from pexpect>4.3->ipython>=6.1.0->ipywidgets->-r requirements.txt (line 4)) (0.7.0)\n",
104
+ "Requirement already satisfied: platformdirs>=2.5.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from pooch>=1.1->librosa->datasets->-r requirements.txt (line 2)) (4.4.0)\n",
105
+ "Requirement already satisfied: wcwidth in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from prompt_toolkit<3.1.0,>=3.0.41->ipython>=6.1.0->ipywidgets->-r requirements.txt (line 4)) (0.2.14)\n",
106
+ "Requirement already satisfied: six>=1.5 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from python-dateutil>=2.8.2->pandas->datasets->-r requirements.txt (line 2)) (1.17.0)\n",
107
+ "Requirement already satisfied: threadpoolctl>=3.1.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from scikit-learn>=1.1.0->librosa->datasets->-r requirements.txt (line 2)) (3.6.0)\n",
108
+ "Requirement already satisfied: asttokens>=2.1.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from stack_data->ipython>=6.1.0->ipywidgets->-r requirements.txt (line 4)) (3.0.1)\n",
109
+ "Requirement already satisfied: pure-eval in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from stack_data->ipython>=6.1.0->ipywidgets->-r requirements.txt (line 4)) (0.2.3)\n",
110
+ "Requirement already satisfied: executing>=1.2.0 in /Users/cminixho/.pyenv/versions/3.10.13/lib/python3.10/site-packages (from stack_data->ipython>=6.1.0->ipywidgets->-r requirements.txt (line 4)) (2.2.1)\n",
111
+ "\n",
112
+ "\u001b[1m[\u001b[0m\u001b[34;49mnotice\u001b[0m\u001b[1;39;49m]\u001b[0m\u001b[39;49m A new release of pip is available: \u001b[0m\u001b[31;49m23.0.1\u001b[0m\u001b[39;49m -> \u001b[0m\u001b[32;49m25.3\u001b[0m\n",
113
+ "\u001b[1m[\u001b[0m\u001b[34;49mnotice\u001b[0m\u001b[1;39;49m]\u001b[0m\u001b[39;49m To update, run: \u001b[0m\u001b[32;49m/Users/cminixho/.pyenv/versions/3.10.13/bin/python -m pip install --upgrade pip\u001b[0m\n",
114
+ "Note: you may need to restart the kernel to use updated packages.\n"
115
+ ]
116
+ }
117
+ ],
118
+ "source": [
119
+ "%pip install -r requirements.txt"
120
+ ]
121
+ },
122
+ {
123
+ "cell_type": "code",
124
+ "execution_count": 2,
125
+ "id": "62334e1e",
126
+ "metadata": {},
127
+ "outputs": [],
128
+ "source": [
129
+ "from datasets import load_dataset, concatenate_datasets\n",
130
+ "\n",
131
+ "ds = load_dataset(\"edinburghcstr/edacc\")"
132
+ ]
133
+ },
134
+ {
135
+ "cell_type": "code",
136
+ "execution_count": 3,
137
+ "id": "695479b7",
138
+ "metadata": {},
139
+ "outputs": [],
140
+ "source": [
141
+ "ds_all = concatenate_datasets([ds[\"validation\"], ds[\"test\"]])"
142
+ ]
143
+ },
144
+ {
145
+ "cell_type": "code",
146
+ "execution_count": 35,
147
+ "id": "afd697e1",
148
+ "metadata": {},
149
+ "outputs": [
150
+ {
151
+ "data": {
152
+ "application/vnd.jupyter.widget-view+json": {
153
+ "model_id": "531f921a89bc45edb1b57e0ffcd6cb25",
154
+ "version_major": 2,
155
+ "version_minor": 0
156
+ },
157
+ "text/plain": [
158
+ " 0%| | 0/19137 [00:00<?, ?it/s]"
159
+ ]
160
+ },
161
+ "metadata": {},
162
+ "output_type": "display_data"
163
+ }
164
+ ],
165
+ "source": [
166
+ "import torch\n",
167
+ "from tqdm.auto import tqdm\n",
168
+ "from transformers import AutoModelForCausalLM, AutoTokenizer\n",
169
+ "from pathlib import Path\n",
170
+ "import torchaudio\n",
171
+ "import json\n",
172
+ "\n",
173
+ "checkpoint = \"HuggingFaceTB/SmolLM2-135M\"\n",
174
+ "\n",
175
+ "tokenizer = AutoTokenizer.from_pretrained(checkpoint)\n",
176
+ "# Use bfloat16 for efficiency\n",
177
+ "model = AutoModelForCausalLM.from_pretrained(checkpoint)\n",
178
+ "\n",
179
+ "# Get the token ID for \"?\"\n",
180
+ "question_mark_token = tokenizer.encode(\"?\", add_special_tokens=False)[0]\n",
181
+ "\n",
182
+ "questions = []\n",
183
+ "\n",
184
+ "Path(\"questions\").mkdir(exist_ok=True)\n",
185
+ "\n",
186
+ "for i in tqdm(range(len(ds_all))):\n",
187
+ " # Use lowercase text as input\n",
188
+ " text = ds_all[i][\"text\"].lower()\n",
189
+ " inputs = tokenizer.encode(text, return_tensors=\"pt\").to(model.device)\n",
190
+ "\n",
191
+ " # Get model predictions for next token\n",
192
+ " with torch.no_grad():\n",
193
+ " outputs = model(inputs)\n",
194
+ " # Get logits for the last position (next token prediction)\n",
195
+ " next_token_logits = outputs.logits[0, -1, :]\n",
196
+ " # Get the most likely next token\n",
197
+ " predicted_token_id = torch.argmax(next_token_logits).item()\n",
198
+ "\n",
199
+ " # Check if predicted next token is a question mark\n",
200
+ " if predicted_token_id == question_mark_token:\n",
201
+ " torchaudio.save(\n",
202
+ " f\"questions/{i}.wav\",\n",
203
+ " torch.from_numpy(ds_all[i][\"audio\"][\"array\"]).unsqueeze(0).float(),\n",
204
+ " ds_all[i][\"audio\"][\"sampling_rate\"],\n",
205
+ " )\n",
206
+ " with open(f\"questions/{i}.json\", \"w\") as f:\n",
207
+ " json_out = ds_all[i].copy()\n",
208
+ " del json_out[\"audio\"]\n",
209
+ " json.dump(json_out, f)"
210
+ ]
211
+ },
212
+ {
213
+ "cell_type": "code",
214
+ "execution_count": 7,
215
+ "id": "5999fd61",
216
+ "metadata": {},
217
+ "outputs": [],
218
+ "source": [
219
+ "from dotenv import load_dotenv\n",
220
+ "from pathlib import Path\n",
221
+ "import json\n",
222
+ "\n",
223
+ "load_dotenv()\n",
224
+ "\n",
225
+ "import replicate"
226
+ ]
227
+ },
228
+ {
229
+ "cell_type": "code",
230
+ "execution_count": null,
231
+ "id": "42d1c793",
232
+ "metadata": {},
233
+ "outputs": [],
234
+ "source": [
235
+ "accents_gender = {}\n",
236
+ "for item in sorted(Path(\"questions\").glob(\"*.json\")):\n",
237
+ " with open(item, \"r\") as f:\n",
238
+ " data = json.load(f)\n",
239
+ " if len(data[\"text\"]) > 50:\n",
240
+ " if data[\"accent\"] == \"Mainstream US English\" and data[\"l1\"] != \"Mainstream US English\":\n",
241
+ " continue\n",
242
+ " if data[\"speaker\"] == \"EDACC-C19-A\":\n",
243
+ " continue\n",
244
+ " accent_gender_key = f\"{data['accent']}-{data['gender']}\"\n",
245
+ " if accent_gender_key not in accents_gender:\n",
246
+ " accents_gender[accent_gender_key] = item.stem\n",
247
+ " else:\n",
248
+ " pass"
249
+ ]
250
+ },
251
+ {
252
+ "cell_type": "code",
253
+ "execution_count": 17,
254
+ "id": "80ed0d46",
255
+ "metadata": {},
256
+ "outputs": [
257
+ {
258
+ "name": "stdout",
259
+ "output_type": "stream",
260
+ "text": [
261
+ "German-female\n",
262
+ "Bulgarian-male\n",
263
+ "Romanian-female\n",
264
+ "Lithuanian-male\n",
265
+ "Ghanain English-female\n",
266
+ "Japanese-female\n",
267
+ "Indonesian-female\n",
268
+ "Egyptian-male\n",
269
+ "Latin-female\n",
270
+ "Brazilian-male\n"
271
+ ]
272
+ }
273
+ ],
274
+ "source": [
275
+ "accent_gender_both = {}\n",
276
+ "for k in accents_gender:\n",
277
+ " if \"-male\" in k:\n",
278
+ " if k.replace(\"-male\", \"-female\") in accents_gender:\n",
279
+ " accent_gender_both[k] = accents_gender[k]\n",
280
+ " else:\n",
281
+ " print(k)\n",
282
+ " elif \"-female\" in k:\n",
283
+ " if k.replace(\"-female\", \"-male\") in accents_gender:\n",
284
+ " accent_gender_both[k] = accents_gender[k]\n",
285
+ " else:\n",
286
+ " print(k)\n",
287
+ "\n"
288
+ ]
289
+ },
290
+ {
291
+ "cell_type": "code",
292
+ "execution_count": 18,
293
+ "id": "3df3e24f",
294
+ "metadata": {},
295
+ "outputs": [
296
+ {
297
+ "name": "stdout",
298
+ "output_type": "stream",
299
+ "text": [
300
+ "{'Irish English', 'Eastern European', 'Catalan', 'Indonesian English', 'Vietnamese', 'French', 'Southern British English', 'Mainstream US English', 'South African English', 'Jamaican English', 'Spanish', 'Latin American', 'Chinese', 'Kenyan English', \"Don't know\", 'European', 'Italian', 'Indian English', 'Nigerian English', 'Scottish English'}\n"
301
+ ]
302
+ }
303
+ ],
304
+ "source": [
305
+ "print(set([x.split(\"-\")[0] for x in accent_gender_both.keys()]))"
306
+ ]
307
+ },
308
+ {
309
+ "cell_type": "code",
310
+ "execution_count": 19,
311
+ "id": "f09bc7e8",
312
+ "metadata": {},
313
+ "outputs": [],
314
+ "source": [
315
+ "keys_to_keep = [\n",
316
+ " \"Mainstream US\",\n",
317
+ " \"Southern British\",\n",
318
+ " \"Indian\",\n",
319
+ " \"Chinese\",\n",
320
+ " \"Latin American\",\n",
321
+ " \"Eastern European\",\n",
322
+ "]\n",
323
+ "\n",
324
+ "final_dict = {}\n",
325
+ "for k in accent_gender_both:\n",
326
+ " if any(x in k for x in keys_to_keep):\n",
327
+ " final_dict[k] = accent_gender_both[k]\n",
328
+ " else:\n",
329
+ " pass\n",
330
+ "\n",
331
+ "\n",
332
+ "\n"
333
+ ]
334
+ },
335
+ {
336
+ "cell_type": "code",
337
+ "execution_count": 20,
338
+ "id": "13247f3b",
339
+ "metadata": {},
340
+ "outputs": [
341
+ {
342
+ "data": {
343
+ "text/plain": [
344
+ "{'Eastern European-male': '10153',\n",
345
+ " 'Southern British English-female': '1028',\n",
346
+ " 'Mainstream US English-female': '10293',\n",
347
+ " 'Chinese-male': '10314',\n",
348
+ " 'Indian English-female': '10892',\n",
349
+ " 'Latin American-female': '11303',\n",
350
+ " 'Latin American-male': '11310',\n",
351
+ " 'Indian English-male': '13170',\n",
352
+ " 'Southern British English-male': '16172',\n",
353
+ " 'Chinese-female': '16385',\n",
354
+ " 'Mainstream US English-male': '18708',\n",
355
+ " 'Eastern European-female': '4252'}"
356
+ ]
357
+ },
358
+ "execution_count": 20,
359
+ "metadata": {},
360
+ "output_type": "execute_result"
361
+ }
362
+ ],
363
+ "source": [
364
+ "final_dict"
365
+ ]
366
+ },
367
+ {
368
+ "cell_type": "code",
369
+ "execution_count": null,
370
+ "id": "b33d29a0",
371
+ "metadata": {},
372
+ "outputs": [
373
+ {
374
+ "name": "stdout",
375
+ "output_type": "stream",
376
+ "text": [
377
+ "Loaded as API: https://cdminix-megatts3-voice-cloning.hf.space ✔\n",
378
+ "Total prompts: 80 (40 with hesitation, 40 without)\n",
379
+ "Total files to generate: 960\n"
380
+ ]
381
+ },
382
+ {
383
+ "data": {
384
+ "application/vnd.jupyter.widget-view+json": {
385
+ "model_id": "e7073169b40f4bb2a947d00dc962fe04",
386
+ "version_major": 2,
387
+ "version_minor": 0
388
+ },
389
+ "text/plain": [
390
+ "Generating audio: 0%| | 0/960 [00:00<?, ?it/s]"
391
+ ]
392
+ },
393
+ "metadata": {},
394
+ "output_type": "display_data"
395
+ }
396
+ ],
397
+ "source": [
398
+ "import pandas as pd\n",
399
+ "from gradio_client import Client, handle_file\n",
400
+ "import shutil\n",
401
+ "from tqdm.auto import tqdm\n",
402
+ "import time\n",
403
+ "\n",
404
+ "# Initialize MegaTTS3 client\n",
405
+ "client = Client(\"cdminix/MegaTTS3-Voice-Cloning\")\n",
406
+ "\n",
407
+ "# Retry helper function with exponential backoff\n",
408
+ "def generate_with_retry(client, speaker_audio_path, text, max_retries=5, initial_delay=1):\n",
409
+ " \"\"\"Generate speech with retry logic and exponential backoff.\"\"\"\n",
410
+ " for attempt in range(max_retries):\n",
411
+ " try:\n",
412
+ " result = client.predict(\n",
413
+ " inp_audio=handle_file(speaker_audio_path),\n",
414
+ " inp_text=text,\n",
415
+ " infer_timestep=32,\n",
416
+ " p_w=1.4,\n",
417
+ " t_w=3,\n",
418
+ " api_name=\"/generate_speech\"\n",
419
+ " )\n",
420
+ " return result\n",
421
+ " except Exception as e:\n",
422
+ " if attempt == max_retries - 1:\n",
423
+ " # Last attempt failed, raise the exception\n",
424
+ " raise\n",
425
+ " # Calculate backoff delay: 1s, 2s, 4s, 8s, 16s\n",
426
+ " delay = initial_delay * (2 ** attempt)\n",
427
+ " print(f\"\\nRetry {attempt + 1}/{max_retries} after error: {str(e)[:100]}\")\n",
428
+ " print(f\"Waiting {delay}s before retrying...\")\n",
429
+ " time.sleep(delay)\n",
430
+ "\n",
431
+ "df = pd.read_csv(\"tts_prompts.csv\")\n",
432
+ "df_hesitation = df[df[\"has_disfluency\"] == True]\n",
433
+ "df_no_hesitation = df[df[\"has_disfluency\"] == False]\n",
434
+ "\n",
435
+ "print(f\"Total prompts: {len(df)} ({len(df_hesitation)} with hesitation, {len(df_no_hesitation)} without)\")\n",
436
+ "\n",
437
+ "result_rows = []\n",
438
+ "\n",
439
+ "# Calculate total number of files to generate\n",
440
+ "total_files = len(final_dict.keys()) * (len(df_hesitation) + len(df_no_hesitation))\n",
441
+ "print(f\"Total files to generate: {total_files}\")\n",
442
+ "\n",
443
+ "pbar = tqdm(total=total_files, desc=\"Generating audio\")\n",
444
+ "\n",
445
+ "for k in sorted(final_dict.keys()):\n",
446
+ " accent = k.split(\"-\")[0]\n",
447
+ " accent = accent.replace(\" \", \"_\").lower()\n",
448
+ " gender = k.split(\"-\")[1]\n",
449
+ " Path(f\"dataset/{accent}/{gender}/hesitation\").mkdir(parents=True, exist_ok=True)\n",
450
+ " \n",
451
+ " speaker_audio_path = \"questions/\" + final_dict[k] + \".wav\"\n",
452
+ "\n",
453
+ " for idx, row in df_hesitation.iterrows():\n",
454
+ " _id = f\"{idx}-{accent}-{gender}-{final_dict[k]}-hesitation\"\n",
455
+ " output_path = Path(f\"dataset/{accent}/{gender}/hesitation/{_id}.wav\")\n",
456
+ " if output_path.exists():\n",
457
+ " result_rows.append(\n",
458
+ " {\n",
459
+ " \"id\": _id,\n",
460
+ " \"accent\": accent,\n",
461
+ " \"gender\": gender,\n",
462
+ " \"speaker_audio\": speaker_audio_path,\n",
463
+ " }\n",
464
+ " )\n",
465
+ " pbar.update(1)\n",
466
+ " continue\n",
467
+ " \n",
468
+ " # Generate with retry logic\n",
469
+ " result = generate_with_retry(client, speaker_audio_path, row[\"text\"])\n",
470
+ " \n",
471
+ " # Copy the result file to output path\n",
472
+ " shutil.copy(result, output_path)\n",
473
+ " \n",
474
+ " result_rows.append(\n",
475
+ " {\n",
476
+ " \"id\": _id,\n",
477
+ " \"accent\": accent,\n",
478
+ " \"gender\": gender,\n",
479
+ " \"speaker_audio\": speaker_audio_path,\n",
480
+ " }\n",
481
+ " )\n",
482
+ " pbar.update(1)\n",
483
+ "\n",
484
+ " Path(f\"dataset/{accent}/{gender}/no_hesitation\").mkdir(parents=True, exist_ok=True)\n",
485
+ "\n",
486
+ " for idx, row in df_no_hesitation.iterrows():\n",
487
+ " _id = f\"{idx}-{accent}-{gender}-{final_dict[k]}-no_hesitation\"\n",
488
+ " output_path = Path(f\"dataset/{accent}/{gender}/no_hesitation/{_id}.wav\")\n",
489
+ " if output_path.exists():\n",
490
+ " result_rows.append(\n",
491
+ " {\n",
492
+ " \"id\": _id,\n",
493
+ " \"accent\": accent,\n",
494
+ " \"gender\": gender,\n",
495
+ " \"speaker_audio\": speaker_audio_path,\n",
496
+ " }\n",
497
+ " )\n",
498
+ " pbar.update(1)\n",
499
+ " continue\n",
500
+ " \n",
501
+ " # Generate with retry logic\n",
502
+ " result = generate_with_retry(client, speaker_audio_path, row[\"text\"])\n",
503
+ " \n",
504
+ " # Copy the result file to output path\n",
505
+ " shutil.copy(result, output_path)\n",
506
+ " \n",
507
+ " result_rows.append(\n",
508
+ " {\n",
509
+ " \"id\": _id,\n",
510
+ " \"accent\": accent,\n",
511
+ " \"gender\": gender,\n",
512
+ " \"speaker_audio\": speaker_audio_path,\n",
513
+ " }\n",
514
+ " )\n",
515
+ " pbar.update(1)\n",
516
+ "\n",
517
+ "pbar.close()\n",
518
+ "\n",
519
+ "df_result = pd.DataFrame(result_rows)\n",
520
+ "df_result.to_csv(\"tts_prompts_result.csv\", index=False)\n",
521
+ "print(f\"Completed! Generated {len(result_rows)} audio files.\")"
522
+ ]
523
+ },
524
+ {
525
+ "cell_type": "code",
526
+ "execution_count": null,
527
+ "id": "6c97f22b",
528
+ "metadata": {},
529
+ "outputs": [],
530
+ "source": []
531
+ }
532
+ ],
533
+ "metadata": {
534
+ "kernelspec": {
535
+ "display_name": "Python 3",
536
+ "language": "python",
537
+ "name": "python3"
538
+ },
539
+ "language_info": {
540
+ "codemirror_mode": {
541
+ "name": "ipython",
542
+ "version": 3
543
+ },
544
+ "file_extension": ".py",
545
+ "mimetype": "text/x-python",
546
+ "name": "python",
547
+ "nbconvert_exporter": "python",
548
+ "pygments_lexer": "ipython3",
549
+ "version": "3.10.13"
550
+ }
551
+ },
552
+ "nbformat": 4,
553
+ "nbformat_minor": 5
554
+ }
dataset/.DS_Store ADDED
Binary file (6.15 kB). View file
 
dataset/chinese/.DS_Store ADDED
Binary file (6.15 kB). View file
 
dataset/chinese/female/hesitation/40-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58269a6da4ae8ac5e6bbcec1b94bfa56f53b3212c00fa14e1e210ba0300627d9
3
+ size 1152044
dataset/chinese/female/hesitation/41-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0503571dce7d0c2b6ed08384882ed2f2c8bcb30882f8681657f7369f2410b94c
3
+ size 1186604
dataset/chinese/female/hesitation/42-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac710badfe973a0116be5a0e6c960c4c4feb984f56b34396bfbeaa9af3905713
3
+ size 963884
dataset/chinese/female/hesitation/43-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6026bb5e60a79a9a5486fac701790cda6d3036b66252f4876c0c435213f6c69a
3
+ size 894764
dataset/chinese/female/hesitation/44-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5566c15cb607e36c3bd01bb528b43c741205eba694f65f500b644f331624c043
3
+ size 887084
dataset/chinese/female/hesitation/45-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4478d98ef7b19febc490327f301eedcb76dd6e8450e16265c2506ab95bdab4fb
3
+ size 1278764
dataset/chinese/female/hesitation/46-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15cfcc3f0ed8907e641ab87356ce0b657ec1c85d3423ec77de9d9ba965c8d4f0
3
+ size 1052204
dataset/chinese/female/hesitation/47-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cff6e98074cf9fa4da0f762fe8c40e9810a6ba2e1bc7c8596421864557b8871
3
+ size 1217324
dataset/chinese/female/hesitation/48-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9339dce6439fd436b31e1691fa59ec497a8d9dc04a87288f13d42bbb2accc144
3
+ size 1159724
dataset/chinese/female/hesitation/49-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe5ae510d4940bb5a19bd4dfb77c98b8e60185893cd6629cee2a5007eec983de
3
+ size 1098284
dataset/chinese/female/hesitation/50-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f93ad87b67c94732800620551ae5683714f154a0746df9ecd9ed2a5c9ab74675
3
+ size 986924
dataset/chinese/female/hesitation/51-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bac31997c86883a2a8eff859217067f5fe38cd8a3911fe448b71762164981c0
3
+ size 841004
dataset/chinese/female/hesitation/52-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d44bf302ab80c325b72165514132f3de4fda12d00010e1e1d8c3e206c3751b76
3
+ size 883244
dataset/chinese/female/hesitation/53-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12777ea41afd71b804c2b0aacc552e8ff560421e7f1ff734d33f69a27647fcea
3
+ size 725804
dataset/chinese/female/hesitation/54-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6721029e11e2f0d2701bf18f6061c6e1f92429d67bca18a830476b923eec5141
3
+ size 933164
dataset/chinese/female/hesitation/55-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fef2eb93f20932bf7210c5449c0ca6ae602876c902fcba51a901275b1ac4a093
3
+ size 1194284
dataset/chinese/female/hesitation/56-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c57b4a7555b68dd726ab04cfce7e1b6f9dde9306290de78864ecb165d65d520d
3
+ size 1059884
dataset/chinese/female/hesitation/57-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4f7c965644da4a525582c014d3fec0095db1d2857638f2c23aef7710161c028
3
+ size 1075244
dataset/chinese/female/hesitation/58-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:695a87be77dfbc37c5ad40f06e267fe24e64119be46ba79e91cab693001b0608
3
+ size 1098284
dataset/chinese/female/hesitation/59-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8289481b6dc9dfadf7b009bb19fb1e82ca905ee07658772d0e02c1d7f7b94599
3
+ size 1098284
dataset/chinese/female/hesitation/60-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3126998789bfb5d92a884ec5cbc69b9a65073b26b3f7b29c423b842cb34e7b7a
3
+ size 975404
dataset/chinese/female/hesitation/61-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c93f45f057b6bc3a3917b0c2326259f0b084ad48ad7509e6b45e750356261686
3
+ size 879404
dataset/chinese/female/hesitation/62-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1b4c958e342ddaa609bed013695d5acbebae0279e295fc5c6c01f2a04392491
3
+ size 929324
dataset/chinese/female/hesitation/63-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eacf7169dfeb7b6c35423304e3afea5ff07f963108d5cf80d5ad12dcac64b44d
3
+ size 1140524
dataset/chinese/female/hesitation/64-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d386f93c3506b1b52d45e4a3cf8bbd247f11efd8c98ba2b3278cd2611a39052
3
+ size 967724
dataset/chinese/female/hesitation/65-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc9306fbd97dd819909b5cd18f8d06deee1bbee4b85d72aa0363a12512ea8b25
3
+ size 967724
dataset/chinese/female/hesitation/66-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f98ad81c644132957c56ead5fd601d2b247920089d1417ef68e10941c735752d
3
+ size 860204
dataset/chinese/female/hesitation/67-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:021b516a8c82bbd95b3713bc0e4f7ff1dde9909d9740883f767332959c809d47
3
+ size 1017644
dataset/chinese/female/hesitation/68-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dd4dd75eb029d05851853e6eda9fe2f5674cf661c215158aaecd5fb70067cef
3
+ size 986924
dataset/chinese/female/hesitation/69-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66cba5ac92867f2791b6ec315339f61d542df7d60c30ec1f8b6e6fcc51d5d445
3
+ size 990764
dataset/chinese/female/hesitation/70-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd4fb48dbedea5418d854b3c669bfb61d00106617624f952650d9b1dce99fc7d
3
+ size 967724
dataset/chinese/female/hesitation/71-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc7f5c08dc6649f10a7890505c1350e2af0273aa2630f74c6fe393d2c048511c
3
+ size 952364
dataset/chinese/female/hesitation/72-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:317f44df5fbd2a29b06a1a61048cce798d40ecbceed9bafc7b5494022ea38343
3
+ size 1198124
dataset/chinese/female/hesitation/73-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1cbb3cd68848053928db1b177e88e5bb156aff5c6b2872cc3d0891b77c1ad386
3
+ size 1136684
dataset/chinese/female/hesitation/74-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3128c10a41faebe77dcf2e20c7aa8f5ab2e458d12440287e32d95b36114c0313
3
+ size 1090604
dataset/chinese/female/hesitation/75-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ae20a3b90a7be5db41f562252a8e54c0ba8444ea5e51b95948f241c02a89b14
3
+ size 979244
dataset/chinese/female/hesitation/76-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7535824a21ef4015ccc02861d93fe3985ade4a56af4987d0c0ce10157c65271
3
+ size 967724
dataset/chinese/female/hesitation/77-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:240afc2f2916fe59937c1425f1556d0fe875f2924f9aebfc0ca454851408ebf7
3
+ size 1082924
dataset/chinese/female/hesitation/78-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:141a5f4f6e2be384d78690102352ecddcc9f04ea14552d75712b439a6628de31
3
+ size 994604
dataset/chinese/female/hesitation/79-chinese-female-16385-hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:688e21d1b86beb3c5e1d09a81e60518d55f741c0ab79548b9362ea28952022ab
3
+ size 975404
dataset/chinese/female/no_hesitation/0-chinese-female-16385-no_hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a812bbdc9cf27e14b99a3c5348281c1d20fa0d38658475bd65055bc91786f57
3
+ size 844844
dataset/chinese/female/no_hesitation/1-chinese-female-16385-no_hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bf091fc7c90a7ebf74cc1fe0405835fc1b2aab62cea9a03d1ec2dad5db74b9b
3
+ size 902444
dataset/chinese/female/no_hesitation/10-chinese-female-16385-no_hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:473fe3a12f03817bb4efe1489e6372c1697cc43a360ba908c9efee1efd4bbe59
3
+ size 890924
dataset/chinese/female/no_hesitation/11-chinese-female-16385-no_hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c2ebed5d3b91a97cfdac0764da3e72d397caf587968dd9ab275286735de8a5a
3
+ size 794924
dataset/chinese/female/no_hesitation/12-chinese-female-16385-no_hesitation.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52b87590d1d081c803deacda0e8d8267527f2326f70f64fe72dccdb29929bcf3
3
+ size 775724