ayf3 commited on
Commit
6f3ccd4
·
verified ·
1 Parent(s): 83604c9

Upload scripts/rvc_train_colab.ipynb with huggingface_hub

Browse files
Files changed (1) hide show
  1. scripts/rvc_train_colab.ipynb +241 -281
scripts/rvc_train_colab.ipynb CHANGED
@@ -1,284 +1,244 @@
1
  {
2
- "nbformat": 4,
3
- "nbformat_minor": 0,
4
- "metadata": {
5
- "colab": {
6
- "provenance": [],
7
- "gpuType": "T4"
 
 
 
 
 
 
 
 
8
  },
9
- "kernelspec": {
10
- "name": "python3",
11
- "display_name": "Python 3"
12
- },
13
- "accelerator": "GPU"
14
- },
15
- "cells": [
16
- {
17
- "cell_type": "markdown",
18
- "metadata": {},
19
- "source": [
20
- "# NumberBlocks One - RVC Voice Cloning Training\n",
21
- "\n",
22
- "一键训练脚本:自动下载 2607 个训练片段,训练 RVC 模型,上传模型到 HuggingFace。\n",
23
- "\n",
24
- "**使用方法**: Runtime → Change runtime type → T4 GPU → Run All\n",
25
- "\n",
26
- "预计训练时间:~30-60分钟 (T4 GPU, 10000 steps)"
27
- ]
28
- },
29
- {
30
- "cell_type": "code",
31
- "execution_count": null,
32
- "metadata": {},
33
- "outputs": [],
34
- "source": [
35
- "#@title 1. 环境检查\n",
36
- "import subprocess\n",
37
- "result = subprocess.run(['nvidia-smi'], capture_output=True, text=True)\n",
38
- "if result.returncode != 0:\n",
39
- " print('⚠️ No GPU detected! Go to Runtime → Change runtime type → T4 GPU')\n",
40
- "else:\n",
41
- " print('✅ GPU detected:')\n",
42
- " print(result.stdout[:500])"
43
- ]
44
- },
45
- {
46
- "cell_type": "code",
47
- "execution_count": null,
48
- "metadata": {},
49
- "outputs": [],
50
- "source": [
51
- "#@title 2. 安装 RVC 和依赖 (~5 min)\n",
52
- "%cd /content\n",
53
- "\n",
54
- "# Install RVC\n",
55
- "!git clone https://github.com/RVC-Project/Retrieval-based-Voice-Conversion.git RVC\n",
56
- "%cd RVC\n",
57
- "!pip install -r requirements.txt -q\n",
58
- "!pip install praat-parselmouth -q\n",
59
- "\n",
60
- "# Install ffmpeg\n",
61
- "!apt-get install -y ffmpeg -qq\n",
62
- "\n",
63
- "print('✅ RVC installed')"
64
- ]
65
- },
66
- {
67
- "cell_type": "code",
68
- "execution_count": null,
69
- "metadata": {},
70
- "outputs": [],
71
- "source": [
72
- "#@title 3. 下载训练数据 (2607 segments from HuggingFace)\n",
73
- "import os\n",
74
- "from huggingface_hub import HfApi, hf_hub_download\n",
75
- "\n",
76
- "DATASET_REPO = 'ayf3/numberblocks-one-voice-dataset'\n",
77
- "WORK_DIR = '/content/RVC/dataset/numberblocks_one'\n",
78
- "os.makedirs(WORK_DIR, exist_ok=True)\n",
79
- "\n",
80
- "api = HfApi()\n",
81
- "files = api.list_repo_files(DATASET_REPO, repo_type='dataset')\n",
82
- "segments = [f for f in files if f.startswith('segments/') and f.endswith('.wav')]\n",
83
- "\n",
84
- "print(f'Downloading {len(segments)} training segments...')\n",
85
- "\n",
86
- "# Download in parallel using huggingface_hub\n",
87
- "from concurrent.futures import ThreadPoolExecutor, as_completed\n",
88
- "\n",
89
- "def download_seg(seg_path):\n",
90
- " local = hf_hub_download(DATASET_REPO, seg_path, repo_type='dataset',\n",
91
- " local_dir='/content/rvc_data')\n",
92
- " return local\n",
93
- "\n",
94
- "# Download all segments to a temp dir, then we'll point RVC to it\n",
95
- "import shutil\n",
96
- "TEMP_DIR = '/content/rvc_data/segments'\n",
97
- "os.makedirs(TEMP_DIR, exist_ok=True)\n",
98
- "\n",
99
- "downloaded = 0\n",
100
- "with ThreadPoolExecutor(max_workers=8) as executor:\n",
101
- " futures = {executor.submit(download_seg, s): s for s in segments}\n",
102
- " for future in as_completed(futures):\n",
103
- " downloaded += 1\n",
104
- " if downloaded % 200 == 0:\n",
105
- " print(f' Downloaded {downloaded}/{len(segments)}...')\n",
106
- "\n",
107
- "print(f'✅ Downloaded {downloaded} segments')\n",
108
- "\n",
109
- "# Copy to RVC dataset dir with clean names\n",
110
- "for f in os.listdir(TEMP_DIR):\n",
111
- " if f.endswith('.wav'):\n",
112
- " src = os.path.join(TEMP_DIR, f)\n",
113
- " dst = os.path.join(WORK_DIR, f)\n",
114
- " if not os.path.exists(dst):\n",
115
- " shutil.copy2(src, dst)\n",
116
- "\n",
117
- "segment_files = [f for f in os.listdir(WORK_DIR) if f.endswith('.wav')]\n",
118
- "print(f'✅ {len(segment_files)} segments ready in {WORK_DIR}')\n",
119
- "\n",
120
- "# Calculate total audio duration\n",
121
- "total_dur = 0\n",
122
- "for f in segment_files[:10]: # Sample to estimate\n",
123
- " result = subprocess.run(\n",
124
- " ['ffprobe', '-v', 'quiet', '-show_entries', 'format=duration',\n",
125
- " '-of', 'csv=p=0', os.path.join(WORK_DIR, f)],\n",
126
- " capture_output=True, text=True\n",
127
- " )\n",
128
- " if result.stdout.strip():\n",
129
- " total_dur += float(result.stdout.strip())\n",
130
- "avg_dur = total_dur / min(10, len(segment_files))\n",
131
- "est_total = avg_dur * len(segment_files) / 60\n",
132
- "print(f'📊 Estimated total audio: {est_total:.1f} minutes')"
133
- ]
134
- },
135
- {
136
- "cell_type": "code",
137
- "execution_count": null,
138
- "metadata": {},
139
- "outputs": [],
140
- "source": [
141
- "#@title 4. 特征提取 (~10 min)\n",
142
- "%cd /content/RVC\n",
143
- "\n",
144
- "MODEL_NAME = 'numberblocks_one'\n",
145
- "DATASET_PATH = '/content/RVC/dataset/numberblocks_one'\n",
146
- "EXP_DIR = f'/content/RVC/logs/{MODEL_NAME}'\n",
147
- "os.makedirs(EXP_DIR, exist_ok=True)\n",
148
- "\n",
149
- "# Run feature extraction\n",
150
- "!python infer/modules/train/extract/extract_f0_print.py \\\n",
151
- " {EXP_DIR} \\\n",
152
- " {DATASET_PATH} \\\n",
153
- " harvest \\\n",
154
- " 0 \\\n",
155
- " 8\n",
156
- "\n",
157
- "print('F0 extraction done')\n",
158
- "\n",
159
- "# Extract features\n",
160
- "!python infer/modules/train/extract/extract_feature_print.py \\\n",
161
- " {EXP_DIR} \\\n",
162
- " v2 \\\n",
163
- " 0 \\\n",
164
- " 8\n",
165
- "\n",
166
- "print('✅ Feature extraction complete')"
167
- ]
168
- },
169
- {
170
- "cell_type": "code",
171
- "execution_count": null,
172
- "metadata": {},
173
- "outputs": [],
174
- "source": [
175
- "#@title 5. 训练 RVC 模型 (~20-40 min)\n",
176
- "import os\n",
177
- "\n",
178
- "MODEL_NAME = 'numberblocks_one'\n",
179
- "EXP_DIR = f'/content/RVC/logs/{MODEL_NAME}'\n",
180
- "\n",
181
- "# Check how many features were extracted\n",
182
- "f0_files = [f for f in os.listdir(EXP_DIR) if f.endswith('.f0.npy') or f.endswith('.f0')]\n",
183
- "feat_files = [f for f in os.listdir(EXP_DIR) if 'feature' in f.lower()]\n",
184
- "print(f'F0 files: {len(f0_files)}, Feature files found')\n",
185
- "\n",
186
- "# List all files in EXP_DIR\n",
187
- "all_files = os.listdir(EXP_DIR)\n",
188
- "print(f'Total files in {EXP_DIR}: {len(all_files)}')\n",
189
- "for f in sorted(all_files)[:20]:\n",
190
- " print(f' {f}')\n",
191
- "\n",
192
- "# Train with config\n",
193
- "SAVE_EVERY = 1000 #@param {type:\"number\"}\n",
194
- "TOTAL_STEPS = 10000 #@param {type:\"number\"}\n",
195
- "BATCH_SIZE = 8 #@param {type:\"number\"}\n",
196
- "\n",
197
- "%cd /content/RVC\n",
198
- "\n",
199
- "!python infer/modules/train/train.py \\\n",
200
- " -e {MODEL_NAME} \\\n",
201
- " -s {TOTAL_STEPS} \\\n",
202
- " -b {BATCH_SIZE} \\\n",
203
- " -sr 40k \\\n",
204
- " -v v2 \\\n",
205
- " -g 0 \\\n",
206
- " -se {SAVE_EVERY}\n",
207
- "\n",
208
- "print('✅ Training complete!')"
209
- ]
210
- },
211
- {
212
- "cell_type": "code",
213
- "execution_count": null,
214
- "metadata": {},
215
- "outputs": [],
216
- "source": [
217
- "#@title 6. 上传模型到 HuggingFace\n",
218
- "import os\n",
219
- "import glob\n",
220
- "from huggingface_hub import HfApi\n",
221
- "\n",
222
- "MODEL_NAME = 'numberblocks_one'\n",
223
- "DATASET_REPO = 'ayf3/numberblocks-one-voice-dataset'\n",
224
- "\n",
225
- "# Find the trained model\n",
226
- "model_dir = f'/content/RVC/weights'\n",
227
- "pth_files = glob.glob(f'{model_dir}/**/*{MODEL_NAME}*.pth', recursive=True)\n",
228
- "\n",
229
- "if not pth_files:\n",
230
- " # Check in logs dir\n",
231
- " pth_files = glob.glob(f'/content/RVC/logs/{MODEL_NAME}/*.pth')\n",
232
- " if not pth_files:\n",
233
- " pth_files = glob.glob(f'/content/RVC/logs/{MODEL_NAME}/**/*.pth', recursive=True)\n",
234
- "\n",
235
- "if pth_files:\n",
236
- " print(f'Found model files:')\n",
237
- " for p in pth_files:\n",
238
- " size_mb = os.path.getsize(p) / 1024 / 1024\n",
239
- " print(f' {p} ({size_mb:.1f} MB)')\n",
240
- " \n",
241
- " # Upload the best model (largest .pth)\n",
242
- " best = max(pth_files, key=lambda x: os.path.getsize(x))\n",
243
- " print(f'\\nUploading {best} to HuggingFace...')\n",
244
- " \n",
245
- " api = HfApi()\n",
246
- " api.upload_file(\n",
247
- " path_or_fileobj=best,\n",
248
- " path_in_repo=f'models/{os.path.basename(best)}',\n",
249
- " repo_id=DATASET_REPO,\n",
250
- " repo_type='dataset',\n",
251
- " )\n",
252
- " print(f'✅ Model uploaded to {DATASET_REPO}/models/{os.path.basename(best)}')\n",
253
- " \n",
254
- " # Also upload feature index if exists\n",
255
- " index_files = glob.glob(f'/content/RVC/logs/{MODEL_NAME}/**/*.index', recursive=True)\n",
256
- " for idx_file in index_files:\n",
257
- " api.upload_file(\n",
258
- " path_or_fileobj=idx_file,\n",
259
- " path_in_repo=f'models/{os.path.basename(idx_file)}',\n",
260
- " repo_id=DATASET_REPO,\n",
261
- " repo_type='dataset',\n",
262
- " )\n",
263
- " print(f'✅ Index uploaded: {os.path.basename(idx_file)}')\n",
264
- "else:\n",
265
- " print('❌ No .pth model files found. Training may have failed.')\n",
266
- " # List all files to debug\n",
267
- " for root, dirs, files_in_dir in os.walk(f'/content/RVC/logs/{MODEL_NAME}'):\n",
268
- " for f in files_in_dir:\n",
269
- " print(f' {os.path.join(root, f)}')"
270
- ]
271
- },
272
- {
273
- "cell_type": "markdown",
274
- "metadata": {},
275
- "source": [
276
- "## 🎉 完成!\n",
277
- "\n",
278
- "训练好的模型已上传到 HuggingFace dataset 的 `models/` 目录。\n",
279
- "\n",
280
- "**下一步**: 使用训练好的模型进行 voice conversion 推理。"
281
- ]
282
- }
283
- ]
284
  }
 
1
  {
2
+ "nbformat": 4,
3
+ "nbformat_minor": 0,
4
+ "metadata": {
5
+ "colab": {
6
+ "provenance": [],
7
+ "gpuType": "T4"
8
+ },
9
+ "kernelspec": {
10
+ "name": "python3",
11
+ "display_name": "Python 3"
12
+ },
13
+ "language_info": {
14
+ "name": "python"
15
+ }
16
  },
17
+ "cells": [
18
+ {
19
+ "cell_type": "markdown",
20
+ "metadata": {},
21
+ "source": [
22
+ "# NumberBlocks One - RVC v2 Voice Cloning Training\n",
23
+ "\n",
24
+ "\u4e00\u952e\u8bad\u7ec3\uff1a\u81ea\u52a8\u4e0b\u8f7d\u6570\u636e \u2192 \u7279\u5f81\u63d0\u53d6 \u2192 \u8bad\u7ec3 RVC v2 \u6a21\u578b \u2192 \u4e0a\u4f20\u5230 HuggingFace\u3002\n",
25
+ "\n",
26
+ "**\u4f7f\u7528**: Runtime \u2192 Change runtime type \u2192 **T4 GPU** \u2192 Run All\n",
27
+ "\n",
28
+ "\u9884\u8ba1: ~30-60min (T4, 5000 steps on top500+augmented)\n",
29
+ "\n",
30
+ "\u6570\u636e\u6e90: 500 \u9ad8\u8d28\u91cf segments + 1500 augmented = **2000 \u8bad\u7ec3\u6837\u672c**\n"
31
+ ]
32
+ },
33
+ {
34
+ "cell_type": "code",
35
+ "metadata": {},
36
+ "source": [
37
+ "#@title 1. \u73af\u5883\u68c0\u67e5\n",
38
+ "import subprocess, os\n",
39
+ "result = subprocess.run(['nvidia-smi'], capture_output=True, text=True)\n",
40
+ "if result.returncode != 0:\n",
41
+ " print('\u26a0\ufe0f No GPU! Runtime \u2192 Change runtime type \u2192 T4 GPU')\n",
42
+ " raise RuntimeError('GPU required')\n",
43
+ "else:\n",
44
+ " print('\u2705 GPU detected')\n",
45
+ " lines = result.stdout.split('\\n')\n",
46
+ " print(lines[8] if len(lines) > 8 else result.stdout[:300])\n"
47
+ ],
48
+ "outputs": [],
49
+ "execution_count": null
50
+ },
51
+ {
52
+ "cell_type": "code",
53
+ "metadata": {},
54
+ "source": [
55
+ "#@title 2. \u5b89\u88c5 RVC \u548c\u4f9d\u8d56 (~5 min)\n",
56
+ "import os, subprocess, sys\n",
57
+ "\n",
58
+ "%cd /content\n",
59
+ "\n",
60
+ "if not os.path.exists('/content/RVC'):\n",
61
+ " !git clone --depth 1 https://github.com/RVC-Project/Retrieval-based-Voice-Conversion.git RVC\n",
62
+ "\n",
63
+ "%cd RVC\n",
64
+ "subprocess.run([sys.executable, '-m', 'pip', 'install', '-r', 'requirements.txt', '-q'], check=False)\n",
65
+ "subprocess.run([sys.executable, '-m', 'pip', 'install', 'praat-parselmouth', '-q'], check=False)\n",
66
+ "subprocess.run(['apt-get', 'install', '-y', 'ffmpeg', '-qq'], check=False)\n",
67
+ "\n",
68
+ "print('\u2705 RVC installed')\n"
69
+ ],
70
+ "outputs": [],
71
+ "execution_count": null
72
+ },
73
+ {
74
+ "cell_type": "code",
75
+ "metadata": {},
76
+ "source": [
77
+ "#@title 3. \u4e0b\u8f7d\u8bad\u7ec3\u6570\u636e (top500 + augmented = ~2000 files)\n",
78
+ "import os, shutil, concurrent.futures\n",
79
+ "from huggingface_hub import HfApi, hf_hub_download\n",
80
+ "\n",
81
+ "DATASET_REPO = 'ayf3/numberblocks-one-voice-dataset'\n",
82
+ "WORK_DIR = '/content/RVC/dataset/numberblocks_one'\n",
83
+ "os.makedirs(WORK_DIR, exist_ok=True)\n",
84
+ "\n",
85
+ "api = HfApi()\n",
86
+ "all_files = list(api.list_repo_files(DATASET_REPO, repo_type='dataset'))\n",
87
+ "\n",
88
+ "train_files = [f for f in all_files\n",
89
+ " if (f.startswith('data/train_top500/') or f.startswith('data/train_augmented/'))\n",
90
+ " and f.endswith('.wav')]\n",
91
+ "\n",
92
+ "print(f'Total training files to download: {len(train_files)}')\n",
93
+ "\n",
94
+ "def download_one(f):\n",
95
+ " try:\n",
96
+ " local = hf_hub_download(DATASET_REPO, f, repo_type='dataset')\n",
97
+ " basename = os.path.basename(f)\n",
98
+ " shutil.copy2(local, os.path.join(WORK_DIR, basename))\n",
99
+ " return True\n",
100
+ " except:\n",
101
+ " return False\n",
102
+ "\n",
103
+ "with concurrent.futures.ThreadPoolExecutor(max_workers=8) as executor:\n",
104
+ " results = list(executor.map(download_one, train_files))\n",
105
+ "\n",
106
+ "success = sum(1 for r in results if r)\n",
107
+ "existing = [f for f in os.listdir(WORK_DIR) if f.endswith('.wav')]\n",
108
+ "print(f'\u2705 Downloaded: {success}/{len(train_files)}, Files in dataset dir: {len(existing)}')\n"
109
+ ],
110
+ "outputs": [],
111
+ "execution_count": null
112
+ },
113
+ {
114
+ "cell_type": "code",
115
+ "metadata": {},
116
+ "source": [
117
+ "#@title 4. \u7279\u5f81\u63d0\u53d6 (~10-15 min)\n",
118
+ "import os, subprocess, sys\n",
119
+ "\n",
120
+ "%cd /content/RVC\n",
121
+ "\n",
122
+ "MODEL_NAME = 'numberblocks_one'\n",
123
+ "DATASET_PATH = '/content/RVC/dataset/numberblocks_one'\n",
124
+ "EXP_DIR = f'/content/RVC/logs/{MODEL_NAME}'\n",
125
+ "os.makedirs(EXP_DIR, exist_ok=True)\n",
126
+ "\n",
127
+ "wav_files = [f for f in os.listdir(DATASET_PATH) if f.endswith('.wav')]\n",
128
+ "print(f'Training WAV files: {len(wav_files)}')\n",
129
+ "\n",
130
+ "# F0 extraction\n",
131
+ "print('\\n--- Extracting F0 (harvest) ---')\n",
132
+ "r = subprocess.run([sys.executable, 'infer/modules/train/extract/extract_f0_print.py',\n",
133
+ " EXP_DIR, DATASET_PATH, 'harvest', '0', '8'], capture_output=True, text=True, timeout=1800)\n",
134
+ "print(r.stdout[-500:] if r.stdout else '')\n",
135
+ "print(f'F0 return code: {r.returncode}')\n",
136
+ "\n",
137
+ "# Feature extraction\n",
138
+ "print('\\n--- Extracting features (v2) ---')\n",
139
+ "r = subprocess.run([sys.executable, 'infer/modules/train/extract/extract_feature_print.py',\n",
140
+ " EXP_DIR, 'v2', DATASET_PATH, '0', '8', 'fp32'], capture_output=True, text=True, timeout=1800)\n",
141
+ "print(r.stdout[-500:] if r.stdout else '')\n",
142
+ "print(f'Feature return code: {r.returncode}')\n",
143
+ "\n",
144
+ "exp_files = os.listdir(EXP_DIR)\n",
145
+ "print(f'\\n\u2705 Files in experiment dir: {len(exp_files)}')\n"
146
+ ],
147
+ "outputs": [],
148
+ "execution_count": null
149
+ },
150
+ {
151
+ "cell_type": "code",
152
+ "metadata": {},
153
+ "source": [
154
+ "#@title 5. \u8bad\u7ec3 RVC v2 \u6a21\u578b (~20-40 min)\n",
155
+ "import os, subprocess, sys, glob\n",
156
+ "\n",
157
+ "%cd /content/RVC\n",
158
+ "\n",
159
+ "MODEL_NAME = 'numberblocks_one'\n",
160
+ "EXP_DIR = f'/content/RVC/logs/{MODEL_NAME}'\n",
161
+ "\n",
162
+ "cmd = [sys.executable, 'infer/modules/train/train.py',\n",
163
+ " '-e', EXP_DIR, '-n', MODEL_NAME,\n",
164
+ " '-sr', '40k', '-bs', '8', '-g', '0',\n",
165
+ " '-te', '5000', '-se', '500',\n",
166
+ " '-v', 'v2', '-l', '1']\n",
167
+ "\n",
168
+ "print(f'Training: {\" \".join(cmd)}')\n",
169
+ "r = subprocess.run(cmd, capture_output=True, text=True, timeout=7200)\n",
170
+ "print(r.stdout[-1000:] if r.stdout else '')\n",
171
+ "if r.returncode != 0:\n",
172
+ " print('STDERR:', r.stderr[-500:])\n",
173
+ "\n",
174
+ "pth_files = glob.glob(f'{EXP_DIR}/**/*.pth', recursive=True)\n",
175
+ "print(f'\\n\u2705 PTH files: {len(pth_files)}')\n",
176
+ "for p in pth_files:\n",
177
+ " print(f' {p} ({os.path.getsize(p)/1024/1024:.1f} MB)')\n"
178
+ ],
179
+ "outputs": [],
180
+ "execution_count": null
181
+ },
182
+ {
183
+ "cell_type": "code",
184
+ "metadata": {},
185
+ "source": [
186
+ "#@title 6. \u4e0a\u4f20\u6a21\u578b\u5230 HuggingFace\n",
187
+ "import os, glob, json\n",
188
+ "from huggingface_hub import HfApi\n",
189
+ "\n",
190
+ "DATASET_REPO = 'ayf3/numberblocks-one-voice-dataset'\n",
191
+ "api = HfApi()\n",
192
+ "\n",
193
+ "pth_files = glob.glob('/content/RVC/**/*.pth', recursive=True)\n",
194
+ "print(f'Found {len(pth_files)} .pth files')\n",
195
+ "\n",
196
+ "if not pth_files:\n",
197
+ " print('\u26a0\ufe0f No model files found!')\n",
198
+ "else:\n",
199
+ " for pth_path in pth_files:\n",
200
+ " fname = os.path.basename(pth_path)\n",
201
+ " size_mb = os.path.getsize(pth_path) / 1024 / 1024\n",
202
+ " print(f'Uploading {fname} ({size_mb:.1f} MB)...')\n",
203
+ " api.upload_file(\n",
204
+ " path_or_fileobj=pth_path,\n",
205
+ " path_in_repo=f'models/{fname}',\n",
206
+ " repo_id=DATASET_REPO,\n",
207
+ " repo_type='dataset',\n",
208
+ " )\n",
209
+ " print(f' \u2705 models/{fname}')\n",
210
+ "\n",
211
+ " # Upload config\n",
212
+ " config = {\n",
213
+ " 'model_name': 'numberblocks_one',\n",
214
+ " 'version': 'v2',\n",
215
+ " 'sample_rate': 40000,\n",
216
+ " 'training_files': 2000,\n",
217
+ " 'data_source': 'top500 + augmented',\n",
218
+ " 'total_steps': 5000,\n",
219
+ " 'batch_size': 8,\n",
220
+ " }\n",
221
+ " with open('/tmp/train_config.json', 'w') as f:\n",
222
+ " json.dump(config, f, indent=2)\n",
223
+ " api.upload_file(\n",
224
+ " path_or_fileobj='/tmp/train_config.json',\n",
225
+ " path_in_repo='models/train_config.json',\n",
226
+ " repo_id=DATASET_REPO,\n",
227
+ " repo_type='dataset',\n",
228
+ " )\n",
229
+ " print('\u2705 Config uploaded')\n"
230
+ ],
231
+ "outputs": [],
232
+ "execution_count": null
233
+ },
234
+ {
235
+ "cell_type": "markdown",
236
+ "metadata": {},
237
+ "source": [
238
+ "## \ud83c\udf89 \u5b8c\u6210\uff01\n",
239
+ "\n",
240
+ "\u6a21\u578b\u5df2\u4e0a\u4f20\u5230 `models/` \u76ee\u5f55\u3002\u4e0b\u4e00\u6b65\uff1avoice conversion \u63a8\u7406\u3002\n"
241
+ ]
242
+ }
243
+ ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
244
  }