AlphJain commited on
Commit
fdabf1d
·
verified ·
1 Parent(s): 189b4b0

Upload folder using huggingface_hub

Browse files
finetuned_model/best_model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a538c78e65ee3d69272552405032f952685de4b3e03a8b295eb8cc476b2db62d
3
+ size 5607927701
finetuned_model/best_model_1773.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a538c78e65ee3d69272552405032f952685de4b3e03a8b295eb8cc476b2db62d
3
+ size 5607927701
finetuned_model/best_model_3546.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8aca7f6cf414b855dcca79d8559b39666b57b482cd15cb4f453c7d3ddd4a660a
3
+ size 5562740736
finetuned_model/colab_kernel_launcher.py ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright 2023 Google Inc.
2
+ #
3
+ # Licensed under the Apache License, Version 2.0 (the "License");
4
+ # you may not use this file except in compliance with the License.
5
+ # You may obtain a copy of the License at
6
+ #
7
+ # http://www.apache.org/licenses/LICENSE-2.0
8
+ #
9
+ # Unless required by applicable law or agreed to in writing, software
10
+ # distributed under the License is distributed on an "AS IS" BASIS,
11
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+ # See the License for the specific language governing permissions and
13
+ # limitations under the License.
14
+ """Custom kernel launcher app to customize socket options."""
15
+
16
+ from ipykernel import kernelapp
17
+ import zmq
18
+
19
+
20
+ # We want to set the high water mark on *all* sockets to 0, as we don't want
21
+ # the backend dropping any messages. We want to set this before any calls to
22
+ # bind or connect.
23
+ #
24
+ # In principle we should override `init_sockets`, but it's hard to set options
25
+ # on the `zmq.Context` there without rewriting the entire method. Instead we
26
+ # settle for only setting this on `iopub`, as that's the most important for our
27
+ # use case.
28
+ class ColabKernelApp(kernelapp.IPKernelApp):
29
+
30
+ def init_iopub(self, context):
31
+ context.setsockopt(zmq.RCVHWM, 0)
32
+ context.setsockopt(zmq.SNDHWM, 0)
33
+ return super().init_iopub(context)
34
+
35
+
36
+ if __name__ == '__main__':
37
+ ColabKernelApp.launch_instance()
finetuned_model/config.json ADDED
@@ -0,0 +1,206 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_path": "tts/run/",
3
+ "logger_uri": null,
4
+ "run_name": "xtts-finetuning",
5
+ "project_name": "VoiceCrafter",
6
+ "run_description": "\n GPT XTTS training\n ",
7
+ "print_step": 250,
8
+ "plot_step": 1000,
9
+ "model_param_stats": false,
10
+ "wandb_entity": null,
11
+ "dashboard_logger": "wandb",
12
+ "save_on_interrupt": true,
13
+ "log_model_step": 1000,
14
+ "save_step": 9999999999,
15
+ "save_n_checkpoints": 1,
16
+ "save_checkpoints": false,
17
+ "save_all_best": false,
18
+ "save_best_after": 0,
19
+ "target_loss": null,
20
+ "print_eval": false,
21
+ "test_delay_epochs": 0,
22
+ "run_eval": true,
23
+ "run_eval_steps": null,
24
+ "distributed_backend": "nccl",
25
+ "distributed_url": "tcp://localhost:54321",
26
+ "mixed_precision": false,
27
+ "precision": "fp16",
28
+ "epochs": 1000,
29
+ "batch_size": 1,
30
+ "eval_batch_size": 1,
31
+ "grad_clip": 0.0,
32
+ "scheduler_after_epoch": true,
33
+ "lr": 5e-06,
34
+ "optimizer": "AdamW",
35
+ "optimizer_params": {
36
+ "betas": [
37
+ 0.9,
38
+ 0.96
39
+ ],
40
+ "eps": 1e-08,
41
+ "weight_decay": 0.01
42
+ },
43
+ "lr_scheduler": "MultiStepLR",
44
+ "lr_scheduler_params": {
45
+ "milestones": [
46
+ 900000,
47
+ 2700000,
48
+ 5400000
49
+ ],
50
+ "gamma": 0.5,
51
+ "last_epoch": -1
52
+ },
53
+ "use_grad_scaler": false,
54
+ "allow_tf32": false,
55
+ "cudnn_enable": true,
56
+ "cudnn_deterministic": false,
57
+ "cudnn_benchmark": false,
58
+ "training_seed": 1,
59
+ "model": "xtts",
60
+ "num_loader_workers": 8,
61
+ "num_eval_loader_workers": 0,
62
+ "use_noise_augment": false,
63
+ "audio": {
64
+ "sample_rate": 24000,
65
+ "output_sample_rate": 24000,
66
+ "dvae_sample_rate": 24000
67
+ },
68
+ "use_phonemes": false,
69
+ "phonemizer": null,
70
+ "phoneme_language": null,
71
+ "compute_input_seq_cache": false,
72
+ "text_cleaner": null,
73
+ "enable_eos_bos_chars": false,
74
+ "test_sentences_file": "",
75
+ "phoneme_cache_path": null,
76
+ "characters": null,
77
+ "add_blank": false,
78
+ "batch_group_size": 1,
79
+ "loss_masking": null,
80
+ "min_audio_len": 1,
81
+ "max_audio_len": Infinity,
82
+ "min_text_len": 1,
83
+ "max_text_len": Infinity,
84
+ "compute_f0": false,
85
+ "compute_energy": false,
86
+ "compute_linear_spec": false,
87
+ "precompute_num_workers": 0,
88
+ "start_by_longest": false,
89
+ "shuffle": false,
90
+ "drop_last": false,
91
+ "datasets": [
92
+ {
93
+ "formatter": "",
94
+ "dataset_name": "",
95
+ "path": "",
96
+ "meta_file_train": "",
97
+ "ignored_speakers": null,
98
+ "language": "",
99
+ "phonemizer": "",
100
+ "meta_file_val": "",
101
+ "meta_file_attn_mask": ""
102
+ }
103
+ ],
104
+ "test_sentences": [
105
+ {
106
+ "text": "It took me quite a long time to develop a voice, and now that I have it I'm not going to be silent.",
107
+ "speaker_wav": "extracted_dataset/output/wavs/d65488ea-a4f2-4c65-98e4-0fb2d4cee3db_1.wav",
108
+ "language": "en"
109
+ },
110
+ {
111
+ "text": "This cake is great. It's so delicious and moist.",
112
+ "speaker_wav": "extracted_dataset/output/wavs/d65488ea-a4f2-4c65-98e4-0fb2d4cee3db_1.wav",
113
+ "language": "en"
114
+ },
115
+ {
116
+ "text": "And soon, nothing more terrible, nothing more true, and specious stuff that says no rational being can fear a thing it will not feel, not seeing that this is what we fear.",
117
+ "speaker_wav": "extracted_dataset/output/wavs/d65488ea-a4f2-4c65-98e4-0fb2d4cee3db_1.wav",
118
+ "language": "en"
119
+ }
120
+ ],
121
+ "eval_split_max_size": 256,
122
+ "eval_split_size": 0.01,
123
+ "use_speaker_weighted_sampler": false,
124
+ "speaker_weighted_sampler_alpha": 1.0,
125
+ "use_language_weighted_sampler": false,
126
+ "language_weighted_sampler_alpha": 1.0,
127
+ "use_length_weighted_sampler": false,
128
+ "length_weighted_sampler_alpha": 1.0,
129
+ "model_args": {
130
+ "gpt_batch_size": 1,
131
+ "enable_redaction": false,
132
+ "kv_cache": true,
133
+ "gpt_checkpoint": "",
134
+ "clvp_checkpoint": null,
135
+ "decoder_checkpoint": null,
136
+ "num_chars": 255,
137
+ "tokenizer_file": "tts/run/XTTS_v2.0_original_model_files/vocab.json",
138
+ "gpt_max_audio_tokens": 605,
139
+ "gpt_max_text_tokens": 402,
140
+ "gpt_max_prompt_tokens": 70,
141
+ "gpt_layers": 30,
142
+ "gpt_n_model_channels": 1024,
143
+ "gpt_n_heads": 16,
144
+ "gpt_number_text_tokens": 6681,
145
+ "gpt_start_text_token": 261,
146
+ "gpt_stop_text_token": 0,
147
+ "gpt_num_audio_tokens": 1026,
148
+ "gpt_start_audio_token": 1024,
149
+ "gpt_stop_audio_token": 1025,
150
+ "gpt_code_stride_len": 1024,
151
+ "gpt_use_masking_gt_prompt_approach": true,
152
+ "gpt_use_perceiver_resampler": true,
153
+ "input_sample_rate": 22050,
154
+ "output_sample_rate": 24000,
155
+ "output_hop_length": 256,
156
+ "decoder_input_dim": 1024,
157
+ "d_vector_dim": 512,
158
+ "cond_d_vector_in_each_upsampling_layer": true,
159
+ "duration_const": 102400,
160
+ "min_conditioning_length": 20640,
161
+ "max_conditioning_length": 720000,
162
+ "gpt_loss_text_ce_weight": 0.01,
163
+ "gpt_loss_mel_ce_weight": 1.0,
164
+ "debug_loading_failures": true,
165
+ "max_wav_length": 720000,
166
+ "max_text_length": 200,
167
+ "mel_norm_file": "tts/run/XTTS_v2.0_original_model_files/mel_stats.pth",
168
+ "dvae_checkpoint": "tts/run/XTTS_v2.0_original_model_files/dvae.pth",
169
+ "xtts_checkpoint": "tts/run/XTTS_v2.0_original_model_files/model.pth",
170
+ "vocoder": ""
171
+ },
172
+ "model_dir": null,
173
+ "languages": [
174
+ "en",
175
+ "es",
176
+ "fr",
177
+ "de",
178
+ "it",
179
+ "pt",
180
+ "pl",
181
+ "tr",
182
+ "ru",
183
+ "nl",
184
+ "cs",
185
+ "ar",
186
+ "zh-cn",
187
+ "hu",
188
+ "ko",
189
+ "ja",
190
+ "hi"
191
+ ],
192
+ "temperature": 0.85,
193
+ "length_penalty": 1.0,
194
+ "repetition_penalty": 2.0,
195
+ "top_k": 50,
196
+ "top_p": 0.85,
197
+ "num_gpt_outputs": 1,
198
+ "gpt_cond_len": 12,
199
+ "gpt_cond_chunk_len": 4,
200
+ "max_ref_len": 10,
201
+ "sound_norm_refs": false,
202
+ "optimizer_wd_only_on_weights": true,
203
+ "weighted_loss_attrs": null,
204
+ "weighted_loss_multipliers": null,
205
+ "github_branch": "inside_docker"
206
+ }
finetuned_model/trainer_0_log.txt ADDED
@@ -0,0 +1,197 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ > Training Environment:
2
+ | > Backend: Torch
3
+ | > Mixed precision: False
4
+ | > Precision: float32
5
+ | > Current device: 0
6
+ | > Num. of GPUs: 1
7
+ | > Num. of CPUs: 4
8
+ | > Num. of Torch Threads: 1
9
+ | > Torch seed: 1
10
+ | > Torch CUDNN: True
11
+ | > Torch CUDNN deterministic: False
12
+ | > Torch CUDNN benchmark: False
13
+ | > Torch TF32 MatMul: False
14
+
15
+ > Model has 518442047 parameters
16
+
17
+  > EPOCH: 0/1000
18
+ --> tts/run/xtts-finetuning-May-28-2025_05+56AM-0000000
19
+
20
+  > EVALUATION 
21
+
22
+
23
+ --> EVAL PERFORMANCE
24
+ | > avg_loader_time: 0.05358801569257464 (+0)
25
+ | > avg_loss_text_ce: 0.02478479796222278 (+0)
26
+ | > avg_loss_mel_ce: 3.741819156919207 (+0)
27
+ | > avg_loss: 3.76660396030971 (+0)
28
+
29
+
30
+  > EPOCH: 1/1000
31
+ --> tts/run/xtts-finetuning-May-28-2025_05+56AM-0000000
32
+
33
+  > TRAINING (2025-05-28 05:57:23) 
34
+
35
+  --> TIME: 2025-05-28 05:57:26 -- STEP: 0/1773 -- GLOBAL_STEP: 0
36
+ | > loss_text_ce: 0.024488264694809914 (0.024488264694809914)
37
+ | > loss_mel_ce: 4.036444187164307 (4.036444187164307)
38
+ | > loss: 0.06345207244157791 (0.06345207244157791)
39
+ | > current_lr: 5e-06
40
+ | > step_time: 0.7383 (0.7382748126983643)
41
+ | > loader_time: 2.1633 (2.1632590293884277)
42
+
43
+
44
+  --> TIME: 2025-05-28 05:59:55 -- STEP: 250/1773 -- GLOBAL_STEP: 250
45
+ | > loss_text_ce: 0.022974064573645592 (0.024572087943553922)
46
+ | > loss_mel_ce: 4.102844715118408 (nan)
47
+ | > loss: 0.06446591764688492 (nan)
48
+ | > current_lr: 5e-06
49
+ | > step_time: 0.5202 (0.5721063699722287)
50
+ | > loader_time: 0.0127 (0.021605122566223153)
51
+
52
+
53
+  --> TIME: 2025-05-28 06:02:25 -- STEP: 500/1773 -- GLOBAL_STEP: 500
54
+ | > loss_text_ce: 0.022191306576132774 (0.024244515029713504)
55
+ | > loss_mel_ce: 3.560326099395752 (nan)
56
+ | > loss: 0.055976834148168564 (nan)
57
+ | > current_lr: 5e-06
58
+ | > step_time: 0.5558 (0.5743874120712283)
59
+ | > loader_time: 0.0145 (0.021079427242279045)
60
+
61
+
62
+  --> TIME: 2025-05-28 06:04:52 -- STEP: 750/1773 -- GLOBAL_STEP: 750
63
+ | > loss_text_ce: 0.023066112771630287 (0.024316282462328673)
64
+ | > loss_mel_ce: 3.4665474891662598 (nan)
65
+ | > loss: 0.054525211453437805 (nan)
66
+ | > current_lr: 5e-06
67
+ | > step_time: 0.6241 (0.5720558166503911)
68
+ | > loader_time: 0.0226 (0.020561660766601552)
69
+
70
+
71
+  --> TIME: 2025-05-28 06:07:18 -- STEP: 1000/1773 -- GLOBAL_STEP: 1000
72
+ | > loss_text_ce: 0.023107804358005524 (0.0242976028714329)
73
+ | > loss_mel_ce: 3.397033214569092 (nan)
74
+ | > loss: 0.053439702838659286 (nan)
75
+ | > current_lr: 5e-06
76
+ | > step_time: 0.631 (0.5697787494659426)
77
+ | > loader_time: 0.0242 (0.020013311624526974)
78
+
79
+
80
+  --> TIME: 2025-05-28 06:09:42 -- STEP: 1250/1773 -- GLOBAL_STEP: 1250
81
+ | > loss_text_ce: 0.019792545586824417 (0.02419562596976758)
82
+ | > loss_mel_ce: 5.3489155769348145 (nan)
83
+ | > loss: 0.08388606458902359 (nan)
84
+ | > current_lr: 5e-06
85
+ | > step_time: 0.5923 (0.5667600589752204)
86
+ | > loader_time: 0.0251 (0.019566329574584956)
87
+
88
+
89
+  --> TIME: 2025-05-28 06:12:06 -- STEP: 1500/1773 -- GLOBAL_STEP: 1500
90
+ | > loss_text_ce: 0.02061852626502514 (0.024097217947865537)
91
+ | > loss_mel_ce: 3.454913377761841 (nan)
92
+ | > loss: 0.05430518463253975 (nan)
93
+ | > current_lr: 5e-06
94
+ | > step_time: 0.5326 (0.5648289133707688)
95
+ | > loader_time: 0.0122 (0.019175546487172453)
96
+
97
+
98
+  --> TIME: 2025-05-28 06:14:31 -- STEP: 1750/1773 -- GLOBAL_STEP: 1750
99
+ | > loss_text_ce: 0.018680287525057793 (0.024025386422872552)
100
+ | > loss_mel_ce: 3.2065491676330566 (nan)
101
+ | > loss: 0.0503942109644413 (nan)
102
+ | > current_lr: 5e-06
103
+ | > step_time: 0.5234 (0.5642632334572935)
104
+ | > loader_time: 0.0126 (0.019005961554391045)
105
+
106
+
107
+  > EVALUATION 
108
+
109
+
110
+ --> EVAL PERFORMANCE
111
+ | > avg_loader_time: 0.04692824227469308 (-0.0066597734178815565)
112
+ | > avg_loss_text_ce: 0.02401351955320154 (-0.0007712784090212407)
113
+ | > avg_loss_mel_ce: 3.0045580523354665 (-0.7372611045837405)
114
+ | > avg_loss: 3.028571551186698 (-0.7380324091230119)
115
+
116
+ > BEST MODEL : tts/run/xtts-finetuning-May-28-2025_05+56AM-0000000/best_model_1773.pth
117
+
118
+  > EPOCH: 2/1000
119
+ --> tts/run/xtts-finetuning-May-28-2025_05+56AM-0000000
120
+
121
+  > TRAINING (2025-05-28 06:15:44) 
122
+
123
+  --> TIME: 2025-05-28 06:17:57 -- STEP: 227/1773 -- GLOBAL_STEP: 2000
124
+ | > loss_text_ce: 0.02634984627366066 (0.02347405103039374)
125
+ | > loss_mel_ce: 2.9743587970733643 (nan)
126
+ | > loss: 0.04688607156276703 (nan)
127
+ | > current_lr: 5e-06
128
+ | > step_time: 0.5701 (0.5613324474133058)
129
+ | > loader_time: 0.024 (0.01852047705965421)
130
+
131
+
132
+  --> TIME: 2025-05-28 06:20:24 -- STEP: 477/1773 -- GLOBAL_STEP: 2250
133
+ | > loss_text_ce: 0.019006451591849327 (0.0235798091996672)
134
+ | > loss_mel_ce: 2.9675517082214355 (nan)
135
+ | > loss: 0.046664971858263016 (nan)
136
+ | > current_lr: 5e-06
137
+ | > step_time: 0.5472 (0.5630224965653334)
138
+ | > loader_time: 0.0132 (0.01841085212035749)
139
+
140
+
141
+  --> TIME: 2025-05-28 06:22:49 -- STEP: 727/1773 -- GLOBAL_STEP: 2500
142
+ | > loss_text_ce: 0.019247999414801598 (0.023683539267624253)
143
+ | > loss_mel_ce: 4.047469615936279 (nan)
144
+ | > loss: 0.06354246288537979 (nan)
145
+ | > current_lr: 5e-06
146
+ | > step_time: 0.6674 (0.5622831017311867)
147
+ | > loader_time: 0.0246 (0.018256587536167733)
148
+
149
+
150
+  --> TIME: 2025-05-28 06:25:14 -- STEP: 977/1773 -- GLOBAL_STEP: 2750
151
+ | > loss_text_ce: 0.019411535933613777 (0.023644839759492817)
152
+ | > loss_mel_ce: 2.550281047821045 (nan)
153
+ | > loss: 0.040151447057724 (nan)
154
+ | > current_lr: 5e-06
155
+ | > step_time: 0.6017 (0.5617610405265011)
156
+ | > loader_time: 0.0232 (0.018097628590516446)
157
+
158
+
159
+  --> TIME: 2025-05-28 06:27:39 -- STEP: 1227/1773 -- GLOBAL_STEP: 3000
160
+ | > loss_text_ce: 0.018359506502747536 (0.023614716428440572)
161
+ | > loss_mel_ce: 2.8278863430023193 (nan)
162
+ | > loss: 0.0444725900888443 (nan)
163
+ | > current_lr: 5e-06
164
+ | > step_time: 0.5529 (0.5614671928666929)
165
+ | > loader_time: 0.014 (0.01804757506845438)
166
+
167
+
168
+  --> TIME: 2025-05-28 06:30:05 -- STEP: 1477/1773 -- GLOBAL_STEP: 3250
169
+ | > loss_text_ce: 0.024424398317933083 (0.02365642989756753)
170
+ | > loss_mel_ce: 2.85211443901062 (nan)
171
+ | > loss: 0.04494591802358627 (nan)
172
+ | > current_lr: 5e-06
173
+ | > step_time: 0.5568 (0.5615530244869896)
174
+ | > loader_time: 0.0127 (0.01798885096225209)
175
+
176
+
177
+  --> TIME: 2025-05-28 06:32:30 -- STEP: 1727/1773 -- GLOBAL_STEP: 3500
178
+ | > loss_text_ce: 0.013657196424901485 (0.023524502163389426)
179
+ | > loss_mel_ce: 2.7606756687164307 (nan)
180
+ | > loss: 0.043348949402570724 (nan)
181
+ | > grad_norm: 0 (0.0)
182
+ | > current_lr: 5e-06
183
+ | > step_time: 0.5705 (0.5614900450935659)
184
+ | > loader_time: 0.011 (0.01798361260282141)
185
+
186
+
187
+  > EVALUATION 
188
+
189
+
190
+ --> EVAL PERFORMANCE
191
+ | > avg_loader_time: 0.047074174880981444 (+0.0001459326062883612)
192
+ | > avg_loss_text_ce: 0.023535797904644693 (-0.0004777216485568475)
193
+ | > avg_loss_mel_ce: 2.8784927810941423 (-0.12606527124132416)
194
+ | > avg_loss: 2.9020285640444072 (-0.12654298714229073)
195
+
196
+ > BEST MODEL : tts/run/xtts-finetuning-May-28-2025_05+56AM-0000000/best_model_3546.pth
197
+ ! Run is kept in tts/run/xtts-finetuning-May-28-2025_05+56AM-0000000