Upload folder using huggingface_hub
Browse files
finetuned_model/best_model.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a538c78e65ee3d69272552405032f952685de4b3e03a8b295eb8cc476b2db62d
|
| 3 |
+
size 5607927701
|
finetuned_model/best_model_1773.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a538c78e65ee3d69272552405032f952685de4b3e03a8b295eb8cc476b2db62d
|
| 3 |
+
size 5607927701
|
finetuned_model/best_model_3546.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8aca7f6cf414b855dcca79d8559b39666b57b482cd15cb4f453c7d3ddd4a660a
|
| 3 |
+
size 5562740736
|
finetuned_model/colab_kernel_launcher.py
ADDED
|
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2023 Google Inc.
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
"""Custom kernel launcher app to customize socket options."""
|
| 15 |
+
|
| 16 |
+
from ipykernel import kernelapp
|
| 17 |
+
import zmq
|
| 18 |
+
|
| 19 |
+
|
| 20 |
+
# We want to set the high water mark on *all* sockets to 0, as we don't want
|
| 21 |
+
# the backend dropping any messages. We want to set this before any calls to
|
| 22 |
+
# bind or connect.
|
| 23 |
+
#
|
| 24 |
+
# In principle we should override `init_sockets`, but it's hard to set options
|
| 25 |
+
# on the `zmq.Context` there without rewriting the entire method. Instead we
|
| 26 |
+
# settle for only setting this on `iopub`, as that's the most important for our
|
| 27 |
+
# use case.
|
| 28 |
+
class ColabKernelApp(kernelapp.IPKernelApp):
|
| 29 |
+
|
| 30 |
+
def init_iopub(self, context):
|
| 31 |
+
context.setsockopt(zmq.RCVHWM, 0)
|
| 32 |
+
context.setsockopt(zmq.SNDHWM, 0)
|
| 33 |
+
return super().init_iopub(context)
|
| 34 |
+
|
| 35 |
+
|
| 36 |
+
if __name__ == '__main__':
|
| 37 |
+
ColabKernelApp.launch_instance()
|
finetuned_model/config.json
ADDED
|
@@ -0,0 +1,206 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"output_path": "tts/run/",
|
| 3 |
+
"logger_uri": null,
|
| 4 |
+
"run_name": "xtts-finetuning",
|
| 5 |
+
"project_name": "VoiceCrafter",
|
| 6 |
+
"run_description": "\n GPT XTTS training\n ",
|
| 7 |
+
"print_step": 250,
|
| 8 |
+
"plot_step": 1000,
|
| 9 |
+
"model_param_stats": false,
|
| 10 |
+
"wandb_entity": null,
|
| 11 |
+
"dashboard_logger": "wandb",
|
| 12 |
+
"save_on_interrupt": true,
|
| 13 |
+
"log_model_step": 1000,
|
| 14 |
+
"save_step": 9999999999,
|
| 15 |
+
"save_n_checkpoints": 1,
|
| 16 |
+
"save_checkpoints": false,
|
| 17 |
+
"save_all_best": false,
|
| 18 |
+
"save_best_after": 0,
|
| 19 |
+
"target_loss": null,
|
| 20 |
+
"print_eval": false,
|
| 21 |
+
"test_delay_epochs": 0,
|
| 22 |
+
"run_eval": true,
|
| 23 |
+
"run_eval_steps": null,
|
| 24 |
+
"distributed_backend": "nccl",
|
| 25 |
+
"distributed_url": "tcp://localhost:54321",
|
| 26 |
+
"mixed_precision": false,
|
| 27 |
+
"precision": "fp16",
|
| 28 |
+
"epochs": 1000,
|
| 29 |
+
"batch_size": 1,
|
| 30 |
+
"eval_batch_size": 1,
|
| 31 |
+
"grad_clip": 0.0,
|
| 32 |
+
"scheduler_after_epoch": true,
|
| 33 |
+
"lr": 5e-06,
|
| 34 |
+
"optimizer": "AdamW",
|
| 35 |
+
"optimizer_params": {
|
| 36 |
+
"betas": [
|
| 37 |
+
0.9,
|
| 38 |
+
0.96
|
| 39 |
+
],
|
| 40 |
+
"eps": 1e-08,
|
| 41 |
+
"weight_decay": 0.01
|
| 42 |
+
},
|
| 43 |
+
"lr_scheduler": "MultiStepLR",
|
| 44 |
+
"lr_scheduler_params": {
|
| 45 |
+
"milestones": [
|
| 46 |
+
900000,
|
| 47 |
+
2700000,
|
| 48 |
+
5400000
|
| 49 |
+
],
|
| 50 |
+
"gamma": 0.5,
|
| 51 |
+
"last_epoch": -1
|
| 52 |
+
},
|
| 53 |
+
"use_grad_scaler": false,
|
| 54 |
+
"allow_tf32": false,
|
| 55 |
+
"cudnn_enable": true,
|
| 56 |
+
"cudnn_deterministic": false,
|
| 57 |
+
"cudnn_benchmark": false,
|
| 58 |
+
"training_seed": 1,
|
| 59 |
+
"model": "xtts",
|
| 60 |
+
"num_loader_workers": 8,
|
| 61 |
+
"num_eval_loader_workers": 0,
|
| 62 |
+
"use_noise_augment": false,
|
| 63 |
+
"audio": {
|
| 64 |
+
"sample_rate": 24000,
|
| 65 |
+
"output_sample_rate": 24000,
|
| 66 |
+
"dvae_sample_rate": 24000
|
| 67 |
+
},
|
| 68 |
+
"use_phonemes": false,
|
| 69 |
+
"phonemizer": null,
|
| 70 |
+
"phoneme_language": null,
|
| 71 |
+
"compute_input_seq_cache": false,
|
| 72 |
+
"text_cleaner": null,
|
| 73 |
+
"enable_eos_bos_chars": false,
|
| 74 |
+
"test_sentences_file": "",
|
| 75 |
+
"phoneme_cache_path": null,
|
| 76 |
+
"characters": null,
|
| 77 |
+
"add_blank": false,
|
| 78 |
+
"batch_group_size": 1,
|
| 79 |
+
"loss_masking": null,
|
| 80 |
+
"min_audio_len": 1,
|
| 81 |
+
"max_audio_len": Infinity,
|
| 82 |
+
"min_text_len": 1,
|
| 83 |
+
"max_text_len": Infinity,
|
| 84 |
+
"compute_f0": false,
|
| 85 |
+
"compute_energy": false,
|
| 86 |
+
"compute_linear_spec": false,
|
| 87 |
+
"precompute_num_workers": 0,
|
| 88 |
+
"start_by_longest": false,
|
| 89 |
+
"shuffle": false,
|
| 90 |
+
"drop_last": false,
|
| 91 |
+
"datasets": [
|
| 92 |
+
{
|
| 93 |
+
"formatter": "",
|
| 94 |
+
"dataset_name": "",
|
| 95 |
+
"path": "",
|
| 96 |
+
"meta_file_train": "",
|
| 97 |
+
"ignored_speakers": null,
|
| 98 |
+
"language": "",
|
| 99 |
+
"phonemizer": "",
|
| 100 |
+
"meta_file_val": "",
|
| 101 |
+
"meta_file_attn_mask": ""
|
| 102 |
+
}
|
| 103 |
+
],
|
| 104 |
+
"test_sentences": [
|
| 105 |
+
{
|
| 106 |
+
"text": "It took me quite a long time to develop a voice, and now that I have it I'm not going to be silent.",
|
| 107 |
+
"speaker_wav": "extracted_dataset/output/wavs/d65488ea-a4f2-4c65-98e4-0fb2d4cee3db_1.wav",
|
| 108 |
+
"language": "en"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"text": "This cake is great. It's so delicious and moist.",
|
| 112 |
+
"speaker_wav": "extracted_dataset/output/wavs/d65488ea-a4f2-4c65-98e4-0fb2d4cee3db_1.wav",
|
| 113 |
+
"language": "en"
|
| 114 |
+
},
|
| 115 |
+
{
|
| 116 |
+
"text": "And soon, nothing more terrible, nothing more true, and specious stuff that says no rational being can fear a thing it will not feel, not seeing that this is what we fear.",
|
| 117 |
+
"speaker_wav": "extracted_dataset/output/wavs/d65488ea-a4f2-4c65-98e4-0fb2d4cee3db_1.wav",
|
| 118 |
+
"language": "en"
|
| 119 |
+
}
|
| 120 |
+
],
|
| 121 |
+
"eval_split_max_size": 256,
|
| 122 |
+
"eval_split_size": 0.01,
|
| 123 |
+
"use_speaker_weighted_sampler": false,
|
| 124 |
+
"speaker_weighted_sampler_alpha": 1.0,
|
| 125 |
+
"use_language_weighted_sampler": false,
|
| 126 |
+
"language_weighted_sampler_alpha": 1.0,
|
| 127 |
+
"use_length_weighted_sampler": false,
|
| 128 |
+
"length_weighted_sampler_alpha": 1.0,
|
| 129 |
+
"model_args": {
|
| 130 |
+
"gpt_batch_size": 1,
|
| 131 |
+
"enable_redaction": false,
|
| 132 |
+
"kv_cache": true,
|
| 133 |
+
"gpt_checkpoint": "",
|
| 134 |
+
"clvp_checkpoint": null,
|
| 135 |
+
"decoder_checkpoint": null,
|
| 136 |
+
"num_chars": 255,
|
| 137 |
+
"tokenizer_file": "tts/run/XTTS_v2.0_original_model_files/vocab.json",
|
| 138 |
+
"gpt_max_audio_tokens": 605,
|
| 139 |
+
"gpt_max_text_tokens": 402,
|
| 140 |
+
"gpt_max_prompt_tokens": 70,
|
| 141 |
+
"gpt_layers": 30,
|
| 142 |
+
"gpt_n_model_channels": 1024,
|
| 143 |
+
"gpt_n_heads": 16,
|
| 144 |
+
"gpt_number_text_tokens": 6681,
|
| 145 |
+
"gpt_start_text_token": 261,
|
| 146 |
+
"gpt_stop_text_token": 0,
|
| 147 |
+
"gpt_num_audio_tokens": 1026,
|
| 148 |
+
"gpt_start_audio_token": 1024,
|
| 149 |
+
"gpt_stop_audio_token": 1025,
|
| 150 |
+
"gpt_code_stride_len": 1024,
|
| 151 |
+
"gpt_use_masking_gt_prompt_approach": true,
|
| 152 |
+
"gpt_use_perceiver_resampler": true,
|
| 153 |
+
"input_sample_rate": 22050,
|
| 154 |
+
"output_sample_rate": 24000,
|
| 155 |
+
"output_hop_length": 256,
|
| 156 |
+
"decoder_input_dim": 1024,
|
| 157 |
+
"d_vector_dim": 512,
|
| 158 |
+
"cond_d_vector_in_each_upsampling_layer": true,
|
| 159 |
+
"duration_const": 102400,
|
| 160 |
+
"min_conditioning_length": 20640,
|
| 161 |
+
"max_conditioning_length": 720000,
|
| 162 |
+
"gpt_loss_text_ce_weight": 0.01,
|
| 163 |
+
"gpt_loss_mel_ce_weight": 1.0,
|
| 164 |
+
"debug_loading_failures": true,
|
| 165 |
+
"max_wav_length": 720000,
|
| 166 |
+
"max_text_length": 200,
|
| 167 |
+
"mel_norm_file": "tts/run/XTTS_v2.0_original_model_files/mel_stats.pth",
|
| 168 |
+
"dvae_checkpoint": "tts/run/XTTS_v2.0_original_model_files/dvae.pth",
|
| 169 |
+
"xtts_checkpoint": "tts/run/XTTS_v2.0_original_model_files/model.pth",
|
| 170 |
+
"vocoder": ""
|
| 171 |
+
},
|
| 172 |
+
"model_dir": null,
|
| 173 |
+
"languages": [
|
| 174 |
+
"en",
|
| 175 |
+
"es",
|
| 176 |
+
"fr",
|
| 177 |
+
"de",
|
| 178 |
+
"it",
|
| 179 |
+
"pt",
|
| 180 |
+
"pl",
|
| 181 |
+
"tr",
|
| 182 |
+
"ru",
|
| 183 |
+
"nl",
|
| 184 |
+
"cs",
|
| 185 |
+
"ar",
|
| 186 |
+
"zh-cn",
|
| 187 |
+
"hu",
|
| 188 |
+
"ko",
|
| 189 |
+
"ja",
|
| 190 |
+
"hi"
|
| 191 |
+
],
|
| 192 |
+
"temperature": 0.85,
|
| 193 |
+
"length_penalty": 1.0,
|
| 194 |
+
"repetition_penalty": 2.0,
|
| 195 |
+
"top_k": 50,
|
| 196 |
+
"top_p": 0.85,
|
| 197 |
+
"num_gpt_outputs": 1,
|
| 198 |
+
"gpt_cond_len": 12,
|
| 199 |
+
"gpt_cond_chunk_len": 4,
|
| 200 |
+
"max_ref_len": 10,
|
| 201 |
+
"sound_norm_refs": false,
|
| 202 |
+
"optimizer_wd_only_on_weights": true,
|
| 203 |
+
"weighted_loss_attrs": null,
|
| 204 |
+
"weighted_loss_multipliers": null,
|
| 205 |
+
"github_branch": "inside_docker"
|
| 206 |
+
}
|
finetuned_model/trainer_0_log.txt
ADDED
|
@@ -0,0 +1,197 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
> Training Environment:
|
| 2 |
+
| > Backend: Torch
|
| 3 |
+
| > Mixed precision: False
|
| 4 |
+
| > Precision: float32
|
| 5 |
+
| > Current device: 0
|
| 6 |
+
| > Num. of GPUs: 1
|
| 7 |
+
| > Num. of CPUs: 4
|
| 8 |
+
| > Num. of Torch Threads: 1
|
| 9 |
+
| > Torch seed: 1
|
| 10 |
+
| > Torch CUDNN: True
|
| 11 |
+
| > Torch CUDNN deterministic: False
|
| 12 |
+
| > Torch CUDNN benchmark: False
|
| 13 |
+
| > Torch TF32 MatMul: False
|
| 14 |
+
|
| 15 |
+
> Model has 518442047 parameters
|
| 16 |
+
|
| 17 |
+
[4m[1m > EPOCH: 0/1000[0m
|
| 18 |
+
--> tts/run/xtts-finetuning-May-28-2025_05+56AM-0000000
|
| 19 |
+
|
| 20 |
+
[1m > EVALUATION [0m
|
| 21 |
+
|
| 22 |
+
|
| 23 |
+
[1m--> EVAL PERFORMANCE[0m
|
| 24 |
+
| > avg_loader_time: 0.05358801569257464 [0m(+0)
|
| 25 |
+
| > avg_loss_text_ce: 0.02478479796222278 [0m(+0)
|
| 26 |
+
| > avg_loss_mel_ce: 3.741819156919207 [0m(+0)
|
| 27 |
+
| > avg_loss: 3.76660396030971 [0m(+0)
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
[4m[1m > EPOCH: 1/1000[0m
|
| 31 |
+
--> tts/run/xtts-finetuning-May-28-2025_05+56AM-0000000
|
| 32 |
+
|
| 33 |
+
[1m > TRAINING (2025-05-28 05:57:23) [0m
|
| 34 |
+
|
| 35 |
+
[1m --> TIME: 2025-05-28 05:57:26 -- STEP: 0/1773 -- GLOBAL_STEP: 0[0m
|
| 36 |
+
| > loss_text_ce: 0.024488264694809914 (0.024488264694809914)
|
| 37 |
+
| > loss_mel_ce: 4.036444187164307 (4.036444187164307)
|
| 38 |
+
| > loss: 0.06345207244157791 (0.06345207244157791)
|
| 39 |
+
| > current_lr: 5e-06
|
| 40 |
+
| > step_time: 0.7383 (0.7382748126983643)
|
| 41 |
+
| > loader_time: 2.1633 (2.1632590293884277)
|
| 42 |
+
|
| 43 |
+
|
| 44 |
+
[1m --> TIME: 2025-05-28 05:59:55 -- STEP: 250/1773 -- GLOBAL_STEP: 250[0m
|
| 45 |
+
| > loss_text_ce: 0.022974064573645592 (0.024572087943553922)
|
| 46 |
+
| > loss_mel_ce: 4.102844715118408 (nan)
|
| 47 |
+
| > loss: 0.06446591764688492 (nan)
|
| 48 |
+
| > current_lr: 5e-06
|
| 49 |
+
| > step_time: 0.5202 (0.5721063699722287)
|
| 50 |
+
| > loader_time: 0.0127 (0.021605122566223153)
|
| 51 |
+
|
| 52 |
+
|
| 53 |
+
[1m --> TIME: 2025-05-28 06:02:25 -- STEP: 500/1773 -- GLOBAL_STEP: 500[0m
|
| 54 |
+
| > loss_text_ce: 0.022191306576132774 (0.024244515029713504)
|
| 55 |
+
| > loss_mel_ce: 3.560326099395752 (nan)
|
| 56 |
+
| > loss: 0.055976834148168564 (nan)
|
| 57 |
+
| > current_lr: 5e-06
|
| 58 |
+
| > step_time: 0.5558 (0.5743874120712283)
|
| 59 |
+
| > loader_time: 0.0145 (0.021079427242279045)
|
| 60 |
+
|
| 61 |
+
|
| 62 |
+
[1m --> TIME: 2025-05-28 06:04:52 -- STEP: 750/1773 -- GLOBAL_STEP: 750[0m
|
| 63 |
+
| > loss_text_ce: 0.023066112771630287 (0.024316282462328673)
|
| 64 |
+
| > loss_mel_ce: 3.4665474891662598 (nan)
|
| 65 |
+
| > loss: 0.054525211453437805 (nan)
|
| 66 |
+
| > current_lr: 5e-06
|
| 67 |
+
| > step_time: 0.6241 (0.5720558166503911)
|
| 68 |
+
| > loader_time: 0.0226 (0.020561660766601552)
|
| 69 |
+
|
| 70 |
+
|
| 71 |
+
[1m --> TIME: 2025-05-28 06:07:18 -- STEP: 1000/1773 -- GLOBAL_STEP: 1000[0m
|
| 72 |
+
| > loss_text_ce: 0.023107804358005524 (0.0242976028714329)
|
| 73 |
+
| > loss_mel_ce: 3.397033214569092 (nan)
|
| 74 |
+
| > loss: 0.053439702838659286 (nan)
|
| 75 |
+
| > current_lr: 5e-06
|
| 76 |
+
| > step_time: 0.631 (0.5697787494659426)
|
| 77 |
+
| > loader_time: 0.0242 (0.020013311624526974)
|
| 78 |
+
|
| 79 |
+
|
| 80 |
+
[1m --> TIME: 2025-05-28 06:09:42 -- STEP: 1250/1773 -- GLOBAL_STEP: 1250[0m
|
| 81 |
+
| > loss_text_ce: 0.019792545586824417 (0.02419562596976758)
|
| 82 |
+
| > loss_mel_ce: 5.3489155769348145 (nan)
|
| 83 |
+
| > loss: 0.08388606458902359 (nan)
|
| 84 |
+
| > current_lr: 5e-06
|
| 85 |
+
| > step_time: 0.5923 (0.5667600589752204)
|
| 86 |
+
| > loader_time: 0.0251 (0.019566329574584956)
|
| 87 |
+
|
| 88 |
+
|
| 89 |
+
[1m --> TIME: 2025-05-28 06:12:06 -- STEP: 1500/1773 -- GLOBAL_STEP: 1500[0m
|
| 90 |
+
| > loss_text_ce: 0.02061852626502514 (0.024097217947865537)
|
| 91 |
+
| > loss_mel_ce: 3.454913377761841 (nan)
|
| 92 |
+
| > loss: 0.05430518463253975 (nan)
|
| 93 |
+
| > current_lr: 5e-06
|
| 94 |
+
| > step_time: 0.5326 (0.5648289133707688)
|
| 95 |
+
| > loader_time: 0.0122 (0.019175546487172453)
|
| 96 |
+
|
| 97 |
+
|
| 98 |
+
[1m --> TIME: 2025-05-28 06:14:31 -- STEP: 1750/1773 -- GLOBAL_STEP: 1750[0m
|
| 99 |
+
| > loss_text_ce: 0.018680287525057793 (0.024025386422872552)
|
| 100 |
+
| > loss_mel_ce: 3.2065491676330566 (nan)
|
| 101 |
+
| > loss: 0.0503942109644413 (nan)
|
| 102 |
+
| > current_lr: 5e-06
|
| 103 |
+
| > step_time: 0.5234 (0.5642632334572935)
|
| 104 |
+
| > loader_time: 0.0126 (0.019005961554391045)
|
| 105 |
+
|
| 106 |
+
|
| 107 |
+
[1m > EVALUATION [0m
|
| 108 |
+
|
| 109 |
+
|
| 110 |
+
[1m--> EVAL PERFORMANCE[0m
|
| 111 |
+
| > avg_loader_time:[92m 0.04692824227469308 [0m(-0.0066597734178815565)
|
| 112 |
+
| > avg_loss_text_ce:[92m 0.02401351955320154 [0m(-0.0007712784090212407)
|
| 113 |
+
| > avg_loss_mel_ce:[92m 3.0045580523354665 [0m(-0.7372611045837405)
|
| 114 |
+
| > avg_loss:[92m 3.028571551186698 [0m(-0.7380324091230119)
|
| 115 |
+
|
| 116 |
+
> BEST MODEL : tts/run/xtts-finetuning-May-28-2025_05+56AM-0000000/best_model_1773.pth
|
| 117 |
+
|
| 118 |
+
[4m[1m > EPOCH: 2/1000[0m
|
| 119 |
+
--> tts/run/xtts-finetuning-May-28-2025_05+56AM-0000000
|
| 120 |
+
|
| 121 |
+
[1m > TRAINING (2025-05-28 06:15:44) [0m
|
| 122 |
+
|
| 123 |
+
[1m --> TIME: 2025-05-28 06:17:57 -- STEP: 227/1773 -- GLOBAL_STEP: 2000[0m
|
| 124 |
+
| > loss_text_ce: 0.02634984627366066 (0.02347405103039374)
|
| 125 |
+
| > loss_mel_ce: 2.9743587970733643 (nan)
|
| 126 |
+
| > loss: 0.04688607156276703 (nan)
|
| 127 |
+
| > current_lr: 5e-06
|
| 128 |
+
| > step_time: 0.5701 (0.5613324474133058)
|
| 129 |
+
| > loader_time: 0.024 (0.01852047705965421)
|
| 130 |
+
|
| 131 |
+
|
| 132 |
+
[1m --> TIME: 2025-05-28 06:20:24 -- STEP: 477/1773 -- GLOBAL_STEP: 2250[0m
|
| 133 |
+
| > loss_text_ce: 0.019006451591849327 (0.0235798091996672)
|
| 134 |
+
| > loss_mel_ce: 2.9675517082214355 (nan)
|
| 135 |
+
| > loss: 0.046664971858263016 (nan)
|
| 136 |
+
| > current_lr: 5e-06
|
| 137 |
+
| > step_time: 0.5472 (0.5630224965653334)
|
| 138 |
+
| > loader_time: 0.0132 (0.01841085212035749)
|
| 139 |
+
|
| 140 |
+
|
| 141 |
+
[1m --> TIME: 2025-05-28 06:22:49 -- STEP: 727/1773 -- GLOBAL_STEP: 2500[0m
|
| 142 |
+
| > loss_text_ce: 0.019247999414801598 (0.023683539267624253)
|
| 143 |
+
| > loss_mel_ce: 4.047469615936279 (nan)
|
| 144 |
+
| > loss: 0.06354246288537979 (nan)
|
| 145 |
+
| > current_lr: 5e-06
|
| 146 |
+
| > step_time: 0.6674 (0.5622831017311867)
|
| 147 |
+
| > loader_time: 0.0246 (0.018256587536167733)
|
| 148 |
+
|
| 149 |
+
|
| 150 |
+
[1m --> TIME: 2025-05-28 06:25:14 -- STEP: 977/1773 -- GLOBAL_STEP: 2750[0m
|
| 151 |
+
| > loss_text_ce: 0.019411535933613777 (0.023644839759492817)
|
| 152 |
+
| > loss_mel_ce: 2.550281047821045 (nan)
|
| 153 |
+
| > loss: 0.040151447057724 (nan)
|
| 154 |
+
| > current_lr: 5e-06
|
| 155 |
+
| > step_time: 0.6017 (0.5617610405265011)
|
| 156 |
+
| > loader_time: 0.0232 (0.018097628590516446)
|
| 157 |
+
|
| 158 |
+
|
| 159 |
+
[1m --> TIME: 2025-05-28 06:27:39 -- STEP: 1227/1773 -- GLOBAL_STEP: 3000[0m
|
| 160 |
+
| > loss_text_ce: 0.018359506502747536 (0.023614716428440572)
|
| 161 |
+
| > loss_mel_ce: 2.8278863430023193 (nan)
|
| 162 |
+
| > loss: 0.0444725900888443 (nan)
|
| 163 |
+
| > current_lr: 5e-06
|
| 164 |
+
| > step_time: 0.5529 (0.5614671928666929)
|
| 165 |
+
| > loader_time: 0.014 (0.01804757506845438)
|
| 166 |
+
|
| 167 |
+
|
| 168 |
+
[1m --> TIME: 2025-05-28 06:30:05 -- STEP: 1477/1773 -- GLOBAL_STEP: 3250[0m
|
| 169 |
+
| > loss_text_ce: 0.024424398317933083 (0.02365642989756753)
|
| 170 |
+
| > loss_mel_ce: 2.85211443901062 (nan)
|
| 171 |
+
| > loss: 0.04494591802358627 (nan)
|
| 172 |
+
| > current_lr: 5e-06
|
| 173 |
+
| > step_time: 0.5568 (0.5615530244869896)
|
| 174 |
+
| > loader_time: 0.0127 (0.01798885096225209)
|
| 175 |
+
|
| 176 |
+
|
| 177 |
+
[1m --> TIME: 2025-05-28 06:32:30 -- STEP: 1727/1773 -- GLOBAL_STEP: 3500[0m
|
| 178 |
+
| > loss_text_ce: 0.013657196424901485 (0.023524502163389426)
|
| 179 |
+
| > loss_mel_ce: 2.7606756687164307 (nan)
|
| 180 |
+
| > loss: 0.043348949402570724 (nan)
|
| 181 |
+
| > grad_norm: 0 (0.0)
|
| 182 |
+
| > current_lr: 5e-06
|
| 183 |
+
| > step_time: 0.5705 (0.5614900450935659)
|
| 184 |
+
| > loader_time: 0.011 (0.01798361260282141)
|
| 185 |
+
|
| 186 |
+
|
| 187 |
+
[1m > EVALUATION [0m
|
| 188 |
+
|
| 189 |
+
|
| 190 |
+
[1m--> EVAL PERFORMANCE[0m
|
| 191 |
+
| > avg_loader_time:[91m 0.047074174880981444 [0m(+0.0001459326062883612)
|
| 192 |
+
| > avg_loss_text_ce:[92m 0.023535797904644693 [0m(-0.0004777216485568475)
|
| 193 |
+
| > avg_loss_mel_ce:[92m 2.8784927810941423 [0m(-0.12606527124132416)
|
| 194 |
+
| > avg_loss:[92m 2.9020285640444072 [0m(-0.12654298714229073)
|
| 195 |
+
|
| 196 |
+
> BEST MODEL : tts/run/xtts-finetuning-May-28-2025_05+56AM-0000000/best_model_3546.pth
|
| 197 |
+
! Run is kept in tts/run/xtts-finetuning-May-28-2025_05+56AM-0000000
|