Upload folder using huggingface_hub
Browse files- .gitattributes +14 -0
- LJSpeech6/config.yml +21 -118
- LJSpeech6/epoch_2nd_00153/epoch_2nd_00159.wav +3 -0
- LJSpeech6/epoch_2nd_00153/epoch_2nd_00160.wav +3 -0
- LJSpeech6/epoch_2nd_00153/epoch_2nd_00161.wav +3 -0
- LJSpeech6/epoch_2nd_00153/epoch_2nd_00162.wav +3 -0
- LJSpeech6/epoch_2nd_00153/epoch_2nd_00163.wav +3 -0
- LJSpeech6/epoch_2nd_00153/epoch_2nd_00164.wav +3 -0
- LJSpeech6/epoch_2nd_00153/epoch_2nd_00165.wav +3 -0
- LJSpeech6/epoch_2nd_00153/epoch_2nd_00166.wav +3 -0
- LJSpeech6/epoch_2nd_00153/epoch_2nd_00167.wav +3 -0
- LJSpeech6/epoch_2nd_00153/epoch_2nd_00168.wav +3 -0
- LJSpeech6/epoch_2nd_00153/epoch_2nd_00169.wav +3 -0
- LJSpeech6/epoch_2nd_00153/epoch_2nd_00170.wav +3 -0
- LJSpeech6/epoch_2nd_00153/epoch_2nd_00171.wav +3 -0
- LJSpeech6/epoch_2nd_00153/epoch_2nd_00172.wav +3 -0
- LJSpeech6/epoch_2nd_00171.pth +3 -0
- LJSpeech6/tensorboard/events.out.tfevents.1745035998.faa8f2272929.163.0 +3 -0
- LJSpeech6/train.log +73 -71
.gitattributes
CHANGED
|
@@ -441,3 +441,17 @@ LJSpeech6/epoch_2nd_00141/epoch_2nd_00152.wav filter=lfs diff=lfs merge=lfs -tex
|
|
| 441 |
LJSpeech6/epoch_2nd_00141/epoch_2nd_00153.wav filter=lfs diff=lfs merge=lfs -text
|
| 442 |
LJSpeech6/epoch_2nd_00141/epoch_2nd_00154.wav filter=lfs diff=lfs merge=lfs -text
|
| 443 |
LJSpeech6/epoch_2nd_00141/epoch_2nd_00155.wav filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 441 |
LJSpeech6/epoch_2nd_00141/epoch_2nd_00153.wav filter=lfs diff=lfs merge=lfs -text
|
| 442 |
LJSpeech6/epoch_2nd_00141/epoch_2nd_00154.wav filter=lfs diff=lfs merge=lfs -text
|
| 443 |
LJSpeech6/epoch_2nd_00141/epoch_2nd_00155.wav filter=lfs diff=lfs merge=lfs -text
|
| 444 |
+
LJSpeech6/epoch_2nd_00153/epoch_2nd_00159.wav filter=lfs diff=lfs merge=lfs -text
|
| 445 |
+
LJSpeech6/epoch_2nd_00153/epoch_2nd_00160.wav filter=lfs diff=lfs merge=lfs -text
|
| 446 |
+
LJSpeech6/epoch_2nd_00153/epoch_2nd_00161.wav filter=lfs diff=lfs merge=lfs -text
|
| 447 |
+
LJSpeech6/epoch_2nd_00153/epoch_2nd_00162.wav filter=lfs diff=lfs merge=lfs -text
|
| 448 |
+
LJSpeech6/epoch_2nd_00153/epoch_2nd_00163.wav filter=lfs diff=lfs merge=lfs -text
|
| 449 |
+
LJSpeech6/epoch_2nd_00153/epoch_2nd_00164.wav filter=lfs diff=lfs merge=lfs -text
|
| 450 |
+
LJSpeech6/epoch_2nd_00153/epoch_2nd_00165.wav filter=lfs diff=lfs merge=lfs -text
|
| 451 |
+
LJSpeech6/epoch_2nd_00153/epoch_2nd_00166.wav filter=lfs diff=lfs merge=lfs -text
|
| 452 |
+
LJSpeech6/epoch_2nd_00153/epoch_2nd_00167.wav filter=lfs diff=lfs merge=lfs -text
|
| 453 |
+
LJSpeech6/epoch_2nd_00153/epoch_2nd_00168.wav filter=lfs diff=lfs merge=lfs -text
|
| 454 |
+
LJSpeech6/epoch_2nd_00153/epoch_2nd_00169.wav filter=lfs diff=lfs merge=lfs -text
|
| 455 |
+
LJSpeech6/epoch_2nd_00153/epoch_2nd_00170.wav filter=lfs diff=lfs merge=lfs -text
|
| 456 |
+
LJSpeech6/epoch_2nd_00153/epoch_2nd_00171.wav filter=lfs diff=lfs merge=lfs -text
|
| 457 |
+
LJSpeech6/epoch_2nd_00153/epoch_2nd_00172.wav filter=lfs diff=lfs merge=lfs -text
|
LJSpeech6/config.yml
CHANGED
|
@@ -1,118 +1,21 @@
|
|
| 1 |
-
|
| 2 |
-
|
| 3 |
-
|
| 4 |
-
|
| 5 |
-
|
| 6 |
-
|
| 7 |
-
|
| 8 |
-
|
| 9 |
-
|
| 10 |
-
|
| 11 |
-
|
| 12 |
-
|
| 13 |
-
|
| 14 |
-
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
|
| 18 |
-
|
| 19 |
-
|
| 20 |
-
|
| 21 |
-
|
| 22 |
-
data_params:
|
| 23 |
-
train_data: "Data/train_list.txt"
|
| 24 |
-
val_data: "Data/val_list.txt"
|
| 25 |
-
root_path: "/content/StyleTTS2/Dataset"
|
| 26 |
-
OOD_data: "Data/OOD_texts.txt"
|
| 27 |
-
min_length: 50 # sample until texts with this size are obtained for OOD texts
|
| 28 |
-
|
| 29 |
-
preprocess_params:
|
| 30 |
-
sr: 24000
|
| 31 |
-
spect_params:
|
| 32 |
-
n_fft: 2048
|
| 33 |
-
win_length: 1200
|
| 34 |
-
hop_length: 300
|
| 35 |
-
|
| 36 |
-
model_params:
|
| 37 |
-
multispeaker: false
|
| 38 |
-
|
| 39 |
-
dim_in: 64
|
| 40 |
-
hidden_dim: 512
|
| 41 |
-
max_conv_dim: 512
|
| 42 |
-
n_layer: 3
|
| 43 |
-
n_mels: 80
|
| 44 |
-
|
| 45 |
-
n_token: 185 # number of phoneme tokens
|
| 46 |
-
max_dur: 50 # maximum duration of a single phoneme
|
| 47 |
-
style_dim: 128 # style vector size
|
| 48 |
-
|
| 49 |
-
dropout: 0.2
|
| 50 |
-
|
| 51 |
-
# config for decoder
|
| 52 |
-
decoder:
|
| 53 |
-
type: 'istftnet' # either hifigan or istftnet
|
| 54 |
-
resblock_kernel_sizes: [3,7,11]
|
| 55 |
-
upsample_rates : [10, 6]
|
| 56 |
-
upsample_initial_channel: 512
|
| 57 |
-
resblock_dilation_sizes: [[1,3,5], [1,3,5], [1,3,5]]
|
| 58 |
-
upsample_kernel_sizes: [20, 12]
|
| 59 |
-
gen_istft_n_fft: 20
|
| 60 |
-
gen_istft_hop_size: 5
|
| 61 |
-
|
| 62 |
-
# speech language model config
|
| 63 |
-
slm:
|
| 64 |
-
model: 'microsoft/wavlm-base-plus'
|
| 65 |
-
sr: 16000 # sampling rate of SLM
|
| 66 |
-
hidden: 768 # hidden size of SLM
|
| 67 |
-
nlayers: 13 # number of layers of SLM
|
| 68 |
-
initial_channel: 64 # initial channels of SLM discriminator head
|
| 69 |
-
|
| 70 |
-
# style diffusion model config
|
| 71 |
-
diffusion:
|
| 72 |
-
embedding_mask_proba: 0.1
|
| 73 |
-
# transformer config
|
| 74 |
-
transformer:
|
| 75 |
-
num_layers: 3
|
| 76 |
-
num_heads: 8
|
| 77 |
-
head_features: 64
|
| 78 |
-
multiplier: 2
|
| 79 |
-
|
| 80 |
-
# diffusion distribution config
|
| 81 |
-
dist:
|
| 82 |
-
sigma_data: 0.2 # placeholder for estimate_sigma_data set to false
|
| 83 |
-
estimate_sigma_data: true # estimate sigma_data from the current batch if set to true
|
| 84 |
-
mean: -3.0
|
| 85 |
-
std: 1.0
|
| 86 |
-
|
| 87 |
-
loss_params:
|
| 88 |
-
lambda_mel: 5. # mel reconstruction loss
|
| 89 |
-
lambda_gen: 1. # generator loss
|
| 90 |
-
lambda_slm: 1. # slm feature matching loss
|
| 91 |
-
|
| 92 |
-
lambda_mono: 1. # monotonic alignment loss (1st stage, TMA)
|
| 93 |
-
lambda_s2s: 1. # sequence-to-sequence loss (1st stage, TMA)
|
| 94 |
-
TMA_epoch: 2 # TMA starting epoch (1st stage)
|
| 95 |
-
|
| 96 |
-
lambda_F0: 1. # F0 reconstruction loss (2nd stage)
|
| 97 |
-
lambda_norm: 1. # norm reconstruction loss (2nd stage)
|
| 98 |
-
lambda_dur: 1. # duration loss (2nd stage)
|
| 99 |
-
lambda_ce: 20. # duration predictor probability output CE loss (2nd stage)
|
| 100 |
-
lambda_sty: 1. # style reconstruction loss (2nd stage)
|
| 101 |
-
lambda_diff: 1. # score matching loss (2nd stage)
|
| 102 |
-
|
| 103 |
-
diff_epoch: 0 # style diffusion starting epoch (2nd stage)
|
| 104 |
-
joint_epoch: 0 # joint training starting epoch (2nd stage)
|
| 105 |
-
|
| 106 |
-
optimizer_params:
|
| 107 |
-
lr: 0.0001 # general learning rate
|
| 108 |
-
bert_lr: 0.00001 # learning rate for PLBERT
|
| 109 |
-
ft_lr: 0.00001 # learning rate for acoustic modules
|
| 110 |
-
|
| 111 |
-
slmadv_params:
|
| 112 |
-
min_len: 400 # minimum length of samples
|
| 113 |
-
max_len: 500 # maximum length of samples
|
| 114 |
-
batch_percentage: 1 # to prevent out of memory, only use half of the original batch size
|
| 115 |
-
iter: 10 # update the discriminator every this iterations of generator update
|
| 116 |
-
thresh: 5 # gradient norm above which the gradient is scaled
|
| 117 |
-
scale: 0.01 # gradient scaling factor for predictors from SLM discriminators
|
| 118 |
-
sig: 1.5 # sigma for differentiable duration modeling
|
|
|
|
| 1 |
+
{ASR_config: Utils/ASR/config.yml, ASR_path: Utils/ASR/epoch_00080.pth, F0_path: Utils/JDC/bst.t7,
|
| 2 |
+
PLBERT_dir: Utils/PLBERT/, batch_size: 4, data_params: {OOD_data: Data/OOD_texts.txt,
|
| 3 |
+
min_length: 50, root_path: /content/StyleTTS2/Dataset, train_data: Data/train_list.txt,
|
| 4 |
+
val_data: Data/val_list.txt}, device: cuda, epochs_1st: 20, epochs_2nd: 300, first_stage_path: first_stage.pth,
|
| 5 |
+
load_only_params: false, log_dir: Models/LJSpeech6, log_interval: 10, loss_params: {
|
| 6 |
+
TMA_epoch: 10000, diff_epoch: 0, joint_epoch: 0, lambda_F0: 1.0, lambda_ce: 20.0,
|
| 7 |
+
lambda_diff: 1.0, lambda_dur: 1.0, lambda_gen: 1.0, lambda_mel: 5.0, lambda_mono: 1.0,
|
| 8 |
+
lambda_norm: 1.0, lambda_s2s: 1.0, lambda_slm: 1.0, lambda_sty: 1.0}, max_len: 295,
|
| 9 |
+
model_params: {decoder: {gen_istft_hop_size: 5, gen_istft_n_fft: 20, resblock_dilation_sizes: [
|
| 10 |
+
[1, 3, 5], [1, 3, 5], [1, 3, 5]], resblock_kernel_sizes: [3, 7, 11], type: istftnet,
|
| 11 |
+
upsample_initial_channel: 512, upsample_kernel_sizes: [20, 12], upsample_rates: [
|
| 12 |
+
10, 6]}, diffusion: {dist: {estimate_sigma_data: true, mean: -3.0, sigma_data: .nan,
|
| 13 |
+
std: 1.0}, embedding_mask_proba: 0.1, transformer: {head_features: 64, multiplier: 2,
|
| 14 |
+
num_heads: 8, num_layers: 3}}, dim_in: 64, dropout: 0.2, hidden_dim: 512,
|
| 15 |
+
max_conv_dim: 512, max_dur: 50, multispeaker: false, n_layer: 3, n_mels: 80, n_token: 185,
|
| 16 |
+
slm: {hidden: 768, initial_channel: 64, model: microsoft/wavlm-base-plus, nlayers: 13,
|
| 17 |
+
sr: 16000}, style_dim: 128}, optimizer_params: {bert_lr: 1.0e-05, ft_lr: 1.0e-05,
|
| 18 |
+
lr: 0.0001}, preprocess_params: {spect_params: {hop_length: 300, n_fft: 2048,
|
| 19 |
+
win_length: 1200}, sr: 24000}, pretrained_model: LJSpeech6/epoch_2nd_00152.pth,
|
| 20 |
+
save_freq: 1, second_stage_load_pretrained: true, slmadv_params: {batch_percentage: 1,
|
| 21 |
+
iter: 10, max_len: 500, min_len: 400, scale: 0.01, sig: 1.5, thresh: 5}}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
LJSpeech6/epoch_2nd_00153/epoch_2nd_00159.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9e41e882cd6dd54b64c1dae59a244d479ac0cffdce76df9affc371a2e400e0db
|
| 3 |
+
size 967144
|
LJSpeech6/epoch_2nd_00153/epoch_2nd_00160.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:039c311d2001b4e6e644d553de28765c1fe2a962c92b4487bcbf2ea5348450cf
|
| 3 |
+
size 895144
|
LJSpeech6/epoch_2nd_00153/epoch_2nd_00161.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e3c9766a7111e3e1a730637927936e98060ec57def4643c52275111473683df4
|
| 3 |
+
size 944344
|
LJSpeech6/epoch_2nd_00153/epoch_2nd_00162.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ff33e56385e2a27a3328bd5b148dc32e3f0dc33d7c6d92982cdcd9efc371a712
|
| 3 |
+
size 933544
|
LJSpeech6/epoch_2nd_00153/epoch_2nd_00163.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3a63ff23b19d6b36ba0f16195de9165200e6cd3df159bf26c3e6b790d6797317
|
| 3 |
+
size 901144
|
LJSpeech6/epoch_2nd_00153/epoch_2nd_00164.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:46073a8777d2019b7b1e65d91cbdd2f3026477e832ab2e20ad8c577cbd354986
|
| 3 |
+
size 927544
|
LJSpeech6/epoch_2nd_00153/epoch_2nd_00165.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:69b161c4e1d159451816c476c2a253ab0a5126cfc3821d921f123e60c7e7fb85
|
| 3 |
+
size 886744
|
LJSpeech6/epoch_2nd_00153/epoch_2nd_00166.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c0b79968267defbd21cfcc7647d482eb22e96ee6bbb2d294ef42159781986e76
|
| 3 |
+
size 880744
|
LJSpeech6/epoch_2nd_00153/epoch_2nd_00167.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f55af91324df6f61e82dec16ebb22ceb46b864c82710fe890e77e7157272c7a5
|
| 3 |
+
size 889144
|
LJSpeech6/epoch_2nd_00153/epoch_2nd_00168.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:81b181d36dc14fd82df9a8fc72c5148d12d5428331fca5a18a70400d0a7ed50c
|
| 3 |
+
size 897544
|
LJSpeech6/epoch_2nd_00153/epoch_2nd_00169.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:98a5cf0baed1bd9501bbc1c44e1ae42ddf2b790c4d2693a56f2aa8d481933b65
|
| 3 |
+
size 859144
|
LJSpeech6/epoch_2nd_00153/epoch_2nd_00170.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9f33f6bca977146ba853720a0d5eaaa14cd6a3122d2a37f3e383f40f2834d75c
|
| 3 |
+
size 883144
|
LJSpeech6/epoch_2nd_00153/epoch_2nd_00171.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c40d22d313ffe67687824d97acdbf515f0dc3ccc3ea501aa0f33ce637a9ac24e
|
| 3 |
+
size 889144
|
LJSpeech6/epoch_2nd_00153/epoch_2nd_00172.wav
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:edc39e1af77666a3547e7756332b485de1dc45a86c4061eff0acd9c7f396083a
|
| 3 |
+
size 923944
|
LJSpeech6/epoch_2nd_00171.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ce547c96db7572be98884a93fc8b5193785c9b0214228c9ff495befd0abfe218
|
| 3 |
+
size 1279664731
|
LJSpeech6/tensorboard/events.out.tfevents.1745035998.faa8f2272929.163.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f4d34c8135929f55627bc70d23b3ab4ad6aefc1a8895f65244eec0b3ec00c7e7
|
| 3 |
+
size 29883
|
LJSpeech6/train.log
CHANGED
|
@@ -1,71 +1,73 @@
|
|
| 1 |
-
INFO:2025-04-
|
| 2 |
-
INFO:2025-04-
|
| 3 |
-
INFO:2025-04-
|
| 4 |
-
INFO:2025-04-
|
| 5 |
-
INFO:2025-04-
|
| 6 |
-
INFO:2025-04-
|
| 7 |
-
INFO:2025-04-
|
| 8 |
-
INFO:2025-04-
|
| 9 |
-
INFO:2025-04-
|
| 10 |
-
INFO:2025-04-
|
| 11 |
-
INFO:2025-04-
|
| 12 |
-
INFO:2025-04-
|
| 13 |
-
INFO:2025-04-
|
| 14 |
-
INFO:2025-04-
|
| 15 |
-
INFO:2025-04-
|
| 16 |
-
INFO:2025-04-
|
| 17 |
-
INFO:2025-04-
|
| 18 |
-
INFO:2025-04-
|
| 19 |
-
INFO:2025-04-
|
| 20 |
-
INFO:2025-04-
|
| 21 |
-
INFO:2025-04-
|
| 22 |
-
INFO:2025-04-
|
| 23 |
-
INFO:2025-04-
|
| 24 |
-
INFO:2025-04-
|
| 25 |
-
INFO:2025-04-
|
| 26 |
-
INFO:2025-04-
|
| 27 |
-
INFO:2025-04-
|
| 28 |
-
INFO:2025-04-
|
| 29 |
-
INFO:2025-04-
|
| 30 |
-
INFO:2025-04-
|
| 31 |
-
INFO:2025-04-
|
| 32 |
-
INFO:2025-04-
|
| 33 |
-
INFO:2025-04-
|
| 34 |
-
INFO:2025-04-
|
| 35 |
-
INFO:2025-04-
|
| 36 |
-
INFO:2025-04-
|
| 37 |
-
INFO:2025-04-
|
| 38 |
-
INFO:2025-04-
|
| 39 |
-
INFO:2025-04-
|
| 40 |
-
INFO:2025-04-
|
| 41 |
-
INFO:2025-04-
|
| 42 |
-
INFO:2025-04-
|
| 43 |
-
INFO:2025-04-
|
| 44 |
-
INFO:2025-04-
|
| 45 |
-
INFO:2025-04-
|
| 46 |
-
INFO:2025-04-
|
| 47 |
-
INFO:2025-04-
|
| 48 |
-
INFO:2025-04-
|
| 49 |
-
INFO:2025-04-
|
| 50 |
-
INFO:2025-04-
|
| 51 |
-
INFO:2025-04-
|
| 52 |
-
INFO:2025-04-
|
| 53 |
-
INFO:2025-04-
|
| 54 |
-
INFO:2025-04-
|
| 55 |
-
INFO:2025-04-
|
| 56 |
-
INFO:2025-04-
|
| 57 |
-
INFO:2025-04-
|
| 58 |
-
INFO:2025-04-
|
| 59 |
-
INFO:2025-04-
|
| 60 |
-
INFO:2025-04-
|
| 61 |
-
INFO:2025-04-
|
| 62 |
-
INFO:2025-04-
|
| 63 |
-
INFO:2025-04-
|
| 64 |
-
INFO:2025-04-
|
| 65 |
-
INFO:2025-04-
|
| 66 |
-
INFO:2025-04-
|
| 67 |
-
INFO:2025-04-
|
| 68 |
-
INFO:2025-04-
|
| 69 |
-
INFO:2025-04-
|
| 70 |
-
INFO:2025-04-
|
| 71 |
-
INFO:2025-04-
|
|
|
|
|
|
|
|
|
| 1 |
+
INFO:2025-04-19 04:15:05,046: Epoch [160/300], Step [10/51], Loss: 0.61517, Disc Loss: 1.92252, Dur Loss: 1.55691, CE Loss: 0.08648, Norm Loss: 1.38256, F0 Loss: 3.81567, LM Loss: 2.43242, Gen Loss: 18.96918
|
| 2 |
+
INFO:2025-04-19 04:15:48,235: Epoch [160/300], Step [20/51], Loss: 0.60993, Disc Loss: 1.49337, Dur Loss: 1.35902, CE Loss: 0.08046, Norm Loss: 1.11747, F0 Loss: 4.96475, LM Loss: 2.48484, Gen Loss: 21.47816
|
| 3 |
+
INFO:2025-04-19 04:16:32,973: Epoch [160/300], Step [30/51], Loss: 0.61335, Disc Loss: 1.47696, Dur Loss: 1.33782, CE Loss: 0.08089, Norm Loss: 1.25249, F0 Loss: 5.07798, LM Loss: 2.50116, Gen Loss: 21.97194
|
| 4 |
+
INFO:2025-04-19 04:17:18,073: Epoch [160/300], Step [40/51], Loss: 0.63685, Disc Loss: 1.62457, Dur Loss: 1.46981, CE Loss: 0.07561, Norm Loss: 1.29563, F0 Loss: 4.51344, LM Loss: 2.55204, Gen Loss: 22.20159
|
| 5 |
+
INFO:2025-04-19 04:18:03,718: Epoch [160/300], Step [50/51], Loss: 0.62789, Disc Loss: 1.49179, Dur Loss: 1.57644, CE Loss: 0.08975, Norm Loss: 1.06932, F0 Loss: 4.46852, LM Loss: 2.45306, Gen Loss: 20.97278
|
| 6 |
+
INFO:2025-04-19 04:19:07,787: Epoch [161/300], Step [10/51], Loss: 0.60670, Disc Loss: 1.46279, Dur Loss: 1.41943, CE Loss: 0.07864, Norm Loss: 1.30389, F0 Loss: 5.29274, LM Loss: 2.45573, Gen Loss: 21.27743
|
| 7 |
+
INFO:2025-04-19 04:19:53,665: Epoch [161/300], Step [20/51], Loss: 0.61867, Disc Loss: 1.49302, Dur Loss: 1.33735, CE Loss: 0.08404, Norm Loss: 1.17754, F0 Loss: 4.37794, LM Loss: 2.46385, Gen Loss: 21.87168
|
| 8 |
+
INFO:2025-04-19 04:20:39,998: Epoch [161/300], Step [30/51], Loss: 0.65000, Disc Loss: 1.41411, Dur Loss: 1.53492, CE Loss: 0.08392, Norm Loss: 1.59322, F0 Loss: 5.80972, LM Loss: 2.71142, Gen Loss: 21.80326
|
| 9 |
+
INFO:2025-04-19 04:21:26,180: Epoch [161/300], Step [40/51], Loss: 0.59957, Disc Loss: 1.32646, Dur Loss: 1.23625, CE Loss: 0.07225, Norm Loss: 1.11785, F0 Loss: 4.67796, LM Loss: 2.47672, Gen Loss: 22.66970
|
| 10 |
+
INFO:2025-04-19 04:22:11,653: Epoch [161/300], Step [50/51], Loss: 0.60770, Disc Loss: 1.72576, Dur Loss: 1.60120, CE Loss: 0.09378, Norm Loss: 1.07553, F0 Loss: 3.35470, LM Loss: 2.22938, Gen Loss: 19.99246
|
| 11 |
+
INFO:2025-04-19 04:23:06,829: Epoch [162/300], Step [10/51], Loss: 0.59449, Disc Loss: 1.90182, Dur Loss: 1.28500, CE Loss: 0.08296, Norm Loss: 1.05478, F0 Loss: 3.02183, LM Loss: 2.29578, Gen Loss: 19.80863
|
| 12 |
+
INFO:2025-04-19 04:23:53,142: Epoch [162/300], Step [20/51], Loss: 0.59802, Disc Loss: 1.46899, Dur Loss: 1.34292, CE Loss: 0.07726, Norm Loss: 1.35144, F0 Loss: 5.03227, LM Loss: 2.52279, Gen Loss: 22.35310
|
| 13 |
+
INFO:2025-04-19 04:24:38,987: Epoch [162/300], Step [30/51], Loss: 0.62154, Disc Loss: 1.61988, Dur Loss: 1.37634, CE Loss: 0.07608, Norm Loss: 1.33180, F0 Loss: 4.81954, LM Loss: 2.42235, Gen Loss: 20.07341
|
| 14 |
+
INFO:2025-04-19 04:25:25,517: Epoch [162/300], Step [40/51], Loss: 0.60578, Disc Loss: 1.51734, Dur Loss: 1.49730, CE Loss: 0.08584, Norm Loss: 1.22663, F0 Loss: 5.24245, LM Loss: 2.40472, Gen Loss: 21.51778
|
| 15 |
+
INFO:2025-04-19 04:26:10,376: Epoch [162/300], Step [50/51], Loss: 0.61167, Disc Loss: 1.42829, Dur Loss: 1.42879, CE Loss: 0.07514, Norm Loss: 1.28961, F0 Loss: 4.68594, LM Loss: 2.45155, Gen Loss: 22.14058
|
| 16 |
+
INFO:2025-04-19 04:27:05,580: Epoch [163/300], Step [10/51], Loss: 0.62750, Disc Loss: 1.34135, Dur Loss: 1.32606, CE Loss: 0.07860, Norm Loss: 1.51975, F0 Loss: 5.08676, LM Loss: 2.62305, Gen Loss: 22.47166
|
| 17 |
+
INFO:2025-04-19 04:27:50,971: Epoch [163/300], Step [20/51], Loss: 0.60802, Disc Loss: 1.43816, Dur Loss: 1.32021, CE Loss: 0.07684, Norm Loss: 1.38140, F0 Loss: 4.45177, LM Loss: 2.61883, Gen Loss: 21.66393
|
| 18 |
+
INFO:2025-04-19 04:28:38,168: Epoch [163/300], Step [30/51], Loss: 0.63305, Disc Loss: 1.53755, Dur Loss: 1.45575, CE Loss: 0.08497, Norm Loss: 1.65932, F0 Loss: 4.93481, LM Loss: 2.71309, Gen Loss: 21.49944
|
| 19 |
+
INFO:2025-04-19 04:29:23,465: Epoch [163/300], Step [40/51], Loss: 0.60740, Disc Loss: 1.59993, Dur Loss: 1.43148, CE Loss: 0.08958, Norm Loss: 1.37899, F0 Loss: 5.06723, LM Loss: 2.45739, Gen Loss: 21.89077
|
| 20 |
+
INFO:2025-04-19 04:30:08,997: Epoch [163/300], Step [50/51], Loss: 0.60843, Disc Loss: 1.35980, Dur Loss: 1.39992, CE Loss: 0.08484, Norm Loss: 1.40781, F0 Loss: 5.63226, LM Loss: 2.55356, Gen Loss: 21.87129
|
| 21 |
+
INFO:2025-04-19 04:31:04,457: Epoch [164/300], Step [10/51], Loss: 0.62133, Disc Loss: 1.68892, Dur Loss: 1.72061, CE Loss: 0.11014, Norm Loss: 1.54832, F0 Loss: 5.59631, LM Loss: 2.52896, Gen Loss: 20.35616
|
| 22 |
+
INFO:2025-04-19 04:31:50,734: Epoch [164/300], Step [20/51], Loss: 0.63549, Disc Loss: 1.62723, Dur Loss: 1.28060, CE Loss: 0.06983, Norm Loss: 0.90681, F0 Loss: 4.72420, LM Loss: 2.41472, Gen Loss: 21.63782
|
| 23 |
+
INFO:2025-04-19 04:32:36,467: Epoch [164/300], Step [30/51], Loss: 0.63139, Disc Loss: 1.54208, Dur Loss: 1.45628, CE Loss: 0.07978, Norm Loss: 1.08133, F0 Loss: 3.85765, LM Loss: 2.40322, Gen Loss: 21.09708
|
| 24 |
+
INFO:2025-04-19 04:33:22,705: Epoch [164/300], Step [40/51], Loss: 0.63169, Disc Loss: 1.36097, Dur Loss: 1.26678, CE Loss: 0.07231, Norm Loss: 0.98724, F0 Loss: 4.42687, LM Loss: 2.42803, Gen Loss: 22.98399
|
| 25 |
+
INFO:2025-04-19 04:34:07,829: Epoch [164/300], Step [50/51], Loss: 0.59573, Disc Loss: 1.68488, Dur Loss: 1.49981, CE Loss: 0.09279, Norm Loss: 1.24993, F0 Loss: 3.90103, LM Loss: 2.46843, Gen Loss: 20.56778
|
| 26 |
+
INFO:2025-04-19 04:35:01,755: Epoch [165/300], Step [10/51], Loss: 0.61903, Disc Loss: 1.44144, Dur Loss: 1.44196, CE Loss: 0.07952, Norm Loss: 1.48892, F0 Loss: 5.69431, LM Loss: 2.55424, Gen Loss: 22.60819
|
| 27 |
+
INFO:2025-04-19 04:35:47,445: Epoch [165/300], Step [20/51], Loss: 0.61879, Disc Loss: 1.34513, Dur Loss: 1.49256, CE Loss: 0.08996, Norm Loss: 1.07571, F0 Loss: 3.97867, LM Loss: 2.43482, Gen Loss: 22.01490
|
| 28 |
+
INFO:2025-04-19 04:36:33,222: Epoch [165/300], Step [30/51], Loss: 0.61091, Disc Loss: 1.73543, Dur Loss: 1.30992, CE Loss: 0.07911, Norm Loss: 1.19167, F0 Loss: 3.84372, LM Loss: 2.48396, Gen Loss: 19.93699
|
| 29 |
+
INFO:2025-04-19 04:37:20,703: Epoch [165/300], Step [40/51], Loss: 0.64627, Disc Loss: 1.60724, Dur Loss: 1.35634, CE Loss: 0.08046, Norm Loss: 1.67871, F0 Loss: 5.72094, LM Loss: 2.69885, Gen Loss: 20.55851
|
| 30 |
+
INFO:2025-04-19 04:38:05,969: Epoch [165/300], Step [50/51], Loss: 0.64173, Disc Loss: 1.43282, Dur Loss: 1.36469, CE Loss: 0.07329, Norm Loss: 1.03667, F0 Loss: 4.10319, LM Loss: 2.35580, Gen Loss: 22.62904
|
| 31 |
+
INFO:2025-04-19 04:39:00,973: Epoch [166/300], Step [10/51], Loss: 0.61561, Disc Loss: 1.87253, Dur Loss: 1.37466, CE Loss: 0.07242, Norm Loss: 1.09918, F0 Loss: 3.50388, LM Loss: 2.39875, Gen Loss: 19.25333
|
| 32 |
+
INFO:2025-04-19 04:39:47,216: Epoch [166/300], Step [20/51], Loss: 0.61092, Disc Loss: 1.46170, Dur Loss: 1.52468, CE Loss: 0.08490, Norm Loss: 1.16615, F0 Loss: 4.27633, LM Loss: 2.47646, Gen Loss: 21.77772
|
| 33 |
+
INFO:2025-04-19 04:40:34,063: Epoch [166/300], Step [30/51], Loss: 0.62690, Disc Loss: 1.41709, Dur Loss: 1.48991, CE Loss: 0.08842, Norm Loss: 1.06107, F0 Loss: 5.37812, LM Loss: 2.54743, Gen Loss: 21.78275
|
| 34 |
+
INFO:2025-04-19 04:41:19,672: Epoch [166/300], Step [40/51], Loss: 0.60128, Disc Loss: 1.55830, Dur Loss: 1.53070, CE Loss: 0.09550, Norm Loss: 1.32685, F0 Loss: 5.06974, LM Loss: 2.33649, Gen Loss: 21.30290
|
| 35 |
+
INFO:2025-04-19 04:42:04,684: Epoch [166/300], Step [50/51], Loss: 0.60404, Disc Loss: 1.60350, Dur Loss: 1.49562, CE Loss: 0.08529, Norm Loss: 0.97772, F0 Loss: 4.22367, LM Loss: 2.45343, Gen Loss: 21.76797
|
| 36 |
+
INFO:2025-04-19 04:42:59,268: Epoch [167/300], Step [10/51], Loss: 0.62652, Disc Loss: 1.59606, Dur Loss: 1.45061, CE Loss: 0.08515, Norm Loss: 1.33134, F0 Loss: 4.95937, LM Loss: 2.55921, Gen Loss: 21.70388
|
| 37 |
+
INFO:2025-04-19 04:43:45,587: Epoch [167/300], Step [20/51], Loss: 0.62103, Disc Loss: 1.58069, Dur Loss: 1.20130, CE Loss: 0.06234, Norm Loss: 1.04771, F0 Loss: 3.92831, LM Loss: 2.44679, Gen Loss: 20.98036
|
| 38 |
+
INFO:2025-04-19 04:44:31,758: Epoch [167/300], Step [30/51], Loss: 0.62442, Disc Loss: 1.39970, Dur Loss: 1.44425, CE Loss: 0.07671, Norm Loss: 1.23832, F0 Loss: 5.44705, LM Loss: 2.44023, Gen Loss: 21.81331
|
| 39 |
+
INFO:2025-04-19 04:45:18,043: Epoch [167/300], Step [40/51], Loss: 0.60617, Disc Loss: 1.52243, Dur Loss: 1.31025, CE Loss: 0.07606, Norm Loss: 1.00748, F0 Loss: 3.43133, LM Loss: 2.27823, Gen Loss: 21.74066
|
| 40 |
+
INFO:2025-04-19 04:46:03,759: Epoch [167/300], Step [50/51], Loss: 0.61772, Disc Loss: 1.87143, Dur Loss: 1.37577, CE Loss: 0.07955, Norm Loss: 1.26402, F0 Loss: 4.39598, LM Loss: 2.41786, Gen Loss: 19.30043
|
| 41 |
+
INFO:2025-04-19 04:46:59,292: Epoch [168/300], Step [10/51], Loss: 0.62147, Disc Loss: 1.71858, Dur Loss: 1.62945, CE Loss: 0.08499, Norm Loss: 1.15239, F0 Loss: 4.21905, LM Loss: 2.42597, Gen Loss: 20.11444
|
| 42 |
+
INFO:2025-04-19 04:47:45,193: Epoch [168/300], Step [20/51], Loss: 0.61465, Disc Loss: 1.57598, Dur Loss: 1.31845, CE Loss: 0.07853, Norm Loss: 1.47415, F0 Loss: 6.14085, LM Loss: 2.49344, Gen Loss: 21.22107
|
| 43 |
+
INFO:2025-04-19 04:48:31,226: Epoch [168/300], Step [30/51], Loss: 0.61503, Disc Loss: 1.67921, Dur Loss: 1.45270, CE Loss: 0.08215, Norm Loss: 1.04884, F0 Loss: 4.25235, LM Loss: 2.59562, Gen Loss: 20.62066
|
| 44 |
+
INFO:2025-04-19 04:49:17,588: Epoch [168/300], Step [40/51], Loss: 0.60911, Disc Loss: 1.51642, Dur Loss: 1.33139, CE Loss: 0.07627, Norm Loss: 1.37323, F0 Loss: 5.07897, LM Loss: 2.48718, Gen Loss: 22.49410
|
| 45 |
+
INFO:2025-04-19 04:50:03,812: Epoch [168/300], Step [50/51], Loss: 0.63618, Disc Loss: 1.48828, Dur Loss: 1.40381, CE Loss: 0.07788, Norm Loss: 1.24790, F0 Loss: 4.26553, LM Loss: 2.46186, Gen Loss: 21.34413
|
| 46 |
+
INFO:2025-04-19 04:50:59,462: Epoch [169/300], Step [10/51], Loss: 0.62396, Disc Loss: 1.63711, Dur Loss: 1.54662, CE Loss: 0.09918, Norm Loss: 1.26701, F0 Loss: 4.57993, LM Loss: 2.33836, Gen Loss: 21.63839
|
| 47 |
+
INFO:2025-04-19 04:51:44,966: Epoch [169/300], Step [20/51], Loss: 0.59275, Disc Loss: 1.67358, Dur Loss: 1.38058, CE Loss: 0.07983, Norm Loss: 1.03154, F0 Loss: 4.17188, LM Loss: 2.37218, Gen Loss: 20.75462
|
| 48 |
+
INFO:2025-04-19 04:52:30,990: Epoch [169/300], Step [30/51], Loss: 0.62219, Disc Loss: 1.41014, Dur Loss: 1.42184, CE Loss: 0.08616, Norm Loss: 1.19359, F0 Loss: 4.57184, LM Loss: 2.47231, Gen Loss: 22.32779
|
| 49 |
+
INFO:2025-04-19 04:53:16,891: Epoch [169/300], Step [40/51], Loss: 0.62640, Disc Loss: 1.78813, Dur Loss: 1.51392, CE Loss: 0.07849, Norm Loss: 1.19950, F0 Loss: 3.81316, LM Loss: 2.39753, Gen Loss: 19.69670
|
| 50 |
+
INFO:2025-04-19 04:54:03,204: Epoch [169/300], Step [50/51], Loss: 0.61943, Disc Loss: 1.83870, Dur Loss: 1.35278, CE Loss: 0.08010, Norm Loss: 1.64048, F0 Loss: 6.35125, LM Loss: 2.64920, Gen Loss: 20.35914
|
| 51 |
+
INFO:2025-04-19 04:54:58,126: Epoch [170/300], Step [10/51], Loss: 0.61891, Disc Loss: 1.72500, Dur Loss: 1.35739, CE Loss: 0.07986, Norm Loss: 1.51714, F0 Loss: 4.93424, LM Loss: 2.56292, Gen Loss: 20.74751
|
| 52 |
+
INFO:2025-04-19 04:55:44,311: Epoch [170/300], Step [20/51], Loss: 0.62611, Disc Loss: 1.38058, Dur Loss: 1.34940, CE Loss: 0.07707, Norm Loss: 1.22574, F0 Loss: 5.10193, LM Loss: 2.52421, Gen Loss: 22.91841
|
| 53 |
+
INFO:2025-04-19 04:56:30,460: Epoch [170/300], Step [30/51], Loss: 0.58498, Disc Loss: 1.50048, Dur Loss: 1.32411, CE Loss: 0.07293, Norm Loss: 1.43924, F0 Loss: 7.07482, LM Loss: 2.56567, Gen Loss: 21.71566
|
| 54 |
+
INFO:2025-04-19 04:57:17,001: Epoch [170/300], Step [40/51], Loss: 0.62961, Disc Loss: 1.42340, Dur Loss: 1.19090, CE Loss: 0.06325, Norm Loss: 1.14504, F0 Loss: 4.27976, LM Loss: 2.44166, Gen Loss: 21.77470
|
| 55 |
+
INFO:2025-04-19 04:58:02,469: Epoch [170/300], Step [50/51], Loss: 0.59666, Disc Loss: 1.46970, Dur Loss: 1.31525, CE Loss: 0.07444, Norm Loss: 0.92673, F0 Loss: 3.70055, LM Loss: 2.45975, Gen Loss: 21.11955
|
| 56 |
+
INFO:2025-04-19 04:58:58,224: Epoch [171/300], Step [10/51], Loss: 0.62408, Disc Loss: 1.42369, Dur Loss: 1.24555, CE Loss: 0.06815, Norm Loss: 1.17164, F0 Loss: 5.08718, LM Loss: 2.51255, Gen Loss: 22.01647
|
| 57 |
+
INFO:2025-04-19 04:59:44,416: Epoch [171/300], Step [20/51], Loss: 0.59565, Disc Loss: 1.73362, Dur Loss: 1.49764, CE Loss: 0.07781, Norm Loss: 1.23784, F0 Loss: 4.39716, LM Loss: 2.28983, Gen Loss: 20.21249
|
| 58 |
+
INFO:2025-04-19 05:00:29,580: Epoch [171/300], Step [30/51], Loss: 0.62176, Disc Loss: 1.83607, Dur Loss: 1.48329, CE Loss: 0.07947, Norm Loss: 1.16265, F0 Loss: 4.27442, LM Loss: 2.46767, Gen Loss: 20.39859
|
| 59 |
+
INFO:2025-04-19 05:01:15,938: Epoch [171/300], Step [40/51], Loss: 0.65056, Disc Loss: 1.58127, Dur Loss: 1.27457, CE Loss: 0.07442, Norm Loss: 0.89664, F0 Loss: 3.90889, LM Loss: 2.42980, Gen Loss: 21.19791
|
| 60 |
+
INFO:2025-04-19 05:02:01,569: Epoch [171/300], Step [50/51], Loss: 0.59957, Disc Loss: 1.43492, Dur Loss: 1.25705, CE Loss: 0.07790, Norm Loss: 0.97807, F0 Loss: 3.93567, LM Loss: 2.42881, Gen Loss: 21.76072
|
| 61 |
+
INFO:2025-04-19 05:02:57,116: Epoch [172/300], Step [10/51], Loss: 0.61922, Disc Loss: 1.65966, Dur Loss: 1.38347, CE Loss: 0.07313, Norm Loss: 1.23423, F0 Loss: 4.85919, LM Loss: 2.43578, Gen Loss: 20.42861
|
| 62 |
+
INFO:2025-04-19 05:03:43,403: Epoch [172/300], Step [20/51], Loss: 0.63013, Disc Loss: 1.33103, Dur Loss: 1.54024, CE Loss: 0.09194, Norm Loss: 1.39962, F0 Loss: 5.42717, LM Loss: 2.65523, Gen Loss: 22.13144
|
| 63 |
+
INFO:2025-04-19 05:04:29,786: Epoch [172/300], Step [30/51], Loss: 0.61422, Disc Loss: 1.64699, Dur Loss: 1.41179, CE Loss: 0.07872, Norm Loss: 1.04084, F0 Loss: 4.04186, LM Loss: 2.41387, Gen Loss: 21.04662
|
| 64 |
+
INFO:2025-04-19 05:05:15,645: Epoch [172/300], Step [40/51], Loss: 0.59061, Disc Loss: 2.09528, Dur Loss: 1.54836, CE Loss: 0.08906, Norm Loss: 1.30317, F0 Loss: 4.47688, LM Loss: 2.41751, Gen Loss: 18.12341
|
| 65 |
+
INFO:2025-04-19 05:06:01,620: Epoch [172/300], Step [50/51], Loss: 0.60923, Disc Loss: 1.55096, Dur Loss: 1.31466, CE Loss: 0.06740, Norm Loss: 0.78310, F0 Loss: 3.73894, LM Loss: 2.43043, Gen Loss: 21.06657
|
| 66 |
+
INFO:2025-04-19 05:06:55,731: Epoch [173/300], Step [10/51], Loss: 0.57302, Disc Loss: 1.71308, Dur Loss: 1.52903, CE Loss: 0.08405, Norm Loss: 0.86238, F0 Loss: 3.99504, LM Loss: 2.38599, Gen Loss: 20.35188
|
| 67 |
+
INFO:2025-04-19 05:07:40,782: Epoch [173/300], Step [20/51], Loss: 0.61565, Disc Loss: 1.69760, Dur Loss: 1.29815, CE Loss: 0.07395, Norm Loss: 1.48307, F0 Loss: 4.94854, LM Loss: 2.62610, Gen Loss: 20.37193
|
| 68 |
+
INFO:2025-04-19 05:08:27,762: Epoch [173/300], Step [30/51], Loss: 0.62734, Disc Loss: 1.46203, Dur Loss: 1.49120, CE Loss: 0.08097, Norm Loss: 1.19858, F0 Loss: 4.82632, LM Loss: 2.64832, Gen Loss: 21.97266
|
| 69 |
+
INFO:2025-04-19 05:09:14,387: Epoch [173/300], Step [40/51], Loss: 0.58837, Disc Loss: 1.48593, Dur Loss: 1.38515, CE Loss: 0.07578, Norm Loss: 1.05050, F0 Loss: 4.22295, LM Loss: 2.47323, Gen Loss: 22.07851
|
| 70 |
+
INFO:2025-04-19 05:10:00,165: Epoch [173/300], Step [50/51], Loss: 0.59451, Disc Loss: 1.55293, Dur Loss: 1.30678, CE Loss: 0.07633, Norm Loss: 0.94875, F0 Loss: 3.54441, LM Loss: 2.45517, Gen Loss: 20.98925
|
| 71 |
+
INFO:2025-04-19 05:10:55,003: Epoch [174/300], Step [10/51], Loss: 0.64914, Disc Loss: 1.43084, Dur Loss: 1.34072, CE Loss: 0.07033, Norm Loss: 0.86738, F0 Loss: 3.53228, LM Loss: 2.36271, Gen Loss: 22.93789
|
| 72 |
+
INFO:2025-04-19 05:11:41,435: Epoch [174/300], Step [20/51], Loss: 0.60155, Disc Loss: 1.65688, Dur Loss: 1.34231, CE Loss: 0.06600, Norm Loss: 1.22494, F0 Loss: 4.79476, LM Loss: 2.47413, Gen Loss: 21.28828
|
| 73 |
+
INFO:2025-04-19 05:12:26,703: Epoch [174/300], Step [30/51], Loss: 0.58329, Disc Loss: 1.57699, Dur Loss: 1.41537, CE Loss: 0.08184, Norm Loss: 1.12702, F0 Loss: 4.09413, LM Loss: 2.39012, Gen Loss: 20.87862
|