SirAB commited on
Commit
d7fd12e
·
verified ·
1 Parent(s): 281fc32

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -441,3 +441,17 @@ LJSpeech6/epoch_2nd_00141/epoch_2nd_00152.wav filter=lfs diff=lfs merge=lfs -tex
441
  LJSpeech6/epoch_2nd_00141/epoch_2nd_00153.wav filter=lfs diff=lfs merge=lfs -text
442
  LJSpeech6/epoch_2nd_00141/epoch_2nd_00154.wav filter=lfs diff=lfs merge=lfs -text
443
  LJSpeech6/epoch_2nd_00141/epoch_2nd_00155.wav filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
441
  LJSpeech6/epoch_2nd_00141/epoch_2nd_00153.wav filter=lfs diff=lfs merge=lfs -text
442
  LJSpeech6/epoch_2nd_00141/epoch_2nd_00154.wav filter=lfs diff=lfs merge=lfs -text
443
  LJSpeech6/epoch_2nd_00141/epoch_2nd_00155.wav filter=lfs diff=lfs merge=lfs -text
444
+ LJSpeech6/epoch_2nd_00153/epoch_2nd_00159.wav filter=lfs diff=lfs merge=lfs -text
445
+ LJSpeech6/epoch_2nd_00153/epoch_2nd_00160.wav filter=lfs diff=lfs merge=lfs -text
446
+ LJSpeech6/epoch_2nd_00153/epoch_2nd_00161.wav filter=lfs diff=lfs merge=lfs -text
447
+ LJSpeech6/epoch_2nd_00153/epoch_2nd_00162.wav filter=lfs diff=lfs merge=lfs -text
448
+ LJSpeech6/epoch_2nd_00153/epoch_2nd_00163.wav filter=lfs diff=lfs merge=lfs -text
449
+ LJSpeech6/epoch_2nd_00153/epoch_2nd_00164.wav filter=lfs diff=lfs merge=lfs -text
450
+ LJSpeech6/epoch_2nd_00153/epoch_2nd_00165.wav filter=lfs diff=lfs merge=lfs -text
451
+ LJSpeech6/epoch_2nd_00153/epoch_2nd_00166.wav filter=lfs diff=lfs merge=lfs -text
452
+ LJSpeech6/epoch_2nd_00153/epoch_2nd_00167.wav filter=lfs diff=lfs merge=lfs -text
453
+ LJSpeech6/epoch_2nd_00153/epoch_2nd_00168.wav filter=lfs diff=lfs merge=lfs -text
454
+ LJSpeech6/epoch_2nd_00153/epoch_2nd_00169.wav filter=lfs diff=lfs merge=lfs -text
455
+ LJSpeech6/epoch_2nd_00153/epoch_2nd_00170.wav filter=lfs diff=lfs merge=lfs -text
456
+ LJSpeech6/epoch_2nd_00153/epoch_2nd_00171.wav filter=lfs diff=lfs merge=lfs -text
457
+ LJSpeech6/epoch_2nd_00153/epoch_2nd_00172.wav filter=lfs diff=lfs merge=lfs -text
LJSpeech6/config.yml CHANGED
@@ -1,118 +1,21 @@
1
- log_dir: "Models/LJSpeech6"
2
- first_stage_path: "first_stage.pth"
3
- save_freq: 1
4
- log_interval: 10
5
- device: "cuda"
6
- epochs_1st: 2000 # number of epochs for first stage training (pre-training)
7
- epochs_2nd: 300 # number of peochs for second stage training (joint training)
8
- # first epoch = 3x650
9
- # second epoch = 2x440
10
- # second stage 1st epoch = 2x300
11
- batch_size: 2
12
- max_len: 550 # maximum number of frames
13
- pretrained_model: "LJSpeech6/epoch_2nd_00151.pth"
14
- second_stage_load_pretrained: true # set to true if the pre-trained model is for 2nd stage
15
- load_only_params: false # set to true if do not want to load epoch numbers and optimizer parameters
16
-
17
- F0_path: "Utils/JDC/bst.t7"
18
- ASR_config: "Utils/ASR/config.yml"
19
- ASR_path: "Utils/ASR/epoch_00080.pth"
20
- PLBERT_dir: 'Utils/PLBERT/'
21
-
22
- data_params:
23
- train_data: "Data/train_list.txt"
24
- val_data: "Data/val_list.txt"
25
- root_path: "/content/StyleTTS2/Dataset"
26
- OOD_data: "Data/OOD_texts.txt"
27
- min_length: 50 # sample until texts with this size are obtained for OOD texts
28
-
29
- preprocess_params:
30
- sr: 24000
31
- spect_params:
32
- n_fft: 2048
33
- win_length: 1200
34
- hop_length: 300
35
-
36
- model_params:
37
- multispeaker: false
38
-
39
- dim_in: 64
40
- hidden_dim: 512
41
- max_conv_dim: 512
42
- n_layer: 3
43
- n_mels: 80
44
-
45
- n_token: 185 # number of phoneme tokens
46
- max_dur: 50 # maximum duration of a single phoneme
47
- style_dim: 128 # style vector size
48
-
49
- dropout: 0.2
50
-
51
- # config for decoder
52
- decoder:
53
- type: 'istftnet' # either hifigan or istftnet
54
- resblock_kernel_sizes: [3,7,11]
55
- upsample_rates : [10, 6]
56
- upsample_initial_channel: 512
57
- resblock_dilation_sizes: [[1,3,5], [1,3,5], [1,3,5]]
58
- upsample_kernel_sizes: [20, 12]
59
- gen_istft_n_fft: 20
60
- gen_istft_hop_size: 5
61
-
62
- # speech language model config
63
- slm:
64
- model: 'microsoft/wavlm-base-plus'
65
- sr: 16000 # sampling rate of SLM
66
- hidden: 768 # hidden size of SLM
67
- nlayers: 13 # number of layers of SLM
68
- initial_channel: 64 # initial channels of SLM discriminator head
69
-
70
- # style diffusion model config
71
- diffusion:
72
- embedding_mask_proba: 0.1
73
- # transformer config
74
- transformer:
75
- num_layers: 3
76
- num_heads: 8
77
- head_features: 64
78
- multiplier: 2
79
-
80
- # diffusion distribution config
81
- dist:
82
- sigma_data: 0.2 # placeholder for estimate_sigma_data set to false
83
- estimate_sigma_data: true # estimate sigma_data from the current batch if set to true
84
- mean: -3.0
85
- std: 1.0
86
-
87
- loss_params:
88
- lambda_mel: 5. # mel reconstruction loss
89
- lambda_gen: 1. # generator loss
90
- lambda_slm: 1. # slm feature matching loss
91
-
92
- lambda_mono: 1. # monotonic alignment loss (1st stage, TMA)
93
- lambda_s2s: 1. # sequence-to-sequence loss (1st stage, TMA)
94
- TMA_epoch: 2 # TMA starting epoch (1st stage)
95
-
96
- lambda_F0: 1. # F0 reconstruction loss (2nd stage)
97
- lambda_norm: 1. # norm reconstruction loss (2nd stage)
98
- lambda_dur: 1. # duration loss (2nd stage)
99
- lambda_ce: 20. # duration predictor probability output CE loss (2nd stage)
100
- lambda_sty: 1. # style reconstruction loss (2nd stage)
101
- lambda_diff: 1. # score matching loss (2nd stage)
102
-
103
- diff_epoch: 0 # style diffusion starting epoch (2nd stage)
104
- joint_epoch: 0 # joint training starting epoch (2nd stage)
105
-
106
- optimizer_params:
107
- lr: 0.0001 # general learning rate
108
- bert_lr: 0.00001 # learning rate for PLBERT
109
- ft_lr: 0.00001 # learning rate for acoustic modules
110
-
111
- slmadv_params:
112
- min_len: 400 # minimum length of samples
113
- max_len: 500 # maximum length of samples
114
- batch_percentage: 1 # to prevent out of memory, only use half of the original batch size
115
- iter: 10 # update the discriminator every this iterations of generator update
116
- thresh: 5 # gradient norm above which the gradient is scaled
117
- scale: 0.01 # gradient scaling factor for predictors from SLM discriminators
118
- sig: 1.5 # sigma for differentiable duration modeling
 
1
+ {ASR_config: Utils/ASR/config.yml, ASR_path: Utils/ASR/epoch_00080.pth, F0_path: Utils/JDC/bst.t7,
2
+ PLBERT_dir: Utils/PLBERT/, batch_size: 4, data_params: {OOD_data: Data/OOD_texts.txt,
3
+ min_length: 50, root_path: /content/StyleTTS2/Dataset, train_data: Data/train_list.txt,
4
+ val_data: Data/val_list.txt}, device: cuda, epochs_1st: 20, epochs_2nd: 300, first_stage_path: first_stage.pth,
5
+ load_only_params: false, log_dir: Models/LJSpeech6, log_interval: 10, loss_params: {
6
+ TMA_epoch: 10000, diff_epoch: 0, joint_epoch: 0, lambda_F0: 1.0, lambda_ce: 20.0,
7
+ lambda_diff: 1.0, lambda_dur: 1.0, lambda_gen: 1.0, lambda_mel: 5.0, lambda_mono: 1.0,
8
+ lambda_norm: 1.0, lambda_s2s: 1.0, lambda_slm: 1.0, lambda_sty: 1.0}, max_len: 295,
9
+ model_params: {decoder: {gen_istft_hop_size: 5, gen_istft_n_fft: 20, resblock_dilation_sizes: [
10
+ [1, 3, 5], [1, 3, 5], [1, 3, 5]], resblock_kernel_sizes: [3, 7, 11], type: istftnet,
11
+ upsample_initial_channel: 512, upsample_kernel_sizes: [20, 12], upsample_rates: [
12
+ 10, 6]}, diffusion: {dist: {estimate_sigma_data: true, mean: -3.0, sigma_data: .nan,
13
+ std: 1.0}, embedding_mask_proba: 0.1, transformer: {head_features: 64, multiplier: 2,
14
+ num_heads: 8, num_layers: 3}}, dim_in: 64, dropout: 0.2, hidden_dim: 512,
15
+ max_conv_dim: 512, max_dur: 50, multispeaker: false, n_layer: 3, n_mels: 80, n_token: 185,
16
+ slm: {hidden: 768, initial_channel: 64, model: microsoft/wavlm-base-plus, nlayers: 13,
17
+ sr: 16000}, style_dim: 128}, optimizer_params: {bert_lr: 1.0e-05, ft_lr: 1.0e-05,
18
+ lr: 0.0001}, preprocess_params: {spect_params: {hop_length: 300, n_fft: 2048,
19
+ win_length: 1200}, sr: 24000}, pretrained_model: LJSpeech6/epoch_2nd_00152.pth,
20
+ save_freq: 1, second_stage_load_pretrained: true, slmadv_params: {batch_percentage: 1,
21
+ iter: 10, max_len: 500, min_len: 400, scale: 0.01, sig: 1.5, thresh: 5}}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
LJSpeech6/epoch_2nd_00153/epoch_2nd_00159.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e41e882cd6dd54b64c1dae59a244d479ac0cffdce76df9affc371a2e400e0db
3
+ size 967144
LJSpeech6/epoch_2nd_00153/epoch_2nd_00160.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:039c311d2001b4e6e644d553de28765c1fe2a962c92b4487bcbf2ea5348450cf
3
+ size 895144
LJSpeech6/epoch_2nd_00153/epoch_2nd_00161.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3c9766a7111e3e1a730637927936e98060ec57def4643c52275111473683df4
3
+ size 944344
LJSpeech6/epoch_2nd_00153/epoch_2nd_00162.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff33e56385e2a27a3328bd5b148dc32e3f0dc33d7c6d92982cdcd9efc371a712
3
+ size 933544
LJSpeech6/epoch_2nd_00153/epoch_2nd_00163.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a63ff23b19d6b36ba0f16195de9165200e6cd3df159bf26c3e6b790d6797317
3
+ size 901144
LJSpeech6/epoch_2nd_00153/epoch_2nd_00164.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46073a8777d2019b7b1e65d91cbdd2f3026477e832ab2e20ad8c577cbd354986
3
+ size 927544
LJSpeech6/epoch_2nd_00153/epoch_2nd_00165.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69b161c4e1d159451816c476c2a253ab0a5126cfc3821d921f123e60c7e7fb85
3
+ size 886744
LJSpeech6/epoch_2nd_00153/epoch_2nd_00166.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0b79968267defbd21cfcc7647d482eb22e96ee6bbb2d294ef42159781986e76
3
+ size 880744
LJSpeech6/epoch_2nd_00153/epoch_2nd_00167.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f55af91324df6f61e82dec16ebb22ceb46b864c82710fe890e77e7157272c7a5
3
+ size 889144
LJSpeech6/epoch_2nd_00153/epoch_2nd_00168.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81b181d36dc14fd82df9a8fc72c5148d12d5428331fca5a18a70400d0a7ed50c
3
+ size 897544
LJSpeech6/epoch_2nd_00153/epoch_2nd_00169.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98a5cf0baed1bd9501bbc1c44e1ae42ddf2b790c4d2693a56f2aa8d481933b65
3
+ size 859144
LJSpeech6/epoch_2nd_00153/epoch_2nd_00170.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f33f6bca977146ba853720a0d5eaaa14cd6a3122d2a37f3e383f40f2834d75c
3
+ size 883144
LJSpeech6/epoch_2nd_00153/epoch_2nd_00171.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c40d22d313ffe67687824d97acdbf515f0dc3ccc3ea501aa0f33ce637a9ac24e
3
+ size 889144
LJSpeech6/epoch_2nd_00153/epoch_2nd_00172.wav ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edc39e1af77666a3547e7756332b485de1dc45a86c4061eff0acd9c7f396083a
3
+ size 923944
LJSpeech6/epoch_2nd_00171.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce547c96db7572be98884a93fc8b5193785c9b0214228c9ff495befd0abfe218
3
+ size 1279664731
LJSpeech6/tensorboard/events.out.tfevents.1745035998.faa8f2272929.163.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4d34c8135929f55627bc70d23b3ab4ad6aefc1a8895f65244eec0b3ec00c7e7
3
+ size 29883
LJSpeech6/train.log CHANGED
@@ -1,71 +1,73 @@
1
- INFO:2025-04-18 18:38:35,588: Epoch [152/2000], Step [10/103], Mel Loss: 0.57166, Gen Loss: 16.83132, Disc Loss: 2.43236, Mono Loss: 0.02193, S2S Loss: 9.49231, SLM Loss: 2.51258
2
- INFO:2025-04-18 18:39:04,034: Epoch [152/2000], Step [20/103], Mel Loss: 0.51311, Gen Loss: 19.25039, Disc Loss: 2.02201, Mono Loss: 0.02076, S2S Loss: 7.03681, SLM Loss: 2.47461
3
- INFO:2025-04-18 18:39:33,757: Epoch [152/2000], Step [30/103], Mel Loss: 0.54050, Gen Loss: 18.68884, Disc Loss: 1.87438, Mono Loss: 0.02688, S2S Loss: 7.64351, SLM Loss: 2.33585
4
- INFO:2025-04-18 18:40:03,206: Epoch [152/2000], Step [40/103], Mel Loss: 0.52822, Gen Loss: 15.79294, Disc Loss: 2.44520, Mono Loss: 0.01991, S2S Loss: 6.78955, SLM Loss: 2.47681
5
- INFO:2025-04-18 18:40:33,769: Epoch [152/2000], Step [50/103], Mel Loss: 0.52643, Gen Loss: 18.56133, Disc Loss: 2.18126, Mono Loss: 0.05364, S2S Loss: 7.77592, SLM Loss: 2.27586
6
- INFO:2025-04-18 18:41:01,775: Epoch [152/2000], Step [60/103], Mel Loss: 0.52666, Gen Loss: 19.28783, Disc Loss: 1.68320, Mono Loss: 0.03253, S2S Loss: 7.00832, SLM Loss: 2.63286
7
- INFO:2025-04-18 18:41:30,907: Epoch [152/2000], Step [70/103], Mel Loss: 0.56361, Gen Loss: 21.19762, Disc Loss: 1.56164, Mono Loss: 0.02666, S2S Loss: 6.74913, SLM Loss: 2.44062
8
- INFO:2025-04-18 18:41:58,641: Epoch [152/2000], Step [80/103], Mel Loss: 0.54404, Gen Loss: 21.41392, Disc Loss: 1.59368, Mono Loss: 0.02451, S2S Loss: 6.44588, SLM Loss: 2.43518
9
- INFO:2025-04-18 18:42:28,632: Epoch [152/2000], Step [90/103], Mel Loss: 0.54803, Gen Loss: 20.44135, Disc Loss: 1.60138, Mono Loss: 0.02891, S2S Loss: 6.20671, SLM Loss: 2.36022
10
- INFO:2025-04-18 18:42:55,979: Epoch [152/2000], Step [100/103], Mel Loss: 0.51911, Gen Loss: 19.80222, Disc Loss: 1.73530, Mono Loss: 0.03127, S2S Loss: 6.11233, SLM Loss: 2.41981
11
- INFO:2025-04-18 18:43:45,418: Epoch [153/2000], Step [10/103], Mel Loss: 0.53442, Gen Loss: 19.31555, Disc Loss: 1.92297, Mono Loss: 0.03056, S2S Loss: 5.87421, SLM Loss: 2.47313
12
- INFO:2025-04-18 18:44:11,518: Epoch [153/2000], Step [20/103], Mel Loss: 0.52366, Gen Loss: 17.86818, Disc Loss: 2.12986, Mono Loss: 0.01898, S2S Loss: 5.93414, SLM Loss: 2.31931
13
- INFO:2025-04-18 18:44:41,410: Epoch [153/2000], Step [30/103], Mel Loss: 0.51928, Gen Loss: 17.22287, Disc Loss: 2.20561, Mono Loss: 0.02667, S2S Loss: 5.58996, SLM Loss: 2.35181
14
- INFO:2025-04-18 18:45:12,509: Epoch [153/2000], Step [40/103], Mel Loss: 0.53984, Gen Loss: 16.62889, Disc Loss: 2.18520, Mono Loss: 0.01845, S2S Loss: 5.16366, SLM Loss: 2.40187
15
- INFO:2025-04-18 18:45:41,992: Epoch [153/2000], Step [50/103], Mel Loss: 0.50245, Gen Loss: 20.47397, Disc Loss: 1.65294, Mono Loss: 0.03118, S2S Loss: 5.36155, SLM Loss: 2.38660
16
- INFO:2025-04-18 18:46:12,663: Epoch [153/2000], Step [60/103], Mel Loss: 0.54741, Gen Loss: 19.52068, Disc Loss: 1.69318, Mono Loss: 0.03479, S2S Loss: 5.16356, SLM Loss: 2.36826
17
- INFO:2025-04-18 18:46:42,598: Epoch [153/2000], Step [70/103], Mel Loss: 0.53083, Gen Loss: 15.74927, Disc Loss: 2.27095, Mono Loss: 0.01687, S2S Loss: 5.33945, SLM Loss: 2.37491
18
- INFO:2025-04-18 18:47:12,098: Epoch [153/2000], Step [80/103], Mel Loss: 0.51057, Gen Loss: 17.70938, Disc Loss: 2.21856, Mono Loss: 0.01817, S2S Loss: 5.04235, SLM Loss: 2.32953
19
- INFO:2025-04-18 18:47:41,770: Epoch [153/2000], Step [90/103], Mel Loss: 0.54683, Gen Loss: 18.34101, Disc Loss: 2.31700, Mono Loss: 0.02445, S2S Loss: 4.88693, SLM Loss: 2.49539
20
- INFO:2025-04-18 18:48:10,147: Epoch [153/2000], Step [100/103], Mel Loss: 0.48502, Gen Loss: 20.11668, Disc Loss: 1.59937, Mono Loss: 0.02834, S2S Loss: 4.89747, SLM Loss: 2.33590
21
- INFO:2025-04-18 18:48:57,135: Epoch [154/2000], Step [10/103], Mel Loss: 0.51520, Gen Loss: 19.16153, Disc Loss: 1.78688, Mono Loss: 0.03530, S2S Loss: 5.05134, SLM Loss: 2.25493
22
- INFO:2025-04-18 18:49:28,176: Epoch [154/2000], Step [20/103], Mel Loss: 0.51373, Gen Loss: 20.48920, Disc Loss: 1.81395, Mono Loss: 0.03313, S2S Loss: 4.39037, SLM Loss: 2.36754
23
- INFO:2025-04-18 18:49:57,619: Epoch [154/2000], Step [30/103], Mel Loss: 0.51367, Gen Loss: 20.78860, Disc Loss: 1.55421, Mono Loss: 0.02592, S2S Loss: 4.66546, SLM Loss: 2.37533
24
- INFO:2025-04-18 18:50:26,162: Epoch [154/2000], Step [40/103], Mel Loss: 0.48840, Gen Loss: 18.98166, Disc Loss: 1.75466, Mono Loss: 0.02780, S2S Loss: 4.46077, SLM Loss: 2.35540
25
- INFO:2025-04-18 18:50:56,240: Epoch [154/2000], Step [50/103], Mel Loss: 0.51536, Gen Loss: 19.84693, Disc Loss: 1.55964, Mono Loss: 0.02405, S2S Loss: 4.34059, SLM Loss: 2.45846
26
- INFO:2025-04-18 18:51:23,055: Epoch [154/2000], Step [60/103], Mel Loss: 0.48315, Gen Loss: 19.20110, Disc Loss: 2.03561, Mono Loss: 0.02107, S2S Loss: 4.72636, SLM Loss: 2.51333
27
- INFO:2025-04-18 18:51:50,856: Epoch [154/2000], Step [70/103], Mel Loss: 0.49882, Gen Loss: 18.70953, Disc Loss: 1.99929, Mono Loss: 0.02237, S2S Loss: 4.20378, SLM Loss: 2.59645
28
- INFO:2025-04-18 18:52:20,380: Epoch [154/2000], Step [80/103], Mel Loss: 0.51202, Gen Loss: 21.04130, Disc Loss: 1.72185, Mono Loss: 0.03234, S2S Loss: 4.22647, SLM Loss: 2.57194
29
- INFO:2025-04-18 18:52:50,500: Epoch [154/2000], Step [90/103], Mel Loss: 0.53054, Gen Loss: 20.40186, Disc Loss: 1.60000, Mono Loss: 0.02985, S2S Loss: 3.72298, SLM Loss: 2.45289
30
- INFO:2025-04-18 18:53:20,814: Epoch [154/2000], Step [100/103], Mel Loss: 0.48371, Gen Loss: 17.23027, Disc Loss: 2.19072, Mono Loss: 0.04925, S2S Loss: 3.92045, SLM Loss: 2.14295
31
- INFO:2025-04-18 18:54:10,091: Epoch [155/2000], Step [10/103], Mel Loss: 0.53092, Gen Loss: 20.17961, Disc Loss: 1.74581, Mono Loss: 0.02264, S2S Loss: 3.88469, SLM Loss: 2.44163
32
- INFO:2025-04-18 18:54:39,300: Epoch [155/2000], Step [20/103], Mel Loss: 0.50307, Gen Loss: 20.08510, Disc Loss: 1.67222, Mono Loss: 0.02759, S2S Loss: 3.83373, SLM Loss: 2.68349
33
- INFO:2025-04-18 18:55:06,278: Epoch [155/2000], Step [30/103], Mel Loss: 0.48582, Gen Loss: 14.90438, Disc Loss: 2.54310, Mono Loss: 0.02342, S2S Loss: 3.81850, SLM Loss: 2.29407
34
- INFO:2025-04-18 18:55:34,677: Epoch [155/2000], Step [40/103], Mel Loss: 0.52121, Gen Loss: 17.88344, Disc Loss: 2.16958, Mono Loss: 0.02339, S2S Loss: 3.99454, SLM Loss: 2.22043
35
- INFO:2025-04-18 18:56:04,731: Epoch [155/2000], Step [50/103], Mel Loss: 0.51297, Gen Loss: 21.15893, Disc Loss: 1.59340, Mono Loss: 0.02427, S2S Loss: 3.91466, SLM Loss: 2.44757
36
- INFO:2025-04-18 18:56:33,168: Epoch [155/2000], Step [60/103], Mel Loss: 0.51488, Gen Loss: 20.56116, Disc Loss: 1.62387, Mono Loss: 0.07581, S2S Loss: 3.72628, SLM Loss: 2.53573
37
- INFO:2025-04-18 18:57:03,274: Epoch [155/2000], Step [70/103], Mel Loss: 0.51163, Gen Loss: 19.04642, Disc Loss: 1.97788, Mono Loss: 0.03178, S2S Loss: 3.92079, SLM Loss: 2.29763
38
- INFO:2025-04-18 18:57:32,256: Epoch [155/2000], Step [80/103], Mel Loss: 0.48156, Gen Loss: 17.80066, Disc Loss: 2.03975, Mono Loss: 0.09422, S2S Loss: 3.61530, SLM Loss: 2.08303
39
- INFO:2025-04-18 18:58:01,877: Epoch [155/2000], Step [90/103], Mel Loss: 0.49917, Gen Loss: 19.46781, Disc Loss: 1.97671, Mono Loss: 0.03159, S2S Loss: 3.89030, SLM Loss: 2.42135
40
- INFO:2025-04-18 18:58:31,197: Epoch [155/2000], Step [100/103], Mel Loss: 0.52138, Gen Loss: 17.70389, Disc Loss: 2.16771, Mono Loss: 0.02996, S2S Loss: 3.75878, SLM Loss: 2.22861
41
- INFO:2025-04-18 18:59:23,555: Epoch [156/2000], Step [10/103], Mel Loss: 0.53710, Gen Loss: 20.34490, Disc Loss: 1.64239, Mono Loss: 0.02757, S2S Loss: 3.57522, SLM Loss: 2.52982
42
- INFO:2025-04-18 18:59:54,688: Epoch [156/2000], Step [20/103], Mel Loss: 0.52933, Gen Loss: 20.51277, Disc Loss: 1.53158, Mono Loss: 0.02881, S2S Loss: 3.81691, SLM Loss: 2.33845
43
- INFO:2025-04-18 19:00:23,692: Epoch [156/2000], Step [30/103], Mel Loss: 0.48438, Gen Loss: 20.05314, Disc Loss: 1.67462, Mono Loss: 0.04089, S2S Loss: 3.74279, SLM Loss: 2.35771
44
- INFO:2025-04-18 19:00:52,510: Epoch [156/2000], Step [40/103], Mel Loss: 0.48853, Gen Loss: 20.01217, Disc Loss: 1.74529, Mono Loss: 0.02816, S2S Loss: 4.09301, SLM Loss: 2.16567
45
- INFO:2025-04-18 19:01:23,575: Epoch [156/2000], Step [50/103], Mel Loss: 0.50674, Gen Loss: 17.09556, Disc Loss: 2.15051, Mono Loss: 0.01906, S2S Loss: 3.72620, SLM Loss: 2.23461
46
- INFO:2025-04-18 19:01:53,482: Epoch [156/2000], Step [60/103], Mel Loss: 0.48390, Gen Loss: 19.77821, Disc Loss: 1.68133, Mono Loss: 0.02758, S2S Loss: 3.57257, SLM Loss: 2.36122
47
- INFO:2025-04-18 19:02:21,649: Epoch [156/2000], Step [70/103], Mel Loss: 0.50470, Gen Loss: 20.81038, Disc Loss: 1.52878, Mono Loss: 0.02826, S2S Loss: 3.44860, SLM Loss: 2.47844
48
- INFO:2025-04-18 19:02:50,768: Epoch [156/2000], Step [80/103], Mel Loss: 0.49182, Gen Loss: 20.31980, Disc Loss: 1.72196, Mono Loss: 0.03408, S2S Loss: 3.69133, SLM Loss: 2.24293
49
- INFO:2025-04-18 19:03:19,862: Epoch [156/2000], Step [90/103], Mel Loss: 0.49856, Gen Loss: 20.31413, Disc Loss: 1.73396, Mono Loss: 0.03428, S2S Loss: 3.30523, SLM Loss: 2.26049
50
- INFO:2025-04-18 19:03:48,753: Epoch [156/2000], Step [100/103], Mel Loss: 0.49692, Gen Loss: 21.32432, Disc Loss: 1.52412, Mono Loss: 0.03133, S2S Loss: 4.03286, SLM Loss: 2.26040
51
- INFO:2025-04-18 19:04:34,720: Epoch [157/2000], Step [10/103], Mel Loss: 0.49485, Gen Loss: 18.80105, Disc Loss: 1.92701, Mono Loss: 0.03421, S2S Loss: 3.49123, SLM Loss: 2.37880
52
- INFO:2025-04-18 19:05:06,454: Epoch [157/2000], Step [20/103], Mel Loss: 0.50804, Gen Loss: 20.36051, Disc Loss: 1.60502, Mono Loss: 0.03054, S2S Loss: 3.55885, SLM Loss: 2.56064
53
- INFO:2025-04-18 19:05:36,030: Epoch [157/2000], Step [30/103], Mel Loss: 0.49093, Gen Loss: 18.35858, Disc Loss: 2.38216, Mono Loss: 0.02324, S2S Loss: 3.31186, SLM Loss: 2.32921
54
- INFO:2025-04-18 19:06:04,886: Epoch [157/2000], Step [40/103], Mel Loss: 0.48760, Gen Loss: 21.24952, Disc Loss: 1.59863, Mono Loss: 0.03000, S2S Loss: 3.19579, SLM Loss: 2.35667
55
- INFO:2025-04-18 19:06:33,446: Epoch [157/2000], Step [50/103], Mel Loss: 0.48245, Gen Loss: 20.34266, Disc Loss: 1.61908, Mono Loss: 0.03342, S2S Loss: 3.19096, SLM Loss: 2.28566
56
- INFO:2025-04-18 19:07:01,777: Epoch [157/2000], Step [60/103], Mel Loss: 0.51093, Gen Loss: 18.96944, Disc Loss: 1.71149, Mono Loss: 0.02480, S2S Loss: 3.52545, SLM Loss: 2.19753
57
- INFO:2025-04-18 19:07:31,210: Epoch [157/2000], Step [70/103], Mel Loss: 0.50472, Gen Loss: 17.25173, Disc Loss: 2.15439, Mono Loss: 0.02294, S2S Loss: 3.23457, SLM Loss: 2.21443
58
- INFO:2025-04-18 19:07:58,028: Epoch [157/2000], Step [80/103], Mel Loss: 0.48006, Gen Loss: 19.54880, Disc Loss: 1.93674, Mono Loss: 0.08029, S2S Loss: 3.54700, SLM Loss: 2.11758
59
- INFO:2025-04-18 19:08:26,002: Epoch [157/2000], Step [90/103], Mel Loss: 0.50211, Gen Loss: 19.64443, Disc Loss: 1.63930, Mono Loss: 0.02386, S2S Loss: 3.22157, SLM Loss: 2.32133
60
- INFO:2025-04-18 19:08:55,848: Epoch [157/2000], Step [100/103], Mel Loss: 0.53463, Gen Loss: 20.53067, Disc Loss: 1.76513, Mono Loss: 0.03127, S2S Loss: 3.36474, SLM Loss: 2.30822
61
- INFO:2025-04-18 19:09:41,707: Epoch [158/2000], Step [10/103], Mel Loss: 0.46765, Gen Loss: 20.96872, Disc Loss: 1.57101, Mono Loss: 0.02670, S2S Loss: 2.94770, SLM Loss: 2.24093
62
- INFO:2025-04-18 19:10:09,406: Epoch [158/2000], Step [20/103], Mel Loss: 0.48450, Gen Loss: 18.45076, Disc Loss: 2.14726, Mono Loss: 0.02265, S2S Loss: 3.13564, SLM Loss: 2.27195
63
- INFO:2025-04-18 19:10:38,828: Epoch [158/2000], Step [30/103], Mel Loss: 0.49821, Gen Loss: 21.41924, Disc Loss: 1.54396, Mono Loss: 0.02530, S2S Loss: 3.42863, SLM Loss: 2.50890
64
- INFO:2025-04-18 19:11:06,831: Epoch [158/2000], Step [40/103], Mel Loss: 0.48802, Gen Loss: 21.30021, Disc Loss: 1.61593, Mono Loss: 0.02610, S2S Loss: 3.16726, SLM Loss: 2.40428
65
- INFO:2025-04-18 19:11:35,629: Epoch [158/2000], Step [50/103], Mel Loss: 0.51217, Gen Loss: 21.48411, Disc Loss: 1.50159, Mono Loss: 0.03615, S2S Loss: 3.14382, SLM Loss: 2.85411
66
- INFO:2025-04-18 19:12:03,773: Epoch [158/2000], Step [60/103], Mel Loss: 0.52681, Gen Loss: 20.43410, Disc Loss: 1.44218, Mono Loss: 0.02475, S2S Loss: 3.29054, SLM Loss: 2.37830
67
- INFO:2025-04-18 19:12:34,164: Epoch [158/2000], Step [70/103], Mel Loss: 0.53476, Gen Loss: 20.74742, Disc Loss: 1.51000, Mono Loss: 0.02568, S2S Loss: 3.54637, SLM Loss: 2.22783
68
- INFO:2025-04-18 19:13:01,928: Epoch [158/2000], Step [80/103], Mel Loss: 0.50271, Gen Loss: 18.49941, Disc Loss: 2.29595, Mono Loss: 0.02611, S2S Loss: 3.16774, SLM Loss: 2.44704
69
- INFO:2025-04-18 19:13:31,729: Epoch [158/2000], Step [90/103], Mel Loss: 0.52324, Gen Loss: 20.88881, Disc Loss: 1.56876, Mono Loss: 0.02646, S2S Loss: 3.25483, SLM Loss: 2.36448
70
- INFO:2025-04-18 19:14:01,835: Epoch [158/2000], Step [100/103], Mel Loss: 0.50185, Gen Loss: 20.03394, Disc Loss: 1.66773, Mono Loss: 0.03087, S2S Loss: 3.23687, SLM Loss: 2.06097
71
- INFO:2025-04-18 19:14:47,942: Epoch [159/2000], Step [10/103], Mel Loss: 0.47812, Gen Loss: 17.33949, Disc Loss: 2.04583, Mono Loss: 0.04497, S2S Loss: 2.56003, SLM Loss: 2.23770
 
 
 
1
+ INFO:2025-04-19 04:15:05,046: Epoch [160/300], Step [10/51], Loss: 0.61517, Disc Loss: 1.92252, Dur Loss: 1.55691, CE Loss: 0.08648, Norm Loss: 1.38256, F0 Loss: 3.81567, LM Loss: 2.43242, Gen Loss: 18.96918
2
+ INFO:2025-04-19 04:15:48,235: Epoch [160/300], Step [20/51], Loss: 0.60993, Disc Loss: 1.49337, Dur Loss: 1.35902, CE Loss: 0.08046, Norm Loss: 1.11747, F0 Loss: 4.96475, LM Loss: 2.48484, Gen Loss: 21.47816
3
+ INFO:2025-04-19 04:16:32,973: Epoch [160/300], Step [30/51], Loss: 0.61335, Disc Loss: 1.47696, Dur Loss: 1.33782, CE Loss: 0.08089, Norm Loss: 1.25249, F0 Loss: 5.07798, LM Loss: 2.50116, Gen Loss: 21.97194
4
+ INFO:2025-04-19 04:17:18,073: Epoch [160/300], Step [40/51], Loss: 0.63685, Disc Loss: 1.62457, Dur Loss: 1.46981, CE Loss: 0.07561, Norm Loss: 1.29563, F0 Loss: 4.51344, LM Loss: 2.55204, Gen Loss: 22.20159
5
+ INFO:2025-04-19 04:18:03,718: Epoch [160/300], Step [50/51], Loss: 0.62789, Disc Loss: 1.49179, Dur Loss: 1.57644, CE Loss: 0.08975, Norm Loss: 1.06932, F0 Loss: 4.46852, LM Loss: 2.45306, Gen Loss: 20.97278
6
+ INFO:2025-04-19 04:19:07,787: Epoch [161/300], Step [10/51], Loss: 0.60670, Disc Loss: 1.46279, Dur Loss: 1.41943, CE Loss: 0.07864, Norm Loss: 1.30389, F0 Loss: 5.29274, LM Loss: 2.45573, Gen Loss: 21.27743
7
+ INFO:2025-04-19 04:19:53,665: Epoch [161/300], Step [20/51], Loss: 0.61867, Disc Loss: 1.49302, Dur Loss: 1.33735, CE Loss: 0.08404, Norm Loss: 1.17754, F0 Loss: 4.37794, LM Loss: 2.46385, Gen Loss: 21.87168
8
+ INFO:2025-04-19 04:20:39,998: Epoch [161/300], Step [30/51], Loss: 0.65000, Disc Loss: 1.41411, Dur Loss: 1.53492, CE Loss: 0.08392, Norm Loss: 1.59322, F0 Loss: 5.80972, LM Loss: 2.71142, Gen Loss: 21.80326
9
+ INFO:2025-04-19 04:21:26,180: Epoch [161/300], Step [40/51], Loss: 0.59957, Disc Loss: 1.32646, Dur Loss: 1.23625, CE Loss: 0.07225, Norm Loss: 1.11785, F0 Loss: 4.67796, LM Loss: 2.47672, Gen Loss: 22.66970
10
+ INFO:2025-04-19 04:22:11,653: Epoch [161/300], Step [50/51], Loss: 0.60770, Disc Loss: 1.72576, Dur Loss: 1.60120, CE Loss: 0.09378, Norm Loss: 1.07553, F0 Loss: 3.35470, LM Loss: 2.22938, Gen Loss: 19.99246
11
+ INFO:2025-04-19 04:23:06,829: Epoch [162/300], Step [10/51], Loss: 0.59449, Disc Loss: 1.90182, Dur Loss: 1.28500, CE Loss: 0.08296, Norm Loss: 1.05478, F0 Loss: 3.02183, LM Loss: 2.29578, Gen Loss: 19.80863
12
+ INFO:2025-04-19 04:23:53,142: Epoch [162/300], Step [20/51], Loss: 0.59802, Disc Loss: 1.46899, Dur Loss: 1.34292, CE Loss: 0.07726, Norm Loss: 1.35144, F0 Loss: 5.03227, LM Loss: 2.52279, Gen Loss: 22.35310
13
+ INFO:2025-04-19 04:24:38,987: Epoch [162/300], Step [30/51], Loss: 0.62154, Disc Loss: 1.61988, Dur Loss: 1.37634, CE Loss: 0.07608, Norm Loss: 1.33180, F0 Loss: 4.81954, LM Loss: 2.42235, Gen Loss: 20.07341
14
+ INFO:2025-04-19 04:25:25,517: Epoch [162/300], Step [40/51], Loss: 0.60578, Disc Loss: 1.51734, Dur Loss: 1.49730, CE Loss: 0.08584, Norm Loss: 1.22663, F0 Loss: 5.24245, LM Loss: 2.40472, Gen Loss: 21.51778
15
+ INFO:2025-04-19 04:26:10,376: Epoch [162/300], Step [50/51], Loss: 0.61167, Disc Loss: 1.42829, Dur Loss: 1.42879, CE Loss: 0.07514, Norm Loss: 1.28961, F0 Loss: 4.68594, LM Loss: 2.45155, Gen Loss: 22.14058
16
+ INFO:2025-04-19 04:27:05,580: Epoch [163/300], Step [10/51], Loss: 0.62750, Disc Loss: 1.34135, Dur Loss: 1.32606, CE Loss: 0.07860, Norm Loss: 1.51975, F0 Loss: 5.08676, LM Loss: 2.62305, Gen Loss: 22.47166
17
+ INFO:2025-04-19 04:27:50,971: Epoch [163/300], Step [20/51], Loss: 0.60802, Disc Loss: 1.43816, Dur Loss: 1.32021, CE Loss: 0.07684, Norm Loss: 1.38140, F0 Loss: 4.45177, LM Loss: 2.61883, Gen Loss: 21.66393
18
+ INFO:2025-04-19 04:28:38,168: Epoch [163/300], Step [30/51], Loss: 0.63305, Disc Loss: 1.53755, Dur Loss: 1.45575, CE Loss: 0.08497, Norm Loss: 1.65932, F0 Loss: 4.93481, LM Loss: 2.71309, Gen Loss: 21.49944
19
+ INFO:2025-04-19 04:29:23,465: Epoch [163/300], Step [40/51], Loss: 0.60740, Disc Loss: 1.59993, Dur Loss: 1.43148, CE Loss: 0.08958, Norm Loss: 1.37899, F0 Loss: 5.06723, LM Loss: 2.45739, Gen Loss: 21.89077
20
+ INFO:2025-04-19 04:30:08,997: Epoch [163/300], Step [50/51], Loss: 0.60843, Disc Loss: 1.35980, Dur Loss: 1.39992, CE Loss: 0.08484, Norm Loss: 1.40781, F0 Loss: 5.63226, LM Loss: 2.55356, Gen Loss: 21.87129
21
+ INFO:2025-04-19 04:31:04,457: Epoch [164/300], Step [10/51], Loss: 0.62133, Disc Loss: 1.68892, Dur Loss: 1.72061, CE Loss: 0.11014, Norm Loss: 1.54832, F0 Loss: 5.59631, LM Loss: 2.52896, Gen Loss: 20.35616
22
+ INFO:2025-04-19 04:31:50,734: Epoch [164/300], Step [20/51], Loss: 0.63549, Disc Loss: 1.62723, Dur Loss: 1.28060, CE Loss: 0.06983, Norm Loss: 0.90681, F0 Loss: 4.72420, LM Loss: 2.41472, Gen Loss: 21.63782
23
+ INFO:2025-04-19 04:32:36,467: Epoch [164/300], Step [30/51], Loss: 0.63139, Disc Loss: 1.54208, Dur Loss: 1.45628, CE Loss: 0.07978, Norm Loss: 1.08133, F0 Loss: 3.85765, LM Loss: 2.40322, Gen Loss: 21.09708
24
+ INFO:2025-04-19 04:33:22,705: Epoch [164/300], Step [40/51], Loss: 0.63169, Disc Loss: 1.36097, Dur Loss: 1.26678, CE Loss: 0.07231, Norm Loss: 0.98724, F0 Loss: 4.42687, LM Loss: 2.42803, Gen Loss: 22.98399
25
+ INFO:2025-04-19 04:34:07,829: Epoch [164/300], Step [50/51], Loss: 0.59573, Disc Loss: 1.68488, Dur Loss: 1.49981, CE Loss: 0.09279, Norm Loss: 1.24993, F0 Loss: 3.90103, LM Loss: 2.46843, Gen Loss: 20.56778
26
+ INFO:2025-04-19 04:35:01,755: Epoch [165/300], Step [10/51], Loss: 0.61903, Disc Loss: 1.44144, Dur Loss: 1.44196, CE Loss: 0.07952, Norm Loss: 1.48892, F0 Loss: 5.69431, LM Loss: 2.55424, Gen Loss: 22.60819
27
+ INFO:2025-04-19 04:35:47,445: Epoch [165/300], Step [20/51], Loss: 0.61879, Disc Loss: 1.34513, Dur Loss: 1.49256, CE Loss: 0.08996, Norm Loss: 1.07571, F0 Loss: 3.97867, LM Loss: 2.43482, Gen Loss: 22.01490
28
+ INFO:2025-04-19 04:36:33,222: Epoch [165/300], Step [30/51], Loss: 0.61091, Disc Loss: 1.73543, Dur Loss: 1.30992, CE Loss: 0.07911, Norm Loss: 1.19167, F0 Loss: 3.84372, LM Loss: 2.48396, Gen Loss: 19.93699
29
+ INFO:2025-04-19 04:37:20,703: Epoch [165/300], Step [40/51], Loss: 0.64627, Disc Loss: 1.60724, Dur Loss: 1.35634, CE Loss: 0.08046, Norm Loss: 1.67871, F0 Loss: 5.72094, LM Loss: 2.69885, Gen Loss: 20.55851
30
+ INFO:2025-04-19 04:38:05,969: Epoch [165/300], Step [50/51], Loss: 0.64173, Disc Loss: 1.43282, Dur Loss: 1.36469, CE Loss: 0.07329, Norm Loss: 1.03667, F0 Loss: 4.10319, LM Loss: 2.35580, Gen Loss: 22.62904
31
+ INFO:2025-04-19 04:39:00,973: Epoch [166/300], Step [10/51], Loss: 0.61561, Disc Loss: 1.87253, Dur Loss: 1.37466, CE Loss: 0.07242, Norm Loss: 1.09918, F0 Loss: 3.50388, LM Loss: 2.39875, Gen Loss: 19.25333
32
+ INFO:2025-04-19 04:39:47,216: Epoch [166/300], Step [20/51], Loss: 0.61092, Disc Loss: 1.46170, Dur Loss: 1.52468, CE Loss: 0.08490, Norm Loss: 1.16615, F0 Loss: 4.27633, LM Loss: 2.47646, Gen Loss: 21.77772
33
+ INFO:2025-04-19 04:40:34,063: Epoch [166/300], Step [30/51], Loss: 0.62690, Disc Loss: 1.41709, Dur Loss: 1.48991, CE Loss: 0.08842, Norm Loss: 1.06107, F0 Loss: 5.37812, LM Loss: 2.54743, Gen Loss: 21.78275
34
+ INFO:2025-04-19 04:41:19,672: Epoch [166/300], Step [40/51], Loss: 0.60128, Disc Loss: 1.55830, Dur Loss: 1.53070, CE Loss: 0.09550, Norm Loss: 1.32685, F0 Loss: 5.06974, LM Loss: 2.33649, Gen Loss: 21.30290
35
+ INFO:2025-04-19 04:42:04,684: Epoch [166/300], Step [50/51], Loss: 0.60404, Disc Loss: 1.60350, Dur Loss: 1.49562, CE Loss: 0.08529, Norm Loss: 0.97772, F0 Loss: 4.22367, LM Loss: 2.45343, Gen Loss: 21.76797
36
+ INFO:2025-04-19 04:42:59,268: Epoch [167/300], Step [10/51], Loss: 0.62652, Disc Loss: 1.59606, Dur Loss: 1.45061, CE Loss: 0.08515, Norm Loss: 1.33134, F0 Loss: 4.95937, LM Loss: 2.55921, Gen Loss: 21.70388
37
+ INFO:2025-04-19 04:43:45,587: Epoch [167/300], Step [20/51], Loss: 0.62103, Disc Loss: 1.58069, Dur Loss: 1.20130, CE Loss: 0.06234, Norm Loss: 1.04771, F0 Loss: 3.92831, LM Loss: 2.44679, Gen Loss: 20.98036
38
+ INFO:2025-04-19 04:44:31,758: Epoch [167/300], Step [30/51], Loss: 0.62442, Disc Loss: 1.39970, Dur Loss: 1.44425, CE Loss: 0.07671, Norm Loss: 1.23832, F0 Loss: 5.44705, LM Loss: 2.44023, Gen Loss: 21.81331
39
+ INFO:2025-04-19 04:45:18,043: Epoch [167/300], Step [40/51], Loss: 0.60617, Disc Loss: 1.52243, Dur Loss: 1.31025, CE Loss: 0.07606, Norm Loss: 1.00748, F0 Loss: 3.43133, LM Loss: 2.27823, Gen Loss: 21.74066
40
+ INFO:2025-04-19 04:46:03,759: Epoch [167/300], Step [50/51], Loss: 0.61772, Disc Loss: 1.87143, Dur Loss: 1.37577, CE Loss: 0.07955, Norm Loss: 1.26402, F0 Loss: 4.39598, LM Loss: 2.41786, Gen Loss: 19.30043
41
+ INFO:2025-04-19 04:46:59,292: Epoch [168/300], Step [10/51], Loss: 0.62147, Disc Loss: 1.71858, Dur Loss: 1.62945, CE Loss: 0.08499, Norm Loss: 1.15239, F0 Loss: 4.21905, LM Loss: 2.42597, Gen Loss: 20.11444
42
+ INFO:2025-04-19 04:47:45,193: Epoch [168/300], Step [20/51], Loss: 0.61465, Disc Loss: 1.57598, Dur Loss: 1.31845, CE Loss: 0.07853, Norm Loss: 1.47415, F0 Loss: 6.14085, LM Loss: 2.49344, Gen Loss: 21.22107
43
+ INFO:2025-04-19 04:48:31,226: Epoch [168/300], Step [30/51], Loss: 0.61503, Disc Loss: 1.67921, Dur Loss: 1.45270, CE Loss: 0.08215, Norm Loss: 1.04884, F0 Loss: 4.25235, LM Loss: 2.59562, Gen Loss: 20.62066
44
+ INFO:2025-04-19 04:49:17,588: Epoch [168/300], Step [40/51], Loss: 0.60911, Disc Loss: 1.51642, Dur Loss: 1.33139, CE Loss: 0.07627, Norm Loss: 1.37323, F0 Loss: 5.07897, LM Loss: 2.48718, Gen Loss: 22.49410
45
+ INFO:2025-04-19 04:50:03,812: Epoch [168/300], Step [50/51], Loss: 0.63618, Disc Loss: 1.48828, Dur Loss: 1.40381, CE Loss: 0.07788, Norm Loss: 1.24790, F0 Loss: 4.26553, LM Loss: 2.46186, Gen Loss: 21.34413
46
+ INFO:2025-04-19 04:50:59,462: Epoch [169/300], Step [10/51], Loss: 0.62396, Disc Loss: 1.63711, Dur Loss: 1.54662, CE Loss: 0.09918, Norm Loss: 1.26701, F0 Loss: 4.57993, LM Loss: 2.33836, Gen Loss: 21.63839
47
+ INFO:2025-04-19 04:51:44,966: Epoch [169/300], Step [20/51], Loss: 0.59275, Disc Loss: 1.67358, Dur Loss: 1.38058, CE Loss: 0.07983, Norm Loss: 1.03154, F0 Loss: 4.17188, LM Loss: 2.37218, Gen Loss: 20.75462
48
+ INFO:2025-04-19 04:52:30,990: Epoch [169/300], Step [30/51], Loss: 0.62219, Disc Loss: 1.41014, Dur Loss: 1.42184, CE Loss: 0.08616, Norm Loss: 1.19359, F0 Loss: 4.57184, LM Loss: 2.47231, Gen Loss: 22.32779
49
+ INFO:2025-04-19 04:53:16,891: Epoch [169/300], Step [40/51], Loss: 0.62640, Disc Loss: 1.78813, Dur Loss: 1.51392, CE Loss: 0.07849, Norm Loss: 1.19950, F0 Loss: 3.81316, LM Loss: 2.39753, Gen Loss: 19.69670
50
+ INFO:2025-04-19 04:54:03,204: Epoch [169/300], Step [50/51], Loss: 0.61943, Disc Loss: 1.83870, Dur Loss: 1.35278, CE Loss: 0.08010, Norm Loss: 1.64048, F0 Loss: 6.35125, LM Loss: 2.64920, Gen Loss: 20.35914
51
+ INFO:2025-04-19 04:54:58,126: Epoch [170/300], Step [10/51], Loss: 0.61891, Disc Loss: 1.72500, Dur Loss: 1.35739, CE Loss: 0.07986, Norm Loss: 1.51714, F0 Loss: 4.93424, LM Loss: 2.56292, Gen Loss: 20.74751
52
+ INFO:2025-04-19 04:55:44,311: Epoch [170/300], Step [20/51], Loss: 0.62611, Disc Loss: 1.38058, Dur Loss: 1.34940, CE Loss: 0.07707, Norm Loss: 1.22574, F0 Loss: 5.10193, LM Loss: 2.52421, Gen Loss: 22.91841
53
+ INFO:2025-04-19 04:56:30,460: Epoch [170/300], Step [30/51], Loss: 0.58498, Disc Loss: 1.50048, Dur Loss: 1.32411, CE Loss: 0.07293, Norm Loss: 1.43924, F0 Loss: 7.07482, LM Loss: 2.56567, Gen Loss: 21.71566
54
+ INFO:2025-04-19 04:57:17,001: Epoch [170/300], Step [40/51], Loss: 0.62961, Disc Loss: 1.42340, Dur Loss: 1.19090, CE Loss: 0.06325, Norm Loss: 1.14504, F0 Loss: 4.27976, LM Loss: 2.44166, Gen Loss: 21.77470
55
+ INFO:2025-04-19 04:58:02,469: Epoch [170/300], Step [50/51], Loss: 0.59666, Disc Loss: 1.46970, Dur Loss: 1.31525, CE Loss: 0.07444, Norm Loss: 0.92673, F0 Loss: 3.70055, LM Loss: 2.45975, Gen Loss: 21.11955
56
+ INFO:2025-04-19 04:58:58,224: Epoch [171/300], Step [10/51], Loss: 0.62408, Disc Loss: 1.42369, Dur Loss: 1.24555, CE Loss: 0.06815, Norm Loss: 1.17164, F0 Loss: 5.08718, LM Loss: 2.51255, Gen Loss: 22.01647
57
+ INFO:2025-04-19 04:59:44,416: Epoch [171/300], Step [20/51], Loss: 0.59565, Disc Loss: 1.73362, Dur Loss: 1.49764, CE Loss: 0.07781, Norm Loss: 1.23784, F0 Loss: 4.39716, LM Loss: 2.28983, Gen Loss: 20.21249
58
+ INFO:2025-04-19 05:00:29,580: Epoch [171/300], Step [30/51], Loss: 0.62176, Disc Loss: 1.83607, Dur Loss: 1.48329, CE Loss: 0.07947, Norm Loss: 1.16265, F0 Loss: 4.27442, LM Loss: 2.46767, Gen Loss: 20.39859
59
+ INFO:2025-04-19 05:01:15,938: Epoch [171/300], Step [40/51], Loss: 0.65056, Disc Loss: 1.58127, Dur Loss: 1.27457, CE Loss: 0.07442, Norm Loss: 0.89664, F0 Loss: 3.90889, LM Loss: 2.42980, Gen Loss: 21.19791
60
+ INFO:2025-04-19 05:02:01,569: Epoch [171/300], Step [50/51], Loss: 0.59957, Disc Loss: 1.43492, Dur Loss: 1.25705, CE Loss: 0.07790, Norm Loss: 0.97807, F0 Loss: 3.93567, LM Loss: 2.42881, Gen Loss: 21.76072
61
+ INFO:2025-04-19 05:02:57,116: Epoch [172/300], Step [10/51], Loss: 0.61922, Disc Loss: 1.65966, Dur Loss: 1.38347, CE Loss: 0.07313, Norm Loss: 1.23423, F0 Loss: 4.85919, LM Loss: 2.43578, Gen Loss: 20.42861
62
+ INFO:2025-04-19 05:03:43,403: Epoch [172/300], Step [20/51], Loss: 0.63013, Disc Loss: 1.33103, Dur Loss: 1.54024, CE Loss: 0.09194, Norm Loss: 1.39962, F0 Loss: 5.42717, LM Loss: 2.65523, Gen Loss: 22.13144
63
+ INFO:2025-04-19 05:04:29,786: Epoch [172/300], Step [30/51], Loss: 0.61422, Disc Loss: 1.64699, Dur Loss: 1.41179, CE Loss: 0.07872, Norm Loss: 1.04084, F0 Loss: 4.04186, LM Loss: 2.41387, Gen Loss: 21.04662
64
+ INFO:2025-04-19 05:05:15,645: Epoch [172/300], Step [40/51], Loss: 0.59061, Disc Loss: 2.09528, Dur Loss: 1.54836, CE Loss: 0.08906, Norm Loss: 1.30317, F0 Loss: 4.47688, LM Loss: 2.41751, Gen Loss: 18.12341
65
+ INFO:2025-04-19 05:06:01,620: Epoch [172/300], Step [50/51], Loss: 0.60923, Disc Loss: 1.55096, Dur Loss: 1.31466, CE Loss: 0.06740, Norm Loss: 0.78310, F0 Loss: 3.73894, LM Loss: 2.43043, Gen Loss: 21.06657
66
+ INFO:2025-04-19 05:06:55,731: Epoch [173/300], Step [10/51], Loss: 0.57302, Disc Loss: 1.71308, Dur Loss: 1.52903, CE Loss: 0.08405, Norm Loss: 0.86238, F0 Loss: 3.99504, LM Loss: 2.38599, Gen Loss: 20.35188
67
+ INFO:2025-04-19 05:07:40,782: Epoch [173/300], Step [20/51], Loss: 0.61565, Disc Loss: 1.69760, Dur Loss: 1.29815, CE Loss: 0.07395, Norm Loss: 1.48307, F0 Loss: 4.94854, LM Loss: 2.62610, Gen Loss: 20.37193
68
+ INFO:2025-04-19 05:08:27,762: Epoch [173/300], Step [30/51], Loss: 0.62734, Disc Loss: 1.46203, Dur Loss: 1.49120, CE Loss: 0.08097, Norm Loss: 1.19858, F0 Loss: 4.82632, LM Loss: 2.64832, Gen Loss: 21.97266
69
+ INFO:2025-04-19 05:09:14,387: Epoch [173/300], Step [40/51], Loss: 0.58837, Disc Loss: 1.48593, Dur Loss: 1.38515, CE Loss: 0.07578, Norm Loss: 1.05050, F0 Loss: 4.22295, LM Loss: 2.47323, Gen Loss: 22.07851
70
+ INFO:2025-04-19 05:10:00,165: Epoch [173/300], Step [50/51], Loss: 0.59451, Disc Loss: 1.55293, Dur Loss: 1.30678, CE Loss: 0.07633, Norm Loss: 0.94875, F0 Loss: 3.54441, LM Loss: 2.45517, Gen Loss: 20.98925
71
+ INFO:2025-04-19 05:10:55,003: Epoch [174/300], Step [10/51], Loss: 0.64914, Disc Loss: 1.43084, Dur Loss: 1.34072, CE Loss: 0.07033, Norm Loss: 0.86738, F0 Loss: 3.53228, LM Loss: 2.36271, Gen Loss: 22.93789
72
+ INFO:2025-04-19 05:11:41,435: Epoch [174/300], Step [20/51], Loss: 0.60155, Disc Loss: 1.65688, Dur Loss: 1.34231, CE Loss: 0.06600, Norm Loss: 1.22494, F0 Loss: 4.79476, LM Loss: 2.47413, Gen Loss: 21.28828
73
+ INFO:2025-04-19 05:12:26,703: Epoch [174/300], Step [30/51], Loss: 0.58329, Disc Loss: 1.57699, Dur Loss: 1.41537, CE Loss: 0.08184, Norm Loss: 1.12702, F0 Loss: 4.09413, LM Loss: 2.39012, Gen Loss: 20.87862