zwa73 commited on
Commit ·
1807e40
1
Parent(s): a6c05ee
sovits 3
Browse files- so-vits-svc-4.0/20230318-SilenusNS-44k/best/D_12800.pth +3 -0
- so-vits-svc-4.0/20230318-SilenusNS-44k/best/G_12800.pth +3 -0
- so-vits-svc-4.0/20230318-SilenusNS-44k/best/config.json +93 -0
- so-vits-svc-4.0/20230318-SilenusNS-44k/config.json +93 -0
- so-vits-svc-4.0/20230318-SilenusNS-44k/kmeans_10000.pt +3 -0
- so-vits-svc-4.0/20230318-SilenusNS-44k/pruned/P_G_12800.pth +3 -0
- so-vits-svc-4.0/20230318-SilenusNS-44k/pruned/config.json +93 -0
- so-vits-svc-4.0/20230318-SilenusNS-44k/train.log +498 -0
- so-vits-svc-4.0/20230324-Juewa-44k/best/D_12800.pth +3 -0
- so-vits-svc-4.0/20230324-Juewa-44k/best/G_12800.pth +3 -0
- so-vits-svc-4.0/20230324-Juewa-44k/best/config.json +93 -0
- so-vits-svc-4.0/20230324-Juewa-44k/config.json +93 -0
- so-vits-svc-4.0/20230324-Juewa-44k/kmeans_10000.pt +3 -0
- so-vits-svc-4.0/20230324-Juewa-44k/pruned/P_G_12800.pth +3 -0
- so-vits-svc-4.0/20230324-Juewa-44k/pruned/config.json +93 -0
- so-vits-svc-4.0/20230324-Juewa-44k/train.log +431 -0
- so-vits-svc-4.0/20230324-JuewaNS-44k/best/D_13600.pth +3 -0
- so-vits-svc-4.0/20230324-JuewaNS-44k/best/G_13600.pth +3 -0
- so-vits-svc-4.0/20230324-JuewaNS-44k/best/config.json +93 -0
- so-vits-svc-4.0/20230324-JuewaNS-44k/config.json +93 -0
- so-vits-svc-4.0/20230324-JuewaNS-44k/kmeans_10000.pt +3 -0
- so-vits-svc-4.0/20230324-JuewaNS-44k/pruned/P_G_13600.pth +3 -0
- so-vits-svc-4.0/20230324-JuewaNS-44k/pruned/config.json +93 -0
- so-vits-svc-4.0/20230324-JuewaNS-44k/train.log +772 -0
so-vits-svc-4.0/20230318-SilenusNS-44k/best/D_12800.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9d833c37330bba865519e043b3c8f652914dcc063e31f3e1df324dc95f7b1cbb
|
| 3 |
+
size 561098185
|
so-vits-svc-4.0/20230318-SilenusNS-44k/best/G_12800.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:12fb4e79b127698c14e337ea09c26b92c1447e1d963c6343220ab83052989cd6
|
| 3 |
+
size 542789405
|
so-vits-svc-4.0/20230318-SilenusNS-44k/best/config.json
ADDED
|
@@ -0,0 +1,93 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"train": {
|
| 3 |
+
"log_interval": 200,
|
| 4 |
+
"eval_interval": 800,
|
| 5 |
+
"seed": 1234,
|
| 6 |
+
"epochs": 10000,
|
| 7 |
+
"learning_rate": 0.0001,
|
| 8 |
+
"betas": [
|
| 9 |
+
0.8,
|
| 10 |
+
0.99
|
| 11 |
+
],
|
| 12 |
+
"eps": 1e-09,
|
| 13 |
+
"batch_size": 6,
|
| 14 |
+
"fp16_run": false,
|
| 15 |
+
"lr_decay": 0.999875,
|
| 16 |
+
"segment_size": 10240,
|
| 17 |
+
"init_lr_ratio": 1,
|
| 18 |
+
"warmup_epochs": 0,
|
| 19 |
+
"c_mel": 45,
|
| 20 |
+
"c_kl": 1.0,
|
| 21 |
+
"use_sr": true,
|
| 22 |
+
"max_speclen": 512,
|
| 23 |
+
"port": "8001",
|
| 24 |
+
"keep_ckpts": 99
|
| 25 |
+
},
|
| 26 |
+
"data": {
|
| 27 |
+
"training_files": "filelists/train.txt",
|
| 28 |
+
"validation_files": "filelists/val.txt",
|
| 29 |
+
"max_wav_value": 32768.0,
|
| 30 |
+
"sampling_rate": 44100,
|
| 31 |
+
"filter_length": 2048,
|
| 32 |
+
"hop_length": 512,
|
| 33 |
+
"win_length": 2048,
|
| 34 |
+
"n_mel_channels": 80,
|
| 35 |
+
"mel_fmin": 0.0,
|
| 36 |
+
"mel_fmax": 22050
|
| 37 |
+
},
|
| 38 |
+
"model": {
|
| 39 |
+
"inter_channels": 192,
|
| 40 |
+
"hidden_channels": 192,
|
| 41 |
+
"filter_channels": 768,
|
| 42 |
+
"n_heads": 2,
|
| 43 |
+
"n_layers": 6,
|
| 44 |
+
"kernel_size": 3,
|
| 45 |
+
"p_dropout": 0.1,
|
| 46 |
+
"resblock": "1",
|
| 47 |
+
"resblock_kernel_sizes": [
|
| 48 |
+
3,
|
| 49 |
+
7,
|
| 50 |
+
11
|
| 51 |
+
],
|
| 52 |
+
"resblock_dilation_sizes": [
|
| 53 |
+
[
|
| 54 |
+
1,
|
| 55 |
+
3,
|
| 56 |
+
5
|
| 57 |
+
],
|
| 58 |
+
[
|
| 59 |
+
1,
|
| 60 |
+
3,
|
| 61 |
+
5
|
| 62 |
+
],
|
| 63 |
+
[
|
| 64 |
+
1,
|
| 65 |
+
3,
|
| 66 |
+
5
|
| 67 |
+
]
|
| 68 |
+
],
|
| 69 |
+
"upsample_rates": [
|
| 70 |
+
8,
|
| 71 |
+
8,
|
| 72 |
+
2,
|
| 73 |
+
2,
|
| 74 |
+
2
|
| 75 |
+
],
|
| 76 |
+
"upsample_initial_channel": 512,
|
| 77 |
+
"upsample_kernel_sizes": [
|
| 78 |
+
16,
|
| 79 |
+
16,
|
| 80 |
+
4,
|
| 81 |
+
4,
|
| 82 |
+
4
|
| 83 |
+
],
|
| 84 |
+
"n_layers_q": 3,
|
| 85 |
+
"use_spectral_norm": false,
|
| 86 |
+
"gin_channels": 256,
|
| 87 |
+
"ssl_dim": 256,
|
| 88 |
+
"n_speakers": 200
|
| 89 |
+
},
|
| 90 |
+
"spk": {
|
| 91 |
+
"SilenusNS": 0
|
| 92 |
+
}
|
| 93 |
+
}
|
so-vits-svc-4.0/20230318-SilenusNS-44k/config.json
ADDED
|
@@ -0,0 +1,93 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"train": {
|
| 3 |
+
"log_interval": 200,
|
| 4 |
+
"eval_interval": 800,
|
| 5 |
+
"seed": 1234,
|
| 6 |
+
"epochs": 10000,
|
| 7 |
+
"learning_rate": 0.0001,
|
| 8 |
+
"betas": [
|
| 9 |
+
0.8,
|
| 10 |
+
0.99
|
| 11 |
+
],
|
| 12 |
+
"eps": 1e-09,
|
| 13 |
+
"batch_size": 6,
|
| 14 |
+
"fp16_run": false,
|
| 15 |
+
"lr_decay": 0.999875,
|
| 16 |
+
"segment_size": 10240,
|
| 17 |
+
"init_lr_ratio": 1,
|
| 18 |
+
"warmup_epochs": 0,
|
| 19 |
+
"c_mel": 45,
|
| 20 |
+
"c_kl": 1.0,
|
| 21 |
+
"use_sr": true,
|
| 22 |
+
"max_speclen": 512,
|
| 23 |
+
"port": "8001",
|
| 24 |
+
"keep_ckpts": 99
|
| 25 |
+
},
|
| 26 |
+
"data": {
|
| 27 |
+
"training_files": "filelists/train.txt",
|
| 28 |
+
"validation_files": "filelists/val.txt",
|
| 29 |
+
"max_wav_value": 32768.0,
|
| 30 |
+
"sampling_rate": 44100,
|
| 31 |
+
"filter_length": 2048,
|
| 32 |
+
"hop_length": 512,
|
| 33 |
+
"win_length": 2048,
|
| 34 |
+
"n_mel_channels": 80,
|
| 35 |
+
"mel_fmin": 0.0,
|
| 36 |
+
"mel_fmax": 22050
|
| 37 |
+
},
|
| 38 |
+
"model": {
|
| 39 |
+
"inter_channels": 192,
|
| 40 |
+
"hidden_channels": 192,
|
| 41 |
+
"filter_channels": 768,
|
| 42 |
+
"n_heads": 2,
|
| 43 |
+
"n_layers": 6,
|
| 44 |
+
"kernel_size": 3,
|
| 45 |
+
"p_dropout": 0.1,
|
| 46 |
+
"resblock": "1",
|
| 47 |
+
"resblock_kernel_sizes": [
|
| 48 |
+
3,
|
| 49 |
+
7,
|
| 50 |
+
11
|
| 51 |
+
],
|
| 52 |
+
"resblock_dilation_sizes": [
|
| 53 |
+
[
|
| 54 |
+
1,
|
| 55 |
+
3,
|
| 56 |
+
5
|
| 57 |
+
],
|
| 58 |
+
[
|
| 59 |
+
1,
|
| 60 |
+
3,
|
| 61 |
+
5
|
| 62 |
+
],
|
| 63 |
+
[
|
| 64 |
+
1,
|
| 65 |
+
3,
|
| 66 |
+
5
|
| 67 |
+
]
|
| 68 |
+
],
|
| 69 |
+
"upsample_rates": [
|
| 70 |
+
8,
|
| 71 |
+
8,
|
| 72 |
+
2,
|
| 73 |
+
2,
|
| 74 |
+
2
|
| 75 |
+
],
|
| 76 |
+
"upsample_initial_channel": 512,
|
| 77 |
+
"upsample_kernel_sizes": [
|
| 78 |
+
16,
|
| 79 |
+
16,
|
| 80 |
+
4,
|
| 81 |
+
4,
|
| 82 |
+
4
|
| 83 |
+
],
|
| 84 |
+
"n_layers_q": 3,
|
| 85 |
+
"use_spectral_norm": false,
|
| 86 |
+
"gin_channels": 256,
|
| 87 |
+
"ssl_dim": 256,
|
| 88 |
+
"n_speakers": 200
|
| 89 |
+
},
|
| 90 |
+
"spk": {
|
| 91 |
+
"SilenusNS": 0
|
| 92 |
+
}
|
| 93 |
+
}
|
so-vits-svc-4.0/20230318-SilenusNS-44k/kmeans_10000.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:32950ebf071b62f9bf6b6240c2c4a18f22535fa51430703104749b568da1b1ec
|
| 3 |
+
size 15446393
|
so-vits-svc-4.0/20230318-SilenusNS-44k/pruned/P_G_12800.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:927581e85ab6d5485819b44b3dcb3c616ed104b15660ef91699c2458a9863116
|
| 3 |
+
size 180885829
|
so-vits-svc-4.0/20230318-SilenusNS-44k/pruned/config.json
ADDED
|
@@ -0,0 +1,93 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"train": {
|
| 3 |
+
"log_interval": 200,
|
| 4 |
+
"eval_interval": 800,
|
| 5 |
+
"seed": 1234,
|
| 6 |
+
"epochs": 10000,
|
| 7 |
+
"learning_rate": 0.0001,
|
| 8 |
+
"betas": [
|
| 9 |
+
0.8,
|
| 10 |
+
0.99
|
| 11 |
+
],
|
| 12 |
+
"eps": 1e-09,
|
| 13 |
+
"batch_size": 6,
|
| 14 |
+
"fp16_run": false,
|
| 15 |
+
"lr_decay": 0.999875,
|
| 16 |
+
"segment_size": 10240,
|
| 17 |
+
"init_lr_ratio": 1,
|
| 18 |
+
"warmup_epochs": 0,
|
| 19 |
+
"c_mel": 45,
|
| 20 |
+
"c_kl": 1.0,
|
| 21 |
+
"use_sr": true,
|
| 22 |
+
"max_speclen": 512,
|
| 23 |
+
"port": "8001",
|
| 24 |
+
"keep_ckpts": 99
|
| 25 |
+
},
|
| 26 |
+
"data": {
|
| 27 |
+
"training_files": "filelists/train.txt",
|
| 28 |
+
"validation_files": "filelists/val.txt",
|
| 29 |
+
"max_wav_value": 32768.0,
|
| 30 |
+
"sampling_rate": 44100,
|
| 31 |
+
"filter_length": 2048,
|
| 32 |
+
"hop_length": 512,
|
| 33 |
+
"win_length": 2048,
|
| 34 |
+
"n_mel_channels": 80,
|
| 35 |
+
"mel_fmin": 0.0,
|
| 36 |
+
"mel_fmax": 22050
|
| 37 |
+
},
|
| 38 |
+
"model": {
|
| 39 |
+
"inter_channels": 192,
|
| 40 |
+
"hidden_channels": 192,
|
| 41 |
+
"filter_channels": 768,
|
| 42 |
+
"n_heads": 2,
|
| 43 |
+
"n_layers": 6,
|
| 44 |
+
"kernel_size": 3,
|
| 45 |
+
"p_dropout": 0.1,
|
| 46 |
+
"resblock": "1",
|
| 47 |
+
"resblock_kernel_sizes": [
|
| 48 |
+
3,
|
| 49 |
+
7,
|
| 50 |
+
11
|
| 51 |
+
],
|
| 52 |
+
"resblock_dilation_sizes": [
|
| 53 |
+
[
|
| 54 |
+
1,
|
| 55 |
+
3,
|
| 56 |
+
5
|
| 57 |
+
],
|
| 58 |
+
[
|
| 59 |
+
1,
|
| 60 |
+
3,
|
| 61 |
+
5
|
| 62 |
+
],
|
| 63 |
+
[
|
| 64 |
+
1,
|
| 65 |
+
3,
|
| 66 |
+
5
|
| 67 |
+
]
|
| 68 |
+
],
|
| 69 |
+
"upsample_rates": [
|
| 70 |
+
8,
|
| 71 |
+
8,
|
| 72 |
+
2,
|
| 73 |
+
2,
|
| 74 |
+
2
|
| 75 |
+
],
|
| 76 |
+
"upsample_initial_channel": 512,
|
| 77 |
+
"upsample_kernel_sizes": [
|
| 78 |
+
16,
|
| 79 |
+
16,
|
| 80 |
+
4,
|
| 81 |
+
4,
|
| 82 |
+
4
|
| 83 |
+
],
|
| 84 |
+
"n_layers_q": 3,
|
| 85 |
+
"use_spectral_norm": false,
|
| 86 |
+
"gin_channels": 256,
|
| 87 |
+
"ssl_dim": 256,
|
| 88 |
+
"n_speakers": 200
|
| 89 |
+
},
|
| 90 |
+
"spk": {
|
| 91 |
+
"SilenusNS": 0
|
| 92 |
+
}
|
| 93 |
+
}
|
so-vits-svc-4.0/20230318-SilenusNS-44k/train.log
ADDED
|
@@ -0,0 +1,498 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2023-03-18 15:25:05,061 44k INFO {'train': {'log_interval': 200, 'eval_interval': 800, 'seed': 1234, 'epochs': 10000, 'learning_rate': 0.0001, 'betas': [0.8, 0.99], 'eps': 1e-09, 'batch_size': 6, 'fp16_run': False, 'lr_decay': 0.999875, 'segment_size': 10240, 'init_lr_ratio': 1, 'warmup_epochs': 0, 'c_mel': 45, 'c_kl': 1.0, 'use_sr': True, 'max_speclen': 512, 'port': '8001', 'keep_ckpts': 99}, 'data': {'training_files': 'filelists/train.txt', 'validation_files': 'filelists/val.txt', 'max_wav_value': 32768.0, 'sampling_rate': 44100, 'filter_length': 2048, 'hop_length': 512, 'win_length': 2048, 'n_mel_channels': 80, 'mel_fmin': 0.0, 'mel_fmax': 22050}, 'model': {'inter_channels': 192, 'hidden_channels': 192, 'filter_channels': 768, 'n_heads': 2, 'n_layers': 6, 'kernel_size': 3, 'p_dropout': 0.1, 'resblock': '1', 'resblock_kernel_sizes': [3, 7, 11], 'resblock_dilation_sizes': [[1, 3, 5], [1, 3, 5], [1, 3, 5]], 'upsample_rates': [8, 8, 2, 2, 2], 'upsample_initial_channel': 512, 'upsample_kernel_sizes': [16, 16, 4, 4, 4], 'n_layers_q': 3, 'use_spectral_norm': False, 'gin_channels': 256, 'ssl_dim': 256, 'n_speakers': 200}, 'spk': {'SilenusNS': 0}, 'model_dir': './logs/44k'}
|
| 2 |
+
2023-03-18 15:25:05,062 44k WARNING /root/so-vits-svc-4.0 is not a git repository, therefore hash value comparison will be ignored.
|
| 3 |
+
2023-03-18 15:25:08,101 44k INFO emb_g.weight is not in the checkpoint
|
| 4 |
+
2023-03-18 15:25:08,170 44k INFO Loaded checkpoint './logs/44k/G_0.pth' (iteration 0)
|
| 5 |
+
2023-03-18 15:25:08,321 44k INFO Loaded checkpoint './logs/44k/D_0.pth' (iteration 0)
|
| 6 |
+
2023-03-18 15:25:15,937 44k INFO Train Epoch: 1 [0%]
|
| 7 |
+
2023-03-18 15:25:15,938 44k INFO Losses: [2.8153839111328125, 2.3640732765197754, 10.997515678405762, 34.97066116333008, 3.6714816093444824], step: 0, lr: 0.0001
|
| 8 |
+
2023-03-18 15:25:20,732 44k INFO Saving model and optimizer state at iteration 1 to ./logs/44k/G_0.pth
|
| 9 |
+
2023-03-18 15:25:21,865 44k INFO Saving model and optimizer state at iteration 1 to ./logs/44k/D_0.pth
|
| 10 |
+
2023-03-18 15:25:42,275 44k INFO ====> Epoch: 1, cost 37.21 s
|
| 11 |
+
2023-03-18 15:26:01,596 44k INFO ====> Epoch: 2, cost 19.32 s
|
| 12 |
+
2023-03-18 15:26:20,960 44k INFO ====> Epoch: 3, cost 19.36 s
|
| 13 |
+
2023-03-18 15:26:40,421 44k INFO ====> Epoch: 4, cost 19.46 s
|
| 14 |
+
2023-03-18 15:26:59,757 44k INFO ====> Epoch: 5, cost 19.34 s
|
| 15 |
+
2023-03-18 15:27:05,274 44k INFO Train Epoch: 6 [13%]
|
| 16 |
+
2023-03-18 15:27:05,275 44k INFO Losses: [2.455998420715332, 2.1098105907440186, 8.791666030883789, 22.87582778930664, 1.493965744972229], step: 200, lr: 9.993751562304699e-05
|
| 17 |
+
2023-03-18 15:27:20,129 44k INFO ====> Epoch: 6, cost 20.37 s
|
| 18 |
+
2023-03-18 15:27:39,462 44k INFO ====> Epoch: 7, cost 19.33 s
|
| 19 |
+
2023-03-18 15:27:58,896 44k INFO ====> Epoch: 8, cost 19.43 s
|
| 20 |
+
2023-03-18 15:28:18,498 44k INFO ====> Epoch: 9, cost 19.60 s
|
| 21 |
+
2023-03-18 15:28:38,294 44k INFO ====> Epoch: 10, cost 19.80 s
|
| 22 |
+
2023-03-18 15:28:45,816 44k INFO Train Epoch: 11 [26%]
|
| 23 |
+
2023-03-18 15:28:45,818 44k INFO Losses: [2.5894227027893066, 2.678623676300049, 10.583251953125, 23.435916900634766, 1.342075228691101], step: 400, lr: 9.987507028906759e-05
|
| 24 |
+
2023-03-18 15:28:58,332 44k INFO ====> Epoch: 11, cost 20.04 s
|
| 25 |
+
2023-03-18 15:29:18,589 44k INFO ====> Epoch: 12, cost 20.26 s
|
| 26 |
+
2023-03-18 15:29:39,694 44k INFO ====> Epoch: 13, cost 21.11 s
|
| 27 |
+
2023-03-18 15:29:59,110 44k INFO ====> Epoch: 14, cost 19.42 s
|
| 28 |
+
2023-03-18 15:30:18,828 44k INFO ====> Epoch: 15, cost 19.72 s
|
| 29 |
+
2023-03-18 15:30:28,571 44k INFO Train Epoch: 16 [38%]
|
| 30 |
+
2023-03-18 15:30:28,572 44k INFO Losses: [2.3110649585723877, 2.5217154026031494, 8.599373817443848, 20.32848358154297, 1.1233292818069458], step: 600, lr: 9.981266397366609e-05
|
| 31 |
+
2023-03-18 15:30:39,261 44k INFO ====> Epoch: 16, cost 20.43 s
|
| 32 |
+
2023-03-18 15:30:58,613 44k INFO ====> Epoch: 17, cost 19.35 s
|
| 33 |
+
2023-03-18 15:31:18,017 44k INFO ====> Epoch: 18, cost 19.40 s
|
| 34 |
+
2023-03-18 15:31:37,554 44k INFO ====> Epoch: 19, cost 19.54 s
|
| 35 |
+
2023-03-18 15:31:57,299 44k INFO ====> Epoch: 20, cost 19.74 s
|
| 36 |
+
2023-03-18 15:32:09,588 44k INFO Train Epoch: 21 [51%]
|
| 37 |
+
2023-03-18 15:32:09,589 44k INFO Losses: [2.4298205375671387, 2.607797145843506, 11.093040466308594, 22.630563735961914, 1.1880390644073486], step: 800, lr: 9.975029665246193e-05
|
| 38 |
+
2023-03-18 15:32:13,823 44k INFO Saving model and optimizer state at iteration 21 to ./logs/44k/G_800.pth
|
| 39 |
+
2023-03-18 15:32:14,917 44k INFO Saving model and optimizer state at iteration 21 to ./logs/44k/D_800.pth
|
| 40 |
+
2023-03-18 15:32:23,333 44k INFO ====> Epoch: 21, cost 26.03 s
|
| 41 |
+
2023-03-18 15:32:43,083 44k INFO ====> Epoch: 22, cost 19.75 s
|
| 42 |
+
2023-03-18 15:33:03,071 44k INFO ====> Epoch: 23, cost 19.99 s
|
| 43 |
+
2023-03-18 15:33:22,903 44k INFO ====> Epoch: 24, cost 19.83 s
|
| 44 |
+
2023-03-18 15:33:42,382 44k INFO ====> Epoch: 25, cost 19.48 s
|
| 45 |
+
2023-03-18 15:33:56,327 44k INFO Train Epoch: 26 [64%]
|
| 46 |
+
2023-03-18 15:33:56,328 44k INFO Losses: [2.0924313068389893, 2.961833953857422, 7.720591068267822, 18.682811737060547, 0.679410457611084], step: 1000, lr: 9.968796830108985e-05
|
| 47 |
+
2023-03-18 15:34:02,813 44k INFO ====> Epoch: 26, cost 20.43 s
|
| 48 |
+
2023-03-18 15:34:22,323 44k INFO ====> Epoch: 27, cost 19.51 s
|
| 49 |
+
2023-03-18 15:34:41,799 44k INFO ====> Epoch: 28, cost 19.48 s
|
| 50 |
+
2023-03-18 15:35:01,307 44k INFO ====> Epoch: 29, cost 19.51 s
|
| 51 |
+
2023-03-18 15:35:20,945 44k INFO ====> Epoch: 30, cost 19.64 s
|
| 52 |
+
2023-03-18 15:35:36,798 44k INFO Train Epoch: 31 [77%]
|
| 53 |
+
2023-03-18 15:35:36,800 44k INFO Losses: [2.3155126571655273, 2.206496477127075, 11.064719200134277, 19.939838409423828, 0.7182945013046265], step: 1200, lr: 9.962567889519979e-05
|
| 54 |
+
2023-03-18 15:35:41,510 44k INFO ====> Epoch: 31, cost 20.57 s
|
| 55 |
+
2023-03-18 15:36:01,618 44k INFO ====> Epoch: 32, cost 20.11 s
|
| 56 |
+
2023-03-18 15:36:21,079 44k INFO ====> Epoch: 33, cost 19.46 s
|
| 57 |
+
2023-03-18 15:36:40,302 44k INFO ====> Epoch: 34, cost 19.22 s
|
| 58 |
+
2023-03-18 15:36:59,771 44k INFO ====> Epoch: 35, cost 19.47 s
|
| 59 |
+
2023-03-18 15:37:17,648 44k INFO Train Epoch: 36 [90%]
|
| 60 |
+
2023-03-18 15:37:17,650 44k INFO Losses: [1.8354809284210205, 2.9983744621276855, 10.637933731079102, 18.277450561523438, 0.9634370803833008], step: 1400, lr: 9.956342841045691e-05
|
| 61 |
+
2023-03-18 15:37:19,971 44k INFO ====> Epoch: 36, cost 20.20 s
|
| 62 |
+
2023-03-18 15:37:39,555 44k INFO ====> Epoch: 37, cost 19.58 s
|
| 63 |
+
2023-03-18 15:37:58,866 44k INFO ====> Epoch: 38, cost 19.31 s
|
| 64 |
+
2023-03-18 15:38:18,557 44k INFO ====> Epoch: 39, cost 19.69 s
|
| 65 |
+
2023-03-18 15:38:38,195 44k INFO ====> Epoch: 40, cost 19.64 s
|
| 66 |
+
2023-03-18 15:38:57,921 44k INFO ====> Epoch: 41, cost 19.73 s
|
| 67 |
+
2023-03-18 15:39:01,847 44k INFO Train Epoch: 42 [3%]
|
| 68 |
+
2023-03-18 15:39:01,848 44k INFO Losses: [2.7251152992248535, 2.1925649642944336, 7.784726142883301, 18.160188674926758, 0.8011401295661926], step: 1600, lr: 9.948877917043875e-05
|
| 69 |
+
2023-03-18 15:39:06,045 44k INFO Saving model and optimizer state at iteration 42 to ./logs/44k/G_1600.pth
|
| 70 |
+
2023-03-18 15:39:07,090 44k INFO Saving model and optimizer state at iteration 42 to ./logs/44k/D_1600.pth
|
| 71 |
+
2023-03-18 15:39:23,376 44k INFO ====> Epoch: 42, cost 25.45 s
|
| 72 |
+
2023-03-18 15:39:43,065 44k INFO ====> Epoch: 43, cost 19.69 s
|
| 73 |
+
2023-03-18 15:40:02,810 44k INFO ====> Epoch: 44, cost 19.75 s
|
| 74 |
+
2023-03-18 15:40:22,260 44k INFO ====> Epoch: 45, cost 19.45 s
|
| 75 |
+
2023-03-18 15:40:42,514 44k INFO ====> Epoch: 46, cost 20.25 s
|
| 76 |
+
2023-03-18 15:40:48,648 44k INFO Train Epoch: 47 [15%]
|
| 77 |
+
2023-03-18 15:40:48,649 44k INFO Losses: [2.5461223125457764, 2.1349644660949707, 6.2196221351623535, 15.844928741455078, 1.0983959436416626], step: 1800, lr: 9.942661422663591e-05
|
| 78 |
+
2023-03-18 15:41:02,946 44k INFO ====> Epoch: 47, cost 20.43 s
|
| 79 |
+
2023-03-18 15:41:23,064 44k INFO ====> Epoch: 48, cost 20.12 s
|
| 80 |
+
2023-03-18 15:41:43,578 44k INFO ====> Epoch: 49, cost 20.51 s
|
| 81 |
+
2023-03-18 15:42:03,255 44k INFO ====> Epoch: 50, cost 19.68 s
|
| 82 |
+
2023-03-18 15:42:23,980 44k INFO ====> Epoch: 51, cost 20.72 s
|
| 83 |
+
2023-03-18 15:42:32,232 44k INFO Train Epoch: 52 [28%]
|
| 84 |
+
2023-03-18 15:42:32,234 44k INFO Losses: [2.4787557125091553, 2.4002110958099365, 7.5291852951049805, 17.007627487182617, 0.9123826026916504], step: 2000, lr: 9.936448812621091e-05
|
| 85 |
+
2023-03-18 15:42:44,654 44k INFO ====> Epoch: 52, cost 20.67 s
|
| 86 |
+
2023-03-18 15:43:04,778 44k INFO ====> Epoch: 53, cost 20.12 s
|
| 87 |
+
2023-03-18 15:43:24,491 44k INFO ====> Epoch: 54, cost 19.71 s
|
| 88 |
+
2023-03-18 15:43:44,206 44k INFO ====> Epoch: 55, cost 19.72 s
|
| 89 |
+
2023-03-18 15:44:03,873 44k INFO ====> Epoch: 56, cost 19.67 s
|
| 90 |
+
2023-03-18 15:44:14,036 44k INFO Train Epoch: 57 [41%]
|
| 91 |
+
2023-03-18 15:44:14,038 44k INFO Losses: [2.3435280323028564, 2.4136173725128174, 8.905498504638672, 21.293319702148438, 0.8389438390731812], step: 2200, lr: 9.930240084489267e-05
|
| 92 |
+
2023-03-18 15:44:24,123 44k INFO ====> Epoch: 57, cost 20.25 s
|
| 93 |
+
2023-03-18 15:44:43,994 44k INFO ====> Epoch: 58, cost 19.87 s
|
| 94 |
+
2023-03-18 15:45:05,109 44k INFO ====> Epoch: 59, cost 21.11 s
|
| 95 |
+
2023-03-18 15:45:24,973 44k INFO ====> Epoch: 60, cost 19.86 s
|
| 96 |
+
2023-03-18 15:45:45,042 44k INFO ====> Epoch: 61, cost 20.07 s
|
| 97 |
+
2023-03-18 15:45:57,802 44k INFO Train Epoch: 62 [54%]
|
| 98 |
+
2023-03-18 15:45:57,803 44k INFO Losses: [2.546645164489746, 2.279296636581421, 8.504755973815918, 22.139497756958008, 0.8844197392463684], step: 2400, lr: 9.924035235842533e-05
|
| 99 |
+
2023-03-18 15:46:02,298 44k INFO Saving model and optimizer state at iteration 62 to ./logs/44k/G_2400.pth
|
| 100 |
+
2023-03-18 15:46:03,340 44k INFO Saving model and optimizer state at iteration 62 to ./logs/44k/D_2400.pth
|
| 101 |
+
2023-03-18 15:46:12,023 44k INFO ====> Epoch: 62, cost 26.98 s
|
| 102 |
+
2023-03-18 15:46:31,951 44k INFO ====> Epoch: 63, cost 19.93 s
|
| 103 |
+
2023-03-18 15:46:51,635 44k INFO ====> Epoch: 64, cost 19.68 s
|
| 104 |
+
2023-03-18 15:47:11,463 44k INFO ====> Epoch: 65, cost 19.83 s
|
| 105 |
+
2023-03-18 15:47:31,393 44k INFO ====> Epoch: 66, cost 19.93 s
|
| 106 |
+
2023-03-18 15:47:45,850 44k INFO Train Epoch: 67 [67%]
|
| 107 |
+
2023-03-18 15:47:45,852 44k INFO Losses: [2.351210594177246, 2.3612356185913086, 7.31257963180542, 14.846061706542969, 1.1602634191513062], step: 2600, lr: 9.917834264256819e-05
|
| 108 |
+
2023-03-18 15:47:52,492 44k INFO ====> Epoch: 67, cost 21.10 s
|
| 109 |
+
2023-03-18 15:48:12,681 44k INFO ====> Epoch: 68, cost 20.19 s
|
| 110 |
+
2023-03-18 15:48:32,852 44k INFO ====> Epoch: 69, cost 20.17 s
|
| 111 |
+
2023-03-18 15:48:53,823 44k INFO ====> Epoch: 70, cost 20.97 s
|
| 112 |
+
2023-03-18 15:49:13,763 44k INFO ====> Epoch: 71, cost 19.94 s
|
| 113 |
+
2023-03-18 15:49:30,248 44k INFO Train Epoch: 72 [79%]
|
| 114 |
+
2023-03-18 15:49:30,250 44k INFO Losses: [2.2914111614227295, 2.856121778488159, 8.181981086730957, 13.763769149780273, 0.6541903614997864], step: 2800, lr: 9.911637167309565e-05
|
| 115 |
+
2023-03-18 15:49:34,249 44k INFO ====> Epoch: 72, cost 20.49 s
|
| 116 |
+
2023-03-18 15:49:54,134 44k INFO ====> Epoch: 73, cost 19.88 s
|
| 117 |
+
2023-03-18 15:50:15,695 44k INFO ====> Epoch: 74, cost 21.56 s
|
| 118 |
+
2023-03-18 15:50:35,823 44k INFO ====> Epoch: 75, cost 20.13 s
|
| 119 |
+
2023-03-18 15:50:56,045 44k INFO ====> Epoch: 76, cost 20.22 s
|
| 120 |
+
2023-03-18 15:51:14,689 44k INFO Train Epoch: 77 [92%]
|
| 121 |
+
2023-03-18 15:51:14,691 44k INFO Losses: [2.3739912509918213, 2.6581947803497314, 9.7283935546875, 19.127307891845703, 0.7439061999320984], step: 3000, lr: 9.905443942579728e-05
|
| 122 |
+
2023-03-18 15:51:16,643 44k INFO ====> Epoch: 77, cost 20.60 s
|
| 123 |
+
2023-03-18 15:51:36,563 44k INFO ====> Epoch: 78, cost 19.92 s
|
| 124 |
+
2023-03-18 15:51:56,622 44k INFO ====> Epoch: 79, cost 20.06 s
|
| 125 |
+
2023-03-18 15:52:16,476 44k INFO ====> Epoch: 80, cost 19.85 s
|
| 126 |
+
2023-03-18 15:52:36,368 44k INFO ====> Epoch: 81, cost 19.89 s
|
| 127 |
+
2023-03-18 15:52:57,128 44k INFO ====> Epoch: 82, cost 20.76 s
|
| 128 |
+
2023-03-18 15:53:01,575 44k INFO Train Epoch: 83 [5%]
|
| 129 |
+
2023-03-18 15:53:01,576 44k INFO Losses: [2.575831413269043, 2.0763652324676514, 10.35012435913086, 21.57923698425293, 0.7908153533935547], step: 3200, lr: 9.89801718082432e-05
|
| 130 |
+
2023-03-18 15:53:06,109 44k INFO Saving model and optimizer state at iteration 83 to ./logs/44k/G_3200.pth
|
| 131 |
+
2023-03-18 15:53:07,151 44k INFO Saving model and optimizer state at iteration 83 to ./logs/44k/D_3200.pth
|
| 132 |
+
2023-03-18 15:53:23,356 44k INFO ====> Epoch: 83, cost 26.23 s
|
| 133 |
+
2023-03-18 15:53:43,274 44k INFO ====> Epoch: 84, cost 19.92 s
|
| 134 |
+
2023-03-18 15:54:04,946 44k INFO ====> Epoch: 85, cost 21.67 s
|
| 135 |
+
2023-03-18 15:54:24,908 44k INFO ====> Epoch: 86, cost 19.96 s
|
| 136 |
+
2023-03-18 15:54:46,074 44k INFO ====> Epoch: 87, cost 21.17 s
|
| 137 |
+
2023-03-18 15:54:52,624 44k INFO Train Epoch: 88 [18%]
|
| 138 |
+
2023-03-18 15:54:52,625 44k INFO Losses: [2.676368236541748, 2.128420829772949, 6.567540645599365, 14.580389022827148, 0.793415904045105], step: 3400, lr: 9.891832466458178e-05
|
| 139 |
+
2023-03-18 15:55:06,763 44k INFO ====> Epoch: 88, cost 20.69 s
|
| 140 |
+
2023-03-18 15:55:27,296 44k INFO ====> Epoch: 89, cost 20.53 s
|
| 141 |
+
2023-03-18 15:55:47,223 44k INFO ====> Epoch: 90, cost 19.93 s
|
| 142 |
+
2023-03-18 15:56:07,621 44k INFO ====> Epoch: 91, cost 20.40 s
|
| 143 |
+
2023-03-18 15:56:27,682 44k INFO ====> Epoch: 92, cost 20.06 s
|
| 144 |
+
2023-03-18 15:56:36,318 44k INFO Train Epoch: 93 [31%]
|
| 145 |
+
2023-03-18 15:56:36,320 44k INFO Losses: [2.152287006378174, 2.4968721866607666, 12.756311416625977, 22.887723922729492, 1.057271122932434], step: 3600, lr: 9.885651616572276e-05
|
| 146 |
+
2023-03-18 15:56:48,140 44k INFO ====> Epoch: 93, cost 20.46 s
|
| 147 |
+
2023-03-18 15:57:07,967 44k INFO ====> Epoch: 94, cost 19.83 s
|
| 148 |
+
2023-03-18 15:57:28,520 44k INFO ====> Epoch: 95, cost 20.55 s
|
| 149 |
+
2023-03-18 15:57:48,340 44k INFO ====> Epoch: 96, cost 19.82 s
|
| 150 |
+
2023-03-18 15:58:08,947 44k INFO ====> Epoch: 97, cost 20.61 s
|
| 151 |
+
2023-03-18 15:58:19,669 44k INFO Train Epoch: 98 [44%]
|
| 152 |
+
2023-03-18 15:58:19,670 44k INFO Losses: [2.427189350128174, 2.0381827354431152, 9.996284484863281, 19.298826217651367, 0.6840081810951233], step: 3800, lr: 9.879474628751914e-05
|
| 153 |
+
2023-03-18 15:58:29,484 44k INFO ====> Epoch: 98, cost 20.54 s
|
| 154 |
+
2023-03-18 15:58:49,521 44k INFO ====> Epoch: 99, cost 20.04 s
|
| 155 |
+
2023-03-18 15:59:10,060 44k INFO ====> Epoch: 100, cost 20.54 s
|
| 156 |
+
2023-03-18 15:59:30,092 44k INFO ====> Epoch: 101, cost 20.03 s
|
| 157 |
+
2023-03-18 15:59:50,195 44k INFO ====> Epoch: 102, cost 20.10 s
|
| 158 |
+
2023-03-18 16:00:03,365 44k INFO Train Epoch: 103 [56%]
|
| 159 |
+
2023-03-18 16:00:03,367 44k INFO Losses: [2.5533149242401123, 2.544842004776001, 10.731807708740234, 20.337570190429688, 0.4659505784511566], step: 4000, lr: 9.873301500583906e-05
|
| 160 |
+
2023-03-18 16:00:08,064 44k INFO Saving model and optimizer state at iteration 103 to ./logs/44k/G_4000.pth
|
| 161 |
+
2023-03-18 16:00:09,144 44k INFO Saving model and optimizer state at iteration 103 to ./logs/44k/D_4000.pth
|
| 162 |
+
2023-03-18 16:00:16,893 44k INFO ====> Epoch: 103, cost 26.70 s
|
| 163 |
+
2023-03-18 16:00:36,789 44k INFO ====> Epoch: 104, cost 19.90 s
|
| 164 |
+
2023-03-18 16:00:58,587 44k INFO ====> Epoch: 105, cost 21.80 s
|
| 165 |
+
2023-03-18 16:01:18,557 44k INFO ====> Epoch: 106, cost 19.97 s
|
| 166 |
+
2023-03-18 16:01:38,667 44k INFO ====> Epoch: 107, cost 20.11 s
|
| 167 |
+
2023-03-18 16:01:53,610 44k INFO Train Epoch: 108 [69%]
|
| 168 |
+
2023-03-18 16:01:53,612 44k INFO Losses: [2.335444211959839, 2.434490203857422, 8.708498001098633, 19.969039916992188, 0.9741002321243286], step: 4200, lr: 9.867132229656573e-05
|
| 169 |
+
2023-03-18 16:01:59,163 44k INFO ====> Epoch: 108, cost 20.50 s
|
| 170 |
+
2023-03-18 16:02:19,113 44k INFO ====> Epoch: 109, cost 19.95 s
|
| 171 |
+
2023-03-18 16:02:39,324 44k INFO ====> Epoch: 110, cost 20.21 s
|
| 172 |
+
2023-03-18 16:02:59,329 44k INFO ====> Epoch: 111, cost 20.00 s
|
| 173 |
+
2023-03-18 16:03:20,073 44k INFO ====> Epoch: 112, cost 20.74 s
|
| 174 |
+
2023-03-18 16:03:36,919 44k INFO Train Epoch: 113 [82%]
|
| 175 |
+
2023-03-18 16:03:36,921 44k INFO Losses: [2.5057225227355957, 2.2557621002197266, 9.105669021606445, 17.700326919555664, 0.6261863112449646], step: 4400, lr: 9.86096681355974e-05
|
| 176 |
+
2023-03-18 16:03:40,487 44k INFO ====> Epoch: 113, cost 20.41 s
|
| 177 |
+
2023-03-18 16:04:00,296 44k INFO ====> Epoch: 114, cost 19.81 s
|
| 178 |
+
2023-03-18 16:04:20,189 44k INFO ====> Epoch: 115, cost 19.89 s
|
| 179 |
+
2023-03-18 16:04:39,901 44k INFO ====> Epoch: 116, cost 19.71 s
|
| 180 |
+
2023-03-18 16:05:00,262 44k INFO ====> Epoch: 117, cost 20.36 s
|
| 181 |
+
2023-03-18 16:05:19,262 44k INFO Train Epoch: 118 [95%]
|
| 182 |
+
2023-03-18 16:05:19,263 44k INFO Losses: [2.282578229904175, 2.4207687377929688, 9.76828384399414, 21.194089889526367, 0.9633945822715759], step: 4600, lr: 9.854805249884741e-05
|
| 183 |
+
2023-03-18 16:05:20,950 44k INFO ====> Epoch: 118, cost 20.69 s
|
| 184 |
+
2023-03-18 16:05:41,036 44k INFO ====> Epoch: 119, cost 20.09 s
|
| 185 |
+
2023-03-18 16:06:01,142 44k INFO ====> Epoch: 120, cost 20.11 s
|
| 186 |
+
2023-03-18 16:06:21,123 44k INFO ====> Epoch: 121, cost 19.98 s
|
| 187 |
+
2023-03-18 16:06:40,925 44k INFO ====> Epoch: 122, cost 19.80 s
|
| 188 |
+
2023-03-18 16:07:00,860 44k INFO ====> Epoch: 123, cost 19.93 s
|
| 189 |
+
2023-03-18 16:07:05,476 44k INFO Train Epoch: 124 [8%]
|
| 190 |
+
2023-03-18 16:07:05,477 44k INFO Losses: [2.2480196952819824, 2.5063650608062744, 10.50140380859375, 22.77250099182129, 0.9855556488037109], step: 4800, lr: 9.847416455282387e-05
|
| 191 |
+
2023-03-18 16:07:09,857 44k INFO Saving model and optimizer state at iteration 124 to ./logs/44k/G_4800.pth
|
| 192 |
+
2023-03-18 16:07:11,047 44k INFO Saving model and optimizer state at iteration 124 to ./logs/44k/D_4800.pth
|
| 193 |
+
2023-03-18 16:07:26,839 44k INFO ====> Epoch: 124, cost 25.98 s
|
| 194 |
+
2023-03-18 16:07:48,107 44k INFO ====> Epoch: 125, cost 21.27 s
|
| 195 |
+
2023-03-18 16:08:07,938 44k INFO ====> Epoch: 126, cost 19.83 s
|
| 196 |
+
2023-03-18 16:08:27,940 44k INFO ====> Epoch: 127, cost 20.00 s
|
| 197 |
+
2023-03-18 16:08:47,904 44k INFO ====> Epoch: 128, cost 19.96 s
|
| 198 |
+
2023-03-18 16:08:54,670 44k INFO Train Epoch: 129 [21%]
|
| 199 |
+
2023-03-18 16:08:54,672 44k INFO Losses: [2.0410420894622803, 3.1921937465667725, 13.629049301147461, 22.29718017578125, 0.7504812479019165], step: 5000, lr: 9.841263358464336e-05
|
| 200 |
+
2023-03-18 16:09:08,185 44k INFO ====> Epoch: 129, cost 20.28 s
|
| 201 |
+
2023-03-18 16:09:28,708 44k INFO ====> Epoch: 130, cost 20.52 s
|
| 202 |
+
2023-03-18 16:09:49,065 44k INFO ====> Epoch: 131, cost 20.36 s
|
| 203 |
+
2023-03-18 16:10:08,965 44k INFO ====> Epoch: 132, cost 19.90 s
|
| 204 |
+
2023-03-18 16:10:28,600 44k INFO ====> Epoch: 133, cost 19.64 s
|
| 205 |
+
2023-03-18 16:10:37,520 44k INFO Train Epoch: 134 [33%]
|
| 206 |
+
2023-03-18 16:10:37,522 44k INFO Losses: [2.0373544692993164, 2.7676761150360107, 9.952956199645996, 19.683595657348633, 0.7579522728919983], step: 5200, lr: 9.835114106370493e-05
|
| 207 |
+
2023-03-18 16:10:49,091 44k INFO ====> Epoch: 134, cost 20.49 s
|
| 208 |
+
2023-03-18 16:11:09,151 44k INFO ====> Epoch: 135, cost 20.06 s
|
| 209 |
+
2023-03-18 16:11:29,063 44k INFO ====> Epoch: 136, cost 19.91 s
|
| 210 |
+
2023-03-18 16:11:48,924 44k INFO ====> Epoch: 137, cost 19.86 s
|
| 211 |
+
2023-03-18 16:12:08,871 44k INFO ====> Epoch: 138, cost 19.95 s
|
| 212 |
+
2023-03-18 16:12:19,890 44k INFO Train Epoch: 139 [46%]
|
| 213 |
+
2023-03-18 16:12:19,891 44k INFO Losses: [2.2704925537109375, 2.290191650390625, 11.153946876525879, 20.080801010131836, 0.7212128043174744], step: 5400, lr: 9.828968696598508e-05
|
| 214 |
+
2023-03-18 16:12:29,307 44k INFO ====> Epoch: 139, cost 20.44 s
|
| 215 |
+
2023-03-18 16:12:49,225 44k INFO ====> Epoch: 140, cost 19.92 s
|
| 216 |
+
2023-03-18 16:13:09,130 44k INFO ====> Epoch: 141, cost 19.90 s
|
| 217 |
+
2023-03-18 16:13:29,144 44k INFO ====> Epoch: 142, cost 20.01 s
|
| 218 |
+
2023-03-18 16:13:49,016 44k INFO ====> Epoch: 143, cost 19.87 s
|
| 219 |
+
2023-03-18 16:14:02,148 44k INFO Train Epoch: 144 [59%]
|
| 220 |
+
2023-03-18 16:14:02,149 44k INFO Losses: [1.7235863208770752, 3.459230899810791, 11.362791061401367, 15.820011138916016, 0.6718958020210266], step: 5600, lr: 9.822827126747529e-05
|
| 221 |
+
2023-03-18 16:14:06,691 44k INFO Saving model and optimizer state at iteration 144 to ./logs/44k/G_5600.pth
|
| 222 |
+
2023-03-18 16:14:07,736 44k INFO Saving model and optimizer state at iteration 144 to ./logs/44k/D_5600.pth
|
| 223 |
+
2023-03-18 16:14:15,101 44k INFO ====> Epoch: 144, cost 26.08 s
|
| 224 |
+
2023-03-18 16:14:35,033 44k INFO ====> Epoch: 145, cost 19.93 s
|
| 225 |
+
2023-03-18 16:14:55,256 44k INFO ====> Epoch: 146, cost 20.22 s
|
| 226 |
+
2023-03-18 16:15:16,052 44k INFO ====> Epoch: 147, cost 20.80 s
|
| 227 |
+
2023-03-18 16:15:35,922 44k INFO ====> Epoch: 148, cost 19.87 s
|
| 228 |
+
2023-03-18 16:15:51,015 44k INFO Train Epoch: 149 [72%]
|
| 229 |
+
2023-03-18 16:15:51,017 44k INFO Losses: [1.8325955867767334, 2.861543655395508, 11.257867813110352, 22.430797576904297, 0.9396995306015015], step: 5800, lr: 9.816689394418209e-05
|
| 230 |
+
2023-03-18 16:15:56,357 44k INFO ====> Epoch: 149, cost 20.44 s
|
| 231 |
+
2023-03-18 16:16:16,123 44k INFO ====> Epoch: 150, cost 19.77 s
|
| 232 |
+
2023-03-18 16:16:36,108 44k INFO ====> Epoch: 151, cost 19.99 s
|
| 233 |
+
2023-03-18 16:16:56,012 44k INFO ====> Epoch: 152, cost 19.90 s
|
| 234 |
+
2023-03-18 16:17:15,691 44k INFO ====> Epoch: 153, cost 19.68 s
|
| 235 |
+
2023-03-18 16:17:32,924 44k INFO Train Epoch: 154 [85%]
|
| 236 |
+
2023-03-18 16:17:32,925 44k INFO Losses: [2.4294025897979736, 2.6236627101898193, 8.439064979553223, 15.59575080871582, 0.777889609336853], step: 6000, lr: 9.810555497212693e-05
|
| 237 |
+
2023-03-18 16:17:36,044 44k INFO ====> Epoch: 154, cost 20.35 s
|
| 238 |
+
2023-03-18 16:17:58,583 44k INFO ====> Epoch: 155, cost 22.54 s
|
| 239 |
+
2023-03-18 16:18:18,616 44k INFO ====> Epoch: 156, cost 20.03 s
|
| 240 |
+
2023-03-18 16:18:38,531 44k INFO ====> Epoch: 157, cost 19.92 s
|
| 241 |
+
2023-03-18 16:18:58,590 44k INFO ====> Epoch: 158, cost 20.06 s
|
| 242 |
+
2023-03-18 16:19:18,098 44k INFO Train Epoch: 159 [97%]
|
| 243 |
+
2023-03-18 16:19:18,100 44k INFO Losses: [2.6305150985717773, 2.01100754737854, 5.256586074829102, 12.10132884979248, 0.5776822566986084], step: 6200, lr: 9.804425432734629e-05
|
| 244 |
+
2023-03-18 16:19:19,423 44k INFO ====> Epoch: 159, cost 20.83 s
|
| 245 |
+
2023-03-18 16:19:39,921 44k INFO ====> Epoch: 160, cost 20.50 s
|
| 246 |
+
2023-03-18 16:19:59,981 44k INFO ====> Epoch: 161, cost 20.06 s
|
| 247 |
+
2023-03-18 16:20:20,378 44k INFO ====> Epoch: 162, cost 20.40 s
|
| 248 |
+
2023-03-18 16:20:40,436 44k INFO ====> Epoch: 163, cost 20.06 s
|
| 249 |
+
2023-03-18 16:21:00,451 44k INFO ====> Epoch: 164, cost 20.01 s
|
| 250 |
+
2023-03-18 16:21:05,832 44k INFO Train Epoch: 165 [10%]
|
| 251 |
+
2023-03-18 16:21:05,834 44k INFO Losses: [2.3310093879699707, 2.7704384326934814, 6.842113018035889, 14.216880798339844, 0.8058108687400818], step: 6400, lr: 9.797074411189339e-05
|
| 252 |
+
2023-03-18 16:21:10,339 44k INFO Saving model and optimizer state at iteration 165 to ./logs/44k/G_6400.pth
|
| 253 |
+
2023-03-18 16:21:11,548 44k INFO Saving model and optimizer state at iteration 165 to ./logs/44k/D_6400.pth
|
| 254 |
+
2023-03-18 16:21:26,986 44k INFO ====> Epoch: 165, cost 26.54 s
|
| 255 |
+
2023-03-18 16:21:47,179 44k INFO ====> Epoch: 166, cost 20.19 s
|
| 256 |
+
2023-03-18 16:22:06,919 44k INFO ====> Epoch: 167, cost 19.74 s
|
| 257 |
+
2023-03-18 16:22:26,894 44k INFO ====> Epoch: 168, cost 19.97 s
|
| 258 |
+
2023-03-18 16:22:46,930 44k INFO ====> Epoch: 169, cost 20.04 s
|
| 259 |
+
2023-03-18 16:22:54,400 44k INFO Train Epoch: 170 [23%]
|
| 260 |
+
2023-03-18 16:22:54,401 44k INFO Losses: [2.2254819869995117, 2.3765923976898193, 11.166699409484863, 22.277254104614258, 0.7468198537826538], step: 6600, lr: 9.790952770283884e-05
|
| 261 |
+
2023-03-18 16:23:07,540 44k INFO ====> Epoch: 170, cost 20.61 s
|
| 262 |
+
2023-03-18 16:23:27,507 44k INFO ====> Epoch: 171, cost 19.97 s
|
| 263 |
+
2023-03-18 16:23:47,511 44k INFO ====> Epoch: 172, cost 20.00 s
|
| 264 |
+
2023-03-18 16:24:07,612 44k INFO ====> Epoch: 173, cost 20.10 s
|
| 265 |
+
2023-03-18 16:24:27,302 44k INFO ====> Epoch: 174, cost 19.69 s
|
| 266 |
+
2023-03-18 16:24:36,709 44k INFO Train Epoch: 175 [36%]
|
| 267 |
+
2023-03-18 16:24:36,711 44k INFO Losses: [2.520063877105713, 2.7583086490631104, 10.505017280578613, 20.471426010131836, 1.0650105476379395], step: 6800, lr: 9.784834954447608e-05
|
| 268 |
+
2023-03-18 16:24:47,651 44k INFO ====> Epoch: 175, cost 20.35 s
|
| 269 |
+
2023-03-18 16:25:07,509 44k INFO ====> Epoch: 176, cost 19.86 s
|
| 270 |
+
2023-03-18 16:25:27,155 44k INFO ====> Epoch: 177, cost 19.65 s
|
| 271 |
+
2023-03-18 16:25:46,776 44k INFO ====> Epoch: 178, cost 19.62 s
|
| 272 |
+
2023-03-18 16:26:06,771 44k INFO ====> Epoch: 179, cost 19.99 s
|
| 273 |
+
2023-03-18 16:26:18,260 44k INFO Train Epoch: 180 [49%]
|
| 274 |
+
2023-03-18 16:26:18,262 44k INFO Losses: [2.593710422515869, 2.4376070499420166, 9.219118118286133, 20.284957885742188, 0.6395339965820312], step: 7000, lr: 9.778720961290439e-05
|
| 275 |
+
2023-03-18 16:26:27,317 44k INFO ====> Epoch: 180, cost 20.55 s
|
| 276 |
+
2023-03-18 16:26:47,792 44k INFO ====> Epoch: 181, cost 20.47 s
|
| 277 |
+
2023-03-18 16:27:07,625 44k INFO ====> Epoch: 182, cost 19.83 s
|
| 278 |
+
2023-03-18 16:27:28,099 44k INFO ====> Epoch: 183, cost 20.47 s
|
| 279 |
+
2023-03-18 16:27:48,074 44k INFO ====> Epoch: 184, cost 19.98 s
|
| 280 |
+
2023-03-18 16:28:01,863 44k INFO Train Epoch: 185 [62%]
|
| 281 |
+
2023-03-18 16:28:01,865 44k INFO Losses: [2.254882574081421, 2.5508973598480225, 11.759598731994629, 22.320451736450195, 0.728751003742218], step: 7200, lr: 9.772610788423802e-05
|
| 282 |
+
2023-03-18 16:28:06,387 44k INFO Saving model and optimizer state at iteration 185 to ./logs/44k/G_7200.pth
|
| 283 |
+
2023-03-18 16:28:07,442 44k INFO Saving model and optimizer state at iteration 185 to ./logs/44k/D_7200.pth
|
| 284 |
+
2023-03-18 16:28:14,594 44k INFO ====> Epoch: 185, cost 26.52 s
|
| 285 |
+
2023-03-18 16:28:34,485 44k INFO ====> Epoch: 186, cost 19.89 s
|
| 286 |
+
2023-03-18 16:28:54,373 44k INFO ====> Epoch: 187, cost 19.89 s
|
| 287 |
+
2023-03-18 16:29:14,629 44k INFO ====> Epoch: 188, cost 20.26 s
|
| 288 |
+
2023-03-18 16:29:34,530 44k INFO ====> Epoch: 189, cost 19.90 s
|
| 289 |
+
2023-03-18 16:29:50,193 44k INFO Train Epoch: 190 [74%]
|
| 290 |
+
2023-03-18 16:29:50,195 44k INFO Losses: [2.3000588417053223, 2.6743597984313965, 12.745715141296387, 21.8736629486084, 1.0124669075012207], step: 7400, lr: 9.766504433460612e-05
|
| 291 |
+
2023-03-18 16:29:55,076 44k INFO ====> Epoch: 190, cost 20.55 s
|
| 292 |
+
2023-03-18 16:30:14,888 44k INFO ====> Epoch: 191, cost 19.81 s
|
| 293 |
+
2023-03-18 16:30:34,745 44k INFO ====> Epoch: 192, cost 19.86 s
|
| 294 |
+
2023-03-18 16:30:54,694 44k INFO ====> Epoch: 193, cost 19.95 s
|
| 295 |
+
2023-03-18 16:31:14,670 44k INFO ====> Epoch: 194, cost 19.98 s
|
| 296 |
+
2023-03-18 16:31:33,818 44k INFO Train Epoch: 195 [87%]
|
| 297 |
+
2023-03-18 16:31:33,820 44k INFO Losses: [2.2755260467529297, 1.8814972639083862, 11.437822341918945, 18.62908172607422, 0.9606003761291504], step: 7600, lr: 9.760401894015275e-05
|
| 298 |
+
2023-03-18 16:31:36,953 44k INFO ====> Epoch: 195, cost 22.28 s
|
| 299 |
+
2023-03-18 16:31:56,908 44k INFO ====> Epoch: 196, cost 19.95 s
|
| 300 |
+
2023-03-18 16:32:16,687 44k INFO ====> Epoch: 197, cost 19.78 s
|
| 301 |
+
2023-03-18 16:32:36,602 44k INFO ====> Epoch: 198, cost 19.92 s
|
| 302 |
+
2023-03-18 16:32:56,834 44k INFO ====> Epoch: 199, cost 20.23 s
|
| 303 |
+
2023-03-18 16:33:17,030 44k INFO ====> Epoch: 200, cost 20.20 s
|
| 304 |
+
2023-03-18 16:33:20,686 44k INFO Train Epoch: 201 [0%]
|
| 305 |
+
2023-03-18 16:33:20,688 44k INFO Losses: [2.6007676124572754, 2.2041707038879395, 11.113945960998535, 23.517358779907227, 0.8698729276657104], step: 7800, lr: 9.753083879807726e-05
|
| 306 |
+
2023-03-18 16:33:37,631 44k INFO ====> Epoch: 201, cost 20.60 s
|
| 307 |
+
2023-03-18 16:33:57,697 44k INFO ====> Epoch: 202, cost 20.07 s
|
| 308 |
+
2023-03-18 16:34:17,632 44k INFO ====> Epoch: 203, cost 19.93 s
|
| 309 |
+
2023-03-18 16:34:37,812 44k INFO ====> Epoch: 204, cost 20.18 s
|
| 310 |
+
2023-03-18 16:34:57,797 44k INFO ====> Epoch: 205, cost 19.99 s
|
| 311 |
+
2023-03-18 16:35:03,451 44k INFO Train Epoch: 206 [13%]
|
| 312 |
+
2023-03-18 16:35:03,453 44k INFO Losses: [2.4772579669952393, 2.2229604721069336, 12.207804679870605, 21.374996185302734, 0.4583211839199066], step: 8000, lr: 9.746989726111722e-05
|
| 313 |
+
2023-03-18 16:35:08,033 44k INFO Saving model and optimizer state at iteration 206 to ./logs/44k/G_8000.pth
|
| 314 |
+
2023-03-18 16:35:09,056 44k INFO Saving model and optimizer state at iteration 206 to ./logs/44k/D_8000.pth
|
| 315 |
+
2023-03-18 16:35:24,005 44k INFO ====> Epoch: 206, cost 26.21 s
|
| 316 |
+
2023-03-18 16:35:44,479 44k INFO ====> Epoch: 207, cost 20.47 s
|
| 317 |
+
2023-03-18 16:36:04,435 44k INFO ====> Epoch: 208, cost 19.96 s
|
| 318 |
+
2023-03-18 16:36:24,516 44k INFO ====> Epoch: 209, cost 20.08 s
|
| 319 |
+
2023-03-18 16:36:44,785 44k INFO ====> Epoch: 210, cost 20.27 s
|
| 320 |
+
2023-03-18 16:36:52,596 44k INFO Train Epoch: 211 [26%]
|
| 321 |
+
2023-03-18 16:36:52,598 44k INFO Losses: [2.1331419944763184, 2.492473602294922, 9.553248405456543, 17.68283462524414, 0.8661839365959167], step: 8200, lr: 9.740899380309685e-05
|
| 322 |
+
2023-03-18 16:37:05,613 44k INFO ====> Epoch: 211, cost 20.83 s
|
| 323 |
+
2023-03-18 16:37:25,466 44k INFO ====> Epoch: 212, cost 19.85 s
|
| 324 |
+
2023-03-18 16:37:45,586 44k INFO ====> Epoch: 213, cost 20.12 s
|
| 325 |
+
2023-03-18 16:38:06,088 44k INFO ====> Epoch: 214, cost 20.50 s
|
| 326 |
+
2023-03-18 16:38:25,877 44k INFO ====> Epoch: 215, cost 19.79 s
|
| 327 |
+
2023-03-18 16:38:35,618 44k INFO Train Epoch: 216 [38%]
|
| 328 |
+
2023-03-18 16:38:35,620 44k INFO Losses: [2.4721179008483887, 2.524754285812378, 8.135478973388672, 17.019596099853516, 0.7745293974876404], step: 8400, lr: 9.734812840022278e-05
|
| 329 |
+
2023-03-18 16:38:46,258 44k INFO ====> Epoch: 216, cost 20.38 s
|
| 330 |
+
2023-03-18 16:39:06,167 44k INFO ====> Epoch: 217, cost 19.91 s
|
| 331 |
+
2023-03-18 16:39:26,299 44k INFO ====> Epoch: 218, cost 20.13 s
|
| 332 |
+
2023-03-18 16:39:46,288 44k INFO ====> Epoch: 219, cost 19.99 s
|
| 333 |
+
2023-03-18 16:40:06,382 44k INFO ====> Epoch: 220, cost 20.09 s
|
| 334 |
+
2023-03-18 16:40:18,584 44k INFO Train Epoch: 221 [51%]
|
| 335 |
+
2023-03-18 16:40:18,586 44k INFO Losses: [2.237769842147827, 3.1392822265625, 14.79629135131836, 22.576732635498047, 0.5992900729179382], step: 8600, lr: 9.728730102871649e-05
|
| 336 |
+
2023-03-18 16:40:27,279 44k INFO ====> Epoch: 221, cost 20.90 s
|
| 337 |
+
2023-03-18 16:40:49,262 44k INFO ====> Epoch: 222, cost 21.98 s
|
| 338 |
+
2023-03-18 16:41:09,403 44k INFO ====> Epoch: 223, cost 20.14 s
|
| 339 |
+
2023-03-18 16:41:29,335 44k INFO ====> Epoch: 224, cost 19.93 s
|
| 340 |
+
2023-03-18 16:41:49,368 44k INFO ====> Epoch: 225, cost 20.03 s
|
| 341 |
+
2023-03-18 16:42:03,483 44k INFO Train Epoch: 226 [64%]
|
| 342 |
+
2023-03-18 16:42:03,484 44k INFO Losses: [2.1953744888305664, 2.8054661750793457, 12.387676239013672, 20.940820693969727, 0.895363450050354], step: 8800, lr: 9.722651166481428e-05
|
| 343 |
+
2023-03-18 16:42:08,135 44k INFO Saving model and optimizer state at iteration 226 to ./logs/44k/G_8800.pth
|
| 344 |
+
2023-03-18 16:42:09,165 44k INFO Saving model and optimizer state at iteration 226 to ./logs/44k/D_8800.pth
|
| 345 |
+
2023-03-18 16:42:15,680 44k INFO ====> Epoch: 226, cost 26.31 s
|
| 346 |
+
2023-03-18 16:42:35,788 44k INFO ====> Epoch: 227, cost 20.11 s
|
| 347 |
+
2023-03-18 16:42:55,674 44k INFO ====> Epoch: 228, cost 19.89 s
|
| 348 |
+
2023-03-18 16:43:15,864 44k INFO ====> Epoch: 229, cost 20.19 s
|
| 349 |
+
2023-03-18 16:43:35,656 44k INFO ====> Epoch: 230, cost 19.79 s
|
| 350 |
+
2023-03-18 16:43:51,625 44k INFO Train Epoch: 231 [77%]
|
| 351 |
+
2023-03-18 16:43:51,626 44k INFO Losses: [2.2360494136810303, 2.6518003940582275, 10.69636344909668, 20.383787155151367, 0.5332014560699463], step: 9000, lr: 9.716576028476738e-05
|
| 352 |
+
2023-03-18 16:43:56,079 44k INFO ====> Epoch: 231, cost 20.42 s
|
| 353 |
+
2023-03-18 16:44:15,806 44k INFO ====> Epoch: 232, cost 19.73 s
|
| 354 |
+
2023-03-18 16:44:35,609 44k INFO ====> Epoch: 233, cost 19.80 s
|
| 355 |
+
2023-03-18 16:44:55,628 44k INFO ====> Epoch: 234, cost 20.02 s
|
| 356 |
+
2023-03-18 16:45:15,712 44k INFO ====> Epoch: 235, cost 20.08 s
|
| 357 |
+
2023-03-18 16:45:34,104 44k INFO Train Epoch: 236 [90%]
|
| 358 |
+
2023-03-18 16:45:34,106 44k INFO Losses: [2.337740421295166, 2.3855791091918945, 10.388580322265625, 17.848812103271484, 0.6865084171295166], step: 9200, lr: 9.710504686484176e-05
|
| 359 |
+
2023-03-18 16:45:36,420 44k INFO ====> Epoch: 236, cost 20.71 s
|
| 360 |
+
2023-03-18 16:45:56,364 44k INFO ====> Epoch: 237, cost 19.94 s
|
| 361 |
+
2023-03-18 16:46:16,415 44k INFO ====> Epoch: 238, cost 20.05 s
|
| 362 |
+
2023-03-18 16:46:36,407 44k INFO ====> Epoch: 239, cost 19.99 s
|
| 363 |
+
2023-03-18 16:46:56,603 44k INFO ====> Epoch: 240, cost 20.20 s
|
| 364 |
+
2023-03-18 16:47:16,624 44k INFO ====> Epoch: 241, cost 20.02 s
|
| 365 |
+
2023-03-18 16:47:20,599 44k INFO Train Epoch: 242 [3%]
|
| 366 |
+
2023-03-18 16:47:20,601 44k INFO Losses: [2.1655261516571045, 2.7926673889160156, 13.443987846374512, 23.91498565673828, 0.7444508075714111], step: 9400, lr: 9.703224083489565e-05
|
| 367 |
+
2023-03-18 16:47:37,437 44k INFO ====> Epoch: 242, cost 20.81 s
|
| 368 |
+
2023-03-18 16:47:58,544 44k INFO ====> Epoch: 243, cost 21.11 s
|
| 369 |
+
2023-03-18 16:48:18,497 44k INFO ====> Epoch: 244, cost 19.95 s
|
| 370 |
+
2023-03-18 16:48:38,517 44k INFO ====> Epoch: 245, cost 20.02 s
|
| 371 |
+
2023-03-18 16:48:58,373 44k INFO ====> Epoch: 246, cost 19.86 s
|
| 372 |
+
2023-03-18 16:49:04,445 44k INFO Train Epoch: 247 [15%]
|
| 373 |
+
2023-03-18 16:49:04,446 44k INFO Losses: [2.8257107734680176, 2.245692253112793, 6.386197090148926, 12.429339408874512, 0.5741127133369446], step: 9600, lr: 9.69716108437664e-05
|
| 374 |
+
2023-03-18 16:49:08,880 44k INFO Saving model and optimizer state at iteration 247 to ./logs/44k/G_9600.pth
|
| 375 |
+
2023-03-18 16:49:09,927 44k INFO Saving model and optimizer state at iteration 247 to ./logs/44k/D_9600.pth
|
| 376 |
+
2023-03-18 16:49:24,184 44k INFO ====> Epoch: 247, cost 25.81 s
|
| 377 |
+
2023-03-18 16:49:44,408 44k INFO ====> Epoch: 248, cost 20.22 s
|
| 378 |
+
2023-03-18 16:50:04,256 44k INFO ====> Epoch: 249, cost 19.85 s
|
| 379 |
+
2023-03-18 16:50:24,296 44k INFO ====> Epoch: 250, cost 20.04 s
|
| 380 |
+
2023-03-18 16:50:44,262 44k INFO ====> Epoch: 251, cost 19.97 s
|
| 381 |
+
2023-03-18 16:50:52,950 44k INFO Train Epoch: 252 [28%]
|
| 382 |
+
2023-03-18 16:50:52,951 44k INFO Losses: [2.4544928073883057, 2.261460781097412, 7.994678974151611, 17.743366241455078, 0.5375484824180603], step: 9800, lr: 9.691101873690936e-05
|
| 383 |
+
2023-03-18 16:51:05,304 44k INFO ====> Epoch: 252, cost 21.04 s
|
| 384 |
+
2023-03-18 16:51:25,125 44k INFO ====> Epoch: 253, cost 19.82 s
|
| 385 |
+
2023-03-18 16:51:44,876 44k INFO ====> Epoch: 254, cost 19.75 s
|
| 386 |
+
2023-03-18 16:52:04,495 44k INFO ====> Epoch: 255, cost 19.62 s
|
| 387 |
+
2023-03-18 16:52:24,331 44k INFO ====> Epoch: 256, cost 19.84 s
|
| 388 |
+
2023-03-18 16:52:34,494 44k INFO Train Epoch: 257 [41%]
|
| 389 |
+
2023-03-18 16:52:34,495 44k INFO Losses: [2.2319655418395996, 2.4747824668884277, 10.411455154418945, 17.785585403442383, 0.6818886995315552], step: 10000, lr: 9.685046449065278e-05
|
| 390 |
+
2023-03-18 16:52:44,934 44k INFO ====> Epoch: 257, cost 20.60 s
|
| 391 |
+
2023-03-18 16:53:05,038 44k INFO ====> Epoch: 258, cost 20.10 s
|
| 392 |
+
2023-03-18 16:53:24,951 44k INFO ====> Epoch: 259, cost 19.91 s
|
| 393 |
+
2023-03-18 16:53:44,824 44k INFO ====> Epoch: 260, cost 19.87 s
|
| 394 |
+
2023-03-18 16:54:04,673 44k INFO ====> Epoch: 261, cost 19.85 s
|
| 395 |
+
2023-03-18 16:54:17,273 44k INFO Train Epoch: 262 [54%]
|
| 396 |
+
2023-03-18 16:54:17,275 44k INFO Losses: [2.368034839630127, 2.4957849979400635, 11.374425888061523, 21.15058135986328, 0.9322990775108337], step: 10200, lr: 9.678994808133967e-05
|
| 397 |
+
2023-03-18 16:54:25,409 44k INFO ====> Epoch: 262, cost 20.74 s
|
| 398 |
+
2023-03-18 16:54:45,150 44k INFO ====> Epoch: 263, cost 19.74 s
|
| 399 |
+
2023-03-18 16:55:05,117 44k INFO ====> Epoch: 264, cost 19.97 s
|
| 400 |
+
2023-03-18 16:55:25,454 44k INFO ====> Epoch: 265, cost 20.34 s
|
| 401 |
+
2023-03-18 16:55:45,558 44k INFO ====> Epoch: 266, cost 20.10 s
|
| 402 |
+
2023-03-18 16:56:00,233 44k INFO Train Epoch: 267 [67%]
|
| 403 |
+
2023-03-18 16:56:00,234 44k INFO Losses: [2.170872926712036, 2.5498008728027344, 12.449470520019531, 20.475133895874023, 0.5462685823440552], step: 10400, lr: 9.67294694853279e-05
|
| 404 |
+
2023-03-18 16:56:04,770 44k INFO Saving model and optimizer state at iteration 267 to ./logs/44k/G_10400.pth
|
| 405 |
+
2023-03-18 16:56:05,822 44k INFO Saving model and optimizer state at iteration 267 to ./logs/44k/D_10400.pth
|
| 406 |
+
2023-03-18 16:56:11,928 44k INFO ====> Epoch: 267, cost 26.37 s
|
| 407 |
+
2023-03-18 16:56:31,938 44k INFO ====> Epoch: 268, cost 20.01 s
|
| 408 |
+
2023-03-18 16:56:51,823 44k INFO ====> Epoch: 269, cost 19.89 s
|
| 409 |
+
2023-03-18 16:57:11,802 44k INFO ====> Epoch: 270, cost 19.98 s
|
| 410 |
+
2023-03-18 16:57:31,690 44k INFO ====> Epoch: 271, cost 19.89 s
|
| 411 |
+
2023-03-18 16:57:48,179 44k INFO Train Epoch: 272 [79%]
|
| 412 |
+
2023-03-18 16:57:48,181 44k INFO Losses: [2.1866960525512695, 2.445249319076538, 10.951682090759277, 19.355749130249023, 0.6413354873657227], step: 10600, lr: 9.666902867899003e-05
|
| 413 |
+
2023-03-18 16:57:52,250 44k INFO ====> Epoch: 272, cost 20.56 s
|
| 414 |
+
2023-03-18 16:58:12,989 44k INFO ====> Epoch: 273, cost 20.74 s
|
| 415 |
+
2023-03-18 16:58:32,781 44k INFO ====> Epoch: 274, cost 19.79 s
|
| 416 |
+
2023-03-18 16:58:52,693 44k INFO ====> Epoch: 275, cost 19.91 s
|
| 417 |
+
2023-03-18 16:59:12,527 44k INFO ====> Epoch: 276, cost 19.83 s
|
| 418 |
+
2023-03-18 16:59:31,113 44k INFO Train Epoch: 277 [92%]
|
| 419 |
+
2023-03-18 16:59:31,115 44k INFO Losses: [2.256988525390625, 2.6625442504882812, 9.01707649230957, 18.851524353027344, 0.5929552316665649], step: 10800, lr: 9.660862563871342e-05
|
| 420 |
+
2023-03-18 16:59:33,039 44k INFO ====> Epoch: 277, cost 20.51 s
|
| 421 |
+
2023-03-18 16:59:53,278 44k INFO ====> Epoch: 278, cost 20.24 s
|
| 422 |
+
2023-03-18 17:00:13,389 44k INFO ====> Epoch: 279, cost 20.11 s
|
| 423 |
+
2023-03-18 17:00:33,239 44k INFO ====> Epoch: 280, cost 19.85 s
|
| 424 |
+
2023-03-18 17:00:53,087 44k INFO ====> Epoch: 281, cost 19.85 s
|
| 425 |
+
2023-03-18 17:01:14,621 44k INFO ====> Epoch: 282, cost 21.53 s
|
| 426 |
+
2023-03-18 17:01:18,980 44k INFO Train Epoch: 283 [5%]
|
| 427 |
+
2023-03-18 17:01:18,981 44k INFO Losses: [2.56038761138916, 2.2084388732910156, 9.195663452148438, 16.996200561523438, 0.7333958745002747], step: 11000, lr: 9.653619180835758e-05
|
| 428 |
+
2023-03-18 17:01:35,126 44k INFO ====> Epoch: 283, cost 20.50 s
|
| 429 |
+
2023-03-18 17:01:54,863 44k INFO ====> Epoch: 284, cost 19.74 s
|
| 430 |
+
2023-03-18 17:02:14,868 44k INFO ====> Epoch: 285, cost 20.01 s
|
| 431 |
+
2023-03-18 17:02:34,792 44k INFO ====> Epoch: 286, cost 19.92 s
|
| 432 |
+
2023-03-18 17:02:54,864 44k INFO ====> Epoch: 287, cost 20.07 s
|
| 433 |
+
2023-03-18 17:03:01,838 44k INFO Train Epoch: 288 [18%]
|
| 434 |
+
2023-03-18 17:03:01,839 44k INFO Losses: [2.502936363220215, 2.2773265838623047, 9.0471773147583, 18.72901725769043, 0.9384689331054688], step: 11200, lr: 9.647587177037196e-05
|
| 435 |
+
2023-03-18 17:03:06,229 44k INFO Saving model and optimizer state at iteration 288 to ./logs/44k/G_11200.pth
|
| 436 |
+
2023-03-18 17:03:07,422 44k INFO Saving model and optimizer state at iteration 288 to ./logs/44k/D_11200.pth
|
| 437 |
+
2023-03-18 17:03:21,525 44k INFO ====> Epoch: 288, cost 26.66 s
|
| 438 |
+
2023-03-18 17:03:41,370 44k INFO ====> Epoch: 289, cost 19.85 s
|
| 439 |
+
2023-03-18 17:04:01,209 44k INFO ====> Epoch: 290, cost 19.84 s
|
| 440 |
+
2023-03-18 17:04:22,790 44k INFO ====> Epoch: 291, cost 21.58 s
|
| 441 |
+
2023-03-18 17:04:42,738 44k INFO ====> Epoch: 292, cost 19.95 s
|
| 442 |
+
2023-03-18 17:04:51,194 44k INFO Train Epoch: 293 [31%]
|
| 443 |
+
2023-03-18 17:04:51,196 44k INFO Losses: [2.4107394218444824, 2.5568552017211914, 11.272177696228027, 19.823131561279297, 0.7836748957633972], step: 11400, lr: 9.641558942298625e-05
|
| 444 |
+
2023-03-18 17:05:03,095 44k INFO ====> Epoch: 293, cost 20.36 s
|
| 445 |
+
2023-03-18 17:05:23,282 44k INFO ====> Epoch: 294, cost 20.19 s
|
| 446 |
+
2023-03-18 17:05:43,171 44k INFO ====> Epoch: 295, cost 19.89 s
|
| 447 |
+
2023-03-18 17:06:03,498 44k INFO ====> Epoch: 296, cost 20.33 s
|
| 448 |
+
2023-03-18 17:06:23,305 44k INFO ====> Epoch: 297, cost 19.81 s
|
| 449 |
+
2023-03-18 17:06:33,864 44k INFO Train Epoch: 298 [44%]
|
| 450 |
+
2023-03-18 17:06:33,865 44k INFO Losses: [2.7009692192077637, 2.125077962875366, 7.26710319519043, 11.231626510620117, 0.8319841623306274], step: 11600, lr: 9.635534474264972e-05
|
| 451 |
+
2023-03-18 17:06:43,833 44k INFO ====> Epoch: 298, cost 20.53 s
|
| 452 |
+
2023-03-18 17:07:03,635 44k INFO ====> Epoch: 299, cost 19.80 s
|
| 453 |
+
2023-03-18 17:07:23,150 44k INFO ====> Epoch: 300, cost 19.51 s
|
| 454 |
+
2023-03-18 17:07:43,090 44k INFO ====> Epoch: 301, cost 19.94 s
|
| 455 |
+
2023-03-18 17:08:03,333 44k INFO ====> Epoch: 302, cost 20.24 s
|
| 456 |
+
2023-03-18 17:08:16,012 44k INFO Train Epoch: 303 [56%]
|
| 457 |
+
2023-03-18 17:08:16,013 44k INFO Losses: [2.552933931350708, 2.472691059112549, 9.617679595947266, 20.058349609375, 0.4773886203765869], step: 11800, lr: 9.629513770582634e-05
|
| 458 |
+
2023-03-18 17:08:23,517 44k INFO ====> Epoch: 303, cost 20.18 s
|
| 459 |
+
2023-03-18 17:08:43,187 44k INFO ====> Epoch: 304, cost 19.67 s
|
| 460 |
+
2023-03-18 17:09:03,569 44k INFO ====> Epoch: 305, cost 20.38 s
|
| 461 |
+
2023-03-18 17:09:23,391 44k INFO ====> Epoch: 306, cost 19.82 s
|
| 462 |
+
2023-03-18 17:09:42,960 44k INFO ====> Epoch: 307, cost 19.57 s
|
| 463 |
+
2023-03-18 17:09:57,696 44k INFO Train Epoch: 308 [69%]
|
| 464 |
+
2023-03-18 17:09:57,697 44k INFO Losses: [2.641923427581787, 2.117219924926758, 10.303349494934082, 20.380632400512695, 0.3930005133152008], step: 12000, lr: 9.62349682889948e-05
|
| 465 |
+
2023-03-18 17:10:01,898 44k INFO Saving model and optimizer state at iteration 308 to ./logs/44k/G_12000.pth
|
| 466 |
+
2023-03-18 17:10:02,834 44k INFO Saving model and optimizer state at iteration 308 to ./logs/44k/D_12000.pth
|
| 467 |
+
2023-03-18 17:10:08,387 44k INFO ====> Epoch: 308, cost 25.43 s
|
| 468 |
+
2023-03-18 17:10:28,754 44k INFO ====> Epoch: 309, cost 20.37 s
|
| 469 |
+
2023-03-18 17:10:48,349 44k INFO ====> Epoch: 310, cost 19.60 s
|
| 470 |
+
2023-03-18 17:11:09,263 44k INFO ====> Epoch: 311, cost 20.91 s
|
| 471 |
+
2023-03-18 17:11:28,821 44k INFO ====> Epoch: 312, cost 19.56 s
|
| 472 |
+
2023-03-18 17:11:45,379 44k INFO Train Epoch: 313 [82%]
|
| 473 |
+
2023-03-18 17:11:45,380 44k INFO Losses: [2.7167956829071045, 2.2329747676849365, 7.7941179275512695, 16.260107040405273, 0.6568765044212341], step: 12200, lr: 9.617483646864849e-05
|
| 474 |
+
2023-03-18 17:11:48,865 44k INFO ====> Epoch: 313, cost 20.04 s
|
| 475 |
+
2023-03-18 17:12:08,448 44k INFO ====> Epoch: 314, cost 19.58 s
|
| 476 |
+
2023-03-18 17:12:28,134 44k INFO ====> Epoch: 315, cost 19.69 s
|
| 477 |
+
2023-03-18 17:12:47,471 44k INFO ====> Epoch: 316, cost 19.34 s
|
| 478 |
+
2023-03-18 17:13:06,821 44k INFO ====> Epoch: 317, cost 19.35 s
|
| 479 |
+
2023-03-18 17:13:25,502 44k INFO Train Epoch: 318 [95%]
|
| 480 |
+
2023-03-18 17:13:25,503 44k INFO Losses: [2.3885555267333984, 2.3887124061584473, 7.18328857421875, 17.944881439208984, 0.5438012480735779], step: 12400, lr: 9.611474222129547e-05
|
| 481 |
+
2023-03-18 17:13:27,095 44k INFO ====> Epoch: 318, cost 20.27 s
|
| 482 |
+
2023-03-18 17:13:46,557 44k INFO ====> Epoch: 319, cost 19.46 s
|
| 483 |
+
2023-03-18 17:14:06,371 44k INFO ====> Epoch: 320, cost 19.81 s
|
| 484 |
+
2023-03-18 17:14:26,017 44k INFO ====> Epoch: 321, cost 19.65 s
|
| 485 |
+
2023-03-18 17:14:45,545 44k INFO ====> Epoch: 322, cost 19.53 s
|
| 486 |
+
2023-03-18 17:15:04,846 44k INFO ====> Epoch: 323, cost 19.30 s
|
| 487 |
+
2023-03-18 17:15:09,498 44k INFO Train Epoch: 324 [8%]
|
| 488 |
+
2023-03-18 17:15:09,499 44k INFO Losses: [2.1702542304992676, 2.652705430984497, 11.307723045349121, 21.379201889038086, 0.8918676972389221], step: 12600, lr: 9.604267868776807e-05
|
| 489 |
+
2023-03-18 17:15:24,846 44k INFO ====> Epoch: 324, cost 20.00 s
|
| 490 |
+
2023-03-18 17:15:44,240 44k INFO ====> Epoch: 325, cost 19.39 s
|
| 491 |
+
2023-03-18 17:16:03,710 44k INFO ====> Epoch: 326, cost 19.47 s
|
| 492 |
+
2023-03-18 17:16:23,110 44k INFO ====> Epoch: 327, cost 19.40 s
|
| 493 |
+
2023-03-18 17:16:42,685 44k INFO ====> Epoch: 328, cost 19.58 s
|
| 494 |
+
2023-03-18 17:16:49,386 44k INFO Train Epoch: 329 [21%]
|
| 495 |
+
2023-03-18 17:16:49,387 44k INFO Losses: [2.395970344543457, 2.3684022426605225, 10.31136417388916, 17.88117790222168, 0.6932646036148071], step: 12800, lr: 9.5982667018381e-05
|
| 496 |
+
2023-03-18 17:16:53,635 44k INFO Saving model and optimizer state at iteration 329 to ./logs/44k/G_12800.pth
|
| 497 |
+
2023-03-18 17:16:54,607 44k INFO Saving model and optimizer state at iteration 329 to ./logs/44k/D_12800.pth
|
| 498 |
+
2023-03-18 17:17:08,139 44k INFO ====> Epoch: 329, cost 25.45 s
|
so-vits-svc-4.0/20230324-Juewa-44k/best/D_12800.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aa3baa70695e03d963e3ac873c87760b0ca6d02b3b9e94c9716ffe258ec6130d
|
| 3 |
+
size 561098185
|
so-vits-svc-4.0/20230324-Juewa-44k/best/G_12800.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5496c02bff57cd0614a70194b170c43767556242d7b0c8fba4d6ac437f7a9715
|
| 3 |
+
size 542789405
|
so-vits-svc-4.0/20230324-Juewa-44k/best/config.json
ADDED
|
@@ -0,0 +1,93 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"train": {
|
| 3 |
+
"log_interval": 200,
|
| 4 |
+
"eval_interval": 800,
|
| 5 |
+
"seed": 1234,
|
| 6 |
+
"epochs": 10000,
|
| 7 |
+
"learning_rate": 0.0001,
|
| 8 |
+
"betas": [
|
| 9 |
+
0.8,
|
| 10 |
+
0.99
|
| 11 |
+
],
|
| 12 |
+
"eps": 1e-09,
|
| 13 |
+
"batch_size": 6,
|
| 14 |
+
"fp16_run": false,
|
| 15 |
+
"lr_decay": 0.999875,
|
| 16 |
+
"segment_size": 10240,
|
| 17 |
+
"init_lr_ratio": 1,
|
| 18 |
+
"warmup_epochs": 0,
|
| 19 |
+
"c_mel": 45,
|
| 20 |
+
"c_kl": 1.0,
|
| 21 |
+
"use_sr": true,
|
| 22 |
+
"max_speclen": 512,
|
| 23 |
+
"port": "8001",
|
| 24 |
+
"keep_ckpts": 99
|
| 25 |
+
},
|
| 26 |
+
"data": {
|
| 27 |
+
"training_files": "filelists/train.txt",
|
| 28 |
+
"validation_files": "filelists/val.txt",
|
| 29 |
+
"max_wav_value": 32768.0,
|
| 30 |
+
"sampling_rate": 44100,
|
| 31 |
+
"filter_length": 2048,
|
| 32 |
+
"hop_length": 512,
|
| 33 |
+
"win_length": 2048,
|
| 34 |
+
"n_mel_channels": 80,
|
| 35 |
+
"mel_fmin": 0.0,
|
| 36 |
+
"mel_fmax": 22050
|
| 37 |
+
},
|
| 38 |
+
"model": {
|
| 39 |
+
"inter_channels": 192,
|
| 40 |
+
"hidden_channels": 192,
|
| 41 |
+
"filter_channels": 768,
|
| 42 |
+
"n_heads": 2,
|
| 43 |
+
"n_layers": 6,
|
| 44 |
+
"kernel_size": 3,
|
| 45 |
+
"p_dropout": 0.1,
|
| 46 |
+
"resblock": "1",
|
| 47 |
+
"resblock_kernel_sizes": [
|
| 48 |
+
3,
|
| 49 |
+
7,
|
| 50 |
+
11
|
| 51 |
+
],
|
| 52 |
+
"resblock_dilation_sizes": [
|
| 53 |
+
[
|
| 54 |
+
1,
|
| 55 |
+
3,
|
| 56 |
+
5
|
| 57 |
+
],
|
| 58 |
+
[
|
| 59 |
+
1,
|
| 60 |
+
3,
|
| 61 |
+
5
|
| 62 |
+
],
|
| 63 |
+
[
|
| 64 |
+
1,
|
| 65 |
+
3,
|
| 66 |
+
5
|
| 67 |
+
]
|
| 68 |
+
],
|
| 69 |
+
"upsample_rates": [
|
| 70 |
+
8,
|
| 71 |
+
8,
|
| 72 |
+
2,
|
| 73 |
+
2,
|
| 74 |
+
2
|
| 75 |
+
],
|
| 76 |
+
"upsample_initial_channel": 512,
|
| 77 |
+
"upsample_kernel_sizes": [
|
| 78 |
+
16,
|
| 79 |
+
16,
|
| 80 |
+
4,
|
| 81 |
+
4,
|
| 82 |
+
4
|
| 83 |
+
],
|
| 84 |
+
"n_layers_q": 3,
|
| 85 |
+
"use_spectral_norm": false,
|
| 86 |
+
"gin_channels": 256,
|
| 87 |
+
"ssl_dim": 256,
|
| 88 |
+
"n_speakers": 200
|
| 89 |
+
},
|
| 90 |
+
"spk": {
|
| 91 |
+
"Juewa": 0
|
| 92 |
+
}
|
| 93 |
+
}
|
so-vits-svc-4.0/20230324-Juewa-44k/config.json
ADDED
|
@@ -0,0 +1,93 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"train": {
|
| 3 |
+
"log_interval": 200,
|
| 4 |
+
"eval_interval": 800,
|
| 5 |
+
"seed": 1234,
|
| 6 |
+
"epochs": 10000,
|
| 7 |
+
"learning_rate": 0.0001,
|
| 8 |
+
"betas": [
|
| 9 |
+
0.8,
|
| 10 |
+
0.99
|
| 11 |
+
],
|
| 12 |
+
"eps": 1e-09,
|
| 13 |
+
"batch_size": 6,
|
| 14 |
+
"fp16_run": false,
|
| 15 |
+
"lr_decay": 0.999875,
|
| 16 |
+
"segment_size": 10240,
|
| 17 |
+
"init_lr_ratio": 1,
|
| 18 |
+
"warmup_epochs": 0,
|
| 19 |
+
"c_mel": 45,
|
| 20 |
+
"c_kl": 1.0,
|
| 21 |
+
"use_sr": true,
|
| 22 |
+
"max_speclen": 512,
|
| 23 |
+
"port": "8001",
|
| 24 |
+
"keep_ckpts": 99
|
| 25 |
+
},
|
| 26 |
+
"data": {
|
| 27 |
+
"training_files": "filelists/train.txt",
|
| 28 |
+
"validation_files": "filelists/val.txt",
|
| 29 |
+
"max_wav_value": 32768.0,
|
| 30 |
+
"sampling_rate": 44100,
|
| 31 |
+
"filter_length": 2048,
|
| 32 |
+
"hop_length": 512,
|
| 33 |
+
"win_length": 2048,
|
| 34 |
+
"n_mel_channels": 80,
|
| 35 |
+
"mel_fmin": 0.0,
|
| 36 |
+
"mel_fmax": 22050
|
| 37 |
+
},
|
| 38 |
+
"model": {
|
| 39 |
+
"inter_channels": 192,
|
| 40 |
+
"hidden_channels": 192,
|
| 41 |
+
"filter_channels": 768,
|
| 42 |
+
"n_heads": 2,
|
| 43 |
+
"n_layers": 6,
|
| 44 |
+
"kernel_size": 3,
|
| 45 |
+
"p_dropout": 0.1,
|
| 46 |
+
"resblock": "1",
|
| 47 |
+
"resblock_kernel_sizes": [
|
| 48 |
+
3,
|
| 49 |
+
7,
|
| 50 |
+
11
|
| 51 |
+
],
|
| 52 |
+
"resblock_dilation_sizes": [
|
| 53 |
+
[
|
| 54 |
+
1,
|
| 55 |
+
3,
|
| 56 |
+
5
|
| 57 |
+
],
|
| 58 |
+
[
|
| 59 |
+
1,
|
| 60 |
+
3,
|
| 61 |
+
5
|
| 62 |
+
],
|
| 63 |
+
[
|
| 64 |
+
1,
|
| 65 |
+
3,
|
| 66 |
+
5
|
| 67 |
+
]
|
| 68 |
+
],
|
| 69 |
+
"upsample_rates": [
|
| 70 |
+
8,
|
| 71 |
+
8,
|
| 72 |
+
2,
|
| 73 |
+
2,
|
| 74 |
+
2
|
| 75 |
+
],
|
| 76 |
+
"upsample_initial_channel": 512,
|
| 77 |
+
"upsample_kernel_sizes": [
|
| 78 |
+
16,
|
| 79 |
+
16,
|
| 80 |
+
4,
|
| 81 |
+
4,
|
| 82 |
+
4
|
| 83 |
+
],
|
| 84 |
+
"n_layers_q": 3,
|
| 85 |
+
"use_spectral_norm": false,
|
| 86 |
+
"gin_channels": 256,
|
| 87 |
+
"ssl_dim": 256,
|
| 88 |
+
"n_speakers": 200
|
| 89 |
+
},
|
| 90 |
+
"spk": {
|
| 91 |
+
"Juewa": 0
|
| 92 |
+
}
|
| 93 |
+
}
|
so-vits-svc-4.0/20230324-Juewa-44k/kmeans_10000.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3c8f7393e26f890587694355d6b45b6b2ee055ec679991d5f3f0f291832c9b87
|
| 3 |
+
size 15441785
|
so-vits-svc-4.0/20230324-Juewa-44k/pruned/P_G_12800.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a2a99b10fe6dfda20fd5073d58b7ea71b06c2e648989e0614fe605869fafb0e5
|
| 3 |
+
size 180885829
|
so-vits-svc-4.0/20230324-Juewa-44k/pruned/config.json
ADDED
|
@@ -0,0 +1,93 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"train": {
|
| 3 |
+
"log_interval": 200,
|
| 4 |
+
"eval_interval": 800,
|
| 5 |
+
"seed": 1234,
|
| 6 |
+
"epochs": 10000,
|
| 7 |
+
"learning_rate": 0.0001,
|
| 8 |
+
"betas": [
|
| 9 |
+
0.8,
|
| 10 |
+
0.99
|
| 11 |
+
],
|
| 12 |
+
"eps": 1e-09,
|
| 13 |
+
"batch_size": 6,
|
| 14 |
+
"fp16_run": false,
|
| 15 |
+
"lr_decay": 0.999875,
|
| 16 |
+
"segment_size": 10240,
|
| 17 |
+
"init_lr_ratio": 1,
|
| 18 |
+
"warmup_epochs": 0,
|
| 19 |
+
"c_mel": 45,
|
| 20 |
+
"c_kl": 1.0,
|
| 21 |
+
"use_sr": true,
|
| 22 |
+
"max_speclen": 512,
|
| 23 |
+
"port": "8001",
|
| 24 |
+
"keep_ckpts": 99
|
| 25 |
+
},
|
| 26 |
+
"data": {
|
| 27 |
+
"training_files": "filelists/train.txt",
|
| 28 |
+
"validation_files": "filelists/val.txt",
|
| 29 |
+
"max_wav_value": 32768.0,
|
| 30 |
+
"sampling_rate": 44100,
|
| 31 |
+
"filter_length": 2048,
|
| 32 |
+
"hop_length": 512,
|
| 33 |
+
"win_length": 2048,
|
| 34 |
+
"n_mel_channels": 80,
|
| 35 |
+
"mel_fmin": 0.0,
|
| 36 |
+
"mel_fmax": 22050
|
| 37 |
+
},
|
| 38 |
+
"model": {
|
| 39 |
+
"inter_channels": 192,
|
| 40 |
+
"hidden_channels": 192,
|
| 41 |
+
"filter_channels": 768,
|
| 42 |
+
"n_heads": 2,
|
| 43 |
+
"n_layers": 6,
|
| 44 |
+
"kernel_size": 3,
|
| 45 |
+
"p_dropout": 0.1,
|
| 46 |
+
"resblock": "1",
|
| 47 |
+
"resblock_kernel_sizes": [
|
| 48 |
+
3,
|
| 49 |
+
7,
|
| 50 |
+
11
|
| 51 |
+
],
|
| 52 |
+
"resblock_dilation_sizes": [
|
| 53 |
+
[
|
| 54 |
+
1,
|
| 55 |
+
3,
|
| 56 |
+
5
|
| 57 |
+
],
|
| 58 |
+
[
|
| 59 |
+
1,
|
| 60 |
+
3,
|
| 61 |
+
5
|
| 62 |
+
],
|
| 63 |
+
[
|
| 64 |
+
1,
|
| 65 |
+
3,
|
| 66 |
+
5
|
| 67 |
+
]
|
| 68 |
+
],
|
| 69 |
+
"upsample_rates": [
|
| 70 |
+
8,
|
| 71 |
+
8,
|
| 72 |
+
2,
|
| 73 |
+
2,
|
| 74 |
+
2
|
| 75 |
+
],
|
| 76 |
+
"upsample_initial_channel": 512,
|
| 77 |
+
"upsample_kernel_sizes": [
|
| 78 |
+
16,
|
| 79 |
+
16,
|
| 80 |
+
4,
|
| 81 |
+
4,
|
| 82 |
+
4
|
| 83 |
+
],
|
| 84 |
+
"n_layers_q": 3,
|
| 85 |
+
"use_spectral_norm": false,
|
| 86 |
+
"gin_channels": 256,
|
| 87 |
+
"ssl_dim": 256,
|
| 88 |
+
"n_speakers": 200
|
| 89 |
+
},
|
| 90 |
+
"spk": {
|
| 91 |
+
"Juewa": 0
|
| 92 |
+
}
|
| 93 |
+
}
|
so-vits-svc-4.0/20230324-Juewa-44k/train.log
ADDED
|
@@ -0,0 +1,431 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2023-03-24 16:10:32,809 44k INFO {'train': {'log_interval': 200, 'eval_interval': 800, 'seed': 1234, 'epochs': 10000, 'learning_rate': 0.0001, 'betas': [0.8, 0.99], 'eps': 1e-09, 'batch_size': 6, 'fp16_run': False, 'lr_decay': 0.999875, 'segment_size': 10240, 'init_lr_ratio': 1, 'warmup_epochs': 0, 'c_mel': 45, 'c_kl': 1.0, 'use_sr': True, 'max_speclen': 512, 'port': '8001', 'keep_ckpts': 99}, 'data': {'training_files': 'filelists/train.txt', 'validation_files': 'filelists/val.txt', 'max_wav_value': 32768.0, 'sampling_rate': 44100, 'filter_length': 2048, 'hop_length': 512, 'win_length': 2048, 'n_mel_channels': 80, 'mel_fmin': 0.0, 'mel_fmax': 22050}, 'model': {'inter_channels': 192, 'hidden_channels': 192, 'filter_channels': 768, 'n_heads': 2, 'n_layers': 6, 'kernel_size': 3, 'p_dropout': 0.1, 'resblock': '1', 'resblock_kernel_sizes': [3, 7, 11], 'resblock_dilation_sizes': [[1, 3, 5], [1, 3, 5], [1, 3, 5]], 'upsample_rates': [8, 8, 2, 2, 2], 'upsample_initial_channel': 512, 'upsample_kernel_sizes': [16, 16, 4, 4, 4], 'n_layers_q': 3, 'use_spectral_norm': False, 'gin_channels': 256, 'ssl_dim': 256, 'n_speakers': 200}, 'spk': {'Juewa': 0}, 'model_dir': './logs/44k'}
|
| 2 |
+
2023-03-24 16:10:32,810 44k WARNING /root/so-vits-svc-4.0 is not a git repository, therefore hash value comparison will be ignored.
|
| 3 |
+
2023-03-24 16:10:35,834 44k INFO emb_g.weight is not in the checkpoint
|
| 4 |
+
2023-03-24 16:10:35,904 44k INFO Loaded checkpoint './logs/44k/G_0.pth' (iteration 0)
|
| 5 |
+
2023-03-24 16:10:36,087 44k INFO Loaded checkpoint './logs/44k/D_0.pth' (iteration 0)
|
| 6 |
+
2023-03-24 16:10:44,069 44k INFO Train Epoch: 1 [0%]
|
| 7 |
+
2023-03-24 16:10:44,070 44k INFO Losses: [2.755509376525879, 2.2334935665130615, 12.809167861938477, 35.51326370239258, 3.423668384552002], step: 0, lr: 0.0001
|
| 8 |
+
2023-03-24 16:10:49,834 44k INFO Saving model and optimizer state at iteration 1 to ./logs/44k/G_0.pth
|
| 9 |
+
2023-03-24 16:10:51,452 44k INFO Saving model and optimizer state at iteration 1 to ./logs/44k/D_0.pth
|
| 10 |
+
2023-03-24 16:11:16,686 44k INFO ====> Epoch: 1, cost 43.88 s
|
| 11 |
+
2023-03-24 16:11:42,464 44k INFO ====> Epoch: 2, cost 25.78 s
|
| 12 |
+
2023-03-24 16:12:07,775 44k INFO ====> Epoch: 3, cost 25.31 s
|
| 13 |
+
2023-03-24 16:12:32,967 44k INFO ====> Epoch: 4, cost 25.19 s
|
| 14 |
+
2023-03-24 16:12:36,634 44k INFO Train Epoch: 5 [0%]
|
| 15 |
+
2023-03-24 16:12:36,636 44k INFO Losses: [2.4600982666015625, 2.0119576454162598, 11.58346939086914, 19.77667236328125, 1.1188390254974365], step: 200, lr: 9.995000937421877e-05
|
| 16 |
+
2023-03-24 16:12:58,844 44k INFO ====> Epoch: 5, cost 25.88 s
|
| 17 |
+
2023-03-24 16:13:23,698 44k INFO ====> Epoch: 6, cost 24.85 s
|
| 18 |
+
2023-03-24 16:13:48,641 44k INFO ====> Epoch: 7, cost 24.94 s
|
| 19 |
+
2023-03-24 16:14:13,609 44k INFO ====> Epoch: 8, cost 24.97 s
|
| 20 |
+
2023-03-24 16:14:17,212 44k INFO Train Epoch: 9 [0%]
|
| 21 |
+
2023-03-24 16:14:17,214 44k INFO Losses: [2.423004627227783, 2.078486919403076, 9.993337631225586, 23.189149856567383, 1.4135551452636719], step: 400, lr: 9.990004373906418e-05
|
| 22 |
+
2023-03-24 16:14:39,134 44k INFO ====> Epoch: 9, cost 25.52 s
|
| 23 |
+
2023-03-24 16:15:04,111 44k INFO ====> Epoch: 10, cost 24.98 s
|
| 24 |
+
2023-03-24 16:15:28,810 44k INFO ====> Epoch: 11, cost 24.70 s
|
| 25 |
+
2023-03-24 16:15:53,523 44k INFO ====> Epoch: 12, cost 24.71 s
|
| 26 |
+
2023-03-24 16:15:57,344 44k INFO Train Epoch: 13 [0%]
|
| 27 |
+
2023-03-24 16:15:57,345 44k INFO Losses: [2.2417097091674805, 2.8413851261138916, 12.686511039733887, 20.66866683959961, 1.215733289718628], step: 600, lr: 9.98501030820433e-05
|
| 28 |
+
2023-03-24 16:16:19,204 44k INFO ====> Epoch: 13, cost 25.68 s
|
| 29 |
+
2023-03-24 16:16:43,853 44k INFO ====> Epoch: 14, cost 24.65 s
|
| 30 |
+
2023-03-24 16:17:08,953 44k INFO ====> Epoch: 15, cost 25.10 s
|
| 31 |
+
2023-03-24 16:17:33,802 44k INFO ====> Epoch: 16, cost 24.85 s
|
| 32 |
+
2023-03-24 16:17:37,469 44k INFO Train Epoch: 17 [0%]
|
| 33 |
+
2023-03-24 16:17:37,471 44k INFO Losses: [2.2982616424560547, 2.2541720867156982, 8.421069145202637, 14.56866455078125, 1.1874828338623047], step: 800, lr: 9.980018739066937e-05
|
| 34 |
+
2023-03-24 16:17:42,347 44k INFO Saving model and optimizer state at iteration 17 to ./logs/44k/G_800.pth
|
| 35 |
+
2023-03-24 16:17:43,943 44k INFO Saving model and optimizer state at iteration 17 to ./logs/44k/D_800.pth
|
| 36 |
+
2023-03-24 16:18:05,588 44k INFO ====> Epoch: 17, cost 31.79 s
|
| 37 |
+
2023-03-24 16:18:30,523 44k INFO ====> Epoch: 18, cost 24.94 s
|
| 38 |
+
2023-03-24 16:18:55,233 44k INFO ====> Epoch: 19, cost 24.71 s
|
| 39 |
+
2023-03-24 16:19:19,823 44k INFO ====> Epoch: 20, cost 24.59 s
|
| 40 |
+
2023-03-24 16:19:23,336 44k INFO Train Epoch: 21 [0%]
|
| 41 |
+
2023-03-24 16:19:23,337 44k INFO Losses: [2.8217248916625977, 2.243264675140381, 8.612906455993652, 21.242691040039062, 1.4278812408447266], step: 1000, lr: 9.975029665246193e-05
|
| 42 |
+
2023-03-24 16:19:44,860 44k INFO ====> Epoch: 21, cost 25.04 s
|
| 43 |
+
2023-03-24 16:20:09,673 44k INFO ====> Epoch: 22, cost 24.81 s
|
| 44 |
+
2023-03-24 16:20:34,476 44k INFO ====> Epoch: 23, cost 24.80 s
|
| 45 |
+
2023-03-24 16:20:59,205 44k INFO ====> Epoch: 24, cost 24.73 s
|
| 46 |
+
2023-03-24 16:21:03,136 44k INFO Train Epoch: 25 [0%]
|
| 47 |
+
2023-03-24 16:21:03,137 44k INFO Losses: [2.3520963191986084, 2.346210479736328, 12.567983627319336, 20.541658401489258, 1.4853627681732178], step: 1200, lr: 9.970043085494672e-05
|
| 48 |
+
2023-03-24 16:21:24,861 44k INFO ====> Epoch: 25, cost 25.66 s
|
| 49 |
+
2023-03-24 16:21:49,922 44k INFO ====> Epoch: 26, cost 25.06 s
|
| 50 |
+
2023-03-24 16:22:15,067 44k INFO ====> Epoch: 27, cost 25.15 s
|
| 51 |
+
2023-03-24 16:22:40,244 44k INFO ====> Epoch: 28, cost 25.18 s
|
| 52 |
+
2023-03-24 16:22:44,262 44k INFO Train Epoch: 29 [0%]
|
| 53 |
+
2023-03-24 16:22:44,263 44k INFO Losses: [2.468628168106079, 2.170118808746338, 7.571216106414795, 20.927648544311523, 1.2605253458023071], step: 1400, lr: 9.965058998565574e-05
|
| 54 |
+
2023-03-24 16:23:06,111 44k INFO ====> Epoch: 29, cost 25.87 s
|
| 55 |
+
2023-03-24 16:23:31,293 44k INFO ====> Epoch: 30, cost 25.18 s
|
| 56 |
+
2023-03-24 16:23:56,555 44k INFO ====> Epoch: 31, cost 25.26 s
|
| 57 |
+
2023-03-24 16:24:21,726 44k INFO ====> Epoch: 32, cost 25.17 s
|
| 58 |
+
2023-03-24 16:24:25,731 44k INFO Train Epoch: 33 [0%]
|
| 59 |
+
2023-03-24 16:24:25,733 44k INFO Losses: [2.46799373626709, 2.3533313274383545, 11.058666229248047, 22.450557708740234, 1.4724372625350952], step: 1600, lr: 9.960077403212722e-05
|
| 60 |
+
2023-03-24 16:24:30,814 44k INFO Saving model and optimizer state at iteration 33 to ./logs/44k/G_1600.pth
|
| 61 |
+
2023-03-24 16:24:32,265 44k INFO Saving model and optimizer state at iteration 33 to ./logs/44k/D_1600.pth
|
| 62 |
+
2023-03-24 16:24:54,314 44k INFO ====> Epoch: 33, cost 32.59 s
|
| 63 |
+
2023-03-24 16:25:19,632 44k INFO ====> Epoch: 34, cost 25.32 s
|
| 64 |
+
2023-03-24 16:25:45,327 44k INFO ====> Epoch: 35, cost 25.70 s
|
| 65 |
+
2023-03-24 16:26:10,644 44k INFO ====> Epoch: 36, cost 25.32 s
|
| 66 |
+
2023-03-24 16:26:14,616 44k INFO Train Epoch: 37 [0%]
|
| 67 |
+
2023-03-24 16:26:14,618 44k INFO Losses: [2.637969970703125, 2.1250932216644287, 8.898857116699219, 19.450668334960938, 1.0567777156829834], step: 1800, lr: 9.95509829819056e-05
|
| 68 |
+
2023-03-24 16:26:36,678 44k INFO ====> Epoch: 37, cost 26.03 s
|
| 69 |
+
2023-03-24 16:27:01,814 44k INFO ====> Epoch: 38, cost 25.14 s
|
| 70 |
+
2023-03-24 16:27:26,722 44k INFO ====> Epoch: 39, cost 24.91 s
|
| 71 |
+
2023-03-24 16:27:51,749 44k INFO ====> Epoch: 40, cost 25.03 s
|
| 72 |
+
2023-03-24 16:27:55,388 44k INFO Train Epoch: 41 [0%]
|
| 73 |
+
2023-03-24 16:27:55,390 44k INFO Losses: [2.1182312965393066, 2.289332151412964, 11.968257904052734, 19.403751373291016, 1.1568026542663574], step: 2000, lr: 9.950121682254156e-05
|
| 74 |
+
2023-03-24 16:28:17,070 44k INFO ====> Epoch: 41, cost 25.32 s
|
| 75 |
+
2023-03-24 16:28:41,874 44k INFO ====> Epoch: 42, cost 24.80 s
|
| 76 |
+
2023-03-24 16:29:06,865 44k INFO ====> Epoch: 43, cost 24.99 s
|
| 77 |
+
2023-03-24 16:29:31,604 44k INFO ====> Epoch: 44, cost 24.74 s
|
| 78 |
+
2023-03-24 16:29:35,467 44k INFO Train Epoch: 45 [0%]
|
| 79 |
+
2023-03-24 16:29:35,468 44k INFO Losses: [2.7747840881347656, 2.2147929668426514, 8.382989883422852, 18.546131134033203, 1.1774940490722656], step: 2200, lr: 9.945147554159202e-05
|
| 80 |
+
2023-03-24 16:29:56,996 44k INFO ====> Epoch: 45, cost 25.39 s
|
| 81 |
+
2023-03-24 16:30:21,956 44k INFO ====> Epoch: 46, cost 24.96 s
|
| 82 |
+
2023-03-24 16:30:46,873 44k INFO ====> Epoch: 47, cost 24.92 s
|
| 83 |
+
2023-03-24 16:31:11,765 44k INFO ====> Epoch: 48, cost 24.89 s
|
| 84 |
+
2023-03-24 16:31:15,566 44k INFO Train Epoch: 49 [0%]
|
| 85 |
+
2023-03-24 16:31:15,567 44k INFO Losses: [2.438877582550049, 2.4737958908081055, 10.301446914672852, 21.869646072387695, 1.067798137664795], step: 2400, lr: 9.940175912662009e-05
|
| 86 |
+
2023-03-24 16:31:20,492 44k INFO Saving model and optimizer state at iteration 49 to ./logs/44k/G_2400.pth
|
| 87 |
+
2023-03-24 16:31:21,957 44k INFO Saving model and optimizer state at iteration 49 to ./logs/44k/D_2400.pth
|
| 88 |
+
2023-03-24 16:31:43,508 44k INFO ====> Epoch: 49, cost 31.74 s
|
| 89 |
+
2023-03-24 16:32:08,428 44k INFO ====> Epoch: 50, cost 24.92 s
|
| 90 |
+
2023-03-24 16:32:33,099 44k INFO ====> Epoch: 51, cost 24.67 s
|
| 91 |
+
2023-03-24 16:32:57,933 44k INFO ====> Epoch: 52, cost 24.83 s
|
| 92 |
+
2023-03-24 16:33:01,799 44k INFO Train Epoch: 53 [0%]
|
| 93 |
+
2023-03-24 16:33:01,801 44k INFO Losses: [2.3304405212402344, 2.397515058517456, 13.642292976379395, 21.823060989379883, 1.1119345426559448], step: 2600, lr: 9.935206756519513e-05
|
| 94 |
+
2023-03-24 16:33:23,375 44k INFO ====> Epoch: 53, cost 25.44 s
|
| 95 |
+
2023-03-24 16:33:49,208 44k INFO ====> Epoch: 54, cost 25.83 s
|
| 96 |
+
2023-03-24 16:34:14,334 44k INFO ====> Epoch: 55, cost 25.13 s
|
| 97 |
+
2023-03-24 16:34:39,142 44k INFO ====> Epoch: 56, cost 24.81 s
|
| 98 |
+
2023-03-24 16:34:43,036 44k INFO Train Epoch: 57 [0%]
|
| 99 |
+
2023-03-24 16:34:43,038 44k INFO Losses: [2.4579014778137207, 2.3200392723083496, 8.339593887329102, 20.653188705444336, 1.1132451295852661], step: 2800, lr: 9.930240084489267e-05
|
| 100 |
+
2023-03-24 16:35:04,540 44k INFO ====> Epoch: 57, cost 25.40 s
|
| 101 |
+
2023-03-24 16:35:29,399 44k INFO ====> Epoch: 58, cost 24.86 s
|
| 102 |
+
2023-03-24 16:35:54,290 44k INFO ====> Epoch: 59, cost 24.89 s
|
| 103 |
+
2023-03-24 16:36:19,055 44k INFO ====> Epoch: 60, cost 24.77 s
|
| 104 |
+
2023-03-24 16:36:22,860 44k INFO Train Epoch: 61 [0%]
|
| 105 |
+
2023-03-24 16:36:22,862 44k INFO Losses: [2.3698558807373047, 2.2452638149261475, 9.925782203674316, 20.825319290161133, 0.6393007040023804], step: 3000, lr: 9.92527589532945e-05
|
| 106 |
+
2023-03-24 16:36:44,782 44k INFO ====> Epoch: 61, cost 25.73 s
|
| 107 |
+
2023-03-24 16:37:09,557 44k INFO ====> Epoch: 62, cost 24.78 s
|
| 108 |
+
2023-03-24 16:37:34,361 44k INFO ====> Epoch: 63, cost 24.80 s
|
| 109 |
+
2023-03-24 16:37:59,183 44k INFO ====> Epoch: 64, cost 24.82 s
|
| 110 |
+
2023-03-24 16:38:02,824 44k INFO Train Epoch: 65 [0%]
|
| 111 |
+
2023-03-24 16:38:02,826 44k INFO Losses: [2.4456441402435303, 2.655974864959717, 11.801567077636719, 19.65692901611328, 0.9295780062675476], step: 3200, lr: 9.92031418779886e-05
|
| 112 |
+
2023-03-24 16:38:07,725 44k INFO Saving model and optimizer state at iteration 65 to ./logs/44k/G_3200.pth
|
| 113 |
+
2023-03-24 16:38:09,193 44k INFO Saving model and optimizer state at iteration 65 to ./logs/44k/D_3200.pth
|
| 114 |
+
2023-03-24 16:38:30,997 44k INFO ====> Epoch: 65, cost 31.81 s
|
| 115 |
+
2023-03-24 16:38:56,138 44k INFO ====> Epoch: 66, cost 25.14 s
|
| 116 |
+
2023-03-24 16:39:21,052 44k INFO ====> Epoch: 67, cost 24.91 s
|
| 117 |
+
2023-03-24 16:39:46,032 44k INFO ====> Epoch: 68, cost 24.98 s
|
| 118 |
+
2023-03-24 16:39:49,805 44k INFO Train Epoch: 69 [0%]
|
| 119 |
+
2023-03-24 16:39:49,807 44k INFO Losses: [2.511517286300659, 2.0972821712493896, 9.085113525390625, 18.71324348449707, 1.1218843460083008], step: 3400, lr: 9.915354960656915e-05
|
| 120 |
+
2023-03-24 16:40:11,500 44k INFO ====> Epoch: 69, cost 25.47 s
|
| 121 |
+
2023-03-24 16:40:36,372 44k INFO ====> Epoch: 70, cost 24.87 s
|
| 122 |
+
2023-03-24 16:41:01,150 44k INFO ====> Epoch: 71, cost 24.78 s
|
| 123 |
+
2023-03-24 16:41:25,926 44k INFO ====> Epoch: 72, cost 24.78 s
|
| 124 |
+
2023-03-24 16:41:29,560 44k INFO Train Epoch: 73 [0%]
|
| 125 |
+
2023-03-24 16:41:29,561 44k INFO Losses: [2.4871113300323486, 2.247950553894043, 9.168888092041016, 21.552160263061523, 0.9532811641693115], step: 3600, lr: 9.910398212663652e-05
|
| 126 |
+
2023-03-24 16:41:51,435 44k INFO ====> Epoch: 73, cost 25.51 s
|
| 127 |
+
2023-03-24 16:42:16,376 44k INFO ====> Epoch: 74, cost 24.94 s
|
| 128 |
+
2023-03-24 16:42:41,110 44k INFO ====> Epoch: 75, cost 24.73 s
|
| 129 |
+
2023-03-24 16:43:06,273 44k INFO ====> Epoch: 76, cost 25.16 s
|
| 130 |
+
2023-03-24 16:43:09,995 44k INFO Train Epoch: 77 [0%]
|
| 131 |
+
2023-03-24 16:43:09,996 44k INFO Losses: [3.0498757362365723, 2.0931332111358643, 9.740159034729004, 20.8754940032959, 0.8113635182380676], step: 3800, lr: 9.905443942579728e-05
|
| 132 |
+
2023-03-24 16:43:31,648 44k INFO ====> Epoch: 77, cost 25.38 s
|
| 133 |
+
2023-03-24 16:43:56,531 44k INFO ====> Epoch: 78, cost 24.88 s
|
| 134 |
+
2023-03-24 16:44:21,788 44k INFO ====> Epoch: 79, cost 25.26 s
|
| 135 |
+
2023-03-24 16:44:46,863 44k INFO ====> Epoch: 80, cost 25.07 s
|
| 136 |
+
2023-03-24 16:44:50,520 44k INFO Train Epoch: 81 [0%]
|
| 137 |
+
2023-03-24 16:44:50,521 44k INFO Losses: [2.3789913654327393, 2.4813644886016846, 10.093775749206543, 17.8828182220459, 0.9886296987533569], step: 4000, lr: 9.900492149166423e-05
|
| 138 |
+
2023-03-24 16:44:55,336 44k INFO Saving model and optimizer state at iteration 81 to ./logs/44k/G_4000.pth
|
| 139 |
+
2023-03-24 16:44:56,786 44k INFO Saving model and optimizer state at iteration 81 to ./logs/44k/D_4000.pth
|
| 140 |
+
2023-03-24 16:45:18,665 44k INFO ====> Epoch: 81, cost 31.80 s
|
| 141 |
+
2023-03-24 16:45:43,655 44k INFO ====> Epoch: 82, cost 24.99 s
|
| 142 |
+
2023-03-24 16:46:08,720 44k INFO ====> Epoch: 83, cost 25.06 s
|
| 143 |
+
2023-03-24 16:46:33,650 44k INFO ====> Epoch: 84, cost 24.93 s
|
| 144 |
+
2023-03-24 16:46:37,358 44k INFO Train Epoch: 85 [0%]
|
| 145 |
+
2023-03-24 16:46:37,360 44k INFO Losses: [2.3212366104125977, 1.990512490272522, 8.910666465759277, 21.229427337646484, 0.4932689964771271], step: 4200, lr: 9.895542831185631e-05
|
| 146 |
+
2023-03-24 16:46:59,141 44k INFO ====> Epoch: 85, cost 25.49 s
|
| 147 |
+
2023-03-24 16:47:24,077 44k INFO ====> Epoch: 86, cost 24.94 s
|
| 148 |
+
2023-03-24 16:47:49,217 44k INFO ====> Epoch: 87, cost 25.14 s
|
| 149 |
+
2023-03-24 16:48:14,071 44k INFO ====> Epoch: 88, cost 24.85 s
|
| 150 |
+
2023-03-24 16:48:18,000 44k INFO Train Epoch: 89 [0%]
|
| 151 |
+
2023-03-24 16:48:18,002 44k INFO Losses: [2.222428560256958, 2.3397562503814697, 15.815887451171875, 22.820709228515625, 0.802300751209259], step: 4400, lr: 9.89059598739987e-05
|
| 152 |
+
2023-03-24 16:48:40,075 44k INFO ====> Epoch: 89, cost 26.00 s
|
| 153 |
+
2023-03-24 16:49:05,197 44k INFO ====> Epoch: 90, cost 25.12 s
|
| 154 |
+
2023-03-24 16:49:30,107 44k INFO ====> Epoch: 91, cost 24.91 s
|
| 155 |
+
2023-03-24 16:49:55,009 44k INFO ====> Epoch: 92, cost 24.90 s
|
| 156 |
+
2023-03-24 16:49:58,820 44k INFO Train Epoch: 93 [0%]
|
| 157 |
+
2023-03-24 16:49:58,822 44k INFO Losses: [2.3582277297973633, 2.0633132457733154, 10.956375122070312, 19.99727439880371, 1.0134141445159912], step: 4600, lr: 9.885651616572276e-05
|
| 158 |
+
2023-03-24 16:50:20,550 44k INFO ====> Epoch: 93, cost 25.54 s
|
| 159 |
+
2023-03-24 16:50:45,626 44k INFO ====> Epoch: 94, cost 25.08 s
|
| 160 |
+
2023-03-24 16:51:10,825 44k INFO ====> Epoch: 95, cost 25.20 s
|
| 161 |
+
2023-03-24 16:51:35,826 44k INFO ====> Epoch: 96, cost 25.00 s
|
| 162 |
+
2023-03-24 16:51:39,583 44k INFO Train Epoch: 97 [0%]
|
| 163 |
+
2023-03-24 16:51:39,584 44k INFO Losses: [2.5545666217803955, 2.3805904388427734, 11.227080345153809, 19.633548736572266, 0.9872961044311523], step: 4800, lr: 9.880709717466598e-05
|
| 164 |
+
2023-03-24 16:51:44,661 44k INFO Saving model and optimizer state at iteration 97 to ./logs/44k/G_4800.pth
|
| 165 |
+
2023-03-24 16:51:46,114 44k INFO Saving model and optimizer state at iteration 97 to ./logs/44k/D_4800.pth
|
| 166 |
+
2023-03-24 16:52:07,684 44k INFO ====> Epoch: 97, cost 31.86 s
|
| 167 |
+
2023-03-24 16:52:32,269 44k INFO ====> Epoch: 98, cost 24.58 s
|
| 168 |
+
2023-03-24 16:52:57,267 44k INFO ====> Epoch: 99, cost 25.00 s
|
| 169 |
+
2023-03-24 16:53:22,366 44k INFO ====> Epoch: 100, cost 25.10 s
|
| 170 |
+
2023-03-24 16:53:26,197 44k INFO Train Epoch: 101 [0%]
|
| 171 |
+
2023-03-24 16:53:26,198 44k INFO Losses: [2.5303902626037598, 2.4984896183013916, 8.828944206237793, 17.48539161682129, 0.9228643178939819], step: 5000, lr: 9.875770288847208e-05
|
| 172 |
+
2023-03-24 16:53:48,025 44k INFO ====> Epoch: 101, cost 25.66 s
|
| 173 |
+
2023-03-24 16:54:12,974 44k INFO ====> Epoch: 102, cost 24.95 s
|
| 174 |
+
2023-03-24 16:54:37,657 44k INFO ====> Epoch: 103, cost 24.68 s
|
| 175 |
+
2023-03-24 16:55:02,369 44k INFO ====> Epoch: 104, cost 24.71 s
|
| 176 |
+
2023-03-24 16:55:06,175 44k INFO Train Epoch: 105 [0%]
|
| 177 |
+
2023-03-24 16:55:06,176 44k INFO Losses: [2.65112042427063, 2.510831832885742, 9.358697891235352, 21.029691696166992, 0.7588745951652527], step: 5200, lr: 9.870833329479095e-05
|
| 178 |
+
2023-03-24 16:55:27,954 44k INFO ====> Epoch: 105, cost 25.58 s
|
| 179 |
+
2023-03-24 16:55:52,961 44k INFO ====> Epoch: 106, cost 25.01 s
|
| 180 |
+
2023-03-24 16:56:18,063 44k INFO ====> Epoch: 107, cost 25.10 s
|
| 181 |
+
2023-03-24 16:56:43,409 44k INFO ====> Epoch: 108, cost 25.35 s
|
| 182 |
+
2023-03-24 16:56:47,248 44k INFO Train Epoch: 109 [0%]
|
| 183 |
+
2023-03-24 16:56:47,250 44k INFO Losses: [2.5201327800750732, 2.3836162090301514, 9.467650413513184, 21.405014038085938, 1.1192044019699097], step: 5400, lr: 9.865898838127865e-05
|
| 184 |
+
2023-03-24 16:57:08,895 44k INFO ====> Epoch: 109, cost 25.49 s
|
| 185 |
+
2023-03-24 16:57:33,807 44k INFO ====> Epoch: 110, cost 24.91 s
|
| 186 |
+
2023-03-24 16:57:58,673 44k INFO ====> Epoch: 111, cost 24.87 s
|
| 187 |
+
2023-03-24 16:58:23,665 44k INFO ====> Epoch: 112, cost 24.99 s
|
| 188 |
+
2023-03-24 16:58:27,541 44k INFO Train Epoch: 113 [0%]
|
| 189 |
+
2023-03-24 16:58:27,542 44k INFO Losses: [2.516652822494507, 1.9198682308197021, 12.64153003692627, 21.67121124267578, 0.9236412644386292], step: 5600, lr: 9.86096681355974e-05
|
| 190 |
+
2023-03-24 16:58:32,390 44k INFO Saving model and optimizer state at iteration 113 to ./logs/44k/G_5600.pth
|
| 191 |
+
2023-03-24 16:58:33,837 44k INFO Saving model and optimizer state at iteration 113 to ./logs/44k/D_5600.pth
|
| 192 |
+
2023-03-24 16:58:55,689 44k INFO ====> Epoch: 113, cost 32.02 s
|
| 193 |
+
2023-03-24 16:59:20,767 44k INFO ====> Epoch: 114, cost 25.08 s
|
| 194 |
+
2023-03-24 16:59:45,581 44k INFO ====> Epoch: 115, cost 24.81 s
|
| 195 |
+
2023-03-24 17:00:10,507 44k INFO ====> Epoch: 116, cost 24.93 s
|
| 196 |
+
2023-03-24 17:00:14,311 44k INFO Train Epoch: 117 [0%]
|
| 197 |
+
2023-03-24 17:00:14,312 44k INFO Losses: [2.473012924194336, 2.149684190750122, 7.098073959350586, 17.53461456298828, 0.6158500909805298], step: 5800, lr: 9.85603725454156e-05
|
| 198 |
+
2023-03-24 17:00:36,112 44k INFO ====> Epoch: 117, cost 25.61 s
|
| 199 |
+
2023-03-24 17:01:01,203 44k INFO ====> Epoch: 118, cost 25.09 s
|
| 200 |
+
2023-03-24 17:01:26,147 44k INFO ====> Epoch: 119, cost 24.94 s
|
| 201 |
+
2023-03-24 17:01:51,183 44k INFO ====> Epoch: 120, cost 25.04 s
|
| 202 |
+
2023-03-24 17:01:55,020 44k INFO Train Epoch: 121 [0%]
|
| 203 |
+
2023-03-24 17:01:55,022 44k INFO Losses: [2.182035446166992, 2.618431568145752, 10.66880989074707, 19.815397262573242, 0.8694823980331421], step: 6000, lr: 9.851110159840781e-05
|
| 204 |
+
2023-03-24 17:02:16,777 44k INFO ====> Epoch: 121, cost 25.59 s
|
| 205 |
+
2023-03-24 17:02:41,956 44k INFO ====> Epoch: 122, cost 25.18 s
|
| 206 |
+
2023-03-24 17:03:06,973 44k INFO ====> Epoch: 123, cost 25.02 s
|
| 207 |
+
2023-03-24 17:03:31,917 44k INFO ====> Epoch: 124, cost 24.94 s
|
| 208 |
+
2023-03-24 17:03:35,819 44k INFO Train Epoch: 125 [0%]
|
| 209 |
+
2023-03-24 17:03:35,821 44k INFO Losses: [2.406432628631592, 2.431051731109619, 8.33769416809082, 16.252138137817383, 1.0498191118240356], step: 6200, lr: 9.846185528225477e-05
|
| 210 |
+
2023-03-24 17:03:57,713 44k INFO ====> Epoch: 125, cost 25.80 s
|
| 211 |
+
2023-03-24 17:04:22,726 44k INFO ====> Epoch: 126, cost 25.01 s
|
| 212 |
+
2023-03-24 17:04:47,506 44k INFO ====> Epoch: 127, cost 24.78 s
|
| 213 |
+
2023-03-24 17:05:12,524 44k INFO ====> Epoch: 128, cost 25.02 s
|
| 214 |
+
2023-03-24 17:05:16,139 44k INFO Train Epoch: 129 [0%]
|
| 215 |
+
2023-03-24 17:05:16,140 44k INFO Losses: [2.168229818344116, 2.5004663467407227, 9.397270202636719, 21.672155380249023, 1.026598334312439], step: 6400, lr: 9.841263358464336e-05
|
| 216 |
+
2023-03-24 17:05:20,927 44k INFO Saving model and optimizer state at iteration 129 to ./logs/44k/G_6400.pth
|
| 217 |
+
2023-03-24 17:05:22,456 44k INFO Saving model and optimizer state at iteration 129 to ./logs/44k/D_6400.pth
|
| 218 |
+
2023-03-24 17:05:44,177 44k INFO ====> Epoch: 129, cost 31.65 s
|
| 219 |
+
2023-03-24 17:06:09,175 44k INFO ====> Epoch: 130, cost 25.00 s
|
| 220 |
+
2023-03-24 17:06:34,161 44k INFO ====> Epoch: 131, cost 24.99 s
|
| 221 |
+
2023-03-24 17:06:59,334 44k INFO ====> Epoch: 132, cost 25.17 s
|
| 222 |
+
2023-03-24 17:07:03,285 44k INFO Train Epoch: 133 [0%]
|
| 223 |
+
2023-03-24 17:07:03,286 44k INFO Losses: [2.4833321571350098, 1.8855023384094238, 8.147689819335938, 16.213708877563477, 0.935891330242157], step: 6600, lr: 9.836343649326659e-05
|
| 224 |
+
2023-03-24 17:07:25,196 44k INFO ====> Epoch: 133, cost 25.86 s
|
| 225 |
+
2023-03-24 17:07:50,174 44k INFO ====> Epoch: 134, cost 24.98 s
|
| 226 |
+
2023-03-24 17:08:15,531 44k INFO ====> Epoch: 135, cost 25.36 s
|
| 227 |
+
2023-03-24 17:08:40,614 44k INFO ====> Epoch: 136, cost 25.08 s
|
| 228 |
+
2023-03-24 17:08:44,405 44k INFO Train Epoch: 137 [0%]
|
| 229 |
+
2023-03-24 17:08:44,406 44k INFO Losses: [2.6166269779205322, 2.219412088394165, 8.7598295211792, 20.433765411376953, 0.8591962456703186], step: 6800, lr: 9.831426399582366e-05
|
| 230 |
+
2023-03-24 17:09:06,198 44k INFO ====> Epoch: 137, cost 25.58 s
|
| 231 |
+
2023-03-24 17:09:31,096 44k INFO ====> Epoch: 138, cost 24.90 s
|
| 232 |
+
2023-03-24 17:09:56,066 44k INFO ====> Epoch: 139, cost 24.97 s
|
| 233 |
+
2023-03-24 17:10:21,002 44k INFO ====> Epoch: 140, cost 24.94 s
|
| 234 |
+
2023-03-24 17:10:24,806 44k INFO Train Epoch: 141 [0%]
|
| 235 |
+
2023-03-24 17:10:24,807 44k INFO Losses: [2.2800958156585693, 2.6092875003814697, 10.905752182006836, 22.769119262695312, 1.0514625310897827], step: 7000, lr: 9.826511608001993e-05
|
| 236 |
+
2023-03-24 17:10:46,563 44k INFO ====> Epoch: 141, cost 25.56 s
|
| 237 |
+
2023-03-24 17:11:11,429 44k INFO ====> Epoch: 142, cost 24.87 s
|
| 238 |
+
2023-03-24 17:11:36,251 44k INFO ====> Epoch: 143, cost 24.82 s
|
| 239 |
+
2023-03-24 17:12:00,907 44k INFO ====> Epoch: 144, cost 24.66 s
|
| 240 |
+
2023-03-24 17:12:04,652 44k INFO Train Epoch: 145 [0%]
|
| 241 |
+
2023-03-24 17:12:04,654 44k INFO Losses: [2.4433958530426025, 2.1577835083007812, 11.530089378356934, 19.273975372314453, 0.5594630837440491], step: 7200, lr: 9.821599273356685e-05
|
| 242 |
+
2023-03-24 17:12:09,668 44k INFO Saving model and optimizer state at iteration 145 to ./logs/44k/G_7200.pth
|
| 243 |
+
2023-03-24 17:12:11,127 44k INFO Saving model and optimizer state at iteration 145 to ./logs/44k/D_7200.pth
|
| 244 |
+
2023-03-24 17:12:32,817 44k INFO ====> Epoch: 145, cost 31.91 s
|
| 245 |
+
2023-03-24 17:12:57,545 44k INFO ====> Epoch: 146, cost 24.73 s
|
| 246 |
+
2023-03-24 17:13:22,383 44k INFO ====> Epoch: 147, cost 24.84 s
|
| 247 |
+
2023-03-24 17:13:47,376 44k INFO ====> Epoch: 148, cost 24.99 s
|
| 248 |
+
2023-03-24 17:13:51,018 44k INFO Train Epoch: 149 [0%]
|
| 249 |
+
2023-03-24 17:13:51,020 44k INFO Losses: [2.541314125061035, 2.248460292816162, 8.099034309387207, 17.675189971923828, 0.6793069243431091], step: 7400, lr: 9.816689394418209e-05
|
| 250 |
+
2023-03-24 17:14:12,901 44k INFO ====> Epoch: 149, cost 25.52 s
|
| 251 |
+
2023-03-24 17:14:37,615 44k INFO ====> Epoch: 150, cost 24.71 s
|
| 252 |
+
2023-03-24 17:15:02,718 44k INFO ====> Epoch: 151, cost 25.10 s
|
| 253 |
+
2023-03-24 17:15:27,505 44k INFO ====> Epoch: 152, cost 24.79 s
|
| 254 |
+
2023-03-24 17:15:31,093 44k INFO Train Epoch: 153 [0%]
|
| 255 |
+
2023-03-24 17:15:31,095 44k INFO Losses: [2.6885759830474854, 1.7870721817016602, 7.810143947601318, 17.377124786376953, 0.9760600328445435], step: 7600, lr: 9.811781969958938e-05
|
| 256 |
+
2023-03-24 17:15:52,655 44k INFO ====> Epoch: 153, cost 25.15 s
|
| 257 |
+
2023-03-24 17:16:17,956 44k INFO ====> Epoch: 154, cost 25.30 s
|
| 258 |
+
2023-03-24 17:16:42,842 44k INFO ====> Epoch: 155, cost 24.89 s
|
| 259 |
+
2023-03-24 17:17:07,750 44k INFO ====> Epoch: 156, cost 24.91 s
|
| 260 |
+
2023-03-24 17:17:11,629 44k INFO Train Epoch: 157 [0%]
|
| 261 |
+
2023-03-24 17:17:11,630 44k INFO Losses: [2.3176026344299316, 2.416011333465576, 11.307724952697754, 21.595233917236328, 0.9688916206359863], step: 7800, lr: 9.806876998751865e-05
|
| 262 |
+
2023-03-24 17:17:33,336 44k INFO ====> Epoch: 157, cost 25.59 s
|
| 263 |
+
2023-03-24 17:17:58,244 44k INFO ====> Epoch: 158, cost 24.91 s
|
| 264 |
+
2023-03-24 17:18:23,111 44k INFO ====> Epoch: 159, cost 24.87 s
|
| 265 |
+
2023-03-24 17:18:47,965 44k INFO ====> Epoch: 160, cost 24.85 s
|
| 266 |
+
2023-03-24 17:18:51,778 44k INFO Train Epoch: 161 [0%]
|
| 267 |
+
2023-03-24 17:18:51,779 44k INFO Losses: [2.335582971572876, 2.200396776199341, 11.696431159973145, 21.497562408447266, 0.9381815791130066], step: 8000, lr: 9.801974479570593e-05
|
| 268 |
+
2023-03-24 17:18:56,755 44k INFO Saving model and optimizer state at iteration 161 to ./logs/44k/G_8000.pth
|
| 269 |
+
2023-03-24 17:18:58,161 44k INFO Saving model and optimizer state at iteration 161 to ./logs/44k/D_8000.pth
|
| 270 |
+
2023-03-24 17:19:19,776 44k INFO ====> Epoch: 161, cost 31.81 s
|
| 271 |
+
2023-03-24 17:19:44,844 44k INFO ====> Epoch: 162, cost 25.07 s
|
| 272 |
+
2023-03-24 17:20:09,764 44k INFO ====> Epoch: 163, cost 24.92 s
|
| 273 |
+
2023-03-24 17:20:34,563 44k INFO ====> Epoch: 164, cost 24.80 s
|
| 274 |
+
2023-03-24 17:20:38,302 44k INFO Train Epoch: 165 [0%]
|
| 275 |
+
2023-03-24 17:20:38,303 44k INFO Losses: [2.2087090015411377, 2.4203712940216064, 10.896903991699219, 20.731460571289062, 0.6834594011306763], step: 8200, lr: 9.797074411189339e-05
|
| 276 |
+
2023-03-24 17:20:59,999 44k INFO ====> Epoch: 165, cost 25.44 s
|
| 277 |
+
2023-03-24 17:21:25,178 44k INFO ====> Epoch: 166, cost 25.18 s
|
| 278 |
+
2023-03-24 17:21:50,298 44k INFO ====> Epoch: 167, cost 25.12 s
|
| 279 |
+
2023-03-24 17:22:15,251 44k INFO ====> Epoch: 168, cost 24.95 s
|
| 280 |
+
2023-03-24 17:22:18,958 44k INFO Train Epoch: 169 [0%]
|
| 281 |
+
2023-03-24 17:22:18,959 44k INFO Losses: [2.3192877769470215, 2.2281627655029297, 7.640979766845703, 18.116220474243164, 0.751641571521759], step: 8400, lr: 9.792176792382932e-05
|
| 282 |
+
2023-03-24 17:22:40,590 44k INFO ====> Epoch: 169, cost 25.34 s
|
| 283 |
+
2023-03-24 17:23:05,450 44k INFO ====> Epoch: 170, cost 24.86 s
|
| 284 |
+
2023-03-24 17:23:30,354 44k INFO ====> Epoch: 171, cost 24.90 s
|
| 285 |
+
2023-03-24 17:23:55,451 44k INFO ====> Epoch: 172, cost 25.10 s
|
| 286 |
+
2023-03-24 17:23:59,357 44k INFO Train Epoch: 173 [0%]
|
| 287 |
+
2023-03-24 17:23:59,358 44k INFO Losses: [2.6158812046051025, 2.3054511547088623, 11.013618469238281, 19.365203857421875, 0.7469444870948792], step: 8600, lr: 9.787281621926815e-05
|
| 288 |
+
2023-03-24 17:24:21,066 44k INFO ====> Epoch: 173, cost 25.61 s
|
| 289 |
+
2023-03-24 17:24:46,070 44k INFO ====> Epoch: 174, cost 25.00 s
|
| 290 |
+
2023-03-24 17:25:11,054 44k INFO ====> Epoch: 175, cost 24.98 s
|
| 291 |
+
2023-03-24 17:25:36,049 44k INFO ====> Epoch: 176, cost 25.00 s
|
| 292 |
+
2023-03-24 17:25:39,843 44k INFO Train Epoch: 177 [0%]
|
| 293 |
+
2023-03-24 17:25:39,844 44k INFO Losses: [2.4385979175567627, 2.112964153289795, 10.049574851989746, 18.21015739440918, 0.4370778203010559], step: 8800, lr: 9.782388898597041e-05
|
| 294 |
+
2023-03-24 17:25:44,627 44k INFO Saving model and optimizer state at iteration 177 to ./logs/44k/G_8800.pth
|
| 295 |
+
2023-03-24 17:25:46,174 44k INFO Saving model and optimizer state at iteration 177 to ./logs/44k/D_8800.pth
|
| 296 |
+
2023-03-24 17:26:07,968 44k INFO ====> Epoch: 177, cost 31.92 s
|
| 297 |
+
2023-03-24 17:26:32,811 44k INFO ====> Epoch: 178, cost 24.84 s
|
| 298 |
+
2023-03-24 17:26:57,954 44k INFO ====> Epoch: 179, cost 25.14 s
|
| 299 |
+
2023-03-24 17:27:22,976 44k INFO ====> Epoch: 180, cost 25.02 s
|
| 300 |
+
2023-03-24 17:27:26,840 44k INFO Train Epoch: 181 [0%]
|
| 301 |
+
2023-03-24 17:27:26,841 44k INFO Losses: [2.577077865600586, 1.9521126747131348, 6.810126304626465, 15.925239562988281, 0.7762280702590942], step: 9000, lr: 9.777498621170277e-05
|
| 302 |
+
2023-03-24 17:27:48,572 44k INFO ====> Epoch: 181, cost 25.60 s
|
| 303 |
+
2023-03-24 17:28:13,664 44k INFO ====> Epoch: 182, cost 25.09 s
|
| 304 |
+
2023-03-24 17:28:38,409 44k INFO ====> Epoch: 183, cost 24.74 s
|
| 305 |
+
2023-03-24 17:29:03,357 44k INFO ====> Epoch: 184, cost 24.95 s
|
| 306 |
+
2023-03-24 17:29:07,154 44k INFO Train Epoch: 185 [0%]
|
| 307 |
+
2023-03-24 17:29:07,155 44k INFO Losses: [2.735002040863037, 2.042769193649292, 6.685863971710205, 18.602622985839844, 0.7408661246299744], step: 9200, lr: 9.772610788423802e-05
|
| 308 |
+
2023-03-24 17:29:29,064 44k INFO ====> Epoch: 185, cost 25.71 s
|
| 309 |
+
2023-03-24 17:29:54,104 44k INFO ====> Epoch: 186, cost 25.04 s
|
| 310 |
+
2023-03-24 17:30:18,964 44k INFO ====> Epoch: 187, cost 24.86 s
|
| 311 |
+
2023-03-24 17:30:44,025 44k INFO ====> Epoch: 188, cost 25.06 s
|
| 312 |
+
2023-03-24 17:30:47,877 44k INFO Train Epoch: 189 [0%]
|
| 313 |
+
2023-03-24 17:30:47,879 44k INFO Losses: [2.4279744625091553, 2.3245105743408203, 9.516021728515625, 18.581796646118164, 0.8581266403198242], step: 9400, lr: 9.767725399135504e-05
|
| 314 |
+
2023-03-24 17:31:09,751 44k INFO ====> Epoch: 189, cost 25.73 s
|
| 315 |
+
2023-03-24 17:31:34,675 44k INFO ====> Epoch: 190, cost 24.92 s
|
| 316 |
+
2023-03-24 17:31:59,606 44k INFO ====> Epoch: 191, cost 24.93 s
|
| 317 |
+
2023-03-24 17:32:24,562 44k INFO ====> Epoch: 192, cost 24.96 s
|
| 318 |
+
2023-03-24 17:32:28,382 44k INFO Train Epoch: 193 [0%]
|
| 319 |
+
2023-03-24 17:32:28,383 44k INFO Losses: [2.2496345043182373, 2.273651123046875, 10.993385314941406, 18.61309051513672, 0.5576750636100769], step: 9600, lr: 9.762842452083883e-05
|
| 320 |
+
2023-03-24 17:32:33,323 44k INFO Saving model and optimizer state at iteration 193 to ./logs/44k/G_9600.pth
|
| 321 |
+
2023-03-24 17:32:34,749 44k INFO Saving model and optimizer state at iteration 193 to ./logs/44k/D_9600.pth
|
| 322 |
+
2023-03-24 17:32:56,434 44k INFO ====> Epoch: 193, cost 31.87 s
|
| 323 |
+
2023-03-24 17:33:21,551 44k INFO ====> Epoch: 194, cost 25.12 s
|
| 324 |
+
2023-03-24 17:33:46,348 44k INFO ====> Epoch: 195, cost 24.80 s
|
| 325 |
+
2023-03-24 17:34:11,094 44k INFO ====> Epoch: 196, cost 24.75 s
|
| 326 |
+
2023-03-24 17:34:15,062 44k INFO Train Epoch: 197 [0%]
|
| 327 |
+
2023-03-24 17:34:15,064 44k INFO Losses: [2.400710105895996, 2.40057373046875, 8.645580291748047, 18.94593048095703, 0.579620897769928], step: 9800, lr: 9.757961946048049e-05
|
| 328 |
+
2023-03-24 17:34:36,942 44k INFO ====> Epoch: 197, cost 25.85 s
|
| 329 |
+
2023-03-24 17:35:01,995 44k INFO ====> Epoch: 198, cost 25.05 s
|
| 330 |
+
2023-03-24 17:35:26,534 44k INFO ====> Epoch: 199, cost 24.54 s
|
| 331 |
+
2023-03-24 17:35:51,658 44k INFO ====> Epoch: 200, cost 25.12 s
|
| 332 |
+
2023-03-24 17:35:55,306 44k INFO Train Epoch: 201 [0%]
|
| 333 |
+
2023-03-24 17:35:55,307 44k INFO Losses: [2.255695104598999, 2.3569161891937256, 11.471579551696777, 21.66715431213379, 0.7947320938110352], step: 10000, lr: 9.753083879807726e-05
|
| 334 |
+
2023-03-24 17:36:16,899 44k INFO ====> Epoch: 201, cost 25.24 s
|
| 335 |
+
2023-03-24 17:36:41,707 44k INFO ====> Epoch: 202, cost 24.81 s
|
| 336 |
+
2023-03-24 17:37:06,526 44k INFO ====> Epoch: 203, cost 24.82 s
|
| 337 |
+
2023-03-24 17:37:31,076 44k INFO ====> Epoch: 204, cost 24.55 s
|
| 338 |
+
2023-03-24 17:37:34,870 44k INFO Train Epoch: 205 [0%]
|
| 339 |
+
2023-03-24 17:37:34,872 44k INFO Losses: [2.64894437789917, 2.250037670135498, 7.354074001312256, 16.611326217651367, 0.5304523706436157], step: 10200, lr: 9.748208252143241e-05
|
| 340 |
+
2023-03-24 17:37:56,565 44k INFO ====> Epoch: 205, cost 25.49 s
|
| 341 |
+
2023-03-24 17:38:21,446 44k INFO ====> Epoch: 206, cost 24.88 s
|
| 342 |
+
2023-03-24 17:38:46,504 44k INFO ====> Epoch: 207, cost 25.06 s
|
| 343 |
+
2023-03-24 17:39:11,285 44k INFO ====> Epoch: 208, cost 24.78 s
|
| 344 |
+
2023-03-24 17:39:15,102 44k INFO Train Epoch: 209 [0%]
|
| 345 |
+
2023-03-24 17:39:15,103 44k INFO Losses: [2.1992905139923096, 2.505424976348877, 9.921688079833984, 21.01839828491211, 1.1307833194732666], step: 10400, lr: 9.743335061835535e-05
|
| 346 |
+
2023-03-24 17:39:20,006 44k INFO Saving model and optimizer state at iteration 209 to ./logs/44k/G_10400.pth
|
| 347 |
+
2023-03-24 17:39:21,392 44k INFO Saving model and optimizer state at iteration 209 to ./logs/44k/D_10400.pth
|
| 348 |
+
2023-03-24 17:39:43,026 44k INFO ====> Epoch: 209, cost 31.74 s
|
| 349 |
+
2023-03-24 17:40:08,180 44k INFO ====> Epoch: 210, cost 25.15 s
|
| 350 |
+
2023-03-24 17:40:33,240 44k INFO ====> Epoch: 211, cost 25.06 s
|
| 351 |
+
2023-03-24 17:40:58,174 44k INFO ====> Epoch: 212, cost 24.93 s
|
| 352 |
+
2023-03-24 17:41:02,009 44k INFO Train Epoch: 213 [0%]
|
| 353 |
+
2023-03-24 17:41:02,010 44k INFO Losses: [2.2102227210998535, 2.369741678237915, 9.036264419555664, 17.1883602142334, 0.4644804894924164], step: 10600, lr: 9.73846430766616e-05
|
| 354 |
+
2023-03-24 17:41:23,687 44k INFO ====> Epoch: 213, cost 25.51 s
|
| 355 |
+
2023-03-24 17:41:48,589 44k INFO ====> Epoch: 214, cost 24.90 s
|
| 356 |
+
2023-03-24 17:42:13,292 44k INFO ====> Epoch: 215, cost 24.70 s
|
| 357 |
+
2023-03-24 17:42:38,126 44k INFO ====> Epoch: 216, cost 24.83 s
|
| 358 |
+
2023-03-24 17:42:42,082 44k INFO Train Epoch: 217 [0%]
|
| 359 |
+
2023-03-24 17:42:42,084 44k INFO Losses: [2.442573070526123, 2.2122700214385986, 8.983601570129395, 17.680198669433594, 0.7081395387649536], step: 10800, lr: 9.733595988417275e-05
|
| 360 |
+
2023-03-24 17:43:03,592 44k INFO ====> Epoch: 217, cost 25.47 s
|
| 361 |
+
2023-03-24 17:43:28,367 44k INFO ====> Epoch: 218, cost 24.78 s
|
| 362 |
+
2023-03-24 17:43:53,025 44k INFO ====> Epoch: 219, cost 24.66 s
|
| 363 |
+
2023-03-24 17:44:17,974 44k INFO ====> Epoch: 220, cost 24.95 s
|
| 364 |
+
2023-03-24 17:44:21,653 44k INFO Train Epoch: 221 [0%]
|
| 365 |
+
2023-03-24 17:44:21,655 44k INFO Losses: [2.626898765563965, 1.8628339767456055, 10.394402503967285, 15.498502731323242, 0.8859735131263733], step: 11000, lr: 9.728730102871649e-05
|
| 366 |
+
2023-03-24 17:44:43,321 44k INFO ====> Epoch: 221, cost 25.35 s
|
| 367 |
+
2023-03-24 17:45:08,220 44k INFO ====> Epoch: 222, cost 24.90 s
|
| 368 |
+
2023-03-24 17:45:33,125 44k INFO ====> Epoch: 223, cost 24.91 s
|
| 369 |
+
2023-03-24 17:45:57,806 44k INFO ====> Epoch: 224, cost 24.68 s
|
| 370 |
+
2023-03-24 17:46:01,570 44k INFO Train Epoch: 225 [0%]
|
| 371 |
+
2023-03-24 17:46:01,571 44k INFO Losses: [2.50315260887146, 2.425022840499878, 10.251396179199219, 19.030975341796875, 0.6795037984848022], step: 11200, lr: 9.723866649812655e-05
|
| 372 |
+
2023-03-24 17:46:06,371 44k INFO Saving model and optimizer state at iteration 225 to ./logs/44k/G_11200.pth
|
| 373 |
+
2023-03-24 17:46:07,942 44k INFO Saving model and optimizer state at iteration 225 to ./logs/44k/D_11200.pth
|
| 374 |
+
2023-03-24 17:46:29,740 44k INFO ====> Epoch: 225, cost 31.93 s
|
| 375 |
+
2023-03-24 17:46:54,684 44k INFO ====> Epoch: 226, cost 24.94 s
|
| 376 |
+
2023-03-24 17:47:19,823 44k INFO ====> Epoch: 227, cost 25.14 s
|
| 377 |
+
2023-03-24 17:47:44,505 44k INFO ====> Epoch: 228, cost 24.68 s
|
| 378 |
+
2023-03-24 17:47:48,475 44k INFO Train Epoch: 229 [0%]
|
| 379 |
+
2023-03-24 17:47:48,476 44k INFO Losses: [2.6375670433044434, 2.0211079120635986, 8.814630508422852, 19.634490966796875, 0.7629879117012024], step: 11400, lr: 9.719005628024282e-05
|
| 380 |
+
2023-03-24 17:48:09,961 44k INFO ====> Epoch: 229, cost 25.46 s
|
| 381 |
+
2023-03-24 17:48:35,039 44k INFO ====> Epoch: 230, cost 25.08 s
|
| 382 |
+
2023-03-24 17:48:59,974 44k INFO ====> Epoch: 231, cost 24.94 s
|
| 383 |
+
2023-03-24 17:49:24,771 44k INFO ====> Epoch: 232, cost 24.80 s
|
| 384 |
+
2023-03-24 17:49:28,612 44k INFO Train Epoch: 233 [0%]
|
| 385 |
+
2023-03-24 17:49:28,613 44k INFO Losses: [2.6957833766937256, 2.0368285179138184, 6.790450096130371, 16.876953125, 0.8912892937660217], step: 11600, lr: 9.714147036291117e-05
|
| 386 |
+
2023-03-24 17:49:50,306 44k INFO ====> Epoch: 233, cost 25.53 s
|
| 387 |
+
2023-03-24 17:50:15,336 44k INFO ====> Epoch: 234, cost 25.03 s
|
| 388 |
+
2023-03-24 17:50:40,138 44k INFO ====> Epoch: 235, cost 24.80 s
|
| 389 |
+
2023-03-24 17:51:04,920 44k INFO ====> Epoch: 236, cost 24.78 s
|
| 390 |
+
2023-03-24 17:51:08,682 44k INFO Train Epoch: 237 [0%]
|
| 391 |
+
2023-03-24 17:51:08,683 44k INFO Losses: [2.3348302841186523, 2.5667877197265625, 7.6211042404174805, 15.862165451049805, 0.9230325818061829], step: 11800, lr: 9.709290873398365e-05
|
| 392 |
+
2023-03-24 17:51:30,569 44k INFO ====> Epoch: 237, cost 25.65 s
|
| 393 |
+
2023-03-24 17:51:55,319 44k INFO ====> Epoch: 238, cost 24.75 s
|
| 394 |
+
2023-03-24 17:52:20,313 44k INFO ====> Epoch: 239, cost 24.99 s
|
| 395 |
+
2023-03-24 17:52:45,066 44k INFO ====> Epoch: 240, cost 24.75 s
|
| 396 |
+
2023-03-24 17:52:48,739 44k INFO Train Epoch: 241 [0%]
|
| 397 |
+
2023-03-24 17:52:48,741 44k INFO Losses: [2.040863513946533, 2.403212547302246, 12.783172607421875, 20.934587478637695, 0.7837891578674316], step: 12000, lr: 9.704437138131832e-05
|
| 398 |
+
2023-03-24 17:52:53,772 44k INFO Saving model and optimizer state at iteration 241 to ./logs/44k/G_12000.pth
|
| 399 |
+
2023-03-24 17:52:55,180 44k INFO Saving model and optimizer state at iteration 241 to ./logs/44k/D_12000.pth
|
| 400 |
+
2023-03-24 17:53:16,867 44k INFO ====> Epoch: 241, cost 31.80 s
|
| 401 |
+
2023-03-24 17:53:41,966 44k INFO ====> Epoch: 242, cost 25.10 s
|
| 402 |
+
2023-03-24 17:54:06,816 44k INFO ====> Epoch: 243, cost 24.85 s
|
| 403 |
+
2023-03-24 17:54:31,779 44k INFO ====> Epoch: 244, cost 24.96 s
|
| 404 |
+
2023-03-24 17:54:35,451 44k INFO Train Epoch: 245 [0%]
|
| 405 |
+
2023-03-24 17:54:35,453 44k INFO Losses: [2.5395002365112305, 2.109894037246704, 11.065472602844238, 20.373876571655273, 0.8642053604125977], step: 12200, lr: 9.699585829277933e-05
|
| 406 |
+
2023-03-24 17:54:57,236 44k INFO ====> Epoch: 245, cost 25.46 s
|
| 407 |
+
2023-03-24 17:55:22,256 44k INFO ====> Epoch: 246, cost 25.02 s
|
| 408 |
+
2023-03-24 17:55:47,152 44k INFO ====> Epoch: 247, cost 24.90 s
|
| 409 |
+
2023-03-24 17:56:12,120 44k INFO ====> Epoch: 248, cost 24.97 s
|
| 410 |
+
2023-03-24 17:56:15,904 44k INFO Train Epoch: 249 [0%]
|
| 411 |
+
2023-03-24 17:56:15,906 44k INFO Losses: [2.2659645080566406, 2.8718552589416504, 13.755086898803711, 21.769315719604492, 0.3315553367137909], step: 12400, lr: 9.694736945623688e-05
|
| 412 |
+
2023-03-24 17:56:37,475 44k INFO ====> Epoch: 249, cost 25.36 s
|
| 413 |
+
2023-03-24 17:57:02,506 44k INFO ====> Epoch: 250, cost 25.03 s
|
| 414 |
+
2023-03-24 17:57:27,327 44k INFO ====> Epoch: 251, cost 24.82 s
|
| 415 |
+
2023-03-24 17:57:52,472 44k INFO ====> Epoch: 252, cost 25.14 s
|
| 416 |
+
2023-03-24 17:57:56,307 44k INFO Train Epoch: 253 [0%]
|
| 417 |
+
2023-03-24 17:57:56,308 44k INFO Losses: [2.3076179027557373, 2.63967227935791, 9.53144645690918, 20.38763427734375, 0.9321761727333069], step: 12600, lr: 9.689890485956725e-05
|
| 418 |
+
2023-03-24 17:58:18,242 44k INFO ====> Epoch: 253, cost 25.77 s
|
| 419 |
+
2023-03-24 17:58:43,283 44k INFO ====> Epoch: 254, cost 25.04 s
|
| 420 |
+
2023-03-24 17:59:08,313 44k INFO ====> Epoch: 255, cost 25.03 s
|
| 421 |
+
2023-03-24 17:59:32,978 44k INFO ====> Epoch: 256, cost 24.66 s
|
| 422 |
+
2023-03-24 17:59:36,789 44k INFO Train Epoch: 257 [0%]
|
| 423 |
+
2023-03-24 17:59:36,790 44k INFO Losses: [2.601497173309326, 2.15515398979187, 10.487128257751465, 20.16332244873047, 0.6738835573196411], step: 12800, lr: 9.685046449065278e-05
|
| 424 |
+
2023-03-24 17:59:41,642 44k INFO Saving model and optimizer state at iteration 257 to ./logs/44k/G_12800.pth
|
| 425 |
+
2023-03-24 17:59:43,224 44k INFO Saving model and optimizer state at iteration 257 to ./logs/44k/D_12800.pth
|
| 426 |
+
2023-03-24 18:00:05,012 44k INFO ====> Epoch: 257, cost 32.03 s
|
| 427 |
+
2023-03-24 18:00:29,964 44k INFO ====> Epoch: 258, cost 24.95 s
|
| 428 |
+
2023-03-24 18:00:55,127 44k INFO ====> Epoch: 259, cost 25.16 s
|
| 429 |
+
2023-03-24 18:01:19,839 44k INFO ====> Epoch: 260, cost 24.71 s
|
| 430 |
+
2023-03-24 18:01:23,558 44k INFO Train Epoch: 261 [0%]
|
| 431 |
+
2023-03-24 18:01:23,559 44k INFO Losses: [2.4871768951416016, 2.240135908126831, 7.533797740936279, 16.42740821838379, 0.8754367828369141], step: 13000, lr: 9.680204833738185e-05
|
so-vits-svc-4.0/20230324-JuewaNS-44k/best/D_13600.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:54660cd48de970b4509401f732850612d08184a8105bcf5049c697efcd923603
|
| 3 |
+
size 561098185
|
so-vits-svc-4.0/20230324-JuewaNS-44k/best/G_13600.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:33b3974ffd10354d567fd4c89a306ad9b3e4654f3da4436954fbff7ab45f76e3
|
| 3 |
+
size 542789405
|
so-vits-svc-4.0/20230324-JuewaNS-44k/best/config.json
ADDED
|
@@ -0,0 +1,93 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"train": {
|
| 3 |
+
"log_interval": 200,
|
| 4 |
+
"eval_interval": 800,
|
| 5 |
+
"seed": 1234,
|
| 6 |
+
"epochs": 10000,
|
| 7 |
+
"learning_rate": 0.0001,
|
| 8 |
+
"betas": [
|
| 9 |
+
0.8,
|
| 10 |
+
0.99
|
| 11 |
+
],
|
| 12 |
+
"eps": 1e-09,
|
| 13 |
+
"batch_size": 6,
|
| 14 |
+
"fp16_run": false,
|
| 15 |
+
"lr_decay": 0.999875,
|
| 16 |
+
"segment_size": 10240,
|
| 17 |
+
"init_lr_ratio": 1,
|
| 18 |
+
"warmup_epochs": 0,
|
| 19 |
+
"c_mel": 45,
|
| 20 |
+
"c_kl": 1.0,
|
| 21 |
+
"use_sr": true,
|
| 22 |
+
"max_speclen": 512,
|
| 23 |
+
"port": "8001",
|
| 24 |
+
"keep_ckpts": 99
|
| 25 |
+
},
|
| 26 |
+
"data": {
|
| 27 |
+
"training_files": "filelists/train.txt",
|
| 28 |
+
"validation_files": "filelists/val.txt",
|
| 29 |
+
"max_wav_value": 32768.0,
|
| 30 |
+
"sampling_rate": 44100,
|
| 31 |
+
"filter_length": 2048,
|
| 32 |
+
"hop_length": 512,
|
| 33 |
+
"win_length": 2048,
|
| 34 |
+
"n_mel_channels": 80,
|
| 35 |
+
"mel_fmin": 0.0,
|
| 36 |
+
"mel_fmax": 22050
|
| 37 |
+
},
|
| 38 |
+
"model": {
|
| 39 |
+
"inter_channels": 192,
|
| 40 |
+
"hidden_channels": 192,
|
| 41 |
+
"filter_channels": 768,
|
| 42 |
+
"n_heads": 2,
|
| 43 |
+
"n_layers": 6,
|
| 44 |
+
"kernel_size": 3,
|
| 45 |
+
"p_dropout": 0.1,
|
| 46 |
+
"resblock": "1",
|
| 47 |
+
"resblock_kernel_sizes": [
|
| 48 |
+
3,
|
| 49 |
+
7,
|
| 50 |
+
11
|
| 51 |
+
],
|
| 52 |
+
"resblock_dilation_sizes": [
|
| 53 |
+
[
|
| 54 |
+
1,
|
| 55 |
+
3,
|
| 56 |
+
5
|
| 57 |
+
],
|
| 58 |
+
[
|
| 59 |
+
1,
|
| 60 |
+
3,
|
| 61 |
+
5
|
| 62 |
+
],
|
| 63 |
+
[
|
| 64 |
+
1,
|
| 65 |
+
3,
|
| 66 |
+
5
|
| 67 |
+
]
|
| 68 |
+
],
|
| 69 |
+
"upsample_rates": [
|
| 70 |
+
8,
|
| 71 |
+
8,
|
| 72 |
+
2,
|
| 73 |
+
2,
|
| 74 |
+
2
|
| 75 |
+
],
|
| 76 |
+
"upsample_initial_channel": 512,
|
| 77 |
+
"upsample_kernel_sizes": [
|
| 78 |
+
16,
|
| 79 |
+
16,
|
| 80 |
+
4,
|
| 81 |
+
4,
|
| 82 |
+
4
|
| 83 |
+
],
|
| 84 |
+
"n_layers_q": 3,
|
| 85 |
+
"use_spectral_norm": false,
|
| 86 |
+
"gin_channels": 256,
|
| 87 |
+
"ssl_dim": 256,
|
| 88 |
+
"n_speakers": 200
|
| 89 |
+
},
|
| 90 |
+
"spk": {
|
| 91 |
+
"JuewaNS": 0
|
| 92 |
+
}
|
| 93 |
+
}
|
so-vits-svc-4.0/20230324-JuewaNS-44k/config.json
ADDED
|
@@ -0,0 +1,93 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"train": {
|
| 3 |
+
"log_interval": 200,
|
| 4 |
+
"eval_interval": 800,
|
| 5 |
+
"seed": 1234,
|
| 6 |
+
"epochs": 10000,
|
| 7 |
+
"learning_rate": 0.0001,
|
| 8 |
+
"betas": [
|
| 9 |
+
0.8,
|
| 10 |
+
0.99
|
| 11 |
+
],
|
| 12 |
+
"eps": 1e-09,
|
| 13 |
+
"batch_size": 6,
|
| 14 |
+
"fp16_run": false,
|
| 15 |
+
"lr_decay": 0.999875,
|
| 16 |
+
"segment_size": 10240,
|
| 17 |
+
"init_lr_ratio": 1,
|
| 18 |
+
"warmup_epochs": 0,
|
| 19 |
+
"c_mel": 45,
|
| 20 |
+
"c_kl": 1.0,
|
| 21 |
+
"use_sr": true,
|
| 22 |
+
"max_speclen": 512,
|
| 23 |
+
"port": "8001",
|
| 24 |
+
"keep_ckpts": 99
|
| 25 |
+
},
|
| 26 |
+
"data": {
|
| 27 |
+
"training_files": "filelists/train.txt",
|
| 28 |
+
"validation_files": "filelists/val.txt",
|
| 29 |
+
"max_wav_value": 32768.0,
|
| 30 |
+
"sampling_rate": 44100,
|
| 31 |
+
"filter_length": 2048,
|
| 32 |
+
"hop_length": 512,
|
| 33 |
+
"win_length": 2048,
|
| 34 |
+
"n_mel_channels": 80,
|
| 35 |
+
"mel_fmin": 0.0,
|
| 36 |
+
"mel_fmax": 22050
|
| 37 |
+
},
|
| 38 |
+
"model": {
|
| 39 |
+
"inter_channels": 192,
|
| 40 |
+
"hidden_channels": 192,
|
| 41 |
+
"filter_channels": 768,
|
| 42 |
+
"n_heads": 2,
|
| 43 |
+
"n_layers": 6,
|
| 44 |
+
"kernel_size": 3,
|
| 45 |
+
"p_dropout": 0.1,
|
| 46 |
+
"resblock": "1",
|
| 47 |
+
"resblock_kernel_sizes": [
|
| 48 |
+
3,
|
| 49 |
+
7,
|
| 50 |
+
11
|
| 51 |
+
],
|
| 52 |
+
"resblock_dilation_sizes": [
|
| 53 |
+
[
|
| 54 |
+
1,
|
| 55 |
+
3,
|
| 56 |
+
5
|
| 57 |
+
],
|
| 58 |
+
[
|
| 59 |
+
1,
|
| 60 |
+
3,
|
| 61 |
+
5
|
| 62 |
+
],
|
| 63 |
+
[
|
| 64 |
+
1,
|
| 65 |
+
3,
|
| 66 |
+
5
|
| 67 |
+
]
|
| 68 |
+
],
|
| 69 |
+
"upsample_rates": [
|
| 70 |
+
8,
|
| 71 |
+
8,
|
| 72 |
+
2,
|
| 73 |
+
2,
|
| 74 |
+
2
|
| 75 |
+
],
|
| 76 |
+
"upsample_initial_channel": 512,
|
| 77 |
+
"upsample_kernel_sizes": [
|
| 78 |
+
16,
|
| 79 |
+
16,
|
| 80 |
+
4,
|
| 81 |
+
4,
|
| 82 |
+
4
|
| 83 |
+
],
|
| 84 |
+
"n_layers_q": 3,
|
| 85 |
+
"use_spectral_norm": false,
|
| 86 |
+
"gin_channels": 256,
|
| 87 |
+
"ssl_dim": 256,
|
| 88 |
+
"n_speakers": 200
|
| 89 |
+
},
|
| 90 |
+
"spk": {
|
| 91 |
+
"JuewaNS": 0
|
| 92 |
+
}
|
| 93 |
+
}
|
so-vits-svc-4.0/20230324-JuewaNS-44k/kmeans_10000.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9601e2be6d3e4ab9adf65ba0b6146e2feff47cc4b058091735c6e474d49df682
|
| 3 |
+
size 15450169
|
so-vits-svc-4.0/20230324-JuewaNS-44k/pruned/P_G_13600.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5ce3ffbbd7ddf1657d95761e1238f6d465afbacdb16be23c70ccfc02cba55e0f
|
| 3 |
+
size 180885829
|
so-vits-svc-4.0/20230324-JuewaNS-44k/pruned/config.json
ADDED
|
@@ -0,0 +1,93 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"train": {
|
| 3 |
+
"log_interval": 200,
|
| 4 |
+
"eval_interval": 800,
|
| 5 |
+
"seed": 1234,
|
| 6 |
+
"epochs": 10000,
|
| 7 |
+
"learning_rate": 0.0001,
|
| 8 |
+
"betas": [
|
| 9 |
+
0.8,
|
| 10 |
+
0.99
|
| 11 |
+
],
|
| 12 |
+
"eps": 1e-09,
|
| 13 |
+
"batch_size": 6,
|
| 14 |
+
"fp16_run": false,
|
| 15 |
+
"lr_decay": 0.999875,
|
| 16 |
+
"segment_size": 10240,
|
| 17 |
+
"init_lr_ratio": 1,
|
| 18 |
+
"warmup_epochs": 0,
|
| 19 |
+
"c_mel": 45,
|
| 20 |
+
"c_kl": 1.0,
|
| 21 |
+
"use_sr": true,
|
| 22 |
+
"max_speclen": 512,
|
| 23 |
+
"port": "8001",
|
| 24 |
+
"keep_ckpts": 99
|
| 25 |
+
},
|
| 26 |
+
"data": {
|
| 27 |
+
"training_files": "filelists/train.txt",
|
| 28 |
+
"validation_files": "filelists/val.txt",
|
| 29 |
+
"max_wav_value": 32768.0,
|
| 30 |
+
"sampling_rate": 44100,
|
| 31 |
+
"filter_length": 2048,
|
| 32 |
+
"hop_length": 512,
|
| 33 |
+
"win_length": 2048,
|
| 34 |
+
"n_mel_channels": 80,
|
| 35 |
+
"mel_fmin": 0.0,
|
| 36 |
+
"mel_fmax": 22050
|
| 37 |
+
},
|
| 38 |
+
"model": {
|
| 39 |
+
"inter_channels": 192,
|
| 40 |
+
"hidden_channels": 192,
|
| 41 |
+
"filter_channels": 768,
|
| 42 |
+
"n_heads": 2,
|
| 43 |
+
"n_layers": 6,
|
| 44 |
+
"kernel_size": 3,
|
| 45 |
+
"p_dropout": 0.1,
|
| 46 |
+
"resblock": "1",
|
| 47 |
+
"resblock_kernel_sizes": [
|
| 48 |
+
3,
|
| 49 |
+
7,
|
| 50 |
+
11
|
| 51 |
+
],
|
| 52 |
+
"resblock_dilation_sizes": [
|
| 53 |
+
[
|
| 54 |
+
1,
|
| 55 |
+
3,
|
| 56 |
+
5
|
| 57 |
+
],
|
| 58 |
+
[
|
| 59 |
+
1,
|
| 60 |
+
3,
|
| 61 |
+
5
|
| 62 |
+
],
|
| 63 |
+
[
|
| 64 |
+
1,
|
| 65 |
+
3,
|
| 66 |
+
5
|
| 67 |
+
]
|
| 68 |
+
],
|
| 69 |
+
"upsample_rates": [
|
| 70 |
+
8,
|
| 71 |
+
8,
|
| 72 |
+
2,
|
| 73 |
+
2,
|
| 74 |
+
2
|
| 75 |
+
],
|
| 76 |
+
"upsample_initial_channel": 512,
|
| 77 |
+
"upsample_kernel_sizes": [
|
| 78 |
+
16,
|
| 79 |
+
16,
|
| 80 |
+
4,
|
| 81 |
+
4,
|
| 82 |
+
4
|
| 83 |
+
],
|
| 84 |
+
"n_layers_q": 3,
|
| 85 |
+
"use_spectral_norm": false,
|
| 86 |
+
"gin_channels": 256,
|
| 87 |
+
"ssl_dim": 256,
|
| 88 |
+
"n_speakers": 200
|
| 89 |
+
},
|
| 90 |
+
"spk": {
|
| 91 |
+
"JuewaNS": 0
|
| 92 |
+
}
|
| 93 |
+
}
|
so-vits-svc-4.0/20230324-JuewaNS-44k/train.log
ADDED
|
@@ -0,0 +1,772 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2023-03-24 18:19:36,643 44k INFO {'train': {'log_interval': 200, 'eval_interval': 800, 'seed': 1234, 'epochs': 10000, 'learning_rate': 0.0001, 'betas': [0.8, 0.99], 'eps': 1e-09, 'batch_size': 6, 'fp16_run': False, 'lr_decay': 0.999875, 'segment_size': 10240, 'init_lr_ratio': 1, 'warmup_epochs': 0, 'c_mel': 45, 'c_kl': 1.0, 'use_sr': True, 'max_speclen': 512, 'port': '8001', 'keep_ckpts': 99}, 'data': {'training_files': 'filelists/train.txt', 'validation_files': 'filelists/val.txt', 'max_wav_value': 32768.0, 'sampling_rate': 44100, 'filter_length': 2048, 'hop_length': 512, 'win_length': 2048, 'n_mel_channels': 80, 'mel_fmin': 0.0, 'mel_fmax': 22050}, 'model': {'inter_channels': 192, 'hidden_channels': 192, 'filter_channels': 768, 'n_heads': 2, 'n_layers': 6, 'kernel_size': 3, 'p_dropout': 0.1, 'resblock': '1', 'resblock_kernel_sizes': [3, 7, 11], 'resblock_dilation_sizes': [[1, 3, 5], [1, 3, 5], [1, 3, 5]], 'upsample_rates': [8, 8, 2, 2, 2], 'upsample_initial_channel': 512, 'upsample_kernel_sizes': [16, 16, 4, 4, 4], 'n_layers_q': 3, 'use_spectral_norm': False, 'gin_channels': 256, 'ssl_dim': 256, 'n_speakers': 200}, 'spk': {'JuewaNS': 0}, 'model_dir': './logs/44k'}
|
| 2 |
+
2023-03-24 18:19:36,643 44k WARNING /root/so-vits-svc-4.0 is not a git repository, therefore hash value comparison will be ignored.
|
| 3 |
+
2023-03-24 18:19:39,837 44k INFO emb_g.weight is not in the checkpoint
|
| 4 |
+
2023-03-24 18:19:39,907 44k INFO Loaded checkpoint './logs/44k/G_0.pth' (iteration 0)
|
| 5 |
+
2023-03-24 18:19:40,077 44k INFO Loaded checkpoint './logs/44k/D_0.pth' (iteration 0)
|
| 6 |
+
2023-03-24 18:19:47,634 44k INFO Train Epoch: 1 [0%]
|
| 7 |
+
2023-03-24 18:19:47,635 44k INFO Losses: [2.8651366233825684, 2.346381187438965, 12.116447448730469, 39.413856506347656, 3.4299936294555664], step: 0, lr: 0.0001
|
| 8 |
+
2023-03-24 18:19:52,952 44k INFO Saving model and optimizer state at iteration 1 to ./logs/44k/G_0.pth
|
| 9 |
+
2023-03-24 18:19:54,318 44k INFO Saving model and optimizer state at iteration 1 to ./logs/44k/D_0.pth
|
| 10 |
+
2023-03-24 18:20:09,532 44k INFO ====> Epoch: 1, cost 32.89 s
|
| 11 |
+
2023-03-24 18:20:24,387 44k INFO ====> Epoch: 2, cost 14.85 s
|
| 12 |
+
2023-03-24 18:20:39,544 44k INFO ====> Epoch: 3, cost 15.16 s
|
| 13 |
+
2023-03-24 18:20:54,488 44k INFO ====> Epoch: 4, cost 14.94 s
|
| 14 |
+
2023-03-24 18:21:09,454 44k INFO ====> Epoch: 5, cost 14.97 s
|
| 15 |
+
2023-03-24 18:21:24,345 44k INFO ====> Epoch: 6, cost 14.89 s
|
| 16 |
+
2023-03-24 18:21:39,167 44k INFO ====> Epoch: 7, cost 14.82 s
|
| 17 |
+
2023-03-24 18:21:44,333 44k INFO Train Epoch: 8 [14%]
|
| 18 |
+
2023-03-24 18:21:44,335 44k INFO Losses: [2.4412031173706055, 2.5478155612945557, 16.34337043762207, 25.992536544799805, 1.6750681400299072], step: 200, lr: 9.991253280566489e-05
|
| 19 |
+
2023-03-24 18:21:54,640 44k INFO ====> Epoch: 8, cost 15.47 s
|
| 20 |
+
2023-03-24 18:22:09,402 44k INFO ====> Epoch: 9, cost 14.76 s
|
| 21 |
+
2023-03-24 18:22:24,340 44k INFO ====> Epoch: 10, cost 14.94 s
|
| 22 |
+
2023-03-24 18:22:39,146 44k INFO ====> Epoch: 11, cost 14.81 s
|
| 23 |
+
2023-03-24 18:22:53,975 44k INFO ====> Epoch: 12, cost 14.83 s
|
| 24 |
+
2023-03-24 18:23:08,849 44k INFO ====> Epoch: 13, cost 14.87 s
|
| 25 |
+
2023-03-24 18:23:23,609 44k INFO ====> Epoch: 14, cost 14.76 s
|
| 26 |
+
2023-03-24 18:23:30,509 44k INFO Train Epoch: 15 [29%]
|
| 27 |
+
2023-03-24 18:23:30,511 44k INFO Losses: [2.853314161300659, 1.969393253326416, 10.559226036071777, 23.596506118774414, 1.1688786745071411], step: 400, lr: 9.982514211643064e-05
|
| 28 |
+
2023-03-24 18:23:39,135 44k INFO ====> Epoch: 15, cost 15.53 s
|
| 29 |
+
2023-03-24 18:23:54,294 44k INFO ====> Epoch: 16, cost 15.16 s
|
| 30 |
+
2023-03-24 18:24:09,139 44k INFO ====> Epoch: 17, cost 14.85 s
|
| 31 |
+
2023-03-24 18:24:23,809 44k INFO ====> Epoch: 18, cost 14.67 s
|
| 32 |
+
2023-03-24 18:24:38,578 44k INFO ====> Epoch: 19, cost 14.77 s
|
| 33 |
+
2023-03-24 18:24:53,441 44k INFO ====> Epoch: 20, cost 14.86 s
|
| 34 |
+
2023-03-24 18:25:08,355 44k INFO ====> Epoch: 21, cost 14.91 s
|
| 35 |
+
2023-03-24 18:25:16,721 44k INFO Train Epoch: 22 [43%]
|
| 36 |
+
2023-03-24 18:25:16,722 44k INFO Losses: [2.245793342590332, 2.530930995941162, 14.867792129516602, 24.568418502807617, 1.0560626983642578], step: 600, lr: 9.973782786538036e-05
|
| 37 |
+
2023-03-24 18:25:23,991 44k INFO ====> Epoch: 22, cost 15.64 s
|
| 38 |
+
2023-03-24 18:25:40,833 44k INFO ====> Epoch: 23, cost 16.84 s
|
| 39 |
+
2023-03-24 18:25:55,605 44k INFO ====> Epoch: 24, cost 14.77 s
|
| 40 |
+
2023-03-24 18:26:10,502 44k INFO ====> Epoch: 25, cost 14.90 s
|
| 41 |
+
2023-03-24 18:26:25,600 44k INFO ====> Epoch: 26, cost 15.10 s
|
| 42 |
+
2023-03-24 18:26:40,526 44k INFO ====> Epoch: 27, cost 14.93 s
|
| 43 |
+
2023-03-24 18:26:55,340 44k INFO ====> Epoch: 28, cost 14.81 s
|
| 44 |
+
2023-03-24 18:27:05,180 44k INFO Train Epoch: 29 [57%]
|
| 45 |
+
2023-03-24 18:27:05,182 44k INFO Losses: [2.613215923309326, 2.27390718460083, 8.487275123596191, 20.264320373535156, 1.1122970581054688], step: 800, lr: 9.965058998565574e-05
|
| 46 |
+
2023-03-24 18:27:09,722 44k INFO Saving model and optimizer state at iteration 29 to ./logs/44k/G_800.pth
|
| 47 |
+
2023-03-24 18:27:11,038 44k INFO Saving model and optimizer state at iteration 29 to ./logs/44k/D_800.pth
|
| 48 |
+
2023-03-24 18:27:16,521 44k INFO ====> Epoch: 29, cost 21.18 s
|
| 49 |
+
2023-03-24 18:27:31,451 44k INFO ====> Epoch: 30, cost 14.93 s
|
| 50 |
+
2023-03-24 18:27:46,263 44k INFO ====> Epoch: 31, cost 14.81 s
|
| 51 |
+
2023-03-24 18:28:01,205 44k INFO ====> Epoch: 32, cost 14.94 s
|
| 52 |
+
2023-03-24 18:28:15,968 44k INFO ====> Epoch: 33, cost 14.76 s
|
| 53 |
+
2023-03-24 18:28:30,790 44k INFO ====> Epoch: 34, cost 14.82 s
|
| 54 |
+
2023-03-24 18:28:46,345 44k INFO ====> Epoch: 35, cost 15.55 s
|
| 55 |
+
2023-03-24 18:28:57,976 44k INFO Train Epoch: 36 [71%]
|
| 56 |
+
2023-03-24 18:28:57,977 44k INFO Losses: [2.115177869796753, 2.6007447242736816, 13.032262802124023, 26.101831436157227, 0.6284221410751343], step: 1000, lr: 9.956342841045691e-05
|
| 57 |
+
2023-03-24 18:29:01,888 44k INFO ====> Epoch: 36, cost 15.54 s
|
| 58 |
+
2023-03-24 18:29:16,747 44k INFO ====> Epoch: 37, cost 14.86 s
|
| 59 |
+
2023-03-24 18:29:31,673 44k INFO ====> Epoch: 38, cost 14.93 s
|
| 60 |
+
2023-03-24 18:29:46,543 44k INFO ====> Epoch: 39, cost 14.87 s
|
| 61 |
+
2023-03-24 18:30:01,582 44k INFO ====> Epoch: 40, cost 15.04 s
|
| 62 |
+
2023-03-24 18:30:16,391 44k INFO ====> Epoch: 41, cost 14.81 s
|
| 63 |
+
2023-03-24 18:30:31,370 44k INFO ====> Epoch: 42, cost 14.98 s
|
| 64 |
+
2023-03-24 18:30:45,071 44k INFO Train Epoch: 43 [86%]
|
| 65 |
+
2023-03-24 18:30:45,072 44k INFO Losses: [2.658895969390869, 2.289456605911255, 4.936427593231201, 20.44791030883789, 0.47966423630714417], step: 1200, lr: 9.947634307304244e-05
|
| 66 |
+
2023-03-24 18:30:47,487 44k INFO ====> Epoch: 43, cost 16.12 s
|
| 67 |
+
2023-03-24 18:31:02,415 44k INFO ====> Epoch: 44, cost 14.93 s
|
| 68 |
+
2023-03-24 18:31:17,262 44k INFO ====> Epoch: 45, cost 14.85 s
|
| 69 |
+
2023-03-24 18:31:31,991 44k INFO ====> Epoch: 46, cost 14.73 s
|
| 70 |
+
2023-03-24 18:31:46,701 44k INFO ====> Epoch: 47, cost 14.71 s
|
| 71 |
+
2023-03-24 18:32:01,612 44k INFO ====> Epoch: 48, cost 14.91 s
|
| 72 |
+
2023-03-24 18:32:17,899 44k INFO ====> Epoch: 49, cost 16.29 s
|
| 73 |
+
2023-03-24 18:32:32,832 44k INFO ====> Epoch: 50, cost 14.93 s
|
| 74 |
+
2023-03-24 18:32:36,346 44k INFO Train Epoch: 51 [0%]
|
| 75 |
+
2023-03-24 18:32:36,347 44k INFO Losses: [2.707481861114502, 2.2357685565948486, 10.195556640625, 20.226890563964844, 0.812960147857666], step: 1400, lr: 9.937691023999092e-05
|
| 76 |
+
2023-03-24 18:32:48,230 44k INFO ====> Epoch: 51, cost 15.40 s
|
| 77 |
+
2023-03-24 18:33:02,841 44k INFO ====> Epoch: 52, cost 14.61 s
|
| 78 |
+
2023-03-24 18:33:17,559 44k INFO ====> Epoch: 53, cost 14.72 s
|
| 79 |
+
2023-03-24 18:33:32,818 44k INFO ====> Epoch: 54, cost 15.26 s
|
| 80 |
+
2023-03-24 18:33:47,747 44k INFO ====> Epoch: 55, cost 14.93 s
|
| 81 |
+
2023-03-24 18:34:02,348 44k INFO ====> Epoch: 56, cost 14.60 s
|
| 82 |
+
2023-03-24 18:34:17,404 44k INFO ====> Epoch: 57, cost 15.06 s
|
| 83 |
+
2023-03-24 18:34:22,685 44k INFO Train Epoch: 58 [14%]
|
| 84 |
+
2023-03-24 18:34:22,686 44k INFO Losses: [2.843576192855835, 2.228142738342285, 14.787569046020508, 23.943769454956055, 0.993866503238678], step: 1600, lr: 9.928998804478705e-05
|
| 85 |
+
2023-03-24 18:34:27,079 44k INFO Saving model and optimizer state at iteration 58 to ./logs/44k/G_1600.pth
|
| 86 |
+
2023-03-24 18:34:28,258 44k INFO Saving model and optimizer state at iteration 58 to ./logs/44k/D_1600.pth
|
| 87 |
+
2023-03-24 18:34:38,568 44k INFO ====> Epoch: 58, cost 21.16 s
|
| 88 |
+
2023-03-24 18:34:55,234 44k INFO ====> Epoch: 59, cost 16.67 s
|
| 89 |
+
2023-03-24 18:35:09,982 44k INFO ====> Epoch: 60, cost 14.75 s
|
| 90 |
+
2023-03-24 18:35:24,679 44k INFO ====> Epoch: 61, cost 14.70 s
|
| 91 |
+
2023-03-24 18:35:39,348 44k INFO ====> Epoch: 62, cost 14.67 s
|
| 92 |
+
2023-03-24 18:35:54,070 44k INFO ====> Epoch: 63, cost 14.72 s
|
| 93 |
+
2023-03-24 18:36:10,703 44k INFO ====> Epoch: 64, cost 16.63 s
|
| 94 |
+
2023-03-24 18:36:17,521 44k INFO Train Epoch: 65 [29%]
|
| 95 |
+
2023-03-24 18:36:17,522 44k INFO Losses: [2.328249931335449, 2.4285080432891846, 12.13663387298584, 19.89910125732422, 1.2783076763153076], step: 1800, lr: 9.92031418779886e-05
|
| 96 |
+
2023-03-24 18:36:26,007 44k INFO ====> Epoch: 65, cost 15.30 s
|
| 97 |
+
2023-03-24 18:36:40,756 44k INFO ====> Epoch: 66, cost 14.75 s
|
| 98 |
+
2023-03-24 18:36:55,574 44k INFO ====> Epoch: 67, cost 14.82 s
|
| 99 |
+
2023-03-24 18:37:10,496 44k INFO ====> Epoch: 68, cost 14.92 s
|
| 100 |
+
2023-03-24 18:37:25,888 44k INFO ====> Epoch: 69, cost 15.39 s
|
| 101 |
+
2023-03-24 18:37:40,640 44k INFO ====> Epoch: 70, cost 14.75 s
|
| 102 |
+
2023-03-24 18:37:55,697 44k INFO ====> Epoch: 71, cost 15.06 s
|
| 103 |
+
2023-03-24 18:38:04,052 44k INFO Train Epoch: 72 [43%]
|
| 104 |
+
2023-03-24 18:38:04,053 44k INFO Losses: [2.5356943607330322, 2.3216590881347656, 12.424395561218262, 23.910093307495117, 0.887087345123291], step: 2000, lr: 9.911637167309565e-05
|
| 105 |
+
2023-03-24 18:38:11,120 44k INFO ====> Epoch: 72, cost 15.42 s
|
| 106 |
+
2023-03-24 18:38:26,070 44k INFO ====> Epoch: 73, cost 14.95 s
|
| 107 |
+
2023-03-24 18:38:41,212 44k INFO ====> Epoch: 74, cost 15.14 s
|
| 108 |
+
2023-03-24 18:38:57,820 44k INFO ====> Epoch: 75, cost 16.61 s
|
| 109 |
+
2023-03-24 18:39:12,632 44k INFO ====> Epoch: 76, cost 14.81 s
|
| 110 |
+
2023-03-24 18:39:29,340 44k INFO ====> Epoch: 77, cost 16.71 s
|
| 111 |
+
2023-03-24 18:39:44,102 44k INFO ====> Epoch: 78, cost 14.76 s
|
| 112 |
+
2023-03-24 18:39:55,219 44k INFO Train Epoch: 79 [57%]
|
| 113 |
+
2023-03-24 18:39:55,220 44k INFO Losses: [2.525271415710449, 2.084883689880371, 9.368517875671387, 17.8083438873291, 0.8916102647781372], step: 2200, lr: 9.902967736366644e-05
|
| 114 |
+
2023-03-24 18:40:00,683 44k INFO ====> Epoch: 79, cost 16.58 s
|
| 115 |
+
2023-03-24 18:40:15,489 44k INFO ====> Epoch: 80, cost 14.81 s
|
| 116 |
+
2023-03-24 18:40:30,311 44k INFO ====> Epoch: 81, cost 14.82 s
|
| 117 |
+
2023-03-24 18:40:45,505 44k INFO ====> Epoch: 82, cost 15.19 s
|
| 118 |
+
2023-03-24 18:41:00,536 44k INFO ====> Epoch: 83, cost 15.03 s
|
| 119 |
+
2023-03-24 18:41:16,174 44k INFO ====> Epoch: 84, cost 15.64 s
|
| 120 |
+
2023-03-24 18:41:30,993 44k INFO ====> Epoch: 85, cost 14.82 s
|
| 121 |
+
2023-03-24 18:41:42,384 44k INFO Train Epoch: 86 [71%]
|
| 122 |
+
2023-03-24 18:41:42,385 44k INFO Losses: [2.3191444873809814, 2.3854219913482666, 11.815009117126465, 19.916088104248047, 0.8946883082389832], step: 2400, lr: 9.894305888331732e-05
|
| 123 |
+
2023-03-24 18:41:46,894 44k INFO Saving model and optimizer state at iteration 86 to ./logs/44k/G_2400.pth
|
| 124 |
+
2023-03-24 18:41:48,085 44k INFO Saving model and optimizer state at iteration 86 to ./logs/44k/D_2400.pth
|
| 125 |
+
2023-03-24 18:41:52,207 44k INFO ====> Epoch: 86, cost 21.21 s
|
| 126 |
+
2023-03-24 18:42:07,069 44k INFO ====> Epoch: 87, cost 14.86 s
|
| 127 |
+
2023-03-24 18:42:21,905 44k INFO ====> Epoch: 88, cost 14.84 s
|
| 128 |
+
2023-03-24 18:42:36,734 44k INFO ====> Epoch: 89, cost 14.83 s
|
| 129 |
+
2023-03-24 18:42:51,437 44k INFO ====> Epoch: 90, cost 14.70 s
|
| 130 |
+
2023-03-24 18:43:06,191 44k INFO ====> Epoch: 91, cost 14.75 s
|
| 131 |
+
2023-03-24 18:43:21,034 44k INFO ====> Epoch: 92, cost 14.84 s
|
| 132 |
+
2023-03-24 18:43:34,239 44k INFO Train Epoch: 93 [86%]
|
| 133 |
+
2023-03-24 18:43:34,240 44k INFO Losses: [2.467679023742676, 2.421851634979248, 10.510832786560059, 21.985858917236328, 1.062198281288147], step: 2600, lr: 9.885651616572276e-05
|
| 134 |
+
2023-03-24 18:43:36,795 44k INFO ====> Epoch: 93, cost 15.76 s
|
| 135 |
+
2023-03-24 18:43:51,767 44k INFO ====> Epoch: 94, cost 14.97 s
|
| 136 |
+
2023-03-24 18:44:06,700 44k INFO ====> Epoch: 95, cost 14.93 s
|
| 137 |
+
2023-03-24 18:44:21,270 44k INFO ====> Epoch: 96, cost 14.57 s
|
| 138 |
+
2023-03-24 18:44:36,401 44k INFO ====> Epoch: 97, cost 15.13 s
|
| 139 |
+
2023-03-24 18:44:53,093 44k INFO ====> Epoch: 98, cost 16.69 s
|
| 140 |
+
2023-03-24 18:45:07,890 44k INFO ====> Epoch: 99, cost 14.80 s
|
| 141 |
+
2023-03-24 18:45:22,480 44k INFO ====> Epoch: 100, cost 14.59 s
|
| 142 |
+
2023-03-24 18:45:26,087 44k INFO Train Epoch: 101 [0%]
|
| 143 |
+
2023-03-24 18:45:26,088 44k INFO Losses: [2.4023709297180176, 2.3098299503326416, 11.050838470458984, 23.734569549560547, 0.7466572523117065], step: 2800, lr: 9.875770288847208e-05
|
| 144 |
+
2023-03-24 18:45:37,857 44k INFO ====> Epoch: 101, cost 15.38 s
|
| 145 |
+
2023-03-24 18:45:52,566 44k INFO ====> Epoch: 102, cost 14.71 s
|
| 146 |
+
2023-03-24 18:46:07,422 44k INFO ====> Epoch: 103, cost 14.86 s
|
| 147 |
+
2023-03-24 18:46:22,264 44k INFO ====> Epoch: 104, cost 14.84 s
|
| 148 |
+
2023-03-24 18:46:36,961 44k INFO ====> Epoch: 105, cost 14.70 s
|
| 149 |
+
2023-03-24 18:46:51,580 44k INFO ====> Epoch: 106, cost 14.62 s
|
| 150 |
+
2023-03-24 18:47:06,263 44k INFO ====> Epoch: 107, cost 14.68 s
|
| 151 |
+
2023-03-24 18:47:11,409 44k INFO Train Epoch: 108 [14%]
|
| 152 |
+
2023-03-24 18:47:11,411 44k INFO Losses: [2.3292808532714844, 2.2167253494262695, 10.71139907836914, 21.37546157836914, 0.8030653595924377], step: 3000, lr: 9.867132229656573e-05
|
| 153 |
+
2023-03-24 18:47:21,503 44k INFO ====> Epoch: 108, cost 15.24 s
|
| 154 |
+
2023-03-24 18:47:36,263 44k INFO ====> Epoch: 109, cost 14.76 s
|
| 155 |
+
2023-03-24 18:47:50,996 44k INFO ====> Epoch: 110, cost 14.73 s
|
| 156 |
+
2023-03-24 18:48:05,795 44k INFO ====> Epoch: 111, cost 14.80 s
|
| 157 |
+
2023-03-24 18:48:20,515 44k INFO ====> Epoch: 112, cost 14.72 s
|
| 158 |
+
2023-03-24 18:48:35,425 44k INFO ====> Epoch: 113, cost 14.91 s
|
| 159 |
+
2023-03-24 18:48:50,175 44k INFO ====> Epoch: 114, cost 14.75 s
|
| 160 |
+
2023-03-24 18:48:57,076 44k INFO Train Epoch: 115 [29%]
|
| 161 |
+
2023-03-24 18:48:57,078 44k INFO Losses: [2.477935314178467, 2.528744697570801, 8.90847110748291, 22.292436599731445, 0.34426790475845337], step: 3200, lr: 9.858501725933955e-05
|
| 162 |
+
2023-03-24 18:49:02,086 44k INFO Saving model and optimizer state at iteration 115 to ./logs/44k/G_3200.pth
|
| 163 |
+
2023-03-24 18:49:03,349 44k INFO Saving model and optimizer state at iteration 115 to ./logs/44k/D_3200.pth
|
| 164 |
+
2023-03-24 18:49:12,380 44k INFO ====> Epoch: 115, cost 22.21 s
|
| 165 |
+
2023-03-24 18:49:27,120 44k INFO ====> Epoch: 116, cost 14.74 s
|
| 166 |
+
2023-03-24 18:49:41,757 44k INFO ====> Epoch: 117, cost 14.64 s
|
| 167 |
+
2023-03-24 18:49:56,948 44k INFO ====> Epoch: 118, cost 15.19 s
|
| 168 |
+
2023-03-24 18:50:11,603 44k INFO ====> Epoch: 119, cost 14.66 s
|
| 169 |
+
2023-03-24 18:50:26,404 44k INFO ====> Epoch: 120, cost 14.80 s
|
| 170 |
+
2023-03-24 18:50:41,127 44k INFO ====> Epoch: 121, cost 14.72 s
|
| 171 |
+
2023-03-24 18:50:49,536 44k INFO Train Epoch: 122 [43%]
|
| 172 |
+
2023-03-24 18:50:49,537 44k INFO Losses: [2.4078025817871094, 2.5715935230255127, 8.99583911895752, 17.1897029876709, 1.0879757404327393], step: 3400, lr: 9.8498787710708e-05
|
| 173 |
+
2023-03-24 18:50:56,712 44k INFO ====> Epoch: 122, cost 15.58 s
|
| 174 |
+
2023-03-24 18:51:11,449 44k INFO ====> Epoch: 123, cost 14.74 s
|
| 175 |
+
2023-03-24 18:51:26,066 44k INFO ====> Epoch: 124, cost 14.62 s
|
| 176 |
+
2023-03-24 18:51:40,977 44k INFO ====> Epoch: 125, cost 14.91 s
|
| 177 |
+
2023-03-24 18:51:56,070 44k INFO ====> Epoch: 126, cost 15.09 s
|
| 178 |
+
2023-03-24 18:52:10,818 44k INFO ====> Epoch: 127, cost 14.75 s
|
| 179 |
+
2023-03-24 18:52:25,534 44k INFO ====> Epoch: 128, cost 14.72 s
|
| 180 |
+
2023-03-24 18:52:35,476 44k INFO Train Epoch: 129 [57%]
|
| 181 |
+
2023-03-24 18:52:35,477 44k INFO Losses: [2.5406367778778076, 2.3820292949676514, 9.032157897949219, 18.73415756225586, 0.952826201915741], step: 3600, lr: 9.841263358464336e-05
|
| 182 |
+
2023-03-24 18:52:40,946 44k INFO ====> Epoch: 129, cost 15.41 s
|
| 183 |
+
2023-03-24 18:52:55,845 44k INFO ====> Epoch: 130, cost 14.90 s
|
| 184 |
+
2023-03-24 18:53:11,000 44k INFO ====> Epoch: 131, cost 15.16 s
|
| 185 |
+
2023-03-24 18:53:25,870 44k INFO ====> Epoch: 132, cost 14.87 s
|
| 186 |
+
2023-03-24 18:53:40,771 44k INFO ====> Epoch: 133, cost 14.90 s
|
| 187 |
+
2023-03-24 18:53:55,438 44k INFO ====> Epoch: 134, cost 14.67 s
|
| 188 |
+
2023-03-24 18:54:10,252 44k INFO ====> Epoch: 135, cost 14.81 s
|
| 189 |
+
2023-03-24 18:54:21,658 44k INFO Train Epoch: 136 [71%]
|
| 190 |
+
2023-03-24 18:54:21,659 44k INFO Losses: [2.567310333251953, 2.164691686630249, 10.801031112670898, 24.128572463989258, 0.7646280527114868], step: 3800, lr: 9.832655481517557e-05
|
| 191 |
+
2023-03-24 18:54:25,608 44k INFO ====> Epoch: 136, cost 15.36 s
|
| 192 |
+
2023-03-24 18:54:40,505 44k INFO ====> Epoch: 137, cost 14.90 s
|
| 193 |
+
2023-03-24 18:54:55,428 44k INFO ====> Epoch: 138, cost 14.92 s
|
| 194 |
+
2023-03-24 18:55:10,583 44k INFO ====> Epoch: 139, cost 15.16 s
|
| 195 |
+
2023-03-24 18:55:25,273 44k INFO ====> Epoch: 140, cost 14.69 s
|
| 196 |
+
2023-03-24 18:55:39,865 44k INFO ====> Epoch: 141, cost 14.59 s
|
| 197 |
+
2023-03-24 18:55:54,556 44k INFO ====> Epoch: 142, cost 14.69 s
|
| 198 |
+
2023-03-24 18:56:07,672 44k INFO Train Epoch: 143 [86%]
|
| 199 |
+
2023-03-24 18:56:07,673 44k INFO Losses: [2.4808027744293213, 2.3600611686706543, 9.740200996398926, 22.769962310791016, 0.9508404731750488], step: 4000, lr: 9.824055133639235e-05
|
| 200 |
+
2023-03-24 18:56:12,373 44k INFO Saving model and optimizer state at iteration 143 to ./logs/44k/G_4000.pth
|
| 201 |
+
2023-03-24 18:56:13,752 44k INFO Saving model and optimizer state at iteration 143 to ./logs/44k/D_4000.pth
|
| 202 |
+
2023-03-24 18:56:16,196 44k INFO ====> Epoch: 143, cost 21.64 s
|
| 203 |
+
2023-03-24 18:56:31,092 44k INFO ====> Epoch: 144, cost 14.90 s
|
| 204 |
+
2023-03-24 18:56:46,318 44k INFO ====> Epoch: 145, cost 15.23 s
|
| 205 |
+
2023-03-24 18:57:01,107 44k INFO ====> Epoch: 146, cost 14.79 s
|
| 206 |
+
2023-03-24 18:57:15,686 44k INFO ====> Epoch: 147, cost 14.58 s
|
| 207 |
+
2023-03-24 18:57:30,896 44k INFO ====> Epoch: 148, cost 15.21 s
|
| 208 |
+
2023-03-24 18:57:45,509 44k INFO ====> Epoch: 149, cost 14.61 s
|
| 209 |
+
2023-03-24 18:58:00,186 44k INFO ====> Epoch: 150, cost 14.68 s
|
| 210 |
+
2023-03-24 18:58:03,715 44k INFO Train Epoch: 151 [0%]
|
| 211 |
+
2023-03-24 18:58:03,716 44k INFO Losses: [2.453944683074951, 2.4591901302337646, 8.528477668762207, 19.629995346069336, 1.1228570938110352], step: 4200, lr: 9.814235375455375e-05
|
| 212 |
+
2023-03-24 18:58:15,478 44k INFO ====> Epoch: 151, cost 15.29 s
|
| 213 |
+
2023-03-24 18:58:30,175 44k INFO ====> Epoch: 152, cost 14.70 s
|
| 214 |
+
2023-03-24 18:58:44,994 44k INFO ====> Epoch: 153, cost 14.82 s
|
| 215 |
+
2023-03-24 18:58:59,953 44k INFO ====> Epoch: 154, cost 14.96 s
|
| 216 |
+
2023-03-24 18:59:14,793 44k INFO ====> Epoch: 155, cost 14.84 s
|
| 217 |
+
2023-03-24 18:59:29,507 44k INFO ====> Epoch: 156, cost 14.71 s
|
| 218 |
+
2023-03-24 18:59:44,303 44k INFO ====> Epoch: 157, cost 14.80 s
|
| 219 |
+
2023-03-24 18:59:49,734 44k INFO Train Epoch: 158 [14%]
|
| 220 |
+
2023-03-24 18:59:49,735 44k INFO Losses: [2.514904260635376, 2.1705262660980225, 9.76323413848877, 18.711702346801758, 0.7594791650772095], step: 4400, lr: 9.80565113912702e-05
|
| 221 |
+
2023-03-24 18:59:59,801 44k INFO ====> Epoch: 158, cost 15.50 s
|
| 222 |
+
2023-03-24 19:00:14,906 44k INFO ====> Epoch: 159, cost 15.10 s
|
| 223 |
+
2023-03-24 19:00:29,873 44k INFO ====> Epoch: 160, cost 14.97 s
|
| 224 |
+
2023-03-24 19:00:44,776 44k INFO ====> Epoch: 161, cost 14.90 s
|
| 225 |
+
2023-03-24 19:00:59,727 44k INFO ====> Epoch: 162, cost 14.95 s
|
| 226 |
+
2023-03-24 19:01:14,734 44k INFO ====> Epoch: 163, cost 15.01 s
|
| 227 |
+
2023-03-24 19:01:29,720 44k INFO ====> Epoch: 164, cost 14.99 s
|
| 228 |
+
2023-03-24 19:01:36,500 44k INFO Train Epoch: 165 [29%]
|
| 229 |
+
2023-03-24 19:01:36,501 44k INFO Losses: [2.6813385486602783, 2.1468124389648438, 7.211049556732178, 16.662872314453125, 0.9376119375228882], step: 4600, lr: 9.797074411189339e-05
|
| 230 |
+
2023-03-24 19:01:45,150 44k INFO ====> Epoch: 165, cost 15.43 s
|
| 231 |
+
2023-03-24 19:02:00,181 44k INFO ====> Epoch: 166, cost 15.03 s
|
| 232 |
+
2023-03-24 19:02:15,094 44k INFO ====> Epoch: 167, cost 14.91 s
|
| 233 |
+
2023-03-24 19:02:30,176 44k INFO ====> Epoch: 168, cost 15.08 s
|
| 234 |
+
2023-03-24 19:02:45,178 44k INFO ====> Epoch: 169, cost 15.00 s
|
| 235 |
+
2023-03-24 19:03:00,258 44k INFO ====> Epoch: 170, cost 15.08 s
|
| 236 |
+
2023-03-24 19:03:15,267 44k INFO ====> Epoch: 171, cost 15.01 s
|
| 237 |
+
2023-03-24 19:03:23,715 44k INFO Train Epoch: 172 [43%]
|
| 238 |
+
2023-03-24 19:03:23,716 44k INFO Losses: [2.2882304191589355, 2.3772926330566406, 9.894192695617676, 22.153350830078125, 0.46218857169151306], step: 4800, lr: 9.78850518507495e-05
|
| 239 |
+
2023-03-24 19:03:28,574 44k INFO Saving model and optimizer state at iteration 172 to ./logs/44k/G_4800.pth
|
| 240 |
+
2023-03-24 19:03:29,794 44k INFO Saving model and optimizer state at iteration 172 to ./logs/44k/D_4800.pth
|
| 241 |
+
2023-03-24 19:03:36,948 44k INFO ====> Epoch: 172, cost 21.68 s
|
| 242 |
+
2023-03-24 19:03:52,073 44k INFO ====> Epoch: 173, cost 15.12 s
|
| 243 |
+
2023-03-24 19:04:06,911 44k INFO ====> Epoch: 174, cost 14.84 s
|
| 244 |
+
2023-03-24 19:04:21,629 44k INFO ====> Epoch: 175, cost 14.72 s
|
| 245 |
+
2023-03-24 19:04:36,595 44k INFO ====> Epoch: 176, cost 14.97 s
|
| 246 |
+
2023-03-24 19:04:51,655 44k INFO ====> Epoch: 177, cost 15.06 s
|
| 247 |
+
2023-03-24 19:05:06,846 44k INFO ====> Epoch: 178, cost 15.19 s
|
| 248 |
+
2023-03-24 19:05:16,796 44k INFO Train Epoch: 179 [57%]
|
| 249 |
+
2023-03-24 19:05:16,798 44k INFO Losses: [1.827094554901123, 2.9733214378356934, 12.0966215133667, 21.781890869140625, 0.7995937466621399], step: 5000, lr: 9.779943454222217e-05
|
| 250 |
+
2023-03-24 19:05:22,243 44k INFO ====> Epoch: 179, cost 15.40 s
|
| 251 |
+
2023-03-24 19:05:37,206 44k INFO ====> Epoch: 180, cost 14.96 s
|
| 252 |
+
2023-03-24 19:05:52,118 44k INFO ====> Epoch: 181, cost 14.91 s
|
| 253 |
+
2023-03-24 19:06:06,759 44k INFO ====> Epoch: 182, cost 14.64 s
|
| 254 |
+
2023-03-24 19:06:21,816 44k INFO ====> Epoch: 183, cost 15.06 s
|
| 255 |
+
2023-03-24 19:06:36,774 44k INFO ====> Epoch: 184, cost 14.96 s
|
| 256 |
+
2023-03-24 19:06:51,756 44k INFO ====> Epoch: 185, cost 14.98 s
|
| 257 |
+
2023-03-24 19:07:03,944 44k INFO Train Epoch: 186 [71%]
|
| 258 |
+
2023-03-24 19:07:03,946 44k INFO Losses: [2.455173969268799, 2.5180628299713135, 8.574335098266602, 22.821983337402344, 0.8425252437591553], step: 5200, lr: 9.771389212075249e-05
|
| 259 |
+
2023-03-24 19:07:07,998 44k INFO ====> Epoch: 186, cost 16.24 s
|
| 260 |
+
2023-03-24 19:07:23,004 44k INFO ====> Epoch: 187, cost 15.01 s
|
| 261 |
+
2023-03-24 19:07:37,921 44k INFO ====> Epoch: 188, cost 14.92 s
|
| 262 |
+
2023-03-24 19:07:52,959 44k INFO ====> Epoch: 189, cost 15.04 s
|
| 263 |
+
2023-03-24 19:08:08,037 44k INFO ====> Epoch: 190, cost 15.08 s
|
| 264 |
+
2023-03-24 19:08:22,948 44k INFO ====> Epoch: 191, cost 14.91 s
|
| 265 |
+
2023-03-24 19:08:38,306 44k INFO ====> Epoch: 192, cost 15.36 s
|
| 266 |
+
2023-03-24 19:08:51,376 44k INFO Train Epoch: 193 [86%]
|
| 267 |
+
2023-03-24 19:08:51,377 44k INFO Losses: [2.3449130058288574, 2.547933578491211, 11.487386703491211, 22.50403594970703, 0.6760184168815613], step: 5400, lr: 9.762842452083883e-05
|
| 268 |
+
2023-03-24 19:08:53,739 44k INFO ====> Epoch: 193, cost 15.43 s
|
| 269 |
+
2023-03-24 19:09:08,709 44k INFO ====> Epoch: 194, cost 14.97 s
|
| 270 |
+
2023-03-24 19:09:23,456 44k INFO ====> Epoch: 195, cost 14.75 s
|
| 271 |
+
2023-03-24 19:09:38,275 44k INFO ====> Epoch: 196, cost 14.82 s
|
| 272 |
+
2023-03-24 19:09:53,149 44k INFO ====> Epoch: 197, cost 14.87 s
|
| 273 |
+
2023-03-24 19:10:07,952 44k INFO ====> Epoch: 198, cost 14.80 s
|
| 274 |
+
2023-03-24 19:10:22,701 44k INFO ====> Epoch: 199, cost 14.75 s
|
| 275 |
+
2023-03-24 19:10:37,606 44k INFO ====> Epoch: 200, cost 14.91 s
|
| 276 |
+
2023-03-24 19:10:41,351 44k INFO Train Epoch: 201 [0%]
|
| 277 |
+
2023-03-24 19:10:41,352 44k INFO Losses: [2.192190647125244, 2.452057123184204, 14.672394752502441, 22.408723831176758, 0.7568256855010986], step: 5600, lr: 9.753083879807726e-05
|
| 278 |
+
2023-03-24 19:10:46,121 44k INFO Saving model and optimizer state at iteration 201 to ./logs/44k/G_5600.pth
|
| 279 |
+
2023-03-24 19:10:47,341 44k INFO Saving model and optimizer state at iteration 201 to ./logs/44k/D_5600.pth
|
| 280 |
+
2023-03-24 19:10:59,171 44k INFO ====> Epoch: 201, cost 21.56 s
|
| 281 |
+
2023-03-24 19:11:14,778 44k INFO ====> Epoch: 202, cost 15.61 s
|
| 282 |
+
2023-03-24 19:11:30,111 44k INFO ====> Epoch: 203, cost 15.33 s
|
| 283 |
+
2023-03-24 19:11:45,174 44k INFO ====> Epoch: 204, cost 15.06 s
|
| 284 |
+
2023-03-24 19:12:00,190 44k INFO ====> Epoch: 205, cost 15.02 s
|
| 285 |
+
2023-03-24 19:12:16,844 44k INFO ====> Epoch: 206, cost 16.65 s
|
| 286 |
+
2023-03-24 19:12:31,854 44k INFO ====> Epoch: 207, cost 15.01 s
|
| 287 |
+
2023-03-24 19:12:37,138 44k INFO Train Epoch: 208 [14%]
|
| 288 |
+
2023-03-24 19:12:37,139 44k INFO Losses: [2.259969472885132, 2.62406063079834, 11.825155258178711, 20.026582717895508, 0.8043803572654724], step: 5800, lr: 9.744553130976908e-05
|
| 289 |
+
2023-03-24 19:12:47,441 44k INFO ====> Epoch: 208, cost 15.59 s
|
| 290 |
+
2023-03-24 19:13:02,643 44k INFO ====> Epoch: 209, cost 15.20 s
|
| 291 |
+
2023-03-24 19:13:17,512 44k INFO ====> Epoch: 210, cost 14.87 s
|
| 292 |
+
2023-03-24 19:13:32,485 44k INFO ====> Epoch: 211, cost 14.97 s
|
| 293 |
+
2023-03-24 19:13:47,340 44k INFO ====> Epoch: 212, cost 14.86 s
|
| 294 |
+
2023-03-24 19:14:02,260 44k INFO ====> Epoch: 213, cost 14.92 s
|
| 295 |
+
2023-03-24 19:14:17,409 44k INFO ====> Epoch: 214, cost 15.15 s
|
| 296 |
+
2023-03-24 19:14:24,344 44k INFO Train Epoch: 215 [29%]
|
| 297 |
+
2023-03-24 19:14:24,345 44k INFO Losses: [2.371037721633911, 2.795213222503662, 12.122397422790527, 23.133861541748047, 0.5996092557907104], step: 6000, lr: 9.736029843752747e-05
|
| 298 |
+
2023-03-24 19:14:32,970 44k INFO ====> Epoch: 215, cost 15.56 s
|
| 299 |
+
2023-03-24 19:14:47,969 44k INFO ====> Epoch: 216, cost 15.00 s
|
| 300 |
+
2023-03-24 19:15:02,915 44k INFO ====> Epoch: 217, cost 14.95 s
|
| 301 |
+
2023-03-24 19:15:17,769 44k INFO ====> Epoch: 218, cost 14.85 s
|
| 302 |
+
2023-03-24 19:15:32,645 44k INFO ====> Epoch: 219, cost 14.88 s
|
| 303 |
+
2023-03-24 19:15:47,617 44k INFO ====> Epoch: 220, cost 14.97 s
|
| 304 |
+
2023-03-24 19:16:02,732 44k INFO ====> Epoch: 221, cost 15.11 s
|
| 305 |
+
2023-03-24 19:16:11,258 44k INFO Train Epoch: 222 [43%]
|
| 306 |
+
2023-03-24 19:16:11,259 44k INFO Losses: [2.0126640796661377, 2.6407198905944824, 16.501447677612305, 22.289186477661133, 0.433843731880188], step: 6200, lr: 9.727514011608789e-05
|
| 307 |
+
2023-03-24 19:16:18,504 44k INFO ====> Epoch: 222, cost 15.77 s
|
| 308 |
+
2023-03-24 19:16:33,328 44k INFO ====> Epoch: 223, cost 14.82 s
|
| 309 |
+
2023-03-24 19:16:48,335 44k INFO ====> Epoch: 224, cost 15.01 s
|
| 310 |
+
2023-03-24 19:17:05,146 44k INFO ====> Epoch: 225, cost 16.81 s
|
| 311 |
+
2023-03-24 19:17:20,324 44k INFO ====> Epoch: 226, cost 15.18 s
|
| 312 |
+
2023-03-24 19:17:35,399 44k INFO ====> Epoch: 227, cost 15.07 s
|
| 313 |
+
2023-03-24 19:17:50,487 44k INFO ====> Epoch: 228, cost 15.09 s
|
| 314 |
+
2023-03-24 19:18:00,538 44k INFO Train Epoch: 229 [57%]
|
| 315 |
+
2023-03-24 19:18:00,540 44k INFO Losses: [2.4733309745788574, 2.2934553623199463, 10.714012145996094, 20.784181594848633, 0.9193792343139648], step: 6400, lr: 9.719005628024282e-05
|
| 316 |
+
2023-03-24 19:18:05,406 44k INFO Saving model and optimizer state at iteration 229 to ./logs/44k/G_6400.pth
|
| 317 |
+
2023-03-24 19:18:06,751 44k INFO Saving model and optimizer state at iteration 229 to ./logs/44k/D_6400.pth
|
| 318 |
+
2023-03-24 19:18:12,417 44k INFO ====> Epoch: 229, cost 21.93 s
|
| 319 |
+
2023-03-24 19:18:27,279 44k INFO ====> Epoch: 230, cost 14.86 s
|
| 320 |
+
2023-03-24 19:18:42,506 44k INFO ====> Epoch: 231, cost 15.23 s
|
| 321 |
+
2023-03-24 19:18:57,429 44k INFO ====> Epoch: 232, cost 14.92 s
|
| 322 |
+
2023-03-24 19:19:12,291 44k INFO ====> Epoch: 233, cost 14.86 s
|
| 323 |
+
2023-03-24 19:19:27,176 44k INFO ====> Epoch: 234, cost 14.89 s
|
| 324 |
+
2023-03-24 19:19:42,112 44k INFO ====> Epoch: 235, cost 14.94 s
|
| 325 |
+
2023-03-24 19:19:53,840 44k INFO Train Epoch: 236 [71%]
|
| 326 |
+
2023-03-24 19:19:53,842 44k INFO Losses: [2.2217257022857666, 2.443718433380127, 9.876315116882324, 23.468894958496094, 0.6284939050674438], step: 6600, lr: 9.710504686484176e-05
|
| 327 |
+
2023-03-24 19:19:57,698 44k INFO ====> Epoch: 236, cost 15.59 s
|
| 328 |
+
2023-03-24 19:20:12,632 44k INFO ====> Epoch: 237, cost 14.93 s
|
| 329 |
+
2023-03-24 19:20:27,643 44k INFO ====> Epoch: 238, cost 15.01 s
|
| 330 |
+
2023-03-24 19:20:42,666 44k INFO ====> Epoch: 239, cost 15.02 s
|
| 331 |
+
2023-03-24 19:20:57,630 44k INFO ====> Epoch: 240, cost 14.96 s
|
| 332 |
+
2023-03-24 19:21:12,570 44k INFO ====> Epoch: 241, cost 14.94 s
|
| 333 |
+
2023-03-24 19:21:27,365 44k INFO ====> Epoch: 242, cost 14.79 s
|
| 334 |
+
2023-03-24 19:21:40,487 44k INFO Train Epoch: 243 [86%]
|
| 335 |
+
2023-03-24 19:21:40,488 44k INFO Losses: [2.344350576400757, 2.5619940757751465, 9.264684677124023, 17.457204818725586, 0.9792990684509277], step: 6800, lr: 9.702011180479129e-05
|
| 336 |
+
2023-03-24 19:21:42,810 44k INFO ====> Epoch: 243, cost 15.45 s
|
| 337 |
+
2023-03-24 19:21:57,984 44k INFO ====> Epoch: 244, cost 15.17 s
|
| 338 |
+
2023-03-24 19:22:13,342 44k INFO ====> Epoch: 245, cost 15.36 s
|
| 339 |
+
2023-03-24 19:22:30,194 44k INFO ====> Epoch: 246, cost 16.85 s
|
| 340 |
+
2023-03-24 19:22:45,246 44k INFO ====> Epoch: 247, cost 15.05 s
|
| 341 |
+
2023-03-24 19:23:00,158 44k INFO ====> Epoch: 248, cost 14.91 s
|
| 342 |
+
2023-03-24 19:23:15,630 44k INFO ====> Epoch: 249, cost 15.47 s
|
| 343 |
+
2023-03-24 19:23:30,865 44k INFO ====> Epoch: 250, cost 15.24 s
|
| 344 |
+
2023-03-24 19:23:34,804 44k INFO Train Epoch: 251 [0%]
|
| 345 |
+
2023-03-24 19:23:34,805 44k INFO Losses: [2.6337828636169434, 2.0596632957458496, 8.9087553024292, 17.73447608947754, 0.9439815878868103], step: 7000, lr: 9.692313412867544e-05
|
| 346 |
+
2023-03-24 19:23:46,586 44k INFO ====> Epoch: 251, cost 15.72 s
|
| 347 |
+
2023-03-24 19:24:01,932 44k INFO ====> Epoch: 252, cost 15.35 s
|
| 348 |
+
2023-03-24 19:24:16,697 44k INFO ====> Epoch: 253, cost 14.77 s
|
| 349 |
+
2023-03-24 19:24:31,744 44k INFO ====> Epoch: 254, cost 15.05 s
|
| 350 |
+
2023-03-24 19:24:46,613 44k INFO ====> Epoch: 255, cost 14.87 s
|
| 351 |
+
2023-03-24 19:25:01,535 44k INFO ====> Epoch: 256, cost 14.92 s
|
| 352 |
+
2023-03-24 19:25:16,411 44k INFO ====> Epoch: 257, cost 14.88 s
|
| 353 |
+
2023-03-24 19:25:21,922 44k INFO Train Epoch: 258 [14%]
|
| 354 |
+
2023-03-24 19:25:21,924 44k INFO Losses: [2.2161927223205566, 2.5849497318267822, 11.678217887878418, 19.252304077148438, 0.6322188973426819], step: 7200, lr: 9.683835818259144e-05
|
| 355 |
+
2023-03-24 19:25:26,865 44k INFO Saving model and optimizer state at iteration 258 to ./logs/44k/G_7200.pth
|
| 356 |
+
2023-03-24 19:25:28,030 44k INFO Saving model and optimizer state at iteration 258 to ./logs/44k/D_7200.pth
|
| 357 |
+
2023-03-24 19:25:38,272 44k INFO ====> Epoch: 258, cost 21.86 s
|
| 358 |
+
2023-03-24 19:25:53,319 44k INFO ====> Epoch: 259, cost 15.05 s
|
| 359 |
+
2023-03-24 19:26:08,204 44k INFO ====> Epoch: 260, cost 14.88 s
|
| 360 |
+
2023-03-24 19:26:23,726 44k INFO ====> Epoch: 261, cost 15.52 s
|
| 361 |
+
2023-03-24 19:26:38,714 44k INFO ====> Epoch: 262, cost 14.99 s
|
| 362 |
+
2023-03-24 19:26:53,504 44k INFO ====> Epoch: 263, cost 14.79 s
|
| 363 |
+
2023-03-24 19:27:08,499 44k INFO ====> Epoch: 264, cost 14.99 s
|
| 364 |
+
2023-03-24 19:27:15,332 44k INFO Train Epoch: 265 [29%]
|
| 365 |
+
2023-03-24 19:27:15,333 44k INFO Losses: [2.4472720623016357, 2.4915881156921387, 8.705090522766113, 18.240211486816406, 1.0029183626174927], step: 7400, lr: 9.675365638764893e-05
|
| 366 |
+
2023-03-24 19:27:24,059 44k INFO ====> Epoch: 265, cost 15.56 s
|
| 367 |
+
2023-03-24 19:27:39,028 44k INFO ====> Epoch: 266, cost 14.97 s
|
| 368 |
+
2023-03-24 19:27:53,939 44k INFO ====> Epoch: 267, cost 14.91 s
|
| 369 |
+
2023-03-24 19:28:08,716 44k INFO ====> Epoch: 268, cost 14.78 s
|
| 370 |
+
2023-03-24 19:28:23,528 44k INFO ====> Epoch: 269, cost 14.81 s
|
| 371 |
+
2023-03-24 19:28:38,338 44k INFO ====> Epoch: 270, cost 14.81 s
|
| 372 |
+
2023-03-24 19:28:53,337 44k INFO ====> Epoch: 271, cost 15.00 s
|
| 373 |
+
2023-03-24 19:29:01,906 44k INFO Train Epoch: 272 [43%]
|
| 374 |
+
2023-03-24 19:29:01,907 44k INFO Losses: [2.0804944038391113, 2.359567403793335, 10.937973022460938, 18.14143943786621, 0.7806757092475891], step: 7600, lr: 9.666902867899003e-05
|
| 375 |
+
2023-03-24 19:29:08,965 44k INFO ====> Epoch: 272, cost 15.63 s
|
| 376 |
+
2023-03-24 19:29:24,083 44k INFO ====> Epoch: 273, cost 15.12 s
|
| 377 |
+
2023-03-24 19:29:39,161 44k INFO ====> Epoch: 274, cost 15.08 s
|
| 378 |
+
2023-03-24 19:29:54,309 44k INFO ====> Epoch: 275, cost 15.15 s
|
| 379 |
+
2023-03-24 19:30:10,333 44k INFO ====> Epoch: 276, cost 16.02 s
|
| 380 |
+
2023-03-24 19:30:25,682 44k INFO ====> Epoch: 277, cost 15.35 s
|
| 381 |
+
2023-03-24 19:30:40,647 44k INFO ====> Epoch: 278, cost 14.97 s
|
| 382 |
+
2023-03-24 19:30:50,698 44k INFO Train Epoch: 279 [57%]
|
| 383 |
+
2023-03-24 19:30:50,699 44k INFO Losses: [2.1999661922454834, 2.28269624710083, 10.674137115478516, 19.649080276489258, 0.7525098323822021], step: 7800, lr: 9.658447499181352e-05
|
| 384 |
+
2023-03-24 19:30:56,292 44k INFO ====> Epoch: 279, cost 15.65 s
|
| 385 |
+
2023-03-24 19:31:11,237 44k INFO ====> Epoch: 280, cost 14.94 s
|
| 386 |
+
2023-03-24 19:31:26,358 44k INFO ====> Epoch: 281, cost 15.12 s
|
| 387 |
+
2023-03-24 19:31:43,106 44k INFO ====> Epoch: 282, cost 16.75 s
|
| 388 |
+
2023-03-24 19:31:58,007 44k INFO ====> Epoch: 283, cost 14.90 s
|
| 389 |
+
2023-03-24 19:32:13,921 44k INFO ====> Epoch: 284, cost 15.91 s
|
| 390 |
+
2023-03-24 19:32:28,943 44k INFO ====> Epoch: 285, cost 15.02 s
|
| 391 |
+
2023-03-24 19:32:40,760 44k INFO Train Epoch: 286 [71%]
|
| 392 |
+
2023-03-24 19:32:40,762 44k INFO Losses: [2.2327651977539062, 2.5476839542388916, 12.686872482299805, 24.49921417236328, 0.5556154251098633], step: 8000, lr: 9.649999526137489e-05
|
| 393 |
+
2023-03-24 19:32:45,305 44k INFO Saving model and optimizer state at iteration 286 to ./logs/44k/G_8000.pth
|
| 394 |
+
2023-03-24 19:32:46,623 44k INFO Saving model and optimizer state at iteration 286 to ./logs/44k/D_8000.pth
|
| 395 |
+
2023-03-24 19:32:50,602 44k INFO ====> Epoch: 286, cost 21.66 s
|
| 396 |
+
2023-03-24 19:33:05,551 44k INFO ====> Epoch: 287, cost 14.95 s
|
| 397 |
+
2023-03-24 19:33:20,539 44k INFO ====> Epoch: 288, cost 14.99 s
|
| 398 |
+
2023-03-24 19:33:35,651 44k INFO ====> Epoch: 289, cost 15.11 s
|
| 399 |
+
2023-03-24 19:33:50,602 44k INFO ====> Epoch: 290, cost 14.95 s
|
| 400 |
+
2023-03-24 19:34:06,275 44k INFO ====> Epoch: 291, cost 15.67 s
|
| 401 |
+
2023-03-24 19:34:22,007 44k INFO ====> Epoch: 292, cost 15.73 s
|
| 402 |
+
2023-03-24 19:34:35,151 44k INFO Train Epoch: 293 [86%]
|
| 403 |
+
2023-03-24 19:34:35,152 44k INFO Losses: [2.3618199825286865, 2.3041739463806152, 10.38995361328125, 22.8865966796875, 0.8566780090332031], step: 8200, lr: 9.641558942298625e-05
|
| 404 |
+
2023-03-24 19:34:37,465 44k INFO ====> Epoch: 293, cost 15.46 s
|
| 405 |
+
2023-03-24 19:34:52,452 44k INFO ====> Epoch: 294, cost 14.99 s
|
| 406 |
+
2023-03-24 19:35:07,703 44k INFO ====> Epoch: 295, cost 15.25 s
|
| 407 |
+
2023-03-24 19:35:22,584 44k INFO ====> Epoch: 296, cost 14.88 s
|
| 408 |
+
2023-03-24 19:35:37,541 44k INFO ====> Epoch: 297, cost 14.96 s
|
| 409 |
+
2023-03-24 19:35:52,483 44k INFO ====> Epoch: 298, cost 14.94 s
|
| 410 |
+
2023-03-24 19:36:07,363 44k INFO ====> Epoch: 299, cost 14.88 s
|
| 411 |
+
2023-03-24 19:36:22,355 44k INFO ====> Epoch: 300, cost 14.99 s
|
| 412 |
+
2023-03-24 19:36:26,061 44k INFO Train Epoch: 301 [0%]
|
| 413 |
+
2023-03-24 19:36:26,063 44k INFO Losses: [2.2221317291259766, 2.401972770690918, 8.367659568786621, 17.427534103393555, 0.6027737259864807], step: 8400, lr: 9.631921600483981e-05
|
| 414 |
+
2023-03-24 19:36:37,917 44k INFO ====> Epoch: 301, cost 15.56 s
|
| 415 |
+
2023-03-24 19:36:52,952 44k INFO ====> Epoch: 302, cost 15.03 s
|
| 416 |
+
2023-03-24 19:37:08,118 44k INFO ====> Epoch: 303, cost 15.17 s
|
| 417 |
+
2023-03-24 19:37:23,732 44k INFO ====> Epoch: 304, cost 15.61 s
|
| 418 |
+
2023-03-24 19:37:38,791 44k INFO ====> Epoch: 305, cost 15.06 s
|
| 419 |
+
2023-03-24 19:37:53,968 44k INFO ====> Epoch: 306, cost 15.18 s
|
| 420 |
+
2023-03-24 19:38:09,012 44k INFO ====> Epoch: 307, cost 15.04 s
|
| 421 |
+
2023-03-24 19:38:14,255 44k INFO Train Epoch: 308 [14%]
|
| 422 |
+
2023-03-24 19:38:14,256 44k INFO Losses: [2.4681646823883057, 2.6585354804992676, 11.240384101867676, 21.601070404052734, 0.6029421091079712], step: 8600, lr: 9.62349682889948e-05
|
| 423 |
+
2023-03-24 19:38:24,415 44k INFO ====> Epoch: 308, cost 15.40 s
|
| 424 |
+
2023-03-24 19:38:39,411 44k INFO ====> Epoch: 309, cost 15.00 s
|
| 425 |
+
2023-03-24 19:38:54,178 44k INFO ====> Epoch: 310, cost 14.77 s
|
| 426 |
+
2023-03-24 19:39:09,389 44k INFO ====> Epoch: 311, cost 15.21 s
|
| 427 |
+
2023-03-24 19:39:25,427 44k INFO ====> Epoch: 312, cost 16.04 s
|
| 428 |
+
2023-03-24 19:39:41,537 44k INFO ====> Epoch: 313, cost 16.11 s
|
| 429 |
+
2023-03-24 19:39:56,927 44k INFO ====> Epoch: 314, cost 15.39 s
|
| 430 |
+
2023-03-24 19:40:03,872 44k INFO Train Epoch: 315 [29%]
|
| 431 |
+
2023-03-24 19:40:03,873 44k INFO Losses: [2.458681583404541, 2.419649839401245, 11.656747817993164, 19.785127639770508, 0.44571802020072937], step: 8800, lr: 9.615079426226314e-05
|
| 432 |
+
2023-03-24 19:40:08,437 44k INFO Saving model and optimizer state at iteration 315 to ./logs/44k/G_8800.pth
|
| 433 |
+
2023-03-24 19:40:09,602 44k INFO Saving model and optimizer state at iteration 315 to ./logs/44k/D_8800.pth
|
| 434 |
+
2023-03-24 19:40:18,502 44k INFO ====> Epoch: 315, cost 21.57 s
|
| 435 |
+
2023-03-24 19:40:33,824 44k INFO ====> Epoch: 316, cost 15.32 s
|
| 436 |
+
2023-03-24 19:40:49,069 44k INFO ====> Epoch: 317, cost 15.24 s
|
| 437 |
+
2023-03-24 19:41:04,213 44k INFO ====> Epoch: 318, cost 15.14 s
|
| 438 |
+
2023-03-24 19:41:19,512 44k INFO ====> Epoch: 319, cost 15.30 s
|
| 439 |
+
2023-03-24 19:41:33,945 44k INFO ====> Epoch: 320, cost 14.43 s
|
| 440 |
+
2023-03-24 19:41:48,414 44k INFO ====> Epoch: 321, cost 14.47 s
|
| 441 |
+
2023-03-24 19:41:56,657 44k INFO Train Epoch: 322 [43%]
|
| 442 |
+
2023-03-24 19:41:56,658 44k INFO Losses: [2.6525814533233643, 2.5892791748046875, 9.840317726135254, 18.47782325744629, 0.5226555466651917], step: 9000, lr: 9.606669386019102e-05
|
| 443 |
+
2023-03-24 19:42:03,437 44k INFO ====> Epoch: 322, cost 15.02 s
|
| 444 |
+
2023-03-24 19:42:18,723 44k INFO ====> Epoch: 323, cost 15.29 s
|
| 445 |
+
2023-03-24 19:42:33,408 44k INFO ====> Epoch: 324, cost 14.68 s
|
| 446 |
+
2023-03-24 19:42:47,755 44k INFO ====> Epoch: 325, cost 14.35 s
|
| 447 |
+
2023-03-24 19:43:04,032 44k INFO ====> Epoch: 326, cost 16.28 s
|
| 448 |
+
2023-03-24 19:43:18,385 44k INFO ====> Epoch: 327, cost 14.35 s
|
| 449 |
+
2023-03-24 19:43:34,615 44k INFO ====> Epoch: 328, cost 16.23 s
|
| 450 |
+
2023-03-24 19:43:44,189 44k INFO Train Epoch: 329 [57%]
|
| 451 |
+
2023-03-24 19:43:44,191 44k INFO Losses: [2.311946153640747, 2.3908088207244873, 10.318487167358398, 20.419300079345703, 0.5420147776603699], step: 9200, lr: 9.5982667018381e-05
|
| 452 |
+
2023-03-24 19:43:49,497 44k INFO ====> Epoch: 329, cost 14.88 s
|
| 453 |
+
2023-03-24 19:44:03,799 44k INFO ====> Epoch: 330, cost 14.30 s
|
| 454 |
+
2023-03-24 19:44:18,860 44k INFO ====> Epoch: 331, cost 15.06 s
|
| 455 |
+
2023-03-24 19:44:33,888 44k INFO ====> Epoch: 332, cost 15.03 s
|
| 456 |
+
2023-03-24 19:44:49,738 44k INFO ====> Epoch: 333, cost 15.85 s
|
| 457 |
+
2023-03-24 19:45:04,196 44k INFO ====> Epoch: 334, cost 14.46 s
|
| 458 |
+
2023-03-24 19:45:18,582 44k INFO ====> Epoch: 335, cost 14.39 s
|
| 459 |
+
2023-03-24 19:45:29,760 44k INFO Train Epoch: 336 [71%]
|
| 460 |
+
2023-03-24 19:45:29,761 44k INFO Losses: [2.1289186477661133, 2.669475555419922, 12.912238121032715, 24.485408782958984, 0.7686963677406311], step: 9400, lr: 9.589871367249203e-05
|
| 461 |
+
2023-03-24 19:45:33,538 44k INFO ====> Epoch: 336, cost 14.96 s
|
| 462 |
+
2023-03-24 19:45:47,998 44k INFO ====> Epoch: 337, cost 14.46 s
|
| 463 |
+
2023-03-24 19:46:02,426 44k INFO ====> Epoch: 338, cost 14.43 s
|
| 464 |
+
2023-03-24 19:46:16,729 44k INFO ====> Epoch: 339, cost 14.30 s
|
| 465 |
+
2023-03-24 19:46:31,129 44k INFO ====> Epoch: 340, cost 14.40 s
|
| 466 |
+
2023-03-24 19:46:45,835 44k INFO ====> Epoch: 341, cost 14.71 s
|
| 467 |
+
2023-03-24 19:47:00,347 44k INFO ====> Epoch: 342, cost 14.51 s
|
| 468 |
+
2023-03-24 19:47:13,112 44k INFO Train Epoch: 343 [86%]
|
| 469 |
+
2023-03-24 19:47:13,114 44k INFO Losses: [2.662191867828369, 2.280200242996216, 5.206782817840576, 16.19651222229004, 1.028341293334961], step: 9600, lr: 9.581483375823925e-05
|
| 470 |
+
2023-03-24 19:47:17,473 44k INFO Saving model and optimizer state at iteration 343 to ./logs/44k/G_9600.pth
|
| 471 |
+
2023-03-24 19:47:18,606 44k INFO Saving model and optimizer state at iteration 343 to ./logs/44k/D_9600.pth
|
| 472 |
+
2023-03-24 19:47:20,890 44k INFO ====> Epoch: 343, cost 20.54 s
|
| 473 |
+
2023-03-24 19:47:35,430 44k INFO ====> Epoch: 344, cost 14.54 s
|
| 474 |
+
2023-03-24 19:47:51,623 44k INFO ====> Epoch: 345, cost 16.19 s
|
| 475 |
+
2023-03-24 19:48:06,057 44k INFO ====> Epoch: 346, cost 14.43 s
|
| 476 |
+
2023-03-24 19:48:20,464 44k INFO ====> Epoch: 347, cost 14.41 s
|
| 477 |
+
2023-03-24 19:48:34,775 44k INFO ====> Epoch: 348, cost 14.31 s
|
| 478 |
+
2023-03-24 19:48:49,433 44k INFO ====> Epoch: 349, cost 14.66 s
|
| 479 |
+
2023-03-24 19:49:03,842 44k INFO ====> Epoch: 350, cost 14.41 s
|
| 480 |
+
2023-03-24 19:49:07,240 44k INFO Train Epoch: 351 [0%]
|
| 481 |
+
2023-03-24 19:49:07,241 44k INFO Losses: [2.543008804321289, 2.433940887451172, 9.013635635375977, 19.387483596801758, 0.9613720178604126], step: 9800, lr: 9.571906083299264e-05
|
| 482 |
+
2023-03-24 19:49:18,951 44k INFO ====> Epoch: 351, cost 15.11 s
|
| 483 |
+
2023-03-24 19:49:33,480 44k INFO ====> Epoch: 352, cost 14.53 s
|
| 484 |
+
2023-03-24 19:49:49,878 44k INFO ====> Epoch: 353, cost 16.40 s
|
| 485 |
+
2023-03-24 19:50:04,253 44k INFO ====> Epoch: 354, cost 14.37 s
|
| 486 |
+
2023-03-24 19:50:18,739 44k INFO ====> Epoch: 355, cost 14.49 s
|
| 487 |
+
2023-03-24 19:50:33,554 44k INFO ====> Epoch: 356, cost 14.81 s
|
| 488 |
+
2023-03-24 19:50:48,130 44k INFO ====> Epoch: 357, cost 14.58 s
|
| 489 |
+
2023-03-24 19:50:53,068 44k INFO Train Epoch: 358 [14%]
|
| 490 |
+
2023-03-24 19:50:53,069 44k INFO Losses: [2.408367156982422, 2.6140289306640625, 10.188366889953613, 17.812231063842773, 0.4619278311729431], step: 10000, lr: 9.56353380560381e-05
|
| 491 |
+
2023-03-24 19:51:02,994 44k INFO ====> Epoch: 358, cost 14.86 s
|
| 492 |
+
2023-03-24 19:51:18,435 44k INFO ====> Epoch: 359, cost 15.44 s
|
| 493 |
+
2023-03-24 19:51:32,740 44k INFO ====> Epoch: 360, cost 14.30 s
|
| 494 |
+
2023-03-24 19:51:47,268 44k INFO ====> Epoch: 361, cost 14.53 s
|
| 495 |
+
2023-03-24 19:52:01,662 44k INFO ====> Epoch: 362, cost 14.39 s
|
| 496 |
+
2023-03-24 19:52:16,120 44k INFO ====> Epoch: 363, cost 14.46 s
|
| 497 |
+
2023-03-24 19:52:30,679 44k INFO ====> Epoch: 364, cost 14.56 s
|
| 498 |
+
2023-03-24 19:52:37,280 44k INFO Train Epoch: 365 [29%]
|
| 499 |
+
2023-03-24 19:52:37,282 44k INFO Losses: [2.3101587295532227, 2.3744843006134033, 8.811784744262695, 20.133663177490234, 0.5683051347732544], step: 10200, lr: 9.555168850904757e-05
|
| 500 |
+
2023-03-24 19:52:45,820 44k INFO ====> Epoch: 365, cost 15.14 s
|
| 501 |
+
2023-03-24 19:53:00,110 44k INFO ====> Epoch: 366, cost 14.29 s
|
| 502 |
+
2023-03-24 19:53:15,540 44k INFO ====> Epoch: 367, cost 15.43 s
|
| 503 |
+
2023-03-24 19:53:30,189 44k INFO ====> Epoch: 368, cost 14.65 s
|
| 504 |
+
2023-03-24 19:53:44,568 44k INFO ====> Epoch: 369, cost 14.38 s
|
| 505 |
+
2023-03-24 19:53:58,908 44k INFO ====> Epoch: 370, cost 14.34 s
|
| 506 |
+
2023-03-24 19:54:13,226 44k INFO ====> Epoch: 371, cost 14.32 s
|
| 507 |
+
2023-03-24 19:54:21,466 44k INFO Train Epoch: 372 [43%]
|
| 508 |
+
2023-03-24 19:54:21,467 44k INFO Losses: [2.4247655868530273, 2.617013454437256, 7.833914756774902, 15.284951210021973, 0.39104482531547546], step: 10400, lr: 9.546811212796888e-05
|
| 509 |
+
2023-03-24 19:54:25,622 44k INFO Saving model and optimizer state at iteration 372 to ./logs/44k/G_10400.pth
|
| 510 |
+
2023-03-24 19:54:26,937 44k INFO Saving model and optimizer state at iteration 372 to ./logs/44k/D_10400.pth
|
| 511 |
+
2023-03-24 19:54:33,855 44k INFO ====> Epoch: 372, cost 20.63 s
|
| 512 |
+
2023-03-24 19:54:48,244 44k INFO ====> Epoch: 373, cost 14.39 s
|
| 513 |
+
2023-03-24 19:55:04,755 44k INFO ====> Epoch: 374, cost 16.51 s
|
| 514 |
+
2023-03-24 19:55:20,840 44k INFO ====> Epoch: 375, cost 16.09 s
|
| 515 |
+
2023-03-24 19:55:35,407 44k INFO ====> Epoch: 376, cost 14.57 s
|
| 516 |
+
2023-03-24 19:55:49,810 44k INFO ====> Epoch: 377, cost 14.40 s
|
| 517 |
+
2023-03-24 19:56:04,253 44k INFO ====> Epoch: 378, cost 14.44 s
|
| 518 |
+
2023-03-24 19:56:14,055 44k INFO Train Epoch: 379 [57%]
|
| 519 |
+
2023-03-24 19:56:14,056 44k INFO Losses: [2.562399387359619, 2.2941503524780273, 9.387338638305664, 20.59601593017578, 0.7333645224571228], step: 10600, lr: 9.538460884880585e-05
|
| 520 |
+
2023-03-24 19:56:19,432 44k INFO ====> Epoch: 379, cost 15.18 s
|
| 521 |
+
2023-03-24 19:56:33,776 44k INFO ====> Epoch: 380, cost 14.34 s
|
| 522 |
+
2023-03-24 19:56:48,194 44k INFO ====> Epoch: 381, cost 14.42 s
|
| 523 |
+
2023-03-24 19:57:02,933 44k INFO ====> Epoch: 382, cost 14.74 s
|
| 524 |
+
2023-03-24 19:57:17,422 44k INFO ====> Epoch: 383, cost 14.49 s
|
| 525 |
+
2023-03-24 19:57:32,443 44k INFO ====> Epoch: 384, cost 15.02 s
|
| 526 |
+
2023-03-24 19:57:48,657 44k INFO ====> Epoch: 385, cost 16.21 s
|
| 527 |
+
2023-03-24 19:58:01,332 44k INFO Train Epoch: 386 [71%]
|
| 528 |
+
2023-03-24 19:58:01,333 44k INFO Losses: [2.2014482021331787, 2.4676952362060547, 14.335899353027344, 22.053157806396484, 0.9596916437149048], step: 10800, lr: 9.530117860761828e-05
|
| 529 |
+
2023-03-24 19:58:05,553 44k INFO ====> Epoch: 386, cost 16.90 s
|
| 530 |
+
2023-03-24 19:58:20,173 44k INFO ====> Epoch: 387, cost 14.62 s
|
| 531 |
+
2023-03-24 19:58:34,640 44k INFO ====> Epoch: 388, cost 14.47 s
|
| 532 |
+
2023-03-24 19:58:48,960 44k INFO ====> Epoch: 389, cost 14.32 s
|
| 533 |
+
2023-03-24 19:59:03,473 44k INFO ====> Epoch: 390, cost 14.51 s
|
| 534 |
+
2023-03-24 19:59:17,969 44k INFO ====> Epoch: 391, cost 14.50 s
|
| 535 |
+
2023-03-24 19:59:32,349 44k INFO ====> Epoch: 392, cost 14.38 s
|
| 536 |
+
2023-03-24 19:59:45,153 44k INFO Train Epoch: 393 [86%]
|
| 537 |
+
2023-03-24 19:59:45,154 44k INFO Losses: [2.2285561561584473, 2.763125419616699, 10.18687629699707, 21.496538162231445, 0.8079171180725098], step: 11000, lr: 9.52178213405219e-05
|
| 538 |
+
2023-03-24 19:59:47,353 44k INFO ====> Epoch: 393, cost 15.00 s
|
| 539 |
+
2023-03-24 20:00:01,806 44k INFO ====> Epoch: 394, cost 14.45 s
|
| 540 |
+
2023-03-24 20:00:16,188 44k INFO ====> Epoch: 395, cost 14.38 s
|
| 541 |
+
2023-03-24 20:00:31,245 44k INFO ====> Epoch: 396, cost 15.06 s
|
| 542 |
+
2023-03-24 20:00:46,297 44k INFO ====> Epoch: 397, cost 15.05 s
|
| 543 |
+
2023-03-24 20:01:01,571 44k INFO ====> Epoch: 398, cost 15.27 s
|
| 544 |
+
2023-03-24 20:01:15,927 44k INFO ====> Epoch: 399, cost 14.36 s
|
| 545 |
+
2023-03-24 20:01:30,224 44k INFO ====> Epoch: 400, cost 14.30 s
|
| 546 |
+
2023-03-24 20:01:33,793 44k INFO Train Epoch: 401 [0%]
|
| 547 |
+
2023-03-24 20:01:33,794 44k INFO Losses: [2.341519594192505, 2.1632702350616455, 10.988146781921387, 21.963043212890625, 0.8887688517570496], step: 11200, lr: 9.512264516656537e-05
|
| 548 |
+
2023-03-24 20:01:38,128 44k INFO Saving model and optimizer state at iteration 401 to ./logs/44k/G_11200.pth
|
| 549 |
+
2023-03-24 20:01:39,259 44k INFO Saving model and optimizer state at iteration 401 to ./logs/44k/D_11200.pth
|
| 550 |
+
2023-03-24 20:01:51,386 44k INFO ====> Epoch: 401, cost 21.16 s
|
| 551 |
+
2023-03-24 20:02:05,793 44k INFO ====> Epoch: 402, cost 14.41 s
|
| 552 |
+
2023-03-24 20:02:20,365 44k INFO ====> Epoch: 403, cost 14.57 s
|
| 553 |
+
2023-03-24 20:02:34,880 44k INFO ====> Epoch: 404, cost 14.51 s
|
| 554 |
+
2023-03-24 20:02:49,330 44k INFO ====> Epoch: 405, cost 14.45 s
|
| 555 |
+
2023-03-24 20:03:03,690 44k INFO ====> Epoch: 406, cost 14.36 s
|
| 556 |
+
2023-03-24 20:03:18,139 44k INFO ====> Epoch: 407, cost 14.45 s
|
| 557 |
+
2023-03-24 20:03:23,289 44k INFO Train Epoch: 408 [14%]
|
| 558 |
+
2023-03-24 20:03:23,290 44k INFO Losses: [2.1299514770507812, 2.5966148376464844, 12.680700302124023, 19.583683013916016, 0.708603024482727], step: 11400, lr: 9.503944405766085e-05
|
| 559 |
+
2023-03-24 20:03:33,270 44k INFO ====> Epoch: 408, cost 15.13 s
|
| 560 |
+
2023-03-24 20:03:47,805 44k INFO ====> Epoch: 409, cost 14.54 s
|
| 561 |
+
2023-03-24 20:04:02,389 44k INFO ====> Epoch: 410, cost 14.58 s
|
| 562 |
+
2023-03-24 20:04:16,789 44k INFO ====> Epoch: 411, cost 14.40 s
|
| 563 |
+
2023-03-24 20:04:31,317 44k INFO ====> Epoch: 412, cost 14.53 s
|
| 564 |
+
2023-03-24 20:04:45,740 44k INFO ====> Epoch: 413, cost 14.42 s
|
| 565 |
+
2023-03-24 20:05:00,307 44k INFO ====> Epoch: 414, cost 14.57 s
|
| 566 |
+
2023-03-24 20:05:07,625 44k INFO Train Epoch: 415 [29%]
|
| 567 |
+
2023-03-24 20:05:07,627 44k INFO Losses: [2.1400105953216553, 2.480010747909546, 14.507291793823242, 21.961090087890625, 1.0544546842575073], step: 11600, lr: 9.495631572243191e-05
|
| 568 |
+
2023-03-24 20:05:17,591 44k INFO ====> Epoch: 415, cost 17.28 s
|
| 569 |
+
2023-03-24 20:05:32,326 44k INFO ====> Epoch: 416, cost 14.73 s
|
| 570 |
+
2023-03-24 20:05:46,878 44k INFO ====> Epoch: 417, cost 14.55 s
|
| 571 |
+
2023-03-24 20:06:01,677 44k INFO ====> Epoch: 418, cost 14.80 s
|
| 572 |
+
2023-03-24 20:06:16,322 44k INFO ====> Epoch: 419, cost 14.64 s
|
| 573 |
+
2023-03-24 20:06:31,265 44k INFO ====> Epoch: 420, cost 14.94 s
|
| 574 |
+
2023-03-24 20:06:46,256 44k INFO ====> Epoch: 421, cost 14.99 s
|
| 575 |
+
2023-03-24 20:06:54,496 44k INFO Train Epoch: 422 [43%]
|
| 576 |
+
2023-03-24 20:06:54,498 44k INFO Losses: [2.521395444869995, 2.023440361022949, 10.221491813659668, 16.555747985839844, 0.37890228629112244], step: 11800, lr: 9.487326009722552e-05
|
| 577 |
+
2023-03-24 20:07:01,372 44k INFO ====> Epoch: 422, cost 15.12 s
|
| 578 |
+
2023-03-24 20:07:17,538 44k INFO ====> Epoch: 423, cost 16.17 s
|
| 579 |
+
2023-03-24 20:07:32,137 44k INFO ====> Epoch: 424, cost 14.60 s
|
| 580 |
+
2023-03-24 20:07:46,367 44k INFO ====> Epoch: 425, cost 14.23 s
|
| 581 |
+
2023-03-24 20:08:00,897 44k INFO ====> Epoch: 426, cost 14.53 s
|
| 582 |
+
2023-03-24 20:08:16,044 44k INFO ====> Epoch: 427, cost 15.15 s
|
| 583 |
+
2023-03-24 20:08:30,473 44k INFO ====> Epoch: 428, cost 14.43 s
|
| 584 |
+
2023-03-24 20:08:41,356 44k INFO Train Epoch: 429 [57%]
|
| 585 |
+
2023-03-24 20:08:41,357 44k INFO Losses: [2.484126329421997, 2.4067256450653076, 9.9367036819458, 19.814332962036133, 0.9535115361213684], step: 12000, lr: 9.479027711844423e-05
|
| 586 |
+
2023-03-24 20:08:45,857 44k INFO Saving model and optimizer state at iteration 429 to ./logs/44k/G_12000.pth
|
| 587 |
+
2023-03-24 20:08:46,998 44k INFO Saving model and optimizer state at iteration 429 to ./logs/44k/D_12000.pth
|
| 588 |
+
2023-03-24 20:08:52,322 44k INFO ====> Epoch: 429, cost 21.85 s
|
| 589 |
+
2023-03-24 20:09:06,682 44k INFO ====> Epoch: 430, cost 14.36 s
|
| 590 |
+
2023-03-24 20:09:21,042 44k INFO ====> Epoch: 431, cost 14.36 s
|
| 591 |
+
2023-03-24 20:09:35,541 44k INFO ====> Epoch: 432, cost 14.50 s
|
| 592 |
+
2023-03-24 20:09:50,036 44k INFO ====> Epoch: 433, cost 14.50 s
|
| 593 |
+
2023-03-24 20:10:04,556 44k INFO ====> Epoch: 434, cost 14.52 s
|
| 594 |
+
2023-03-24 20:10:18,950 44k INFO ====> Epoch: 435, cost 14.39 s
|
| 595 |
+
2023-03-24 20:10:30,293 44k INFO Train Epoch: 436 [71%]
|
| 596 |
+
2023-03-24 20:10:30,294 44k INFO Losses: [2.1555025577545166, 2.3229312896728516, 7.884840488433838, 16.179161071777344, 0.5915942192077637], step: 12200, lr: 9.470736672254626e-05
|
| 597 |
+
2023-03-24 20:10:34,099 44k INFO ====> Epoch: 436, cost 15.15 s
|
| 598 |
+
2023-03-24 20:10:48,609 44k INFO ====> Epoch: 437, cost 14.51 s
|
| 599 |
+
2023-03-24 20:11:02,959 44k INFO ====> Epoch: 438, cost 14.35 s
|
| 600 |
+
2023-03-24 20:11:17,503 44k INFO ====> Epoch: 439, cost 14.54 s
|
| 601 |
+
2023-03-24 20:11:32,290 44k INFO ====> Epoch: 440, cost 14.79 s
|
| 602 |
+
2023-03-24 20:11:47,198 44k INFO ====> Epoch: 441, cost 14.91 s
|
| 603 |
+
2023-03-24 20:12:01,679 44k INFO ====> Epoch: 442, cost 14.48 s
|
| 604 |
+
2023-03-24 20:12:14,509 44k INFO Train Epoch: 443 [86%]
|
| 605 |
+
2023-03-24 20:12:14,510 44k INFO Losses: [2.2263095378875732, 2.628852367401123, 10.052955627441406, 17.378814697265625, 0.6211011409759521], step: 12400, lr: 9.46245288460454e-05
|
| 606 |
+
2023-03-24 20:12:16,691 44k INFO ====> Epoch: 443, cost 15.01 s
|
| 607 |
+
2023-03-24 20:12:31,424 44k INFO ====> Epoch: 444, cost 14.73 s
|
| 608 |
+
2023-03-24 20:12:45,819 44k INFO ====> Epoch: 445, cost 14.39 s
|
| 609 |
+
2023-03-24 20:13:00,366 44k INFO ====> Epoch: 446, cost 14.55 s
|
| 610 |
+
2023-03-24 20:13:15,743 44k INFO ====> Epoch: 447, cost 15.38 s
|
| 611 |
+
2023-03-24 20:13:30,449 44k INFO ====> Epoch: 448, cost 14.71 s
|
| 612 |
+
2023-03-24 20:13:45,006 44k INFO ====> Epoch: 449, cost 14.56 s
|
| 613 |
+
2023-03-24 20:13:59,691 44k INFO ====> Epoch: 450, cost 14.69 s
|
| 614 |
+
2023-03-24 20:14:03,302 44k INFO Train Epoch: 451 [0%]
|
| 615 |
+
2023-03-24 20:14:03,304 44k INFO Losses: [2.2450530529022217, 2.55692720413208, 9.89914321899414, 16.104248046875, 0.7876774668693542], step: 12600, lr: 9.452994570508276e-05
|
| 616 |
+
2023-03-24 20:14:15,011 44k INFO ====> Epoch: 451, cost 15.32 s
|
| 617 |
+
2023-03-24 20:14:29,690 44k INFO ====> Epoch: 452, cost 14.68 s
|
| 618 |
+
2023-03-24 20:14:44,367 44k INFO ====> Epoch: 453, cost 14.68 s
|
| 619 |
+
2023-03-24 20:14:58,869 44k INFO ====> Epoch: 454, cost 14.50 s
|
| 620 |
+
2023-03-24 20:15:13,461 44k INFO ====> Epoch: 455, cost 14.59 s
|
| 621 |
+
2023-03-24 20:15:27,995 44k INFO ====> Epoch: 456, cost 14.53 s
|
| 622 |
+
2023-03-24 20:15:42,342 44k INFO ====> Epoch: 457, cost 14.35 s
|
| 623 |
+
2023-03-24 20:15:47,531 44k INFO Train Epoch: 458 [14%]
|
| 624 |
+
2023-03-24 20:15:47,532 44k INFO Losses: [2.1057448387145996, 2.3418726921081543, 14.067486763000488, 20.251667022705078, 0.6368907690048218], step: 12800, lr: 9.4447263013768e-05
|
| 625 |
+
2023-03-24 20:15:51,836 44k INFO Saving model and optimizer state at iteration 458 to ./logs/44k/G_12800.pth
|
| 626 |
+
2023-03-24 20:15:53,108 44k INFO Saving model and optimizer state at iteration 458 to ./logs/44k/D_12800.pth
|
| 627 |
+
2023-03-24 20:16:03,015 44k INFO ====> Epoch: 458, cost 20.67 s
|
| 628 |
+
2023-03-24 20:16:17,577 44k INFO ====> Epoch: 459, cost 14.56 s
|
| 629 |
+
2023-03-24 20:16:31,875 44k INFO ====> Epoch: 460, cost 14.30 s
|
| 630 |
+
2023-03-24 20:16:46,440 44k INFO ====> Epoch: 461, cost 14.57 s
|
| 631 |
+
2023-03-24 20:17:00,950 44k INFO ====> Epoch: 462, cost 14.51 s
|
| 632 |
+
2023-03-24 20:17:15,290 44k INFO ====> Epoch: 463, cost 14.34 s
|
| 633 |
+
2023-03-24 20:17:29,633 44k INFO ====> Epoch: 464, cost 14.34 s
|
| 634 |
+
2023-03-24 20:17:36,274 44k INFO Train Epoch: 465 [29%]
|
| 635 |
+
2023-03-24 20:17:36,275 44k INFO Losses: [2.522308349609375, 2.7243692874908447, 11.208446502685547, 20.988624572753906, 0.5713735818862915], step: 13000, lr: 9.436465264268356e-05
|
| 636 |
+
2023-03-24 20:17:44,555 44k INFO ====> Epoch: 465, cost 14.92 s
|
| 637 |
+
2023-03-24 20:17:59,035 44k INFO ====> Epoch: 466, cost 14.48 s
|
| 638 |
+
2023-03-24 20:18:13,574 44k INFO ====> Epoch: 467, cost 14.54 s
|
| 639 |
+
2023-03-24 20:18:28,138 44k INFO ====> Epoch: 468, cost 14.56 s
|
| 640 |
+
2023-03-24 20:18:42,637 44k INFO ====> Epoch: 469, cost 14.50 s
|
| 641 |
+
2023-03-24 20:18:57,185 44k INFO ====> Epoch: 470, cost 14.55 s
|
| 642 |
+
2023-03-24 20:19:11,671 44k INFO ====> Epoch: 471, cost 14.49 s
|
| 643 |
+
2023-03-24 20:19:19,879 44k INFO Train Epoch: 472 [43%]
|
| 644 |
+
2023-03-24 20:19:19,880 44k INFO Losses: [2.4117305278778076, 2.0530574321746826, 8.800864219665527, 12.694659233093262, 0.7633107900619507], step: 13200, lr: 9.428211452857292e-05
|
| 645 |
+
2023-03-24 20:19:26,821 44k INFO ====> Epoch: 472, cost 15.15 s
|
| 646 |
+
2023-03-24 20:19:41,477 44k INFO ====> Epoch: 473, cost 14.66 s
|
| 647 |
+
2023-03-24 20:19:57,261 44k INFO ====> Epoch: 474, cost 15.78 s
|
| 648 |
+
2023-03-24 20:20:11,794 44k INFO ====> Epoch: 475, cost 14.53 s
|
| 649 |
+
2023-03-24 20:20:26,224 44k INFO ====> Epoch: 476, cost 14.43 s
|
| 650 |
+
2023-03-24 20:20:41,168 44k INFO ====> Epoch: 477, cost 14.94 s
|
| 651 |
+
2023-03-24 20:20:55,667 44k INFO ====> Epoch: 478, cost 14.50 s
|
| 652 |
+
2023-03-24 20:21:05,331 44k INFO Train Epoch: 479 [57%]
|
| 653 |
+
2023-03-24 20:21:05,332 44k INFO Losses: [2.286837100982666, 2.894473075866699, 11.783750534057617, 20.42172622680664, 0.7042444944381714], step: 13400, lr: 9.419964860823498e-05
|
| 654 |
+
2023-03-24 20:21:10,655 44k INFO ====> Epoch: 479, cost 14.99 s
|
| 655 |
+
2023-03-24 20:21:25,167 44k INFO ====> Epoch: 480, cost 14.51 s
|
| 656 |
+
2023-03-24 20:21:39,608 44k INFO ====> Epoch: 481, cost 14.44 s
|
| 657 |
+
2023-03-24 20:21:53,961 44k INFO ====> Epoch: 482, cost 14.35 s
|
| 658 |
+
2023-03-24 20:22:09,678 44k INFO ====> Epoch: 483, cost 15.72 s
|
| 659 |
+
2023-03-24 20:22:24,111 44k INFO ====> Epoch: 484, cost 14.43 s
|
| 660 |
+
2023-03-24 20:22:38,527 44k INFO ====> Epoch: 485, cost 14.42 s
|
| 661 |
+
2023-03-24 20:22:49,768 44k INFO Train Epoch: 486 [71%]
|
| 662 |
+
2023-03-24 20:22:49,769 44k INFO Losses: [2.4765379428863525, 2.358840227127075, 9.544150352478027, 17.993436813354492, 0.4492684304714203], step: 13600, lr: 9.411725481852385e-05
|
| 663 |
+
2023-03-24 20:22:54,040 44k INFO Saving model and optimizer state at iteration 486 to ./logs/44k/G_13600.pth
|
| 664 |
+
2023-03-24 20:22:55,162 44k INFO Saving model and optimizer state at iteration 486 to ./logs/44k/D_13600.pth
|
| 665 |
+
2023-03-24 20:22:58,945 44k INFO ====> Epoch: 486, cost 20.42 s
|
| 666 |
+
2023-03-24 20:23:13,211 44k INFO ====> Epoch: 487, cost 14.27 s
|
| 667 |
+
2023-03-24 20:23:29,478 44k INFO ====> Epoch: 488, cost 16.27 s
|
| 668 |
+
2023-03-24 20:23:43,887 44k INFO ====> Epoch: 489, cost 14.41 s
|
| 669 |
+
2023-03-24 20:23:58,565 44k INFO ====> Epoch: 490, cost 14.68 s
|
| 670 |
+
2023-03-24 20:24:13,027 44k INFO ====> Epoch: 491, cost 14.46 s
|
| 671 |
+
2023-03-24 20:24:27,298 44k INFO ====> Epoch: 492, cost 14.27 s
|
| 672 |
+
2023-03-24 20:24:40,013 44k INFO Train Epoch: 493 [86%]
|
| 673 |
+
2023-03-24 20:24:40,014 44k INFO Losses: [2.4640090465545654, 2.574191093444824, 11.566976547241211, 20.15541648864746, 0.44430384039878845], step: 13800, lr: 9.403493309634886e-05
|
| 674 |
+
2023-03-24 20:24:42,220 44k INFO ====> Epoch: 493, cost 14.92 s
|
| 675 |
+
2023-03-24 20:24:56,688 44k INFO ====> Epoch: 494, cost 14.47 s
|
| 676 |
+
2023-03-24 20:25:11,153 44k INFO ====> Epoch: 495, cost 14.47 s
|
| 677 |
+
2023-03-24 20:25:25,566 44k INFO ====> Epoch: 496, cost 14.41 s
|
| 678 |
+
2023-03-24 20:25:39,817 44k INFO ====> Epoch: 497, cost 14.25 s
|
| 679 |
+
2023-03-24 20:25:54,155 44k INFO ====> Epoch: 498, cost 14.34 s
|
| 680 |
+
2023-03-24 20:26:08,438 44k INFO ====> Epoch: 499, cost 14.28 s
|
| 681 |
+
2023-03-24 20:26:22,930 44k INFO ====> Epoch: 500, cost 14.49 s
|
| 682 |
+
2023-03-24 20:26:26,528 44k INFO Train Epoch: 501 [0%]
|
| 683 |
+
2023-03-24 20:26:26,529 44k INFO Losses: [2.4643354415893555, 2.21096134185791, 12.568891525268555, 17.2219181060791, 0.6204159259796143], step: 14000, lr: 9.394093929325224e-05
|
| 684 |
+
2023-03-24 20:26:38,142 44k INFO ====> Epoch: 501, cost 15.21 s
|
| 685 |
+
2023-03-24 20:26:52,567 44k INFO ====> Epoch: 502, cost 14.43 s
|
| 686 |
+
2023-03-24 20:27:07,141 44k INFO ====> Epoch: 503, cost 14.57 s
|
| 687 |
+
2023-03-24 20:27:21,499 44k INFO ====> Epoch: 504, cost 14.36 s
|
| 688 |
+
2023-03-24 20:27:35,990 44k INFO ====> Epoch: 505, cost 14.49 s
|
| 689 |
+
2023-03-24 20:27:50,442 44k INFO ====> Epoch: 506, cost 14.45 s
|
| 690 |
+
2023-03-24 20:28:04,909 44k INFO ====> Epoch: 507, cost 14.47 s
|
| 691 |
+
2023-03-24 20:28:09,895 44k INFO Train Epoch: 508 [14%]
|
| 692 |
+
2023-03-24 20:28:09,896 44k INFO Losses: [2.190594434738159, 2.6167070865631104, 19.042556762695312, 21.837785720825195, 0.6685327291488647], step: 14200, lr: 9.385877178932038e-05
|
| 693 |
+
2023-03-24 20:28:19,888 44k INFO ====> Epoch: 508, cost 14.98 s
|
| 694 |
+
2023-03-24 20:28:34,282 44k INFO ====> Epoch: 509, cost 14.39 s
|
| 695 |
+
2023-03-24 20:28:48,842 44k INFO ====> Epoch: 510, cost 14.56 s
|
| 696 |
+
2023-03-24 20:29:03,258 44k INFO ====> Epoch: 511, cost 14.42 s
|
| 697 |
+
2023-03-24 20:29:17,509 44k INFO ====> Epoch: 512, cost 14.25 s
|
| 698 |
+
2023-03-24 20:29:32,003 44k INFO ====> Epoch: 513, cost 14.49 s
|
| 699 |
+
2023-03-24 20:29:46,533 44k INFO ====> Epoch: 514, cost 14.53 s
|
| 700 |
+
2023-03-24 20:29:53,094 44k INFO Train Epoch: 515 [29%]
|
| 701 |
+
2023-03-24 20:29:53,095 44k INFO Losses: [2.2295584678649902, 2.565598964691162, 12.276839256286621, 20.938312530517578, 0.5135537385940552], step: 14400, lr: 9.377667615499888e-05
|
| 702 |
+
2023-03-24 20:29:57,584 44k INFO Saving model and optimizer state at iteration 515 to ./logs/44k/G_14400.pth
|
| 703 |
+
2023-03-24 20:29:58,746 44k INFO Saving model and optimizer state at iteration 515 to ./logs/44k/D_14400.pth
|
| 704 |
+
2023-03-24 20:30:07,099 44k INFO ====> Epoch: 515, cost 20.57 s
|
| 705 |
+
2023-03-24 20:30:21,493 44k INFO ====> Epoch: 516, cost 14.39 s
|
| 706 |
+
2023-03-24 20:30:35,841 44k INFO ====> Epoch: 517, cost 14.35 s
|
| 707 |
+
2023-03-24 20:30:50,339 44k INFO ====> Epoch: 518, cost 14.50 s
|
| 708 |
+
2023-03-24 20:31:04,644 44k INFO ====> Epoch: 519, cost 14.31 s
|
| 709 |
+
2023-03-24 20:31:19,043 44k INFO ====> Epoch: 520, cost 14.40 s
|
| 710 |
+
2023-03-24 20:31:33,479 44k INFO ====> Epoch: 521, cost 14.44 s
|
| 711 |
+
2023-03-24 20:31:41,552 44k INFO Train Epoch: 522 [43%]
|
| 712 |
+
2023-03-24 20:31:41,553 44k INFO Losses: [2.4139742851257324, 2.372035264968872, 8.859540939331055, 14.895163536071777, 0.356738805770874], step: 14600, lr: 9.36946523274254e-05
|
| 713 |
+
2023-03-24 20:31:48,526 44k INFO ====> Epoch: 522, cost 15.05 s
|
| 714 |
+
2023-03-24 20:32:03,063 44k INFO ====> Epoch: 523, cost 14.54 s
|
| 715 |
+
2023-03-24 20:32:17,498 44k INFO ====> Epoch: 524, cost 14.43 s
|
| 716 |
+
2023-03-24 20:32:31,840 44k INFO ====> Epoch: 525, cost 14.34 s
|
| 717 |
+
2023-03-24 20:32:46,313 44k INFO ====> Epoch: 526, cost 14.47 s
|
| 718 |
+
2023-03-24 20:33:02,532 44k INFO ====> Epoch: 527, cost 16.22 s
|
| 719 |
+
2023-03-24 20:33:17,116 44k INFO ====> Epoch: 528, cost 14.58 s
|
| 720 |
+
2023-03-24 20:33:26,788 44k INFO Train Epoch: 529 [57%]
|
| 721 |
+
2023-03-24 20:33:26,789 44k INFO Losses: [2.420297861099243, 2.644667148590088, 11.601181983947754, 20.18661880493164, 0.6209089159965515], step: 14800, lr: 9.361270024379255e-05
|
| 722 |
+
2023-03-24 20:33:32,068 44k INFO ====> Epoch: 529, cost 14.95 s
|
| 723 |
+
2023-03-24 20:33:46,379 44k INFO ====> Epoch: 530, cost 14.31 s
|
| 724 |
+
2023-03-24 20:34:00,960 44k INFO ====> Epoch: 531, cost 14.58 s
|
| 725 |
+
2023-03-24 20:34:15,291 44k INFO ====> Epoch: 532, cost 14.33 s
|
| 726 |
+
2023-03-24 20:34:29,582 44k INFO ====> Epoch: 533, cost 14.29 s
|
| 727 |
+
2023-03-24 20:34:44,315 44k INFO ====> Epoch: 534, cost 14.73 s
|
| 728 |
+
2023-03-24 20:34:58,766 44k INFO ====> Epoch: 535, cost 14.45 s
|
| 729 |
+
2023-03-24 20:35:10,040 44k INFO Train Epoch: 536 [71%]
|
| 730 |
+
2023-03-24 20:35:10,042 44k INFO Losses: [2.2677812576293945, 2.56514310836792, 10.31601333618164, 19.94939613342285, 0.9051366448402405], step: 15000, lr: 9.353081984134796e-05
|
| 731 |
+
2023-03-24 20:35:13,887 44k INFO ====> Epoch: 536, cost 15.12 s
|
| 732 |
+
2023-03-24 20:35:28,286 44k INFO ====> Epoch: 537, cost 14.40 s
|
| 733 |
+
2023-03-24 20:35:42,698 44k INFO ====> Epoch: 538, cost 14.41 s
|
| 734 |
+
2023-03-24 20:35:57,094 44k INFO ====> Epoch: 539, cost 14.40 s
|
| 735 |
+
2023-03-24 20:36:11,458 44k INFO ====> Epoch: 540, cost 14.36 s
|
| 736 |
+
2023-03-24 20:36:25,885 44k INFO ====> Epoch: 541, cost 14.43 s
|
| 737 |
+
2023-03-24 20:36:40,508 44k INFO ====> Epoch: 542, cost 14.62 s
|
| 738 |
+
2023-03-24 20:36:53,595 44k INFO Train Epoch: 543 [86%]
|
| 739 |
+
2023-03-24 20:36:53,596 44k INFO Losses: [2.5094823837280273, 2.4331843852996826, 6.562588691711426, 18.474308013916016, 0.5865221619606018], step: 15200, lr: 9.344901105739411e-05
|
| 740 |
+
2023-03-24 20:36:57,995 44k INFO Saving model and optimizer state at iteration 543 to ./logs/44k/G_15200.pth
|
| 741 |
+
2023-03-24 20:36:59,235 44k INFO Saving model and optimizer state at iteration 543 to ./logs/44k/D_15200.pth
|
| 742 |
+
2023-03-24 20:37:01,473 44k INFO ====> Epoch: 543, cost 20.96 s
|
| 743 |
+
2023-03-24 20:37:15,632 44k INFO ====> Epoch: 544, cost 14.16 s
|
| 744 |
+
2023-03-24 20:37:29,945 44k INFO ====> Epoch: 545, cost 14.31 s
|
| 745 |
+
2023-03-24 20:37:44,127 44k INFO ====> Epoch: 546, cost 14.18 s
|
| 746 |
+
2023-03-24 20:37:58,227 44k INFO ====> Epoch: 547, cost 14.10 s
|
| 747 |
+
2023-03-24 20:38:12,354 44k INFO ====> Epoch: 548, cost 14.13 s
|
| 748 |
+
2023-03-24 20:38:26,594 44k INFO ====> Epoch: 549, cost 14.24 s
|
| 749 |
+
2023-03-24 20:38:41,002 44k INFO ====> Epoch: 550, cost 14.41 s
|
| 750 |
+
2023-03-24 20:38:44,471 44k INFO Train Epoch: 551 [0%]
|
| 751 |
+
2023-03-24 20:38:44,473 44k INFO Losses: [2.3259267807006836, 2.6285746097564697, 11.082757949829102, 20.842479705810547, 0.6062003374099731], step: 15400, lr: 9.335560292005964e-05
|
| 752 |
+
2023-03-24 20:38:55,883 44k INFO ====> Epoch: 551, cost 14.88 s
|
| 753 |
+
2023-03-24 20:39:10,186 44k INFO ====> Epoch: 552, cost 14.30 s
|
| 754 |
+
2023-03-24 20:39:24,305 44k INFO ====> Epoch: 553, cost 14.12 s
|
| 755 |
+
2023-03-24 20:39:38,479 44k INFO ====> Epoch: 554, cost 14.17 s
|
| 756 |
+
2023-03-24 20:39:52,592 44k INFO ====> Epoch: 555, cost 14.11 s
|
| 757 |
+
2023-03-24 20:40:08,191 44k INFO ====> Epoch: 556, cost 15.60 s
|
| 758 |
+
2023-03-24 20:40:22,344 44k INFO ====> Epoch: 557, cost 14.15 s
|
| 759 |
+
2023-03-24 20:40:27,327 44k INFO Train Epoch: 558 [14%]
|
| 760 |
+
2023-03-24 20:40:27,328 44k INFO Losses: [2.343079090118408, 2.480746030807495, 14.084281921386719, 19.00463104248047, 0.4086940586566925], step: 15600, lr: 9.327394739343082e-05
|
| 761 |
+
2023-03-24 20:40:37,225 44k INFO ====> Epoch: 558, cost 14.88 s
|
| 762 |
+
2023-03-24 20:40:51,779 44k INFO ====> Epoch: 559, cost 14.55 s
|
| 763 |
+
2023-03-24 20:41:05,973 44k INFO ====> Epoch: 560, cost 14.19 s
|
| 764 |
+
2023-03-24 20:41:20,119 44k INFO ====> Epoch: 561, cost 14.15 s
|
| 765 |
+
2023-03-24 20:41:34,410 44k INFO ====> Epoch: 562, cost 14.29 s
|
| 766 |
+
2023-03-24 20:41:48,570 44k INFO ====> Epoch: 563, cost 14.16 s
|
| 767 |
+
2023-03-24 20:42:02,932 44k INFO ====> Epoch: 564, cost 14.36 s
|
| 768 |
+
2023-03-24 20:42:09,543 44k INFO Train Epoch: 565 [29%]
|
| 769 |
+
2023-03-24 20:42:09,545 44k INFO Losses: [2.1794023513793945, 2.599058151245117, 14.124211311340332, 20.959125518798828, 0.5926719307899475], step: 15800, lr: 9.319236328860017e-05
|
| 770 |
+
2023-03-24 20:42:17,885 44k INFO ====> Epoch: 565, cost 14.95 s
|
| 771 |
+
2023-03-24 20:42:32,242 44k INFO ====> Epoch: 566, cost 14.36 s
|
| 772 |
+
2023-03-24 20:42:46,728 44k INFO ====> Epoch: 567, cost 14.49 s
|