Japanese
zwa73 commited on
Commit
983c6fa
·
1 Parent(s): 1807e40
so-vits-svc-4.0/20230325-Dreizehn-44k/best/D_13600.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba7a99daaf96c141fa25a929f2fb5219925113e628481dd415b19704c1d72b50
3
+ size 561098185
so-vits-svc-4.0/20230325-Dreizehn-44k/best/G_13600.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3080cc0577b0947cd0c3b28fb99ef315efda3d78f8e380e2f0f65bf6607d6c6d
3
+ size 542789405
so-vits-svc-4.0/20230325-Dreizehn-44k/best/config.json ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "train": {
3
+ "log_interval": 200,
4
+ "eval_interval": 800,
5
+ "seed": 1234,
6
+ "epochs": 10000,
7
+ "learning_rate": 0.0001,
8
+ "betas": [
9
+ 0.8,
10
+ 0.99
11
+ ],
12
+ "eps": 1e-09,
13
+ "batch_size": 6,
14
+ "fp16_run": false,
15
+ "lr_decay": 0.999875,
16
+ "segment_size": 10240,
17
+ "init_lr_ratio": 1,
18
+ "warmup_epochs": 0,
19
+ "c_mel": 45,
20
+ "c_kl": 1.0,
21
+ "use_sr": true,
22
+ "max_speclen": 512,
23
+ "port": "8001",
24
+ "keep_ckpts": 99
25
+ },
26
+ "data": {
27
+ "training_files": "filelists/train.txt",
28
+ "validation_files": "filelists/val.txt",
29
+ "max_wav_value": 32768.0,
30
+ "sampling_rate": 44100,
31
+ "filter_length": 2048,
32
+ "hop_length": 512,
33
+ "win_length": 2048,
34
+ "n_mel_channels": 80,
35
+ "mel_fmin": 0.0,
36
+ "mel_fmax": 22050
37
+ },
38
+ "model": {
39
+ "inter_channels": 192,
40
+ "hidden_channels": 192,
41
+ "filter_channels": 768,
42
+ "n_heads": 2,
43
+ "n_layers": 6,
44
+ "kernel_size": 3,
45
+ "p_dropout": 0.1,
46
+ "resblock": "1",
47
+ "resblock_kernel_sizes": [
48
+ 3,
49
+ 7,
50
+ 11
51
+ ],
52
+ "resblock_dilation_sizes": [
53
+ [
54
+ 1,
55
+ 3,
56
+ 5
57
+ ],
58
+ [
59
+ 1,
60
+ 3,
61
+ 5
62
+ ],
63
+ [
64
+ 1,
65
+ 3,
66
+ 5
67
+ ]
68
+ ],
69
+ "upsample_rates": [
70
+ 8,
71
+ 8,
72
+ 2,
73
+ 2,
74
+ 2
75
+ ],
76
+ "upsample_initial_channel": 512,
77
+ "upsample_kernel_sizes": [
78
+ 16,
79
+ 16,
80
+ 4,
81
+ 4,
82
+ 4
83
+ ],
84
+ "n_layers_q": 3,
85
+ "use_spectral_norm": false,
86
+ "gin_channels": 256,
87
+ "ssl_dim": 256,
88
+ "n_speakers": 200
89
+ },
90
+ "spk": {
91
+ "Dreizehn": 0
92
+ }
93
+ }
so-vits-svc-4.0/20230325-Dreizehn-44k/config.json ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "train": {
3
+ "log_interval": 200,
4
+ "eval_interval": 800,
5
+ "seed": 1234,
6
+ "epochs": 10000,
7
+ "learning_rate": 0.0001,
8
+ "betas": [
9
+ 0.8,
10
+ 0.99
11
+ ],
12
+ "eps": 1e-09,
13
+ "batch_size": 6,
14
+ "fp16_run": false,
15
+ "lr_decay": 0.999875,
16
+ "segment_size": 10240,
17
+ "init_lr_ratio": 1,
18
+ "warmup_epochs": 0,
19
+ "c_mel": 45,
20
+ "c_kl": 1.0,
21
+ "use_sr": true,
22
+ "max_speclen": 512,
23
+ "port": "8001",
24
+ "keep_ckpts": 99
25
+ },
26
+ "data": {
27
+ "training_files": "filelists/train.txt",
28
+ "validation_files": "filelists/val.txt",
29
+ "max_wav_value": 32768.0,
30
+ "sampling_rate": 44100,
31
+ "filter_length": 2048,
32
+ "hop_length": 512,
33
+ "win_length": 2048,
34
+ "n_mel_channels": 80,
35
+ "mel_fmin": 0.0,
36
+ "mel_fmax": 22050
37
+ },
38
+ "model": {
39
+ "inter_channels": 192,
40
+ "hidden_channels": 192,
41
+ "filter_channels": 768,
42
+ "n_heads": 2,
43
+ "n_layers": 6,
44
+ "kernel_size": 3,
45
+ "p_dropout": 0.1,
46
+ "resblock": "1",
47
+ "resblock_kernel_sizes": [
48
+ 3,
49
+ 7,
50
+ 11
51
+ ],
52
+ "resblock_dilation_sizes": [
53
+ [
54
+ 1,
55
+ 3,
56
+ 5
57
+ ],
58
+ [
59
+ 1,
60
+ 3,
61
+ 5
62
+ ],
63
+ [
64
+ 1,
65
+ 3,
66
+ 5
67
+ ]
68
+ ],
69
+ "upsample_rates": [
70
+ 8,
71
+ 8,
72
+ 2,
73
+ 2,
74
+ 2
75
+ ],
76
+ "upsample_initial_channel": 512,
77
+ "upsample_kernel_sizes": [
78
+ 16,
79
+ 16,
80
+ 4,
81
+ 4,
82
+ 4
83
+ ],
84
+ "n_layers_q": 3,
85
+ "use_spectral_norm": false,
86
+ "gin_channels": 256,
87
+ "ssl_dim": 256,
88
+ "n_speakers": 200
89
+ },
90
+ "spk": {
91
+ "Dreizehn": 0
92
+ }
93
+ }
so-vits-svc-4.0/20230325-Dreizehn-44k/kmeans_10000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c0f201ce3cf292c9d7fc36bd8fc6a4b19626d4975c29c705a153aeccf8d996c
3
+ size 15447609
so-vits-svc-4.0/20230325-Dreizehn-44k/pruned/P_G_13600.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60b2d6d3272113095017126fd69b4298d53428c66168986c245d4b8875e6fd7d
3
+ size 180885829
so-vits-svc-4.0/20230325-Dreizehn-44k/pruned/config.json ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "train": {
3
+ "log_interval": 200,
4
+ "eval_interval": 800,
5
+ "seed": 1234,
6
+ "epochs": 10000,
7
+ "learning_rate": 0.0001,
8
+ "betas": [
9
+ 0.8,
10
+ 0.99
11
+ ],
12
+ "eps": 1e-09,
13
+ "batch_size": 6,
14
+ "fp16_run": false,
15
+ "lr_decay": 0.999875,
16
+ "segment_size": 10240,
17
+ "init_lr_ratio": 1,
18
+ "warmup_epochs": 0,
19
+ "c_mel": 45,
20
+ "c_kl": 1.0,
21
+ "use_sr": true,
22
+ "max_speclen": 512,
23
+ "port": "8001",
24
+ "keep_ckpts": 99
25
+ },
26
+ "data": {
27
+ "training_files": "filelists/train.txt",
28
+ "validation_files": "filelists/val.txt",
29
+ "max_wav_value": 32768.0,
30
+ "sampling_rate": 44100,
31
+ "filter_length": 2048,
32
+ "hop_length": 512,
33
+ "win_length": 2048,
34
+ "n_mel_channels": 80,
35
+ "mel_fmin": 0.0,
36
+ "mel_fmax": 22050
37
+ },
38
+ "model": {
39
+ "inter_channels": 192,
40
+ "hidden_channels": 192,
41
+ "filter_channels": 768,
42
+ "n_heads": 2,
43
+ "n_layers": 6,
44
+ "kernel_size": 3,
45
+ "p_dropout": 0.1,
46
+ "resblock": "1",
47
+ "resblock_kernel_sizes": [
48
+ 3,
49
+ 7,
50
+ 11
51
+ ],
52
+ "resblock_dilation_sizes": [
53
+ [
54
+ 1,
55
+ 3,
56
+ 5
57
+ ],
58
+ [
59
+ 1,
60
+ 3,
61
+ 5
62
+ ],
63
+ [
64
+ 1,
65
+ 3,
66
+ 5
67
+ ]
68
+ ],
69
+ "upsample_rates": [
70
+ 8,
71
+ 8,
72
+ 2,
73
+ 2,
74
+ 2
75
+ ],
76
+ "upsample_initial_channel": 512,
77
+ "upsample_kernel_sizes": [
78
+ 16,
79
+ 16,
80
+ 4,
81
+ 4,
82
+ 4
83
+ ],
84
+ "n_layers_q": 3,
85
+ "use_spectral_norm": false,
86
+ "gin_channels": 256,
87
+ "ssl_dim": 256,
88
+ "n_speakers": 200
89
+ },
90
+ "spk": {
91
+ "Dreizehn": 0
92
+ }
93
+ }
so-vits-svc-4.0/20230325-Dreizehn-44k/train.log ADDED
@@ -0,0 +1,405 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2023-03-25 14:44:39,506 44k INFO {'train': {'log_interval': 200, 'eval_interval': 800, 'seed': 1234, 'epochs': 10000, 'learning_rate': 0.0001, 'betas': [0.8, 0.99], 'eps': 1e-09, 'batch_size': 6, 'fp16_run': False, 'lr_decay': 0.999875, 'segment_size': 10240, 'init_lr_ratio': 1, 'warmup_epochs': 0, 'c_mel': 45, 'c_kl': 1.0, 'use_sr': True, 'max_speclen': 512, 'port': '8001', 'keep_ckpts': 99}, 'data': {'training_files': 'filelists/train.txt', 'validation_files': 'filelists/val.txt', 'max_wav_value': 32768.0, 'sampling_rate': 44100, 'filter_length': 2048, 'hop_length': 512, 'win_length': 2048, 'n_mel_channels': 80, 'mel_fmin': 0.0, 'mel_fmax': 22050}, 'model': {'inter_channels': 192, 'hidden_channels': 192, 'filter_channels': 768, 'n_heads': 2, 'n_layers': 6, 'kernel_size': 3, 'p_dropout': 0.1, 'resblock': '1', 'resblock_kernel_sizes': [3, 7, 11], 'resblock_dilation_sizes': [[1, 3, 5], [1, 3, 5], [1, 3, 5]], 'upsample_rates': [8, 8, 2, 2, 2], 'upsample_initial_channel': 512, 'upsample_kernel_sizes': [16, 16, 4, 4, 4], 'n_layers_q': 3, 'use_spectral_norm': False, 'gin_channels': 256, 'ssl_dim': 256, 'n_speakers': 200}, 'spk': {'Dreizehn': 0}, 'model_dir': './logs/44k'}
2
+ 2023-03-25 14:44:39,506 44k WARNING /root/so-vits-svc-4.0 is not a git repository, therefore hash value comparison will be ignored.
3
+ 2023-03-25 14:44:42,455 44k INFO emb_g.weight is not in the checkpoint
4
+ 2023-03-25 14:44:42,528 44k INFO Loaded checkpoint './logs/44k/G_0.pth' (iteration 0)
5
+ 2023-03-25 14:44:42,698 44k INFO Loaded checkpoint './logs/44k/D_0.pth' (iteration 0)
6
+ 2023-03-25 14:44:49,806 44k INFO Train Epoch: 1 [0%]
7
+ 2023-03-25 14:44:49,807 44k INFO Losses: [2.940145492553711, 2.266042709350586, 7.757871150970459, 30.300212860107422, 4.057180881500244], step: 0, lr: 0.0001
8
+ 2023-03-25 14:44:54,651 44k INFO Saving model and optimizer state at iteration 1 to ./logs/44k/G_0.pth
9
+ 2023-03-25 14:44:56,000 44k INFO Saving model and optimizer state at iteration 1 to ./logs/44k/D_0.pth
10
+ 2023-03-25 14:45:24,127 44k INFO ====> Epoch: 1, cost 44.62 s
11
+ 2023-03-25 14:45:50,654 44k INFO ====> Epoch: 2, cost 26.53 s
12
+ 2023-03-25 14:46:17,848 44k INFO ====> Epoch: 3, cost 27.19 s
13
+ 2023-03-25 14:46:29,013 44k INFO Train Epoch: 4 [33%]
14
+ 2023-03-25 14:46:29,013 44k INFO Losses: [2.512408494949341, 2.084155321121216, 12.769187927246094, 23.224327087402344, 1.4044638872146606], step: 200, lr: 9.996250468730469e-05
15
+ 2023-03-25 14:46:45,744 44k INFO ====> Epoch: 4, cost 27.90 s
16
+ 2023-03-25 14:47:13,993 44k INFO ====> Epoch: 5, cost 28.25 s
17
+ 2023-03-25 14:47:42,175 44k INFO ====> Epoch: 6, cost 28.18 s
18
+ 2023-03-25 14:48:01,068 44k INFO Train Epoch: 7 [67%]
19
+ 2023-03-25 14:48:01,069 44k INFO Losses: [2.7341504096984863, 2.1355395317077637, 7.289322853088379, 17.03099250793457, 1.2272381782531738], step: 400, lr: 9.99250234335941e-05
20
+ 2023-03-25 14:48:09,326 44k INFO ====> Epoch: 7, cost 27.15 s
21
+ 2023-03-25 14:48:36,038 44k INFO ====> Epoch: 8, cost 26.71 s
22
+ 2023-03-25 14:49:03,725 44k INFO ====> Epoch: 9, cost 27.69 s
23
+ 2023-03-25 14:49:31,705 44k INFO ====> Epoch: 10, cost 27.98 s
24
+ 2023-03-25 14:49:35,219 44k INFO Train Epoch: 11 [0%]
25
+ 2023-03-25 14:49:35,220 44k INFO Losses: [2.4548428058624268, 2.215726137161255, 9.75304889678955, 19.071210861206055, 1.1653872728347778], step: 600, lr: 9.987507028906759e-05
26
+ 2023-03-25 14:49:59,796 44k INFO ====> Epoch: 11, cost 28.09 s
27
+ 2023-03-25 14:50:26,328 44k INFO ====> Epoch: 12, cost 26.53 s
28
+ 2023-03-25 14:50:53,108 44k INFO ====> Epoch: 13, cost 26.78 s
29
+ 2023-03-25 14:51:04,486 44k INFO Train Epoch: 14 [33%]
30
+ 2023-03-25 14:51:04,487 44k INFO Losses: [2.5523459911346436, 2.3060622215270996, 13.709832191467285, 22.02556610107422, 1.4591848850250244], step: 800, lr: 9.983762181915804e-05
31
+ 2023-03-25 14:51:08,756 44k INFO Saving model and optimizer state at iteration 14 to ./logs/44k/G_800.pth
32
+ 2023-03-25 14:51:09,936 44k INFO Saving model and optimizer state at iteration 14 to ./logs/44k/D_800.pth
33
+ 2023-03-25 14:51:25,899 44k INFO ====> Epoch: 14, cost 32.79 s
34
+ 2023-03-25 14:51:52,420 44k INFO ====> Epoch: 15, cost 26.52 s
35
+ 2023-03-25 14:52:19,045 44k INFO ====> Epoch: 16, cost 26.62 s
36
+ 2023-03-25 14:52:37,909 44k INFO Train Epoch: 17 [67%]
37
+ 2023-03-25 14:52:37,910 44k INFO Losses: [2.1840286254882812, 2.7090253829956055, 14.09952449798584, 20.535053253173828, 0.9362862706184387], step: 1000, lr: 9.980018739066937e-05
38
+ 2023-03-25 14:52:46,301 44k INFO ====> Epoch: 17, cost 27.26 s
39
+ 2023-03-25 14:53:12,974 44k INFO ====> Epoch: 18, cost 26.67 s
40
+ 2023-03-25 14:53:39,790 44k INFO ====> Epoch: 19, cost 26.82 s
41
+ 2023-03-25 14:54:06,380 44k INFO ====> Epoch: 20, cost 26.59 s
42
+ 2023-03-25 14:54:09,784 44k INFO Train Epoch: 21 [0%]
43
+ 2023-03-25 14:54:09,785 44k INFO Losses: [2.469074010848999, 2.314445972442627, 7.137829303741455, 14.518792152404785, 1.2171446084976196], step: 1200, lr: 9.975029665246193e-05
44
+ 2023-03-25 14:54:33,576 44k INFO ====> Epoch: 21, cost 27.20 s
45
+ 2023-03-25 14:55:00,354 44k INFO ====> Epoch: 22, cost 26.78 s
46
+ 2023-03-25 14:55:27,049 44k INFO ====> Epoch: 23, cost 26.70 s
47
+ 2023-03-25 14:55:38,268 44k INFO Train Epoch: 24 [33%]
48
+ 2023-03-25 14:55:38,269 44k INFO Losses: [2.2463409900665283, 2.576353073120117, 12.585086822509766, 20.897579193115234, 1.2720139026641846], step: 1400, lr: 9.971289496681757e-05
49
+ 2023-03-25 14:55:54,212 44k INFO ====> Epoch: 24, cost 27.16 s
50
+ 2023-03-25 14:56:20,732 44k INFO ====> Epoch: 25, cost 26.52 s
51
+ 2023-03-25 14:56:47,308 44k INFO ====> Epoch: 26, cost 26.58 s
52
+ 2023-03-25 14:57:06,114 44k INFO Train Epoch: 27 [67%]
53
+ 2023-03-25 14:57:06,115 44k INFO Losses: [2.5702061653137207, 2.543508768081665, 8.91149616241455, 18.796783447265625, 0.8504246473312378], step: 1600, lr: 9.967550730505221e-05
54
+ 2023-03-25 14:57:10,496 44k INFO Saving model and optimizer state at iteration 27 to ./logs/44k/G_1600.pth
55
+ 2023-03-25 14:57:11,683 44k INFO Saving model and optimizer state at iteration 27 to ./logs/44k/D_1600.pth
56
+ 2023-03-25 14:57:20,013 44k INFO ====> Epoch: 27, cost 32.71 s
57
+ 2023-03-25 14:57:46,724 44k INFO ====> Epoch: 28, cost 26.71 s
58
+ 2023-03-25 14:58:13,413 44k INFO ====> Epoch: 29, cost 26.69 s
59
+ 2023-03-25 14:58:39,922 44k INFO ====> Epoch: 30, cost 26.51 s
60
+ 2023-03-25 14:58:43,385 44k INFO Train Epoch: 31 [0%]
61
+ 2023-03-25 14:58:43,386 44k INFO Losses: [2.177131175994873, 2.665175437927246, 12.00287914276123, 19.802133560180664, 0.9337674379348755], step: 1800, lr: 9.962567889519979e-05
62
+ 2023-03-25 14:59:07,188 44k INFO ====> Epoch: 31, cost 27.27 s
63
+ 2023-03-25 14:59:33,991 44k INFO ====> Epoch: 32, cost 26.80 s
64
+ 2023-03-25 15:00:00,700 44k INFO ====> Epoch: 33, cost 26.71 s
65
+ 2023-03-25 15:00:11,887 44k INFO Train Epoch: 34 [33%]
66
+ 2023-03-25 15:00:11,888 44k INFO Losses: [2.2040858268737793, 2.457867383956909, 12.169000625610352, 20.236183166503906, 1.0930821895599365], step: 2000, lr: 9.95883239353732e-05
67
+ 2023-03-25 15:00:27,890 44k INFO ====> Epoch: 34, cost 27.19 s
68
+ 2023-03-25 15:00:54,577 44k INFO ====> Epoch: 35, cost 26.69 s
69
+ 2023-03-25 15:01:21,275 44k INFO ====> Epoch: 36, cost 26.70 s
70
+ 2023-03-25 15:01:40,077 44k INFO Train Epoch: 37 [67%]
71
+ 2023-03-25 15:01:40,078 44k INFO Losses: [2.5743024349212646, 2.2537405490875244, 6.321502208709717, 14.697388648986816, 0.5774912238121033], step: 2200, lr: 9.95509829819056e-05
72
+ 2023-03-25 15:01:48,399 44k INFO ====> Epoch: 37, cost 27.12 s
73
+ 2023-03-25 15:02:14,993 44k INFO ====> Epoch: 38, cost 26.59 s
74
+ 2023-03-25 15:02:41,823 44k INFO ====> Epoch: 39, cost 26.83 s
75
+ 2023-03-25 15:03:08,943 44k INFO ====> Epoch: 40, cost 27.12 s
76
+ 2023-03-25 15:03:12,327 44k INFO Train Epoch: 41 [0%]
77
+ 2023-03-25 15:03:12,328 44k INFO Losses: [2.5939323902130127, 2.1589159965515137, 5.888221263885498, 14.312878608703613, 0.8646751046180725], step: 2400, lr: 9.950121682254156e-05
78
+ 2023-03-25 15:03:16,742 44k INFO Saving model and optimizer state at iteration 41 to ./logs/44k/G_2400.pth
79
+ 2023-03-25 15:03:17,938 44k INFO Saving model and optimizer state at iteration 41 to ./logs/44k/D_2400.pth
80
+ 2023-03-25 15:03:41,768 44k INFO ====> Epoch: 41, cost 32.82 s
81
+ 2023-03-25 15:04:08,550 44k INFO ====> Epoch: 42, cost 26.78 s
82
+ 2023-03-25 15:04:37,085 44k INFO ====> Epoch: 43, cost 28.53 s
83
+ 2023-03-25 15:04:49,635 44k INFO Train Epoch: 44 [33%]
84
+ 2023-03-25 15:04:49,636 44k INFO Losses: [2.496485710144043, 2.323936700820923, 9.929926872253418, 19.4571533203125, 0.8525826334953308], step: 2600, lr: 9.94639085301583e-05
85
+ 2023-03-25 15:05:08,311 44k INFO ====> Epoch: 44, cost 31.23 s
86
+ 2023-03-25 15:05:38,820 44k INFO ====> Epoch: 45, cost 30.51 s
87
+ 2023-03-25 15:06:09,384 44k INFO ====> Epoch: 46, cost 30.56 s
88
+ 2023-03-25 15:06:29,222 44k INFO Train Epoch: 47 [67%]
89
+ 2023-03-25 15:06:29,223 44k INFO Losses: [2.4428462982177734, 2.2649857997894287, 11.198545455932617, 16.910449981689453, 0.8590737581253052], step: 2800, lr: 9.942661422663591e-05
90
+ 2023-03-25 15:06:37,524 44k INFO ====> Epoch: 47, cost 28.14 s
91
+ 2023-03-25 15:07:04,582 44k INFO ====> Epoch: 48, cost 27.06 s
92
+ 2023-03-25 15:07:31,254 44k INFO ====> Epoch: 49, cost 26.67 s
93
+ 2023-03-25 15:08:01,616 44k INFO ====> Epoch: 50, cost 30.36 s
94
+ 2023-03-25 15:08:04,934 44k INFO Train Epoch: 51 [0%]
95
+ 2023-03-25 15:08:04,936 44k INFO Losses: [2.133681535720825, 2.7440037727355957, 10.468965530395508, 16.335269927978516, 0.6788036227226257], step: 3000, lr: 9.937691023999092e-05
96
+ 2023-03-25 15:08:28,670 44k INFO ====> Epoch: 51, cost 27.05 s
97
+ 2023-03-25 15:08:55,226 44k INFO ====> Epoch: 52, cost 26.56 s
98
+ 2023-03-25 15:09:25,896 44k INFO ====> Epoch: 53, cost 30.67 s
99
+ 2023-03-25 15:09:37,230 44k INFO Train Epoch: 54 [33%]
100
+ 2023-03-25 15:09:37,231 44k INFO Losses: [2.2836685180664062, 2.4881534576416016, 14.617388725280762, 21.667917251586914, 0.8856773972511292], step: 3200, lr: 9.933964855674948e-05
101
+ 2023-03-25 15:09:41,465 44k INFO Saving model and optimizer state at iteration 54 to ./logs/44k/G_3200.pth
102
+ 2023-03-25 15:09:42,753 44k INFO Saving model and optimizer state at iteration 54 to ./logs/44k/D_3200.pth
103
+ 2023-03-25 15:09:58,674 44k INFO ====> Epoch: 54, cost 32.78 s
104
+ 2023-03-25 15:10:25,308 44k INFO ====> Epoch: 55, cost 26.63 s
105
+ 2023-03-25 15:10:52,009 44k INFO ====> Epoch: 56, cost 26.70 s
106
+ 2023-03-25 15:11:10,887 44k INFO Train Epoch: 57 [67%]
107
+ 2023-03-25 15:11:10,888 44k INFO Losses: [1.9383201599121094, 2.8536083698272705, 8.996301651000977, 15.366815567016602, 0.528129518032074], step: 3400, lr: 9.930240084489267e-05
108
+ 2023-03-25 15:11:19,170 44k INFO ====> Epoch: 57, cost 27.16 s
109
+ 2023-03-25 15:11:45,974 44k INFO ====> Epoch: 58, cost 26.80 s
110
+ 2023-03-25 15:12:12,588 44k INFO ====> Epoch: 59, cost 26.61 s
111
+ 2023-03-25 15:12:39,155 44k INFO ====> Epoch: 60, cost 26.57 s
112
+ 2023-03-25 15:12:42,558 44k INFO Train Epoch: 61 [0%]
113
+ 2023-03-25 15:12:42,559 44k INFO Losses: [2.4830005168914795, 2.3222198486328125, 9.070521354675293, 17.854820251464844, 1.0685895681381226], step: 3600, lr: 9.92527589532945e-05
114
+ 2023-03-25 15:13:06,196 44k INFO ====> Epoch: 61, cost 27.04 s
115
+ 2023-03-25 15:13:32,981 44k INFO ====> Epoch: 62, cost 26.79 s
116
+ 2023-03-25 15:13:59,808 44k INFO ====> Epoch: 63, cost 26.83 s
117
+ 2023-03-25 15:14:12,306 44k INFO Train Epoch: 64 [33%]
118
+ 2023-03-25 15:14:12,307 44k INFO Losses: [2.3387303352355957, 2.5378148555755615, 11.396452903747559, 21.508935928344727, 0.9471877217292786], step: 3800, lr: 9.921554382096622e-05
119
+ 2023-03-25 15:14:30,820 44k INFO ====> Epoch: 64, cost 31.01 s
120
+ 2023-03-25 15:14:57,504 44k INFO ====> Epoch: 65, cost 26.68 s
121
+ 2023-03-25 15:15:24,355 44k INFO ====> Epoch: 66, cost 26.85 s
122
+ 2023-03-25 15:15:43,113 44k INFO Train Epoch: 67 [67%]
123
+ 2023-03-25 15:15:43,114 44k INFO Losses: [2.4443154335021973, 2.425118923187256, 9.618949890136719, 19.272972106933594, 0.7614860534667969], step: 4000, lr: 9.917834264256819e-05
124
+ 2023-03-25 15:15:47,233 44k INFO Saving model and optimizer state at iteration 67 to ./logs/44k/G_4000.pth
125
+ 2023-03-25 15:15:48,506 44k INFO Saving model and optimizer state at iteration 67 to ./logs/44k/D_4000.pth
126
+ 2023-03-25 15:15:56,835 44k INFO ====> Epoch: 67, cost 32.48 s
127
+ 2023-03-25 15:16:23,426 44k INFO ====> Epoch: 68, cost 26.59 s
128
+ 2023-03-25 15:16:50,090 44k INFO ====> Epoch: 69, cost 26.66 s
129
+ 2023-03-25 15:17:16,871 44k INFO ====> Epoch: 70, cost 26.78 s
130
+ 2023-03-25 15:17:20,255 44k INFO Train Epoch: 71 [0%]
131
+ 2023-03-25 15:17:20,256 44k INFO Losses: [2.5252299308776855, 2.3594260215759277, 8.788890838623047, 17.564353942871094, 0.23537257313728333], step: 4200, lr: 9.912876276844171e-05
132
+ 2023-03-25 15:17:43,990 44k INFO ====> Epoch: 71, cost 27.12 s
133
+ 2023-03-25 15:18:10,744 44k INFO ====> Epoch: 72, cost 26.75 s
134
+ 2023-03-25 15:18:37,327 44k INFO ====> Epoch: 73, cost 26.58 s
135
+ 2023-03-25 15:18:48,595 44k INFO Train Epoch: 74 [33%]
136
+ 2023-03-25 15:18:48,596 44k INFO Losses: [2.375034809112549, 2.3302433490753174, 10.43783950805664, 13.910572052001953, 0.8278229236602783], step: 4400, lr: 9.909159412887068e-05
137
+ 2023-03-25 15:19:04,613 44k INFO ====> Epoch: 74, cost 27.29 s
138
+ 2023-03-25 15:19:31,231 44k INFO ====> Epoch: 75, cost 26.62 s
139
+ 2023-03-25 15:19:57,771 44k INFO ====> Epoch: 76, cost 26.54 s
140
+ 2023-03-25 15:20:16,714 44k INFO Train Epoch: 77 [67%]
141
+ 2023-03-25 15:20:16,715 44k INFO Losses: [2.7217864990234375, 1.882816195487976, 4.675782203674316, 8.256917953491211, 0.8678979277610779], step: 4600, lr: 9.905443942579728e-05
142
+ 2023-03-25 15:20:25,166 44k INFO ====> Epoch: 77, cost 27.40 s
143
+ 2023-03-25 15:20:51,789 44k INFO ====> Epoch: 78, cost 26.62 s
144
+ 2023-03-25 15:21:18,445 44k INFO ====> Epoch: 79, cost 26.66 s
145
+ 2023-03-25 15:21:46,733 44k INFO ====> Epoch: 80, cost 28.29 s
146
+ 2023-03-25 15:21:50,085 44k INFO Train Epoch: 81 [0%]
147
+ 2023-03-25 15:21:50,086 44k INFO Losses: [2.6070265769958496, 2.4579570293426514, 6.947027206420898, 16.77017593383789, 0.5974054932594299], step: 4800, lr: 9.900492149166423e-05
148
+ 2023-03-25 15:21:54,255 44k INFO Saving model and optimizer state at iteration 81 to ./logs/44k/G_4800.pth
149
+ 2023-03-25 15:21:55,531 44k INFO Saving model and optimizer state at iteration 81 to ./logs/44k/D_4800.pth
150
+ 2023-03-25 15:22:19,263 44k INFO ====> Epoch: 81, cost 32.53 s
151
+ 2023-03-25 15:22:46,437 44k INFO ====> Epoch: 82, cost 27.17 s
152
+ 2023-03-25 15:23:13,154 44k INFO ====> Epoch: 83, cost 26.72 s
153
+ 2023-03-25 15:23:24,324 44k INFO Train Epoch: 84 [33%]
154
+ 2023-03-25 15:23:24,325 44k INFO Losses: [2.5022735595703125, 2.247422456741333, 10.291221618652344, 19.483619689941406, 0.974785566329956], step: 5000, lr: 9.896779928676716e-05
155
+ 2023-03-25 15:23:40,217 44k INFO ====> Epoch: 84, cost 27.06 s
156
+ 2023-03-25 15:24:06,814 44k INFO ====> Epoch: 85, cost 26.60 s
157
+ 2023-03-25 15:24:33,777 44k INFO ====> Epoch: 86, cost 26.96 s
158
+ 2023-03-25 15:24:52,565 44k INFO Train Epoch: 87 [67%]
159
+ 2023-03-25 15:24:52,566 44k INFO Losses: [2.3922224044799805, 2.1541266441345215, 9.643497467041016, 14.642343521118164, 1.0375347137451172], step: 5200, lr: 9.89306910009569e-05
160
+ 2023-03-25 15:25:00,791 44k INFO ====> Epoch: 87, cost 27.01 s
161
+ 2023-03-25 15:25:27,394 44k INFO ====> Epoch: 88, cost 26.60 s
162
+ 2023-03-25 15:25:53,905 44k INFO ====> Epoch: 89, cost 26.51 s
163
+ 2023-03-25 15:26:20,528 44k INFO ====> Epoch: 90, cost 26.62 s
164
+ 2023-03-25 15:26:23,949 44k INFO Train Epoch: 91 [0%]
165
+ 2023-03-25 15:26:23,950 44k INFO Losses: [2.1768221855163574, 2.630232334136963, 11.107681274414062, 23.186044692993164, 0.9275988340377808], step: 5400, lr: 9.888123492943583e-05
166
+ 2023-03-25 15:26:47,862 44k INFO ====> Epoch: 91, cost 27.33 s
167
+ 2023-03-25 15:27:14,428 44k INFO ====> Epoch: 92, cost 26.57 s
168
+ 2023-03-25 15:27:41,017 44k INFO ====> Epoch: 93, cost 26.59 s
169
+ 2023-03-25 15:27:52,175 44k INFO Train Epoch: 94 [33%]
170
+ 2023-03-25 15:27:52,175 44k INFO Losses: [2.031240463256836, 2.6001627445220947, 12.831860542297363, 17.43545150756836, 0.6656062006950378], step: 5600, lr: 9.884415910120204e-05
171
+ 2023-03-25 15:27:56,288 44k INFO Saving model and optimizer state at iteration 94 to ./logs/44k/G_5600.pth
172
+ 2023-03-25 15:27:57,451 44k INFO Saving model and optimizer state at iteration 94 to ./logs/44k/D_5600.pth
173
+ 2023-03-25 15:28:13,565 44k INFO ====> Epoch: 94, cost 32.55 s
174
+ 2023-03-25 15:28:40,077 44k INFO ====> Epoch: 95, cost 26.51 s
175
+ 2023-03-25 15:29:06,692 44k INFO ====> Epoch: 96, cost 26.61 s
176
+ 2023-03-25 15:29:25,411 44k INFO Train Epoch: 97 [67%]
177
+ 2023-03-25 15:29:25,412 44k INFO Losses: [2.6305477619171143, 2.0891010761260986, 7.081687927246094, 14.060254096984863, 0.7252136468887329], step: 5800, lr: 9.880709717466598e-05
178
+ 2023-03-25 15:29:33,688 44k INFO ====> Epoch: 97, cost 27.00 s
179
+ 2023-03-25 15:30:00,268 44k INFO ====> Epoch: 98, cost 26.58 s
180
+ 2023-03-25 15:30:26,927 44k INFO ====> Epoch: 99, cost 26.66 s
181
+ 2023-03-25 15:30:53,739 44k INFO ====> Epoch: 100, cost 26.81 s
182
+ 2023-03-25 15:30:57,158 44k INFO Train Epoch: 101 [0%]
183
+ 2023-03-25 15:30:57,159 44k INFO Losses: [2.57724928855896, 2.4419920444488525, 9.626336097717285, 20.45102882385254, 0.8625250458717346], step: 6000, lr: 9.875770288847208e-05
184
+ 2023-03-25 15:31:20,916 44k INFO ====> Epoch: 101, cost 27.18 s
185
+ 2023-03-25 15:31:47,606 44k INFO ====> Epoch: 102, cost 26.69 s
186
+ 2023-03-25 15:32:14,189 44k INFO ====> Epoch: 103, cost 26.58 s
187
+ 2023-03-25 15:32:25,493 44k INFO Train Epoch: 104 [33%]
188
+ 2023-03-25 15:32:25,494 44k INFO Losses: [2.366825819015503, 2.278965711593628, 9.511296272277832, 17.809585571289062, 0.773294985294342], step: 6200, lr: 9.872067337896332e-05
189
+ 2023-03-25 15:32:41,421 44k INFO ====> Epoch: 104, cost 27.23 s
190
+ 2023-03-25 15:33:10,300 44k INFO ====> Epoch: 105, cost 28.88 s
191
+ 2023-03-25 15:33:37,022 44k INFO ====> Epoch: 106, cost 26.72 s
192
+ 2023-03-25 15:33:56,020 44k INFO Train Epoch: 107 [67%]
193
+ 2023-03-25 15:33:56,021 44k INFO Losses: [2.5079104900360107, 2.207770347595215, 8.176128387451172, 14.970914840698242, 0.7647779583930969], step: 6400, lr: 9.868365775378495e-05
194
+ 2023-03-25 15:34:00,185 44k INFO Saving model and optimizer state at iteration 107 to ./logs/44k/G_6400.pth
195
+ 2023-03-25 15:34:01,302 44k INFO Saving model and optimizer state at iteration 107 to ./logs/44k/D_6400.pth
196
+ 2023-03-25 15:34:09,595 44k INFO ====> Epoch: 107, cost 32.57 s
197
+ 2023-03-25 15:34:36,317 44k INFO ====> Epoch: 108, cost 26.72 s
198
+ 2023-03-25 15:35:03,186 44k INFO ====> Epoch: 109, cost 26.87 s
199
+ 2023-03-25 15:35:30,271 44k INFO ====> Epoch: 110, cost 27.09 s
200
+ 2023-03-25 15:35:33,805 44k INFO Train Epoch: 111 [0%]
201
+ 2023-03-25 15:35:33,806 44k INFO Losses: [2.3396925926208496, 2.208421230316162, 8.759860038757324, 13.542329788208008, 0.7297613620758057], step: 6600, lr: 9.863432517573002e-05
202
+ 2023-03-25 15:35:58,905 44k INFO ====> Epoch: 111, cost 28.63 s
203
+ 2023-03-25 15:36:26,578 44k INFO ====> Epoch: 112, cost 27.67 s
204
+ 2023-03-25 15:36:54,361 44k INFO ====> Epoch: 113, cost 27.78 s
205
+ 2023-03-25 15:37:05,864 44k INFO Train Epoch: 114 [33%]
206
+ 2023-03-25 15:37:05,865 44k INFO Losses: [2.2866804599761963, 2.403698444366455, 14.162016868591309, 20.087860107421875, 0.8307042121887207], step: 6800, lr: 9.859734192708044e-05
207
+ 2023-03-25 15:37:22,493 44k INFO ====> Epoch: 114, cost 28.13 s
208
+ 2023-03-25 15:37:50,437 44k INFO ====> Epoch: 115, cost 27.94 s
209
+ 2023-03-25 15:38:17,891 44k INFO ====> Epoch: 116, cost 27.45 s
210
+ 2023-03-25 15:38:37,530 44k INFO Train Epoch: 117 [67%]
211
+ 2023-03-25 15:38:37,531 44k INFO Losses: [2.4838879108428955, 2.1345486640930176, 9.153286933898926, 19.457439422607422, 0.8745102882385254], step: 7000, lr: 9.85603725454156e-05
212
+ 2023-03-25 15:38:45,975 44k INFO ====> Epoch: 117, cost 28.08 s
213
+ 2023-03-25 15:39:13,778 44k INFO ====> Epoch: 118, cost 27.80 s
214
+ 2023-03-25 15:39:41,699 44k INFO ====> Epoch: 119, cost 27.92 s
215
+ 2023-03-25 15:40:09,333 44k INFO ====> Epoch: 120, cost 27.63 s
216
+ 2023-03-25 15:40:12,966 44k INFO Train Epoch: 121 [0%]
217
+ 2023-03-25 15:40:12,967 44k INFO Losses: [2.2715530395507812, 2.6235780715942383, 12.198266983032227, 21.405956268310547, 1.187570333480835], step: 7200, lr: 9.851110159840781e-05
218
+ 2023-03-25 15:40:17,172 44k INFO Saving model and optimizer state at iteration 121 to ./logs/44k/G_7200.pth
219
+ 2023-03-25 15:40:18,433 44k INFO Saving model and optimizer state at iteration 121 to ./logs/44k/D_7200.pth
220
+ 2023-03-25 15:40:43,478 44k INFO ====> Epoch: 121, cost 34.15 s
221
+ 2023-03-25 15:41:11,318 44k INFO ====> Epoch: 122, cost 27.84 s
222
+ 2023-03-25 15:41:38,933 44k INFO ====> Epoch: 123, cost 27.62 s
223
+ 2023-03-25 15:41:50,572 44k INFO Train Epoch: 124 [33%]
224
+ 2023-03-25 15:41:50,573 44k INFO Losses: [2.6159331798553467, 1.9919030666351318, 10.321922302246094, 17.80287742614746, 0.8138943910598755], step: 7400, lr: 9.847416455282387e-05
225
+ 2023-03-25 15:42:07,225 44k INFO ====> Epoch: 124, cost 28.29 s
226
+ 2023-03-25 15:42:35,119 44k INFO ====> Epoch: 125, cost 27.89 s
227
+ 2023-03-25 15:43:02,554 44k INFO ====> Epoch: 126, cost 27.43 s
228
+ 2023-03-25 15:43:22,175 44k INFO Train Epoch: 127 [67%]
229
+ 2023-03-25 15:43:22,176 44k INFO Losses: [2.354686737060547, 2.4857425689697266, 8.663324356079102, 13.985154151916504, 0.27259084582328796], step: 7600, lr: 9.84372413569007e-05
230
+ 2023-03-25 15:43:30,787 44k INFO ====> Epoch: 127, cost 28.23 s
231
+ 2023-03-25 15:43:58,989 44k INFO ====> Epoch: 128, cost 28.20 s
232
+ 2023-03-25 15:44:26,506 44k INFO ====> Epoch: 129, cost 27.52 s
233
+ 2023-03-25 15:44:54,120 44k INFO ====> Epoch: 130, cost 27.61 s
234
+ 2023-03-25 15:44:57,800 44k INFO Train Epoch: 131 [0%]
235
+ 2023-03-25 15:44:57,801 44k INFO Losses: [2.362338066101074, 2.521932363510132, 10.726180076599121, 18.021610260009766, 0.8564594984054565], step: 7800, lr: 9.838803196394459e-05
236
+ 2023-03-25 15:45:22,279 44k INFO ====> Epoch: 131, cost 28.16 s
237
+ 2023-03-25 15:45:50,028 44k INFO ====> Epoch: 132, cost 27.75 s
238
+ 2023-03-25 15:46:17,576 44k INFO ====> Epoch: 133, cost 27.55 s
239
+ 2023-03-25 15:46:29,163 44k INFO Train Epoch: 134 [33%]
240
+ 2023-03-25 15:46:29,164 44k INFO Losses: [2.1983914375305176, 2.453049898147583, 10.3521089553833, 19.192773818969727, 1.035571575164795], step: 8000, lr: 9.835114106370493e-05
241
+ 2023-03-25 15:46:33,555 44k INFO Saving model and optimizer state at iteration 134 to ./logs/44k/G_8000.pth
242
+ 2023-03-25 15:46:34,883 44k INFO Saving model and optimizer state at iteration 134 to ./logs/44k/D_8000.pth
243
+ 2023-03-25 15:46:51,651 44k INFO ====> Epoch: 134, cost 34.07 s
244
+ 2023-03-25 15:47:19,182 44k INFO ====> Epoch: 135, cost 27.53 s
245
+ 2023-03-25 15:47:47,310 44k INFO ====> Epoch: 136, cost 28.13 s
246
+ 2023-03-25 15:48:07,405 44k INFO Train Epoch: 137 [67%]
247
+ 2023-03-25 15:48:07,406 44k INFO Losses: [2.5634498596191406, 2.1999950408935547, 9.676583290100098, 19.181659698486328, 0.4880225360393524], step: 8200, lr: 9.831426399582366e-05
248
+ 2023-03-25 15:48:16,176 44k INFO ====> Epoch: 137, cost 28.87 s
249
+ 2023-03-25 15:48:44,491 44k INFO ====> Epoch: 138, cost 28.32 s
250
+ 2023-03-25 15:49:12,145 44k INFO ====> Epoch: 139, cost 27.65 s
251
+ 2023-03-25 15:49:39,691 44k INFO ====> Epoch: 140, cost 27.55 s
252
+ 2023-03-25 15:49:43,562 44k INFO Train Epoch: 141 [0%]
253
+ 2023-03-25 15:49:43,563 44k INFO Losses: [2.1003589630126953, 2.6904215812683105, 11.157882690429688, 18.755420684814453, 0.7207870483398438], step: 8400, lr: 9.826511608001993e-05
254
+ 2023-03-25 15:50:08,579 44k INFO ====> Epoch: 141, cost 28.89 s
255
+ 2023-03-25 15:50:36,778 44k INFO ====> Epoch: 142, cost 28.20 s
256
+ 2023-03-25 15:51:04,578 44k INFO ====> Epoch: 143, cost 27.80 s
257
+ 2023-03-25 15:51:16,196 44k INFO Train Epoch: 144 [33%]
258
+ 2023-03-25 15:51:16,197 44k INFO Losses: [2.6339874267578125, 2.1276330947875977, 10.583199501037598, 17.78839874267578, 0.9267172813415527], step: 8600, lr: 9.822827126747529e-05
259
+ 2023-03-25 15:51:33,003 44k INFO ====> Epoch: 144, cost 28.43 s
260
+ 2023-03-25 15:52:01,405 44k INFO ====> Epoch: 145, cost 28.40 s
261
+ 2023-03-25 15:52:28,990 44k INFO ====> Epoch: 146, cost 27.58 s
262
+ 2023-03-25 15:52:48,861 44k INFO Train Epoch: 147 [67%]
263
+ 2023-03-25 15:52:48,862 44k INFO Losses: [2.5725722312927246, 2.5007855892181396, 7.138617515563965, 13.22614860534668, 0.8027871251106262], step: 8800, lr: 9.819144027000834e-05
264
+ 2023-03-25 15:52:53,303 44k INFO Saving model and optimizer state at iteration 147 to ./logs/44k/G_8800.pth
265
+ 2023-03-25 15:52:54,482 44k INFO Saving model and optimizer state at iteration 147 to ./logs/44k/D_8800.pth
266
+ 2023-03-25 15:53:03,077 44k INFO ====> Epoch: 147, cost 34.09 s
267
+ 2023-03-25 15:53:30,602 44k INFO ====> Epoch: 148, cost 27.53 s
268
+ 2023-03-25 15:53:58,172 44k INFO ====> Epoch: 149, cost 27.57 s
269
+ 2023-03-25 15:54:25,707 44k INFO ====> Epoch: 150, cost 27.54 s
270
+ 2023-03-25 15:54:29,284 44k INFO Train Epoch: 151 [0%]
271
+ 2023-03-25 15:54:29,285 44k INFO Losses: [2.169490337371826, 2.8044533729553223, 11.827469825744629, 18.990873336791992, 0.9448695778846741], step: 9000, lr: 9.814235375455375e-05
272
+ 2023-03-25 15:54:54,505 44k INFO ====> Epoch: 151, cost 28.80 s
273
+ 2023-03-25 15:55:22,519 44k INFO ====> Epoch: 152, cost 28.01 s
274
+ 2023-03-25 15:55:50,002 44k INFO ====> Epoch: 153, cost 27.48 s
275
+ 2023-03-25 15:56:01,731 44k INFO Train Epoch: 154 [33%]
276
+ 2023-03-25 15:56:01,732 44k INFO Losses: [2.2980504035949707, 2.498497247695923, 9.984493255615234, 16.98595428466797, 0.7363108992576599], step: 9200, lr: 9.810555497212693e-05
277
+ 2023-03-25 15:56:18,105 44k INFO ====> Epoch: 154, cost 28.10 s
278
+ 2023-03-25 15:56:45,928 44k INFO ====> Epoch: 155, cost 27.82 s
279
+ 2023-03-25 15:57:13,862 44k INFO ====> Epoch: 156, cost 27.93 s
280
+ 2023-03-25 15:57:33,710 44k INFO Train Epoch: 157 [67%]
281
+ 2023-03-25 15:57:33,711 44k INFO Losses: [2.4687323570251465, 2.260420083999634, 7.67680025100708, 14.242744445800781, 0.7204187512397766], step: 9400, lr: 9.806876998751865e-05
282
+ 2023-03-25 15:57:42,365 44k INFO ====> Epoch: 157, cost 28.50 s
283
+ 2023-03-25 15:58:10,840 44k INFO ====> Epoch: 158, cost 28.47 s
284
+ 2023-03-25 15:58:38,494 44k INFO ====> Epoch: 159, cost 27.65 s
285
+ 2023-03-25 15:59:06,177 44k INFO ====> Epoch: 160, cost 27.68 s
286
+ 2023-03-25 15:59:10,017 44k INFO Train Epoch: 161 [0%]
287
+ 2023-03-25 15:59:10,018 44k INFO Losses: [2.2403976917266846, 2.689331531524658, 10.701247215270996, 18.250370025634766, 0.5913345217704773], step: 9600, lr: 9.801974479570593e-05
288
+ 2023-03-25 15:59:14,725 44k INFO Saving model and optimizer state at iteration 161 to ./logs/44k/G_9600.pth
289
+ 2023-03-25 15:59:15,941 44k INFO Saving model and optimizer state at iteration 161 to ./logs/44k/D_9600.pth
290
+ 2023-03-25 15:59:40,465 44k INFO ====> Epoch: 161, cost 34.29 s
291
+ 2023-03-25 16:00:08,135 44k INFO ====> Epoch: 162, cost 27.67 s
292
+ 2023-03-25 16:00:36,280 44k INFO ====> Epoch: 163, cost 28.15 s
293
+ 2023-03-25 16:00:47,950 44k INFO Train Epoch: 164 [33%]
294
+ 2023-03-25 16:00:47,951 44k INFO Losses: [2.291417360305786, 2.2027976512908936, 9.914422035217285, 14.098004341125488, 0.9870694875717163], step: 9800, lr: 9.798299198589162e-05
295
+ 2023-03-25 16:01:04,546 44k INFO ====> Epoch: 164, cost 28.27 s
296
+ 2023-03-25 16:01:32,268 44k INFO ====> Epoch: 165, cost 27.72 s
297
+ 2023-03-25 16:02:00,779 44k INFO ====> Epoch: 166, cost 28.51 s
298
+ 2023-03-25 16:02:20,253 44k INFO Train Epoch: 167 [67%]
299
+ 2023-03-25 16:02:20,255 44k INFO Losses: [2.5774083137512207, 2.229320764541626, 11.070889472961426, 18.494190216064453, 0.5609847903251648], step: 10000, lr: 9.794625295665828e-05
300
+ 2023-03-25 16:02:29,005 44k INFO ====> Epoch: 167, cost 28.23 s
301
+ 2023-03-25 16:02:57,331 44k INFO ====> Epoch: 168, cost 28.33 s
302
+ 2023-03-25 16:03:25,540 44k INFO ====> Epoch: 169, cost 28.21 s
303
+ 2023-03-25 16:03:53,268 44k INFO ====> Epoch: 170, cost 27.73 s
304
+ 2023-03-25 16:03:56,836 44k INFO Train Epoch: 171 [0%]
305
+ 2023-03-25 16:03:56,837 44k INFO Losses: [2.506603717803955, 2.397548198699951, 10.223786354064941, 18.886367797851562, 0.70171058177948], step: 10200, lr: 9.789728901187598e-05
306
+ 2023-03-25 16:04:21,639 44k INFO ====> Epoch: 171, cost 28.37 s
307
+ 2023-03-25 16:04:49,952 44k INFO ====> Epoch: 172, cost 28.31 s
308
+ 2023-03-25 16:05:18,292 44k INFO ====> Epoch: 173, cost 28.34 s
309
+ 2023-03-25 16:05:30,033 44k INFO Train Epoch: 174 [33%]
310
+ 2023-03-25 16:05:30,035 44k INFO Losses: [2.3508377075195312, 2.1366045475006104, 10.397187232971191, 17.74555015563965, 0.7728736996650696], step: 10400, lr: 9.786058211724074e-05
311
+ 2023-03-25 16:05:34,367 44k INFO Saving model and optimizer state at iteration 174 to ./logs/44k/G_10400.pth
312
+ 2023-03-25 16:05:35,703 44k INFO Saving model and optimizer state at iteration 174 to ./logs/44k/D_10400.pth
313
+ 2023-03-25 16:05:52,335 44k INFO ====> Epoch: 174, cost 34.04 s
314
+ 2023-03-25 16:06:20,613 44k INFO ====> Epoch: 175, cost 28.28 s
315
+ 2023-03-25 16:06:48,223 44k INFO ====> Epoch: 176, cost 27.61 s
316
+ 2023-03-25 16:07:07,774 44k INFO Train Epoch: 177 [67%]
317
+ 2023-03-25 16:07:07,776 44k INFO Losses: [2.3116414546966553, 2.3159842491149902, 11.632226943969727, 18.547286987304688, 0.9227203726768494], step: 10600, lr: 9.782388898597041e-05
318
+ 2023-03-25 16:07:16,338 44k INFO ====> Epoch: 177, cost 28.11 s
319
+ 2023-03-25 16:07:44,185 44k INFO ====> Epoch: 178, cost 27.85 s
320
+ 2023-03-25 16:08:11,900 44k INFO ====> Epoch: 179, cost 27.71 s
321
+ 2023-03-25 16:08:40,228 44k INFO ====> Epoch: 180, cost 28.33 s
322
+ 2023-03-25 16:08:43,879 44k INFO Train Epoch: 181 [0%]
323
+ 2023-03-25 16:08:43,880 44k INFO Losses: [2.40667724609375, 2.235522985458374, 9.322698593139648, 17.760761260986328, 0.6746690273284912], step: 10800, lr: 9.777498621170277e-05
324
+ 2023-03-25 16:09:08,780 44k INFO ====> Epoch: 181, cost 28.55 s
325
+ 2023-03-25 16:09:35,925 44k INFO ====> Epoch: 182, cost 27.15 s
326
+ 2023-03-25 16:10:02,656 44k INFO ====> Epoch: 183, cost 26.73 s
327
+ 2023-03-25 16:10:14,074 44k INFO Train Epoch: 184 [33%]
328
+ 2023-03-25 16:10:14,075 44k INFO Losses: [2.315093755722046, 2.372858762741089, 11.837750434875488, 19.7078800201416, 0.28122562170028687], step: 11000, lr: 9.773832517488488e-05
329
+ 2023-03-25 16:10:30,227 44k INFO ====> Epoch: 184, cost 27.57 s
330
+ 2023-03-25 16:10:56,929 44k INFO ====> Epoch: 185, cost 26.70 s
331
+ 2023-03-25 16:11:23,751 44k INFO ====> Epoch: 186, cost 26.82 s
332
+ 2023-03-25 16:11:44,391 44k INFO Train Epoch: 187 [67%]
333
+ 2023-03-25 16:11:44,392 44k INFO Losses: [2.32973575592041, 2.6208066940307617, 10.395020484924316, 16.26879119873047, 0.376319944858551], step: 11200, lr: 9.77016778842374e-05
334
+ 2023-03-25 16:11:48,550 44k INFO Saving model and optimizer state at iteration 187 to ./logs/44k/G_11200.pth
335
+ 2023-03-25 16:11:49,725 44k INFO Saving model and optimizer state at iteration 187 to ./logs/44k/D_11200.pth
336
+ 2023-03-25 16:11:58,048 44k INFO ====> Epoch: 187, cost 34.30 s
337
+ 2023-03-25 16:12:24,750 44k INFO ====> Epoch: 188, cost 26.70 s
338
+ 2023-03-25 16:12:51,376 44k INFO ====> Epoch: 189, cost 26.63 s
339
+ 2023-03-25 16:13:18,199 44k INFO ====> Epoch: 190, cost 26.82 s
340
+ 2023-03-25 16:13:21,590 44k INFO Train Epoch: 191 [0%]
341
+ 2023-03-25 16:13:21,591 44k INFO Losses: [2.3605268001556396, 2.5424208641052246, 7.856480121612549, 14.165212631225586, 0.7268055081367493], step: 11400, lr: 9.765283620406429e-05
342
+ 2023-03-25 16:13:45,845 44k INFO ====> Epoch: 191, cost 27.65 s
343
+ 2023-03-25 16:14:12,604 44k INFO ====> Epoch: 192, cost 26.76 s
344
+ 2023-03-25 16:14:43,116 44k INFO ====> Epoch: 193, cost 30.51 s
345
+ 2023-03-25 16:14:54,428 44k INFO Train Epoch: 194 [33%]
346
+ 2023-03-25 16:14:54,429 44k INFO Losses: [2.7298760414123535, 1.8387227058410645, 6.033370494842529, 13.283061981201172, 0.9616314768791199], step: 11600, lr: 9.761622096777372e-05
347
+ 2023-03-25 16:15:10,675 44k INFO ====> Epoch: 194, cost 27.56 s
348
+ 2023-03-25 16:15:37,838 44k INFO ====> Epoch: 195, cost 27.16 s
349
+ 2023-03-25 16:16:05,901 44k INFO ====> Epoch: 196, cost 28.06 s
350
+ 2023-03-25 16:16:25,641 44k INFO Train Epoch: 197 [67%]
351
+ 2023-03-25 16:16:25,642 44k INFO Losses: [2.3175127506256104, 2.385716199874878, 11.155668258666992, 17.777097702026367, 0.3992789685726166], step: 11800, lr: 9.757961946048049e-05
352
+ 2023-03-25 16:16:34,258 44k INFO ====> Epoch: 197, cost 28.36 s
353
+ 2023-03-25 16:17:02,036 44k INFO ====> Epoch: 198, cost 27.78 s
354
+ 2023-03-25 16:17:29,701 44k INFO ====> Epoch: 199, cost 27.67 s
355
+ 2023-03-25 16:17:56,622 44k INFO ====> Epoch: 200, cost 26.92 s
356
+ 2023-03-25 16:18:00,038 44k INFO Train Epoch: 201 [0%]
357
+ 2023-03-25 16:18:00,039 44k INFO Losses: [2.126671314239502, 2.953439712524414, 10.01817798614502, 18.810077667236328, 0.6492265462875366], step: 12000, lr: 9.753083879807726e-05
358
+ 2023-03-25 16:18:04,307 44k INFO Saving model and optimizer state at iteration 201 to ./logs/44k/G_12000.pth
359
+ 2023-03-25 16:18:05,443 44k INFO Saving model and optimizer state at iteration 201 to ./logs/44k/D_12000.pth
360
+ 2023-03-25 16:18:30,206 44k INFO ====> Epoch: 201, cost 33.58 s
361
+ 2023-03-25 16:18:58,235 44k INFO ====> Epoch: 202, cost 28.03 s
362
+ 2023-03-25 16:19:26,346 44k INFO ====> Epoch: 203, cost 28.11 s
363
+ 2023-03-25 16:19:38,537 44k INFO Train Epoch: 204 [33%]
364
+ 2023-03-25 16:19:38,538 44k INFO Losses: [2.341325521469116, 2.3583366870880127, 9.117012023925781, 14.948786735534668, 0.8733038902282715], step: 12200, lr: 9.749426930509556e-05
365
+ 2023-03-25 16:19:55,293 44k INFO ====> Epoch: 204, cost 28.95 s
366
+ 2023-03-25 16:20:24,093 44k INFO ====> Epoch: 205, cost 28.80 s
367
+ 2023-03-25 16:20:52,270 44k INFO ====> Epoch: 206, cost 28.18 s
368
+ 2023-03-25 16:21:12,828 44k INFO Train Epoch: 207 [67%]
369
+ 2023-03-25 16:21:12,829 44k INFO Losses: [2.518537759780884, 2.0252599716186523, 8.098270416259766, 14.455513000488281, 0.4925019145011902], step: 12400, lr: 9.745771352395957e-05
370
+ 2023-03-25 16:21:21,530 44k INFO ====> Epoch: 207, cost 29.26 s
371
+ 2023-03-25 16:21:49,885 44k INFO ====> Epoch: 208, cost 28.36 s
372
+ 2023-03-25 16:22:18,738 44k INFO ====> Epoch: 209, cost 28.85 s
373
+ 2023-03-25 16:22:46,546 44k INFO ====> Epoch: 210, cost 27.81 s
374
+ 2023-03-25 16:22:50,385 44k INFO Train Epoch: 211 [0%]
375
+ 2023-03-25 16:22:50,386 44k INFO Losses: [2.324171781539917, 2.519193649291992, 9.139647483825684, 19.630090713500977, 0.3635931611061096], step: 12600, lr: 9.740899380309685e-05
376
+ 2023-03-25 16:23:15,533 44k INFO ====> Epoch: 211, cost 28.99 s
377
+ 2023-03-25 16:23:44,387 44k INFO ====> Epoch: 212, cost 28.85 s
378
+ 2023-03-25 16:24:12,486 44k INFO ====> Epoch: 213, cost 28.10 s
379
+ 2023-03-25 16:24:24,567 44k INFO Train Epoch: 214 [33%]
380
+ 2023-03-25 16:24:24,569 44k INFO Losses: [2.4054665565490723, 2.550511121749878, 12.541123390197754, 19.724668502807617, 1.1664975881576538], step: 12800, lr: 9.7372469996277e-05
381
+ 2023-03-25 16:24:29,234 44k INFO Saving model and optimizer state at iteration 214 to ./logs/44k/G_12800.pth
382
+ 2023-03-25 16:24:30,661 44k INFO Saving model and optimizer state at iteration 214 to ./logs/44k/D_12800.pth
383
+ 2023-03-25 16:24:47,567 44k INFO ====> Epoch: 214, cost 35.08 s
384
+ 2023-03-25 16:25:16,212 44k INFO ====> Epoch: 215, cost 28.65 s
385
+ 2023-03-25 16:25:44,316 44k INFO ====> Epoch: 216, cost 28.10 s
386
+ 2023-03-25 16:26:04,228 44k INFO Train Epoch: 217 [67%]
387
+ 2023-03-25 16:26:04,229 44k INFO Losses: [2.41851544380188, 2.3247714042663574, 9.172240257263184, 16.376054763793945, 0.6352742910385132], step: 13000, lr: 9.733595988417275e-05
388
+ 2023-03-25 16:26:12,954 44k INFO ====> Epoch: 217, cost 28.64 s
389
+ 2023-03-25 16:26:41,838 44k INFO ====> Epoch: 218, cost 28.88 s
390
+ 2023-03-25 16:27:10,745 44k INFO ====> Epoch: 219, cost 28.91 s
391
+ 2023-03-25 16:27:38,859 44k INFO ====> Epoch: 220, cost 28.11 s
392
+ 2023-03-25 16:27:42,563 44k INFO Train Epoch: 221 [0%]
393
+ 2023-03-25 16:27:42,564 44k INFO Losses: [2.5801851749420166, 2.454096555709839, 8.478157043457031, 17.067672729492188, 0.8212155103683472], step: 13200, lr: 9.728730102871649e-05
394
+ 2023-03-25 16:28:07,414 44k INFO ====> Epoch: 221, cost 28.55 s
395
+ 2023-03-25 16:28:35,465 44k INFO ====> Epoch: 222, cost 28.05 s
396
+ 2023-03-25 16:29:04,152 44k INFO ====> Epoch: 223, cost 28.69 s
397
+ 2023-03-25 16:29:16,136 44k INFO Train Epoch: 224 [33%]
398
+ 2023-03-25 16:29:16,137 44k INFO Losses: [2.606571674346924, 2.4211103916168213, 9.568647384643555, 16.520126342773438, 0.5133330225944519], step: 13400, lr: 9.725082285098293e-05
399
+ 2023-03-25 16:29:33,043 44k INFO ====> Epoch: 224, cost 28.89 s
400
+ 2023-03-25 16:30:01,279 44k INFO ====> Epoch: 225, cost 28.24 s
401
+ 2023-03-25 16:30:29,534 44k INFO ====> Epoch: 226, cost 28.26 s
402
+ 2023-03-25 16:30:49,224 44k INFO Train Epoch: 227 [67%]
403
+ 2023-03-25 16:30:49,226 44k INFO Losses: [2.2993545532226562, 2.6229958534240723, 10.669507026672363, 18.169801712036133, 0.5818897485733032], step: 13600, lr: 9.721435835085619e-05
404
+ 2023-03-25 16:30:53,905 44k INFO Saving model and optimizer state at iteration 227 to ./logs/44k/G_13600.pth
405
+ 2023-03-25 16:30:55,146 44k INFO Saving model and optimizer state at iteration 227 to ./logs/44k/D_13600.pth
so-vits-svc-4.0/20230325-DreizehnNS-44k/best/D_13600.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa8a86282eef397e4078906c7447c91bd9b29ddefb6aef61cb261ec0bd10fc30
3
+ size 561098185
so-vits-svc-4.0/20230325-DreizehnNS-44k/best/G_13600.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24849a2d9adcd52bfea6cb2d8b24cd9b2a3cb27dac1b3afec79c2e46b2415721
3
+ size 542789405
so-vits-svc-4.0/20230325-DreizehnNS-44k/best/config.json ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "train": {
3
+ "log_interval": 200,
4
+ "eval_interval": 800,
5
+ "seed": 1234,
6
+ "epochs": 10000,
7
+ "learning_rate": 0.0001,
8
+ "betas": [
9
+ 0.8,
10
+ 0.99
11
+ ],
12
+ "eps": 1e-09,
13
+ "batch_size": 6,
14
+ "fp16_run": false,
15
+ "lr_decay": 0.999875,
16
+ "segment_size": 10240,
17
+ "init_lr_ratio": 1,
18
+ "warmup_epochs": 0,
19
+ "c_mel": 45,
20
+ "c_kl": 1.0,
21
+ "use_sr": true,
22
+ "max_speclen": 512,
23
+ "port": "8001",
24
+ "keep_ckpts": 99
25
+ },
26
+ "data": {
27
+ "training_files": "filelists/train.txt",
28
+ "validation_files": "filelists/val.txt",
29
+ "max_wav_value": 32768.0,
30
+ "sampling_rate": 44100,
31
+ "filter_length": 2048,
32
+ "hop_length": 512,
33
+ "win_length": 2048,
34
+ "n_mel_channels": 80,
35
+ "mel_fmin": 0.0,
36
+ "mel_fmax": 22050
37
+ },
38
+ "model": {
39
+ "inter_channels": 192,
40
+ "hidden_channels": 192,
41
+ "filter_channels": 768,
42
+ "n_heads": 2,
43
+ "n_layers": 6,
44
+ "kernel_size": 3,
45
+ "p_dropout": 0.1,
46
+ "resblock": "1",
47
+ "resblock_kernel_sizes": [
48
+ 3,
49
+ 7,
50
+ 11
51
+ ],
52
+ "resblock_dilation_sizes": [
53
+ [
54
+ 1,
55
+ 3,
56
+ 5
57
+ ],
58
+ [
59
+ 1,
60
+ 3,
61
+ 5
62
+ ],
63
+ [
64
+ 1,
65
+ 3,
66
+ 5
67
+ ]
68
+ ],
69
+ "upsample_rates": [
70
+ 8,
71
+ 8,
72
+ 2,
73
+ 2,
74
+ 2
75
+ ],
76
+ "upsample_initial_channel": 512,
77
+ "upsample_kernel_sizes": [
78
+ 16,
79
+ 16,
80
+ 4,
81
+ 4,
82
+ 4
83
+ ],
84
+ "n_layers_q": 3,
85
+ "use_spectral_norm": false,
86
+ "gin_channels": 256,
87
+ "ssl_dim": 256,
88
+ "n_speakers": 200
89
+ },
90
+ "spk": {
91
+ "DreizehnNS": 0
92
+ }
93
+ }
so-vits-svc-4.0/20230325-DreizehnNS-44k/config.json ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "train": {
3
+ "log_interval": 200,
4
+ "eval_interval": 800,
5
+ "seed": 1234,
6
+ "epochs": 10000,
7
+ "learning_rate": 0.0001,
8
+ "betas": [
9
+ 0.8,
10
+ 0.99
11
+ ],
12
+ "eps": 1e-09,
13
+ "batch_size": 6,
14
+ "fp16_run": false,
15
+ "lr_decay": 0.999875,
16
+ "segment_size": 10240,
17
+ "init_lr_ratio": 1,
18
+ "warmup_epochs": 0,
19
+ "c_mel": 45,
20
+ "c_kl": 1.0,
21
+ "use_sr": true,
22
+ "max_speclen": 512,
23
+ "port": "8001",
24
+ "keep_ckpts": 99
25
+ },
26
+ "data": {
27
+ "training_files": "filelists/train.txt",
28
+ "validation_files": "filelists/val.txt",
29
+ "max_wav_value": 32768.0,
30
+ "sampling_rate": 44100,
31
+ "filter_length": 2048,
32
+ "hop_length": 512,
33
+ "win_length": 2048,
34
+ "n_mel_channels": 80,
35
+ "mel_fmin": 0.0,
36
+ "mel_fmax": 22050
37
+ },
38
+ "model": {
39
+ "inter_channels": 192,
40
+ "hidden_channels": 192,
41
+ "filter_channels": 768,
42
+ "n_heads": 2,
43
+ "n_layers": 6,
44
+ "kernel_size": 3,
45
+ "p_dropout": 0.1,
46
+ "resblock": "1",
47
+ "resblock_kernel_sizes": [
48
+ 3,
49
+ 7,
50
+ 11
51
+ ],
52
+ "resblock_dilation_sizes": [
53
+ [
54
+ 1,
55
+ 3,
56
+ 5
57
+ ],
58
+ [
59
+ 1,
60
+ 3,
61
+ 5
62
+ ],
63
+ [
64
+ 1,
65
+ 3,
66
+ 5
67
+ ]
68
+ ],
69
+ "upsample_rates": [
70
+ 8,
71
+ 8,
72
+ 2,
73
+ 2,
74
+ 2
75
+ ],
76
+ "upsample_initial_channel": 512,
77
+ "upsample_kernel_sizes": [
78
+ 16,
79
+ 16,
80
+ 4,
81
+ 4,
82
+ 4
83
+ ],
84
+ "n_layers_q": 3,
85
+ "use_spectral_norm": false,
86
+ "gin_channels": 256,
87
+ "ssl_dim": 256,
88
+ "n_speakers": 200
89
+ },
90
+ "spk": {
91
+ "DreizehnNS": 0
92
+ }
93
+ }
so-vits-svc-4.0/20230325-DreizehnNS-44k/kmeans_10000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:742a89428e88e07a30f3bcb88e5a24444b3b628cf17130e15b3017f9c60b8272
3
+ size 15457593
so-vits-svc-4.0/20230325-DreizehnNS-44k/pruned/P_G_13600.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d603d1a052d95afc6ecab31f9f633701a72f05c621f40ede6f3d74054cc9fce9
3
+ size 180885829
so-vits-svc-4.0/20230325-DreizehnNS-44k/pruned/config.json ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "train": {
3
+ "log_interval": 200,
4
+ "eval_interval": 800,
5
+ "seed": 1234,
6
+ "epochs": 10000,
7
+ "learning_rate": 0.0001,
8
+ "betas": [
9
+ 0.8,
10
+ 0.99
11
+ ],
12
+ "eps": 1e-09,
13
+ "batch_size": 6,
14
+ "fp16_run": false,
15
+ "lr_decay": 0.999875,
16
+ "segment_size": 10240,
17
+ "init_lr_ratio": 1,
18
+ "warmup_epochs": 0,
19
+ "c_mel": 45,
20
+ "c_kl": 1.0,
21
+ "use_sr": true,
22
+ "max_speclen": 512,
23
+ "port": "8001",
24
+ "keep_ckpts": 99
25
+ },
26
+ "data": {
27
+ "training_files": "filelists/train.txt",
28
+ "validation_files": "filelists/val.txt",
29
+ "max_wav_value": 32768.0,
30
+ "sampling_rate": 44100,
31
+ "filter_length": 2048,
32
+ "hop_length": 512,
33
+ "win_length": 2048,
34
+ "n_mel_channels": 80,
35
+ "mel_fmin": 0.0,
36
+ "mel_fmax": 22050
37
+ },
38
+ "model": {
39
+ "inter_channels": 192,
40
+ "hidden_channels": 192,
41
+ "filter_channels": 768,
42
+ "n_heads": 2,
43
+ "n_layers": 6,
44
+ "kernel_size": 3,
45
+ "p_dropout": 0.1,
46
+ "resblock": "1",
47
+ "resblock_kernel_sizes": [
48
+ 3,
49
+ 7,
50
+ 11
51
+ ],
52
+ "resblock_dilation_sizes": [
53
+ [
54
+ 1,
55
+ 3,
56
+ 5
57
+ ],
58
+ [
59
+ 1,
60
+ 3,
61
+ 5
62
+ ],
63
+ [
64
+ 1,
65
+ 3,
66
+ 5
67
+ ]
68
+ ],
69
+ "upsample_rates": [
70
+ 8,
71
+ 8,
72
+ 2,
73
+ 2,
74
+ 2
75
+ ],
76
+ "upsample_initial_channel": 512,
77
+ "upsample_kernel_sizes": [
78
+ 16,
79
+ 16,
80
+ 4,
81
+ 4,
82
+ 4
83
+ ],
84
+ "n_layers_q": 3,
85
+ "use_spectral_norm": false,
86
+ "gin_channels": 256,
87
+ "ssl_dim": 256,
88
+ "n_speakers": 200
89
+ },
90
+ "spk": {
91
+ "DreizehnNS": 0
92
+ }
93
+ }
so-vits-svc-4.0/20230325-DreizehnNS-44k/train.log ADDED
@@ -0,0 +1,592 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2023-03-25 16:47:27,269 44k INFO {'train': {'log_interval': 200, 'eval_interval': 800, 'seed': 1234, 'epochs': 10000, 'learning_rate': 0.0001, 'betas': [0.8, 0.99], 'eps': 1e-09, 'batch_size': 6, 'fp16_run': False, 'lr_decay': 0.999875, 'segment_size': 10240, 'init_lr_ratio': 1, 'warmup_epochs': 0, 'c_mel': 45, 'c_kl': 1.0, 'use_sr': True, 'max_speclen': 512, 'port': '8001', 'keep_ckpts': 99}, 'data': {'training_files': 'filelists/train.txt', 'validation_files': 'filelists/val.txt', 'max_wav_value': 32768.0, 'sampling_rate': 44100, 'filter_length': 2048, 'hop_length': 512, 'win_length': 2048, 'n_mel_channels': 80, 'mel_fmin': 0.0, 'mel_fmax': 22050}, 'model': {'inter_channels': 192, 'hidden_channels': 192, 'filter_channels': 768, 'n_heads': 2, 'n_layers': 6, 'kernel_size': 3, 'p_dropout': 0.1, 'resblock': '1', 'resblock_kernel_sizes': [3, 7, 11], 'resblock_dilation_sizes': [[1, 3, 5], [1, 3, 5], [1, 3, 5]], 'upsample_rates': [8, 8, 2, 2, 2], 'upsample_initial_channel': 512, 'upsample_kernel_sizes': [16, 16, 4, 4, 4], 'n_layers_q': 3, 'use_spectral_norm': False, 'gin_channels': 256, 'ssl_dim': 256, 'n_speakers': 200}, 'spk': {'DreizehnNS': 0}, 'model_dir': './logs/44k'}
2
+ 2023-03-25 16:47:27,269 44k WARNING /root/so-vits-svc-4.0 is not a git repository, therefore hash value comparison will be ignored.
3
+ 2023-03-25 16:47:30,338 44k INFO emb_g.weight is not in the checkpoint
4
+ 2023-03-25 16:47:30,394 44k INFO Loaded checkpoint './logs/44k/G_0.pth' (iteration 0)
5
+ 2023-03-25 16:47:30,551 44k INFO Loaded checkpoint './logs/44k/D_0.pth' (iteration 0)
6
+ 2023-03-25 16:47:38,222 44k INFO Train Epoch: 1 [0%]
7
+ 2023-03-25 16:47:38,223 44k INFO Losses: [2.6974782943725586, 2.4092516899108887, 14.672739028930664, 37.46104431152344, 4.290388584136963], step: 0, lr: 0.0001
8
+ 2023-03-25 16:47:43,503 44k INFO Saving model and optimizer state at iteration 1 to ./logs/44k/G_0.pth
9
+ 2023-03-25 16:47:44,765 44k INFO Saving model and optimizer state at iteration 1 to ./logs/44k/D_0.pth
10
+ 2023-03-25 16:48:02,181 44k INFO ====> Epoch: 1, cost 34.91 s
11
+ 2023-03-25 16:48:19,046 44k INFO ====> Epoch: 2, cost 16.86 s
12
+ 2023-03-25 16:48:35,900 44k INFO ====> Epoch: 3, cost 16.85 s
13
+ 2023-03-25 16:48:53,235 44k INFO ====> Epoch: 4, cost 17.33 s
14
+ 2023-03-25 16:49:09,884 44k INFO ====> Epoch: 5, cost 16.65 s
15
+ 2023-03-25 16:49:26,410 44k INFO ====> Epoch: 6, cost 16.53 s
16
+ 2023-03-25 16:49:30,788 44k INFO Train Epoch: 7 [6%]
17
+ 2023-03-25 16:49:30,790 44k INFO Losses: [2.2324411869049072, 2.718120574951172, 8.101375579833984, 15.547173500061035, 0.7784972190856934], step: 200, lr: 9.99250234335941e-05
18
+ 2023-03-25 16:49:43,790 44k INFO ====> Epoch: 7, cost 17.38 s
19
+ 2023-03-25 16:50:00,579 44k INFO ====> Epoch: 8, cost 16.79 s
20
+ 2023-03-25 16:50:17,010 44k INFO ====> Epoch: 9, cost 16.43 s
21
+ 2023-03-25 16:50:33,841 44k INFO ====> Epoch: 10, cost 16.83 s
22
+ 2023-03-25 16:50:50,880 44k INFO ====> Epoch: 11, cost 17.04 s
23
+ 2023-03-25 16:51:07,417 44k INFO ====> Epoch: 12, cost 16.54 s
24
+ 2023-03-25 16:51:12,576 44k INFO Train Epoch: 13 [12%]
25
+ 2023-03-25 16:51:12,577 44k INFO Losses: [2.325521469116211, 2.3794703483581543, 8.109458923339844, 15.86378002166748, 1.0891475677490234], step: 400, lr: 9.98501030820433e-05
26
+ 2023-03-25 16:51:25,360 44k INFO ====> Epoch: 13, cost 17.94 s
27
+ 2023-03-25 16:51:42,186 44k INFO ====> Epoch: 14, cost 16.83 s
28
+ 2023-03-25 16:51:59,000 44k INFO ====> Epoch: 15, cost 16.81 s
29
+ 2023-03-25 16:52:15,819 44k INFO ====> Epoch: 16, cost 16.82 s
30
+ 2023-03-25 16:52:32,638 44k INFO ====> Epoch: 17, cost 16.82 s
31
+ 2023-03-25 16:52:49,372 44k INFO ====> Epoch: 18, cost 16.73 s
32
+ 2023-03-25 16:52:55,578 44k INFO Train Epoch: 19 [18%]
33
+ 2023-03-25 16:52:55,579 44k INFO Losses: [2.597564220428467, 2.6969094276428223, 12.184596061706543, 19.04696273803711, 0.6627763509750366], step: 600, lr: 9.977523890319963e-05
34
+ 2023-03-25 16:53:06,934 44k INFO ====> Epoch: 19, cost 17.56 s
35
+ 2023-03-25 16:53:23,846 44k INFO ====> Epoch: 20, cost 16.91 s
36
+ 2023-03-25 16:53:40,397 44k INFO ====> Epoch: 21, cost 16.55 s
37
+ 2023-03-25 16:53:57,546 44k INFO ====> Epoch: 22, cost 17.15 s
38
+ 2023-03-25 16:54:14,372 44k INFO ====> Epoch: 23, cost 16.83 s
39
+ 2023-03-25 16:54:31,052 44k INFO ====> Epoch: 24, cost 16.68 s
40
+ 2023-03-25 16:54:37,717 44k INFO Train Epoch: 25 [24%]
41
+ 2023-03-25 16:54:37,718 44k INFO Losses: [2.5308003425598145, 2.7162580490112305, 9.232569694519043, 15.944351196289062, 0.7880488038063049], step: 800, lr: 9.970043085494672e-05
42
+ 2023-03-25 16:54:42,256 44k INFO Saving model and optimizer state at iteration 25 to ./logs/44k/G_800.pth
43
+ 2023-03-25 16:54:43,576 44k INFO Saving model and optimizer state at iteration 25 to ./logs/44k/D_800.pth
44
+ 2023-03-25 16:54:54,478 44k INFO ====> Epoch: 25, cost 23.43 s
45
+ 2023-03-25 16:55:11,205 44k INFO ====> Epoch: 26, cost 16.73 s
46
+ 2023-03-25 16:55:28,258 44k INFO ====> Epoch: 27, cost 17.05 s
47
+ 2023-03-25 16:55:44,954 44k INFO ====> Epoch: 28, cost 16.70 s
48
+ 2023-03-25 16:56:01,987 44k INFO ====> Epoch: 29, cost 17.03 s
49
+ 2023-03-25 16:56:18,844 44k INFO ====> Epoch: 30, cost 16.86 s
50
+ 2023-03-25 16:56:26,243 44k INFO Train Epoch: 31 [30%]
51
+ 2023-03-25 16:56:26,244 44k INFO Losses: [2.428475856781006, 2.539283037185669, 9.996885299682617, 21.852493286132812, 0.8732186555862427], step: 1000, lr: 9.962567889519979e-05
52
+ 2023-03-25 16:56:36,533 44k INFO ====> Epoch: 31, cost 17.69 s
53
+ 2023-03-25 16:56:53,197 44k INFO ====> Epoch: 32, cost 16.66 s
54
+ 2023-03-25 16:57:09,628 44k INFO ====> Epoch: 33, cost 16.43 s
55
+ 2023-03-25 16:57:26,475 44k INFO ====> Epoch: 34, cost 16.85 s
56
+ 2023-03-25 16:57:42,775 44k INFO ====> Epoch: 35, cost 16.30 s
57
+ 2023-03-25 16:57:59,365 44k INFO ====> Epoch: 36, cost 16.59 s
58
+ 2023-03-25 16:58:07,766 44k INFO Train Epoch: 37 [36%]
59
+ 2023-03-25 16:58:07,767 44k INFO Losses: [2.5228097438812256, 2.288203001022339, 11.588765144348145, 20.175125122070312, 0.6351476311683655], step: 1200, lr: 9.95509829819056e-05
60
+ 2023-03-25 16:58:16,719 44k INFO ====> Epoch: 37, cost 17.35 s
61
+ 2023-03-25 16:58:33,397 44k INFO ====> Epoch: 38, cost 16.68 s
62
+ 2023-03-25 16:58:49,877 44k INFO ====> Epoch: 39, cost 16.48 s
63
+ 2023-03-25 16:59:06,535 44k INFO ====> Epoch: 40, cost 16.66 s
64
+ 2023-03-25 16:59:23,000 44k INFO ====> Epoch: 41, cost 16.47 s
65
+ 2023-03-25 16:59:40,216 44k INFO ====> Epoch: 42, cost 17.22 s
66
+ 2023-03-25 16:59:49,295 44k INFO Train Epoch: 43 [42%]
67
+ 2023-03-25 16:59:49,297 44k INFO Losses: [2.5152947902679443, 2.0511679649353027, 7.733269214630127, 15.723326683044434, 0.8497045040130615], step: 1400, lr: 9.947634307304244e-05
68
+ 2023-03-25 16:59:58,173 44k INFO ====> Epoch: 43, cost 17.96 s
69
+ 2023-03-25 17:00:14,901 44k INFO ====> Epoch: 44, cost 16.73 s
70
+ 2023-03-25 17:00:31,479 44k INFO ====> Epoch: 45, cost 16.58 s
71
+ 2023-03-25 17:00:48,089 44k INFO ====> Epoch: 46, cost 16.61 s
72
+ 2023-03-25 17:01:04,848 44k INFO ====> Epoch: 47, cost 16.76 s
73
+ 2023-03-25 17:01:21,377 44k INFO ====> Epoch: 48, cost 16.53 s
74
+ 2023-03-25 17:01:31,234 44k INFO Train Epoch: 49 [48%]
75
+ 2023-03-25 17:01:31,235 44k INFO Losses: [2.0633249282836914, 2.451162576675415, 13.993906021118164, 21.134328842163086, 0.9099781513214111], step: 1600, lr: 9.940175912662009e-05
76
+ 2023-03-25 17:01:35,874 44k INFO Saving model and optimizer state at iteration 49 to ./logs/44k/G_1600.pth
77
+ 2023-03-25 17:01:37,075 44k INFO Saving model and optimizer state at iteration 49 to ./logs/44k/D_1600.pth
78
+ 2023-03-25 17:01:44,622 44k INFO ====> Epoch: 49, cost 23.24 s
79
+ 2023-03-25 17:02:01,509 44k INFO ====> Epoch: 50, cost 16.89 s
80
+ 2023-03-25 17:02:18,009 44k INFO ====> Epoch: 51, cost 16.50 s
81
+ 2023-03-25 17:02:35,156 44k INFO ====> Epoch: 52, cost 17.15 s
82
+ 2023-03-25 17:02:51,799 44k INFO ====> Epoch: 53, cost 16.64 s
83
+ 2023-03-25 17:03:08,333 44k INFO ====> Epoch: 54, cost 16.53 s
84
+ 2023-03-25 17:03:18,881 44k INFO Train Epoch: 55 [55%]
85
+ 2023-03-25 17:03:18,882 44k INFO Losses: [2.7328269481658936, 2.0261311531066895, 10.961967468261719, 17.05063819885254, 1.2199965715408325], step: 1800, lr: 9.932723110067987e-05
86
+ 2023-03-25 17:03:25,395 44k INFO ====> Epoch: 55, cost 17.06 s
87
+ 2023-03-25 17:03:41,918 44k INFO ====> Epoch: 56, cost 16.52 s
88
+ 2023-03-25 17:03:58,504 44k INFO ====> Epoch: 57, cost 16.59 s
89
+ 2023-03-25 17:04:15,265 44k INFO ====> Epoch: 58, cost 16.76 s
90
+ 2023-03-25 17:04:31,725 44k INFO ====> Epoch: 59, cost 16.46 s
91
+ 2023-03-25 17:04:48,454 44k INFO ====> Epoch: 60, cost 16.73 s
92
+ 2023-03-25 17:04:59,685 44k INFO Train Epoch: 61 [61%]
93
+ 2023-03-25 17:04:59,686 44k INFO Losses: [2.453200101852417, 2.5324313640594482, 10.589628219604492, 22.162248611450195, 1.0994495153427124], step: 2000, lr: 9.92527589532945e-05
94
+ 2023-03-25 17:05:05,700 44k INFO ====> Epoch: 61, cost 17.25 s
95
+ 2023-03-25 17:05:22,609 44k INFO ====> Epoch: 62, cost 16.91 s
96
+ 2023-03-25 17:05:39,301 44k INFO ====> Epoch: 63, cost 16.69 s
97
+ 2023-03-25 17:05:55,873 44k INFO ====> Epoch: 64, cost 16.57 s
98
+ 2023-03-25 17:06:13,560 44k INFO ====> Epoch: 65, cost 17.69 s
99
+ 2023-03-25 17:06:30,219 44k INFO ====> Epoch: 66, cost 16.66 s
100
+ 2023-03-25 17:06:42,692 44k INFO Train Epoch: 67 [67%]
101
+ 2023-03-25 17:06:42,693 44k INFO Losses: [2.295553684234619, 2.4008660316467285, 9.892032623291016, 17.572246551513672, 0.5077621936798096], step: 2200, lr: 9.917834264256819e-05
102
+ 2023-03-25 17:06:47,857 44k INFO ====> Epoch: 67, cost 17.64 s
103
+ 2023-03-25 17:07:04,741 44k INFO ====> Epoch: 68, cost 16.88 s
104
+ 2023-03-25 17:07:21,623 44k INFO ====> Epoch: 69, cost 16.88 s
105
+ 2023-03-25 17:07:38,193 44k INFO ====> Epoch: 70, cost 16.57 s
106
+ 2023-03-25 17:07:54,870 44k INFO ====> Epoch: 71, cost 16.68 s
107
+ 2023-03-25 17:08:11,664 44k INFO ====> Epoch: 72, cost 16.79 s
108
+ 2023-03-25 17:08:25,468 44k INFO Train Epoch: 73 [73%]
109
+ 2023-03-25 17:08:25,470 44k INFO Losses: [2.599912405014038, 2.0814945697784424, 6.905517101287842, 13.569595336914062, 0.7708091735839844], step: 2400, lr: 9.910398212663652e-05
110
+ 2023-03-25 17:08:29,876 44k INFO Saving model and optimizer state at iteration 73 to ./logs/44k/G_2400.pth
111
+ 2023-03-25 17:08:31,092 44k INFO Saving model and optimizer state at iteration 73 to ./logs/44k/D_2400.pth
112
+ 2023-03-25 17:08:35,432 44k INFO ====> Epoch: 73, cost 23.77 s
113
+ 2023-03-25 17:08:51,850 44k INFO ====> Epoch: 74, cost 16.42 s
114
+ 2023-03-25 17:09:08,371 44k INFO ====> Epoch: 75, cost 16.52 s
115
+ 2023-03-25 17:09:25,135 44k INFO ====> Epoch: 76, cost 16.76 s
116
+ 2023-03-25 17:09:41,551 44k INFO ====> Epoch: 77, cost 16.42 s
117
+ 2023-03-25 17:09:58,751 44k INFO ====> Epoch: 78, cost 17.20 s
118
+ 2023-03-25 17:10:12,408 44k INFO Train Epoch: 79 [79%]
119
+ 2023-03-25 17:10:12,409 44k INFO Losses: [2.381700038909912, 2.7567107677459717, 7.031135082244873, 11.056401252746582, 0.745064914226532], step: 2600, lr: 9.902967736366644e-05
120
+ 2023-03-25 17:10:15,809 44k INFO ====> Epoch: 79, cost 17.06 s
121
+ 2023-03-25 17:10:32,728 44k INFO ====> Epoch: 80, cost 16.92 s
122
+ 2023-03-25 17:10:49,486 44k INFO ====> Epoch: 81, cost 16.76 s
123
+ 2023-03-25 17:11:06,316 44k INFO ====> Epoch: 82, cost 16.83 s
124
+ 2023-03-25 17:11:23,200 44k INFO ====> Epoch: 83, cost 16.88 s
125
+ 2023-03-25 17:11:39,866 44k INFO ====> Epoch: 84, cost 16.67 s
126
+ 2023-03-25 17:11:54,418 44k INFO Train Epoch: 85 [85%]
127
+ 2023-03-25 17:11:54,419 44k INFO Losses: [2.4440665245056152, 2.255263328552246, 6.086838722229004, 10.995237350463867, 0.8823586106300354], step: 2800, lr: 9.895542831185631e-05
128
+ 2023-03-25 17:11:57,170 44k INFO ====> Epoch: 85, cost 17.30 s
129
+ 2023-03-25 17:12:14,015 44k INFO ====> Epoch: 86, cost 16.84 s
130
+ 2023-03-25 17:12:31,117 44k INFO ====> Epoch: 87, cost 17.10 s
131
+ 2023-03-25 17:12:47,612 44k INFO ====> Epoch: 88, cost 16.50 s
132
+ 2023-03-25 17:13:04,423 44k INFO ====> Epoch: 89, cost 16.81 s
133
+ 2023-03-25 17:13:21,588 44k INFO ====> Epoch: 90, cost 17.17 s
134
+ 2023-03-25 17:13:36,796 44k INFO Train Epoch: 91 [91%]
135
+ 2023-03-25 17:13:36,797 44k INFO Losses: [2.305187940597534, 2.6297731399536133, 10.468490600585938, 22.139997482299805, 0.8177512288093567], step: 3000, lr: 9.888123492943583e-05
136
+ 2023-03-25 17:13:38,623 44k INFO ====> Epoch: 91, cost 17.04 s
137
+ 2023-03-25 17:13:55,148 44k INFO ====> Epoch: 92, cost 16.52 s
138
+ 2023-03-25 17:14:12,020 44k INFO ====> Epoch: 93, cost 16.87 s
139
+ 2023-03-25 17:14:28,756 44k INFO ====> Epoch: 94, cost 16.74 s
140
+ 2023-03-25 17:14:45,594 44k INFO ====> Epoch: 95, cost 16.84 s
141
+ 2023-03-25 17:15:02,370 44k INFO ====> Epoch: 96, cost 16.78 s
142
+ 2023-03-25 17:15:18,311 44k INFO Train Epoch: 97 [97%]
143
+ 2023-03-25 17:15:18,312 44k INFO Losses: [2.6312460899353027, 2.329620838165283, 7.526452541351318, 11.057599067687988, 0.6342429518699646], step: 3200, lr: 9.880709717466598e-05
144
+ 2023-03-25 17:15:22,926 44k INFO Saving model and optimizer state at iteration 97 to ./logs/44k/G_3200.pth
145
+ 2023-03-25 17:15:24,087 44k INFO Saving model and optimizer state at iteration 97 to ./logs/44k/D_3200.pth
146
+ 2023-03-25 17:15:25,306 44k INFO ====> Epoch: 97, cost 22.94 s
147
+ 2023-03-25 17:15:42,129 44k INFO ====> Epoch: 98, cost 16.82 s
148
+ 2023-03-25 17:15:58,976 44k INFO ====> Epoch: 99, cost 16.85 s
149
+ 2023-03-25 17:16:15,653 44k INFO ====> Epoch: 100, cost 16.68 s
150
+ 2023-03-25 17:16:32,914 44k INFO ====> Epoch: 101, cost 17.26 s
151
+ 2023-03-25 17:16:50,033 44k INFO ====> Epoch: 102, cost 17.12 s
152
+ 2023-03-25 17:17:06,781 44k INFO ====> Epoch: 103, cost 16.75 s
153
+ 2023-03-25 17:17:10,819 44k INFO Train Epoch: 104 [3%]
154
+ 2023-03-25 17:17:10,821 44k INFO Losses: [2.338289737701416, 2.391996145248413, 10.017964363098145, 17.78423309326172, 0.8379142880439758], step: 3400, lr: 9.872067337896332e-05
155
+ 2023-03-25 17:17:24,042 44k INFO ====> Epoch: 104, cost 17.26 s
156
+ 2023-03-25 17:17:40,807 44k INFO ====> Epoch: 105, cost 16.76 s
157
+ 2023-03-25 17:17:57,366 44k INFO ====> Epoch: 106, cost 16.56 s
158
+ 2023-03-25 17:18:14,377 44k INFO ====> Epoch: 107, cost 17.01 s
159
+ 2023-03-25 17:18:31,046 44k INFO ====> Epoch: 108, cost 16.67 s
160
+ 2023-03-25 17:18:47,635 44k INFO ====> Epoch: 109, cost 16.59 s
161
+ 2023-03-25 17:18:52,568 44k INFO Train Epoch: 110 [9%]
162
+ 2023-03-25 17:18:52,569 44k INFO Losses: [2.822373390197754, 2.349123239517212, 4.166832447052002, 13.0591459274292, 1.1201118230819702], step: 3600, lr: 9.864665600773098e-05
163
+ 2023-03-25 17:19:04,926 44k INFO ====> Epoch: 110, cost 17.29 s
164
+ 2023-03-25 17:19:21,696 44k INFO ====> Epoch: 111, cost 16.77 s
165
+ 2023-03-25 17:19:38,208 44k INFO ====> Epoch: 112, cost 16.51 s
166
+ 2023-03-25 17:19:54,872 44k INFO ====> Epoch: 113, cost 16.66 s
167
+ 2023-03-25 17:20:11,555 44k INFO ====> Epoch: 114, cost 16.68 s
168
+ 2023-03-25 17:20:28,509 44k INFO ====> Epoch: 115, cost 16.95 s
169
+ 2023-03-25 17:20:34,295 44k INFO Train Epoch: 116 [15%]
170
+ 2023-03-25 17:20:34,296 44k INFO Losses: [2.072659492492676, 2.792006015777588, 11.296175003051758, 15.94166374206543, 0.21819163858890533], step: 3800, lr: 9.857269413218213e-05
171
+ 2023-03-25 17:20:46,053 44k INFO ====> Epoch: 116, cost 17.54 s
172
+ 2023-03-25 17:21:03,288 44k INFO ====> Epoch: 117, cost 17.24 s
173
+ 2023-03-25 17:21:19,708 44k INFO ====> Epoch: 118, cost 16.42 s
174
+ 2023-03-25 17:21:36,111 44k INFO ====> Epoch: 119, cost 16.40 s
175
+ 2023-03-25 17:21:52,543 44k INFO ====> Epoch: 120, cost 16.43 s
176
+ 2023-03-25 17:22:09,435 44k INFO ====> Epoch: 121, cost 16.89 s
177
+ 2023-03-25 17:22:15,757 44k INFO Train Epoch: 122 [21%]
178
+ 2023-03-25 17:22:15,758 44k INFO Losses: [2.796874523162842, 2.1552910804748535, 7.83256196975708, 11.650614738464355, 0.48031163215637207], step: 4000, lr: 9.8498787710708e-05
179
+ 2023-03-25 17:22:20,248 44k INFO Saving model and optimizer state at iteration 122 to ./logs/44k/G_4000.pth
180
+ 2023-03-25 17:22:21,400 44k INFO Saving model and optimizer state at iteration 122 to ./logs/44k/D_4000.pth
181
+ 2023-03-25 17:22:33,009 44k INFO ====> Epoch: 122, cost 23.57 s
182
+ 2023-03-25 17:22:50,276 44k INFO ====> Epoch: 123, cost 17.27 s
183
+ 2023-03-25 17:23:06,706 44k INFO ====> Epoch: 124, cost 16.43 s
184
+ 2023-03-25 17:23:23,424 44k INFO ====> Epoch: 125, cost 16.72 s
185
+ 2023-03-25 17:23:39,997 44k INFO ====> Epoch: 126, cost 16.57 s
186
+ 2023-03-25 17:23:56,588 44k INFO ====> Epoch: 127, cost 16.59 s
187
+ 2023-03-25 17:24:03,701 44k INFO Train Epoch: 128 [27%]
188
+ 2023-03-25 17:24:03,702 44k INFO Losses: [2.1375842094421387, 3.02132511138916, 10.848799705505371, 14.873762130737305, 0.668673574924469], step: 4200, lr: 9.842493670173108e-05
189
+ 2023-03-25 17:24:13,908 44k INFO ====> Epoch: 128, cost 17.32 s
190
+ 2023-03-25 17:24:30,679 44k INFO ====> Epoch: 129, cost 16.77 s
191
+ 2023-03-25 17:24:47,504 44k INFO ====> Epoch: 130, cost 16.82 s
192
+ 2023-03-25 17:25:04,152 44k INFO ====> Epoch: 131, cost 16.65 s
193
+ 2023-03-25 17:25:21,742 44k INFO ====> Epoch: 132, cost 17.59 s
194
+ 2023-03-25 17:25:39,064 44k INFO ====> Epoch: 133, cost 17.32 s
195
+ 2023-03-25 17:25:46,939 44k INFO Train Epoch: 134 [33%]
196
+ 2023-03-25 17:25:46,940 44k INFO Losses: [2.4770865440368652, 2.198981761932373, 5.717620849609375, 12.555487632751465, 0.7245680093765259], step: 4400, lr: 9.835114106370493e-05
197
+ 2023-03-25 17:25:56,287 44k INFO ====> Epoch: 134, cost 17.22 s
198
+ 2023-03-25 17:26:13,283 44k INFO ====> Epoch: 135, cost 17.00 s
199
+ 2023-03-25 17:26:29,684 44k INFO ====> Epoch: 136, cost 16.40 s
200
+ 2023-03-25 17:26:46,292 44k INFO ====> Epoch: 137, cost 16.61 s
201
+ 2023-03-25 17:27:03,013 44k INFO ====> Epoch: 138, cost 16.72 s
202
+ 2023-03-25 17:27:19,649 44k INFO ====> Epoch: 139, cost 16.64 s
203
+ 2023-03-25 17:27:28,415 44k INFO Train Epoch: 140 [39%]
204
+ 2023-03-25 17:27:28,417 44k INFO Losses: [2.2845919132232666, 3.015265941619873, 8.828485488891602, 16.861722946166992, 1.0159695148468018], step: 4600, lr: 9.827740075511432e-05
205
+ 2023-03-25 17:27:37,169 44k INFO ====> Epoch: 140, cost 17.52 s
206
+ 2023-03-25 17:27:53,821 44k INFO ====> Epoch: 141, cost 16.65 s
207
+ 2023-03-25 17:28:10,387 44k INFO ====> Epoch: 142, cost 16.57 s
208
+ 2023-03-25 17:28:27,169 44k INFO ====> Epoch: 143, cost 16.78 s
209
+ 2023-03-25 17:28:43,398 44k INFO ====> Epoch: 144, cost 16.23 s
210
+ 2023-03-25 17:28:59,928 44k INFO ====> Epoch: 145, cost 16.53 s
211
+ 2023-03-25 17:29:09,400 44k INFO Train Epoch: 146 [45%]
212
+ 2023-03-25 17:29:09,401 44k INFO Losses: [2.016758918762207, 2.6839516162872314, 10.109888076782227, 12.951353073120117, 0.9170368909835815], step: 4800, lr: 9.820371573447515e-05
213
+ 2023-03-25 17:29:13,880 44k INFO Saving model and optimizer state at iteration 146 to ./logs/44k/G_4800.pth
214
+ 2023-03-25 17:29:15,181 44k INFO Saving model and optimizer state at iteration 146 to ./logs/44k/D_4800.pth
215
+ 2023-03-25 17:29:23,012 44k INFO ====> Epoch: 146, cost 23.08 s
216
+ 2023-03-25 17:29:39,655 44k INFO ====> Epoch: 147, cost 16.64 s
217
+ 2023-03-25 17:29:56,259 44k INFO ====> Epoch: 148, cost 16.60 s
218
+ 2023-03-25 17:30:13,013 44k INFO ====> Epoch: 149, cost 16.75 s
219
+ 2023-03-25 17:30:29,995 44k INFO ====> Epoch: 150, cost 16.98 s
220
+ 2023-03-25 17:30:46,364 44k INFO ====> Epoch: 151, cost 16.37 s
221
+ 2023-03-25 17:30:56,410 44k INFO Train Epoch: 152 [52%]
222
+ 2023-03-25 17:30:56,411 44k INFO Losses: [2.8300862312316895, 2.162816286087036, 7.240455150604248, 13.705415725708008, 0.9281750321388245], step: 5000, lr: 9.813008596033443e-05
223
+ 2023-03-25 17:31:03,427 44k INFO ====> Epoch: 152, cost 17.06 s
224
+ 2023-03-25 17:31:20,293 44k INFO ====> Epoch: 153, cost 16.87 s
225
+ 2023-03-25 17:31:36,853 44k INFO ====> Epoch: 154, cost 16.56 s
226
+ 2023-03-25 17:31:53,575 44k INFO ====> Epoch: 155, cost 16.72 s
227
+ 2023-03-25 17:32:10,656 44k INFO ====> Epoch: 156, cost 17.08 s
228
+ 2023-03-25 17:32:27,958 44k INFO ====> Epoch: 157, cost 17.30 s
229
+ 2023-03-25 17:32:39,382 44k INFO Train Epoch: 158 [58%]
230
+ 2023-03-25 17:32:39,383 44k INFO Losses: [2.3936424255371094, 2.3674676418304443, 9.325252532958984, 17.687271118164062, 0.44196417927742004], step: 5200, lr: 9.80565113912702e-05
231
+ 2023-03-25 17:32:45,648 44k INFO ====> Epoch: 158, cost 17.69 s
232
+ 2023-03-25 17:33:02,438 44k INFO ====> Epoch: 159, cost 16.79 s
233
+ 2023-03-25 17:33:19,033 44k INFO ====> Epoch: 160, cost 16.60 s
234
+ 2023-03-25 17:33:35,733 44k INFO ====> Epoch: 161, cost 16.70 s
235
+ 2023-03-25 17:33:52,506 44k INFO ====> Epoch: 162, cost 16.77 s
236
+ 2023-03-25 17:34:09,468 44k INFO ====> Epoch: 163, cost 16.96 s
237
+ 2023-03-25 17:34:21,265 44k INFO Train Epoch: 164 [64%]
238
+ 2023-03-25 17:34:21,267 44k INFO Losses: [2.301771879196167, 2.4344451427459717, 9.849881172180176, 19.910600662231445, 0.5363461971282959], step: 5400, lr: 9.798299198589162e-05
239
+ 2023-03-25 17:34:26,576 44k INFO ====> Epoch: 164, cost 17.11 s
240
+ 2023-03-25 17:34:43,835 44k INFO ====> Epoch: 165, cost 17.26 s
241
+ 2023-03-25 17:35:00,502 44k INFO ====> Epoch: 166, cost 16.67 s
242
+ 2023-03-25 17:35:17,098 44k INFO ====> Epoch: 167, cost 16.60 s
243
+ 2023-03-25 17:35:33,473 44k INFO ====> Epoch: 168, cost 16.38 s
244
+ 2023-03-25 17:35:50,479 44k INFO ====> Epoch: 169, cost 17.01 s
245
+ 2023-03-25 17:36:02,978 44k INFO Train Epoch: 170 [70%]
246
+ 2023-03-25 17:36:02,979 44k INFO Losses: [2.5066232681274414, 2.6114039421081543, 5.712583065032959, 10.28527545928955, 0.8232187628746033], step: 5600, lr: 9.790952770283884e-05
247
+ 2023-03-25 17:36:07,446 44k INFO Saving model and optimizer state at iteration 170 to ./logs/44k/G_5600.pth
248
+ 2023-03-25 17:36:08,606 44k INFO Saving model and optimizer state at iteration 170 to ./logs/44k/D_5600.pth
249
+ 2023-03-25 17:36:13,349 44k INFO ====> Epoch: 170, cost 22.87 s
250
+ 2023-03-25 17:36:30,174 44k INFO ====> Epoch: 171, cost 16.82 s
251
+ 2023-03-25 17:36:46,923 44k INFO ====> Epoch: 172, cost 16.75 s
252
+ 2023-03-25 17:37:03,606 44k INFO ====> Epoch: 173, cost 16.68 s
253
+ 2023-03-25 17:37:20,121 44k INFO ====> Epoch: 174, cost 16.51 s
254
+ 2023-03-25 17:37:36,799 44k INFO ====> Epoch: 175, cost 16.68 s
255
+ 2023-03-25 17:37:50,758 44k INFO Train Epoch: 176 [76%]
256
+ 2023-03-25 17:37:50,759 44k INFO Losses: [2.715217351913452, 2.0549123287200928, 6.569155693054199, 15.182696342468262, 0.7629079222679138], step: 5800, lr: 9.783611850078301e-05
257
+ 2023-03-25 17:37:54,542 44k INFO ====> Epoch: 176, cost 17.74 s
258
+ 2023-03-25 17:38:11,174 44k INFO ====> Epoch: 177, cost 16.63 s
259
+ 2023-03-25 17:38:28,217 44k INFO ====> Epoch: 178, cost 17.04 s
260
+ 2023-03-25 17:38:45,162 44k INFO ====> Epoch: 179, cost 16.95 s
261
+ 2023-03-25 17:39:01,783 44k INFO ====> Epoch: 180, cost 16.62 s
262
+ 2023-03-25 17:39:18,572 44k INFO ====> Epoch: 181, cost 16.79 s
263
+ 2023-03-25 17:39:32,891 44k INFO Train Epoch: 182 [82%]
264
+ 2023-03-25 17:39:32,893 44k INFO Losses: [2.1133475303649902, 2.5383996963500977, 11.810857772827148, 19.861572265625, 0.6742956042289734], step: 6000, lr: 9.776276433842631e-05
265
+ 2023-03-25 17:39:36,059 44k INFO ====> Epoch: 182, cost 17.49 s
266
+ 2023-03-25 17:39:53,021 44k INFO ====> Epoch: 183, cost 16.96 s
267
+ 2023-03-25 17:40:09,474 44k INFO ====> Epoch: 184, cost 16.45 s
268
+ 2023-03-25 17:40:25,931 44k INFO ====> Epoch: 185, cost 16.46 s
269
+ 2023-03-25 17:40:42,785 44k INFO ====> Epoch: 186, cost 16.85 s
270
+ 2023-03-25 17:40:59,892 44k INFO ====> Epoch: 187, cost 17.11 s
271
+ 2023-03-25 17:41:14,757 44k INFO Train Epoch: 188 [88%]
272
+ 2023-03-25 17:41:14,759 44k INFO Losses: [2.27668833732605, 2.766618490219116, 12.279425621032715, 20.261022567749023, 0.8307116031646729], step: 6200, lr: 9.768946517450186e-05
273
+ 2023-03-25 17:41:16,971 44k INFO ====> Epoch: 188, cost 17.08 s
274
+ 2023-03-25 17:41:34,110 44k INFO ====> Epoch: 189, cost 17.14 s
275
+ 2023-03-25 17:41:50,411 44k INFO ====> Epoch: 190, cost 16.30 s
276
+ 2023-03-25 17:42:07,037 44k INFO ====> Epoch: 191, cost 16.63 s
277
+ 2023-03-25 17:42:24,065 44k INFO ====> Epoch: 192, cost 17.03 s
278
+ 2023-03-25 17:42:40,495 44k INFO ====> Epoch: 193, cost 16.43 s
279
+ 2023-03-25 17:42:56,123 44k INFO Train Epoch: 194 [94%]
280
+ 2023-03-25 17:42:56,124 44k INFO Losses: [2.5047290325164795, 2.0337984561920166, 10.720749855041504, 17.247852325439453, 0.6880492568016052], step: 6400, lr: 9.761622096777372e-05
281
+ 2023-03-25 17:43:00,648 44k INFO Saving model and optimizer state at iteration 194 to ./logs/44k/G_6400.pth
282
+ 2023-03-25 17:43:01,842 44k INFO Saving model and optimizer state at iteration 194 to ./logs/44k/D_6400.pth
283
+ 2023-03-25 17:43:03,376 44k INFO ====> Epoch: 194, cost 22.88 s
284
+ 2023-03-25 17:43:19,858 44k INFO ====> Epoch: 195, cost 16.48 s
285
+ 2023-03-25 17:43:36,398 44k INFO ====> Epoch: 196, cost 16.54 s
286
+ 2023-03-25 17:43:53,393 44k INFO ====> Epoch: 197, cost 16.99 s
287
+ 2023-03-25 17:44:10,373 44k INFO ====> Epoch: 198, cost 16.98 s
288
+ 2023-03-25 17:44:27,149 44k INFO ====> Epoch: 199, cost 16.78 s
289
+ 2023-03-25 17:44:43,969 44k INFO ====> Epoch: 200, cost 16.82 s
290
+ 2023-03-25 17:44:47,513 44k INFO Train Epoch: 201 [0%]
291
+ 2023-03-25 17:44:47,514 44k INFO Losses: [2.6873631477355957, 2.1986629962921143, 6.34331750869751, 10.858728408813477, 0.6867925524711609], step: 6600, lr: 9.753083879807726e-05
292
+ 2023-03-25 17:45:01,125 44k INFO ====> Epoch: 201, cost 17.16 s
293
+ 2023-03-25 17:45:17,892 44k INFO ====> Epoch: 202, cost 16.77 s
294
+ 2023-03-25 17:45:34,752 44k INFO ====> Epoch: 203, cost 16.86 s
295
+ 2023-03-25 17:45:51,161 44k INFO ====> Epoch: 204, cost 16.41 s
296
+ 2023-03-25 17:46:08,008 44k INFO ====> Epoch: 205, cost 16.85 s
297
+ 2023-03-25 17:46:25,001 44k INFO ====> Epoch: 206, cost 16.99 s
298
+ 2023-03-25 17:46:29,316 44k INFO Train Epoch: 207 [6%]
299
+ 2023-03-25 17:46:29,317 44k INFO Losses: [2.3735992908477783, 2.3111236095428467, 6.5680999755859375, 11.549568176269531, 0.12051205337047577], step: 6800, lr: 9.745771352395957e-05
300
+ 2023-03-25 17:46:42,105 44k INFO ====> Epoch: 207, cost 17.10 s
301
+ 2023-03-25 17:46:58,695 44k INFO ====> Epoch: 208, cost 16.59 s
302
+ 2023-03-25 17:47:15,241 44k INFO ====> Epoch: 209, cost 16.55 s
303
+ 2023-03-25 17:47:31,924 44k INFO ====> Epoch: 210, cost 16.68 s
304
+ 2023-03-25 17:47:48,389 44k INFO ====> Epoch: 211, cost 16.47 s
305
+ 2023-03-25 17:48:05,112 44k INFO ====> Epoch: 212, cost 16.72 s
306
+ 2023-03-25 17:48:10,310 44k INFO Train Epoch: 213 [12%]
307
+ 2023-03-25 17:48:10,311 44k INFO Losses: [2.338576316833496, 2.1111223697662354, 9.061708450317383, 16.787063598632812, 0.529293954372406], step: 7000, lr: 9.73846430766616e-05
308
+ 2023-03-25 17:48:22,335 44k INFO ====> Epoch: 213, cost 17.22 s
309
+ 2023-03-25 17:48:38,741 44k INFO ====> Epoch: 214, cost 16.41 s
310
+ 2023-03-25 17:48:55,306 44k INFO ====> Epoch: 215, cost 16.56 s
311
+ 2023-03-25 17:49:12,085 44k INFO ====> Epoch: 216, cost 16.78 s
312
+ 2023-03-25 17:49:28,552 44k INFO ====> Epoch: 217, cost 16.47 s
313
+ 2023-03-25 17:49:44,519 44k INFO ====> Epoch: 218, cost 15.97 s
314
+ 2023-03-25 17:49:50,073 44k INFO Train Epoch: 219 [18%]
315
+ 2023-03-25 17:49:50,074 44k INFO Losses: [2.538870334625244, 1.9039925336837769, 8.920069694519043, 13.447583198547363, 0.2965475022792816], step: 7200, lr: 9.731162741507607e-05
316
+ 2023-03-25 17:49:54,374 44k INFO Saving model and optimizer state at iteration 219 to ./logs/44k/G_7200.pth
317
+ 2023-03-25 17:49:55,499 44k INFO Saving model and optimizer state at iteration 219 to ./logs/44k/D_7200.pth
318
+ 2023-03-25 17:50:06,295 44k INFO ====> Epoch: 219, cost 21.78 s
319
+ 2023-03-25 17:50:21,994 44k INFO ====> Epoch: 220, cost 15.70 s
320
+ 2023-03-25 17:50:37,662 44k INFO ====> Epoch: 221, cost 15.67 s
321
+ 2023-03-25 17:50:54,290 44k INFO ====> Epoch: 222, cost 16.63 s
322
+ 2023-03-25 17:51:10,823 44k INFO ====> Epoch: 223, cost 16.53 s
323
+ 2023-03-25 17:51:27,627 44k INFO ====> Epoch: 224, cost 16.80 s
324
+ 2023-03-25 17:51:33,887 44k INFO Train Epoch: 225 [24%]
325
+ 2023-03-25 17:51:33,888 44k INFO Losses: [2.269322395324707, 2.707637310028076, 11.994312286376953, 19.770584106445312, 0.7525858283042908], step: 7400, lr: 9.723866649812655e-05
326
+ 2023-03-25 17:51:44,009 44k INFO ====> Epoch: 225, cost 16.38 s
327
+ 2023-03-25 17:52:00,202 44k INFO ====> Epoch: 226, cost 16.19 s
328
+ 2023-03-25 17:52:16,282 44k INFO ====> Epoch: 227, cost 16.08 s
329
+ 2023-03-25 17:52:32,042 44k INFO ====> Epoch: 228, cost 15.76 s
330
+ 2023-03-25 17:52:47,741 44k INFO ====> Epoch: 229, cost 15.70 s
331
+ 2023-03-25 17:53:03,442 44k INFO ====> Epoch: 230, cost 15.70 s
332
+ 2023-03-25 17:53:10,435 44k INFO Train Epoch: 231 [30%]
333
+ 2023-03-25 17:53:10,436 44k INFO Losses: [2.1911444664001465, 2.909498691558838, 12.862323760986328, 18.82267951965332, 0.7391169667243958], step: 7600, lr: 9.716576028476738e-05
334
+ 2023-03-25 17:53:19,687 44k INFO ====> Epoch: 231, cost 16.25 s
335
+ 2023-03-25 17:53:37,561 44k INFO ====> Epoch: 232, cost 17.87 s
336
+ 2023-03-25 17:53:53,361 44k INFO ====> Epoch: 233, cost 15.80 s
337
+ 2023-03-25 17:54:11,201 44k INFO ====> Epoch: 234, cost 17.84 s
338
+ 2023-03-25 17:54:26,804 44k INFO ====> Epoch: 235, cost 15.60 s
339
+ 2023-03-25 17:54:42,860 44k INFO ====> Epoch: 236, cost 16.06 s
340
+ 2023-03-25 17:54:50,779 44k INFO Train Epoch: 237 [36%]
341
+ 2023-03-25 17:54:50,780 44k INFO Losses: [2.5400569438934326, 2.370675802230835, 7.975701332092285, 14.911909103393555, 0.5070781111717224], step: 7800, lr: 9.709290873398365e-05
342
+ 2023-03-25 17:54:59,921 44k INFO ====> Epoch: 237, cost 17.06 s
343
+ 2023-03-25 17:55:15,649 44k INFO ====> Epoch: 238, cost 15.73 s
344
+ 2023-03-25 17:55:31,290 44k INFO ====> Epoch: 239, cost 15.64 s
345
+ 2023-03-25 17:55:47,042 44k INFO ====> Epoch: 240, cost 15.75 s
346
+ 2023-03-25 17:56:02,659 44k INFO ====> Epoch: 241, cost 15.62 s
347
+ 2023-03-25 17:56:18,393 44k INFO ====> Epoch: 242, cost 15.73 s
348
+ 2023-03-25 17:56:26,769 44k INFO Train Epoch: 243 [42%]
349
+ 2023-03-25 17:56:26,770 44k INFO Losses: [2.5333597660064697, 2.1053004264831543, 8.916841506958008, 15.705123901367188, 0.7188340425491333], step: 8000, lr: 9.702011180479129e-05
350
+ 2023-03-25 17:56:30,936 44k INFO Saving model and optimizer state at iteration 243 to ./logs/44k/G_8000.pth
351
+ 2023-03-25 17:56:32,159 44k INFO Saving model and optimizer state at iteration 243 to ./logs/44k/D_8000.pth
352
+ 2023-03-25 17:56:39,973 44k INFO ====> Epoch: 243, cost 21.58 s
353
+ 2023-03-25 17:56:55,890 44k INFO ====> Epoch: 244, cost 15.92 s
354
+ 2023-03-25 17:57:11,491 44k INFO ====> Epoch: 245, cost 15.60 s
355
+ 2023-03-25 17:57:27,164 44k INFO ====> Epoch: 246, cost 15.67 s
356
+ 2023-03-25 17:57:43,077 44k INFO ====> Epoch: 247, cost 15.91 s
357
+ 2023-03-25 17:57:59,349 44k INFO ====> Epoch: 248, cost 16.27 s
358
+ 2023-03-25 17:58:08,985 44k INFO Train Epoch: 249 [48%]
359
+ 2023-03-25 17:58:08,987 44k INFO Losses: [2.553239345550537, 1.8566803932189941, 6.609829902648926, 16.236770629882812, 0.5468897819519043], step: 8200, lr: 9.694736945623688e-05
360
+ 2023-03-25 17:58:16,350 44k INFO ====> Epoch: 249, cost 17.00 s
361
+ 2023-03-25 17:58:32,729 44k INFO ====> Epoch: 250, cost 16.38 s
362
+ 2023-03-25 17:58:49,065 44k INFO ====> Epoch: 251, cost 16.34 s
363
+ 2023-03-25 17:59:05,556 44k INFO ====> Epoch: 252, cost 16.49 s
364
+ 2023-03-25 17:59:21,096 44k INFO ====> Epoch: 253, cost 15.54 s
365
+ 2023-03-25 17:59:36,732 44k INFO ====> Epoch: 254, cost 15.64 s
366
+ 2023-03-25 17:59:46,629 44k INFO Train Epoch: 255 [55%]
367
+ 2023-03-25 17:59:46,630 44k INFO Losses: [2.565453290939331, 1.9224674701690674, 7.469606876373291, 16.19024658203125, 0.5049387812614441], step: 8400, lr: 9.687468164739773e-05
368
+ 2023-03-25 17:59:53,023 44k INFO ====> Epoch: 255, cost 16.29 s
369
+ 2023-03-25 18:00:08,717 44k INFO ====> Epoch: 256, cost 15.69 s
370
+ 2023-03-25 18:00:24,316 44k INFO ====> Epoch: 257, cost 15.60 s
371
+ 2023-03-25 18:00:40,441 44k INFO ====> Epoch: 258, cost 16.13 s
372
+ 2023-03-25 18:00:57,101 44k INFO ====> Epoch: 259, cost 16.66 s
373
+ 2023-03-25 18:01:13,862 44k INFO ====> Epoch: 260, cost 16.76 s
374
+ 2023-03-25 18:01:25,107 44k INFO Train Epoch: 261 [61%]
375
+ 2023-03-25 18:01:25,109 44k INFO Losses: [2.4458837509155273, 2.323704719543457, 10.821569442749023, 17.358577728271484, 0.8134444952011108], step: 8600, lr: 9.680204833738185e-05
376
+ 2023-03-25 18:01:30,953 44k INFO ====> Epoch: 261, cost 17.09 s
377
+ 2023-03-25 18:01:47,414 44k INFO ====> Epoch: 262, cost 16.46 s
378
+ 2023-03-25 18:02:03,338 44k INFO ====> Epoch: 263, cost 15.92 s
379
+ 2023-03-25 18:02:21,478 44k INFO ====> Epoch: 264, cost 18.14 s
380
+ 2023-03-25 18:02:37,738 44k INFO ====> Epoch: 265, cost 16.26 s
381
+ 2023-03-25 18:02:54,410 44k INFO ====> Epoch: 266, cost 16.67 s
382
+ 2023-03-25 18:03:06,556 44k INFO Train Epoch: 267 [67%]
383
+ 2023-03-25 18:03:06,557 44k INFO Losses: [2.6201024055480957, 2.173233985900879, 5.443298816680908, 13.764752388000488, 0.6760918498039246], step: 8800, lr: 9.67294694853279e-05
384
+ 2023-03-25 18:03:11,333 44k INFO Saving model and optimizer state at iteration 267 to ./logs/44k/G_8800.pth
385
+ 2023-03-25 18:03:12,666 44k INFO Saving model and optimizer state at iteration 267 to ./logs/44k/D_8800.pth
386
+ 2023-03-25 18:03:17,845 44k INFO ====> Epoch: 267, cost 23.43 s
387
+ 2023-03-25 18:03:34,712 44k INFO ====> Epoch: 268, cost 16.87 s
388
+ 2023-03-25 18:03:51,312 44k INFO ====> Epoch: 269, cost 16.60 s
389
+ 2023-03-25 18:04:07,990 44k INFO ====> Epoch: 270, cost 16.68 s
390
+ 2023-03-25 18:04:24,400 44k INFO ====> Epoch: 271, cost 16.41 s
391
+ 2023-03-25 18:04:41,096 44k INFO ====> Epoch: 272, cost 16.70 s
392
+ 2023-03-25 18:04:54,136 44k INFO Train Epoch: 273 [73%]
393
+ 2023-03-25 18:04:54,136 44k INFO Losses: [2.3441848754882812, 2.227895736694336, 10.538599967956543, 14.83042049407959, 0.6608119010925293], step: 9000, lr: 9.665694505040515e-05
394
+ 2023-03-25 18:04:58,449 44k INFO ====> Epoch: 273, cost 17.35 s
395
+ 2023-03-25 18:05:14,970 44k INFO ====> Epoch: 274, cost 16.52 s
396
+ 2023-03-25 18:05:31,373 44k INFO ====> Epoch: 275, cost 16.40 s
397
+ 2023-03-25 18:05:47,843 44k INFO ====> Epoch: 276, cost 16.47 s
398
+ 2023-03-25 18:06:04,185 44k INFO ====> Epoch: 277, cost 16.34 s
399
+ 2023-03-25 18:06:20,638 44k INFO ====> Epoch: 278, cost 16.45 s
400
+ 2023-03-25 18:06:34,148 44k INFO Train Epoch: 279 [79%]
401
+ 2023-03-25 18:06:34,149 44k INFO Losses: [2.3495030403137207, 2.2592837810516357, 14.851377487182617, 18.4000244140625, 0.6196873188018799], step: 9200, lr: 9.658447499181352e-05
402
+ 2023-03-25 18:06:37,606 44k INFO ====> Epoch: 279, cost 16.97 s
403
+ 2023-03-25 18:06:54,350 44k INFO ====> Epoch: 280, cost 16.74 s
404
+ 2023-03-25 18:07:11,093 44k INFO ====> Epoch: 281, cost 16.74 s
405
+ 2023-03-25 18:07:27,706 44k INFO ====> Epoch: 282, cost 16.61 s
406
+ 2023-03-25 18:07:44,093 44k INFO ====> Epoch: 283, cost 16.39 s
407
+ 2023-03-25 18:08:00,560 44k INFO ====> Epoch: 284, cost 16.47 s
408
+ 2023-03-25 18:08:14,883 44k INFO Train Epoch: 285 [85%]
409
+ 2023-03-25 18:08:14,884 44k INFO Losses: [2.4534430503845215, 2.458757162094116, 7.7290263175964355, 14.926005363464355, 0.39300334453582764], step: 9400, lr: 9.651205926878348e-05
410
+ 2023-03-25 18:08:17,638 44k INFO ====> Epoch: 285, cost 17.08 s
411
+ 2023-03-25 18:08:34,163 44k INFO ====> Epoch: 286, cost 16.52 s
412
+ 2023-03-25 18:08:50,591 44k INFO ====> Epoch: 287, cost 16.43 s
413
+ 2023-03-25 18:09:07,061 44k INFO ====> Epoch: 288, cost 16.47 s
414
+ 2023-03-25 18:09:23,434 44k INFO ====> Epoch: 289, cost 16.37 s
415
+ 2023-03-25 18:09:39,824 44k INFO ====> Epoch: 290, cost 16.39 s
416
+ 2023-03-25 18:09:54,922 44k INFO Train Epoch: 291 [91%]
417
+ 2023-03-25 18:09:54,923 44k INFO Losses: [2.393284797668457, 2.443169116973877, 8.165253639221191, 16.708036422729492, 0.9939846396446228], step: 9600, lr: 9.643969784057613e-05
418
+ 2023-03-25 18:09:59,308 44k INFO Saving model and optimizer state at iteration 291 to ./logs/44k/G_9600.pth
419
+ 2023-03-25 18:10:00,733 44k INFO Saving model and optimizer state at iteration 291 to ./logs/44k/D_9600.pth
420
+ 2023-03-25 18:10:02,769 44k INFO ====> Epoch: 291, cost 22.95 s
421
+ 2023-03-25 18:10:19,195 44k INFO ====> Epoch: 292, cost 16.43 s
422
+ 2023-03-25 18:10:35,513 44k INFO ====> Epoch: 293, cost 16.32 s
423
+ 2023-03-25 18:10:52,152 44k INFO ====> Epoch: 294, cost 16.64 s
424
+ 2023-03-25 18:11:08,493 44k INFO ====> Epoch: 295, cost 16.34 s
425
+ 2023-03-25 18:11:24,789 44k INFO ====> Epoch: 296, cost 16.30 s
426
+ 2023-03-25 18:11:41,201 44k INFO Train Epoch: 297 [97%]
427
+ 2023-03-25 18:11:41,203 44k INFO Losses: [2.896554708480835, 2.316603660583496, 5.762091159820557, 10.353036880493164, 0.30002638697624207], step: 9800, lr: 9.636739066648303e-05
428
+ 2023-03-25 18:11:42,365 44k INFO ====> Epoch: 297, cost 17.58 s
429
+ 2023-03-25 18:11:58,961 44k INFO ====> Epoch: 298, cost 16.60 s
430
+ 2023-03-25 18:12:15,138 44k INFO ====> Epoch: 299, cost 16.18 s
431
+ 2023-03-25 18:12:31,489 44k INFO ====> Epoch: 300, cost 16.35 s
432
+ 2023-03-25 18:12:47,799 44k INFO ====> Epoch: 301, cost 16.31 s
433
+ 2023-03-25 18:13:03,973 44k INFO ====> Epoch: 302, cost 16.17 s
434
+ 2023-03-25 18:13:20,685 44k INFO ====> Epoch: 303, cost 16.71 s
435
+ 2023-03-25 18:13:24,919 44k INFO Train Epoch: 304 [3%]
436
+ 2023-03-25 18:13:24,920 44k INFO Losses: [2.515075922012329, 2.2078559398651123, 9.70125675201416, 15.418092727661133, 0.8266330361366272], step: 10000, lr: 9.628310081361311e-05
437
+ 2023-03-25 18:13:38,068 44k INFO ====> Epoch: 304, cost 17.38 s
438
+ 2023-03-25 18:13:54,461 44k INFO ====> Epoch: 305, cost 16.39 s
439
+ 2023-03-25 18:14:11,062 44k INFO ====> Epoch: 306, cost 16.60 s
440
+ 2023-03-25 18:14:27,536 44k INFO ====> Epoch: 307, cost 16.47 s
441
+ 2023-03-25 18:14:43,967 44k INFO ====> Epoch: 308, cost 16.43 s
442
+ 2023-03-25 18:15:00,505 44k INFO ====> Epoch: 309, cost 16.54 s
443
+ 2023-03-25 18:15:05,371 44k INFO Train Epoch: 310 [9%]
444
+ 2023-03-25 18:15:05,372 44k INFO Losses: [2.3045690059661865, 2.395820140838623, 7.718233108520508, 16.340299606323242, 0.6959915161132812], step: 10200, lr: 9.621091105059392e-05
445
+ 2023-03-25 18:15:17,643 44k INFO ====> Epoch: 310, cost 17.14 s
446
+ 2023-03-25 18:15:34,607 44k INFO ====> Epoch: 311, cost 16.96 s
447
+ 2023-03-25 18:15:50,909 44k INFO ====> Epoch: 312, cost 16.30 s
448
+ 2023-03-25 18:16:07,296 44k INFO ====> Epoch: 313, cost 16.39 s
449
+ 2023-03-25 18:16:23,498 44k INFO ====> Epoch: 314, cost 16.20 s
450
+ 2023-03-25 18:16:39,983 44k INFO ====> Epoch: 315, cost 16.48 s
451
+ 2023-03-25 18:16:45,429 44k INFO Train Epoch: 316 [15%]
452
+ 2023-03-25 18:16:45,430 44k INFO Losses: [2.462087392807007, 2.2239341735839844, 9.908827781677246, 15.508355140686035, 0.837641716003418], step: 10400, lr: 9.613877541298036e-05
453
+ 2023-03-25 18:16:49,915 44k INFO Saving model and optimizer state at iteration 316 to ./logs/44k/G_10400.pth
454
+ 2023-03-25 18:16:51,219 44k INFO Saving model and optimizer state at iteration 316 to ./logs/44k/D_10400.pth
455
+ 2023-03-25 18:17:02,702 44k INFO ====> Epoch: 316, cost 22.72 s
456
+ 2023-03-25 18:17:18,983 44k INFO ====> Epoch: 317, cost 16.28 s
457
+ 2023-03-25 18:17:35,311 44k INFO ====> Epoch: 318, cost 16.33 s
458
+ 2023-03-25 18:17:51,528 44k INFO ====> Epoch: 319, cost 16.22 s
459
+ 2023-03-25 18:18:07,950 44k INFO ====> Epoch: 320, cost 16.42 s
460
+ 2023-03-25 18:18:24,626 44k INFO ====> Epoch: 321, cost 16.68 s
461
+ 2023-03-25 18:18:30,769 44k INFO Train Epoch: 322 [21%]
462
+ 2023-03-25 18:18:30,770 44k INFO Losses: [2.3425683975219727, 2.628889322280884, 12.140241622924805, 20.59270477294922, 0.054636288434267044], step: 10600, lr: 9.606669386019102e-05
463
+ 2023-03-25 18:18:41,415 44k INFO ====> Epoch: 322, cost 16.79 s
464
+ 2023-03-25 18:18:57,759 44k INFO ====> Epoch: 323, cost 16.34 s
465
+ 2023-03-25 18:19:14,078 44k INFO ====> Epoch: 324, cost 16.32 s
466
+ 2023-03-25 18:19:30,347 44k INFO ====> Epoch: 325, cost 16.27 s
467
+ 2023-03-25 18:19:47,250 44k INFO ====> Epoch: 326, cost 16.90 s
468
+ 2023-03-25 18:20:03,686 44k INFO ====> Epoch: 327, cost 16.44 s
469
+ 2023-03-25 18:20:10,686 44k INFO Train Epoch: 328 [27%]
470
+ 2023-03-25 18:20:10,687 44k INFO Losses: [2.4458670616149902, 2.2881720066070557, 10.13757610321045, 14.627115249633789, 0.47115644812583923], step: 10800, lr: 9.599466635167497e-05
471
+ 2023-03-25 18:20:20,765 44k INFO ====> Epoch: 328, cost 17.08 s
472
+ 2023-03-25 18:20:37,407 44k INFO ====> Epoch: 329, cost 16.64 s
473
+ 2023-03-25 18:20:53,950 44k INFO ====> Epoch: 330, cost 16.54 s
474
+ 2023-03-25 18:21:10,322 44k INFO ====> Epoch: 331, cost 16.37 s
475
+ 2023-03-25 18:21:26,633 44k INFO ====> Epoch: 332, cost 16.31 s
476
+ 2023-03-25 18:21:42,887 44k INFO ====> Epoch: 333, cost 16.25 s
477
+ 2023-03-25 18:21:50,532 44k INFO Train Epoch: 334 [33%]
478
+ 2023-03-25 18:21:50,533 44k INFO Losses: [2.60351824760437, 2.2693228721618652, 7.4690985679626465, 13.879813194274902, 0.5584508776664734], step: 11000, lr: 9.592269284691169e-05
479
+ 2023-03-25 18:22:00,284 44k INFO ====> Epoch: 334, cost 17.40 s
480
+ 2023-03-25 18:22:16,678 44k INFO ====> Epoch: 335, cost 16.39 s
481
+ 2023-03-25 18:22:33,068 44k INFO ====> Epoch: 336, cost 16.39 s
482
+ 2023-03-25 18:22:49,695 44k INFO ====> Epoch: 337, cost 16.63 s
483
+ 2023-03-25 18:23:06,120 44k INFO ====> Epoch: 338, cost 16.42 s
484
+ 2023-03-25 18:23:22,651 44k INFO ====> Epoch: 339, cost 16.53 s
485
+ 2023-03-25 18:23:31,178 44k INFO Train Epoch: 340 [39%]
486
+ 2023-03-25 18:23:31,179 44k INFO Losses: [2.6228575706481934, 2.4648935794830322, 8.234376907348633, 15.624024391174316, 0.521338939666748], step: 11200, lr: 9.5850773305411e-05
487
+ 2023-03-25 18:23:35,801 44k INFO Saving model and optimizer state at iteration 340 to ./logs/44k/G_11200.pth
488
+ 2023-03-25 18:23:37,093 44k INFO Saving model and optimizer state at iteration 340 to ./logs/44k/D_11200.pth
489
+ 2023-03-25 18:23:45,697 44k INFO ====> Epoch: 340, cost 23.05 s
490
+ 2023-03-25 18:24:01,807 44k INFO ====> Epoch: 341, cost 16.11 s
491
+ 2023-03-25 18:24:18,320 44k INFO ====> Epoch: 342, cost 16.51 s
492
+ 2023-03-25 18:24:34,861 44k INFO ====> Epoch: 343, cost 16.54 s
493
+ 2023-03-25 18:24:51,178 44k INFO ====> Epoch: 344, cost 16.32 s
494
+ 2023-03-25 18:25:07,875 44k INFO ====> Epoch: 345, cost 16.70 s
495
+ 2023-03-25 18:25:16,986 44k INFO Train Epoch: 346 [45%]
496
+ 2023-03-25 18:25:16,987 44k INFO Losses: [2.8093314170837402, 2.031580686569214, 7.040884971618652, 13.01242446899414, 0.9223231673240662], step: 11400, lr: 9.577890768671308e-05
497
+ 2023-03-25 18:25:25,186 44k INFO ====> Epoch: 346, cost 17.31 s
498
+ 2023-03-25 18:25:41,721 44k INFO ====> Epoch: 347, cost 16.53 s
499
+ 2023-03-25 18:25:57,898 44k INFO ====> Epoch: 348, cost 16.18 s
500
+ 2023-03-25 18:26:14,619 44k INFO ====> Epoch: 349, cost 16.72 s
501
+ 2023-03-25 18:26:31,532 44k INFO ====> Epoch: 350, cost 16.91 s
502
+ 2023-03-25 18:26:47,928 44k INFO ====> Epoch: 351, cost 16.40 s
503
+ 2023-03-25 18:26:57,859 44k INFO Train Epoch: 352 [52%]
504
+ 2023-03-25 18:26:57,861 44k INFO Losses: [2.3367810249328613, 2.234591484069824, 11.762563705444336, 19.299331665039062, 0.5701344013214111], step: 11600, lr: 9.570709595038851e-05
505
+ 2023-03-25 18:27:04,732 44k INFO ====> Epoch: 352, cost 16.80 s
506
+ 2023-03-25 18:27:21,403 44k INFO ====> Epoch: 353, cost 16.67 s
507
+ 2023-03-25 18:27:37,737 44k INFO ====> Epoch: 354, cost 16.33 s
508
+ 2023-03-25 18:27:54,743 44k INFO ====> Epoch: 355, cost 17.01 s
509
+ 2023-03-25 18:28:11,397 44k INFO ====> Epoch: 356, cost 16.65 s
510
+ 2023-03-25 18:28:27,983 44k INFO ====> Epoch: 357, cost 16.59 s
511
+ 2023-03-25 18:28:38,827 44k INFO Train Epoch: 358 [58%]
512
+ 2023-03-25 18:28:38,828 44k INFO Losses: [2.4830808639526367, 2.3789222240448, 9.961544036865234, 20.941343307495117, 0.4926968812942505], step: 11800, lr: 9.56353380560381e-05
513
+ 2023-03-25 18:28:45,245 44k INFO ====> Epoch: 358, cost 17.26 s
514
+ 2023-03-25 18:29:01,674 44k INFO ====> Epoch: 359, cost 16.43 s
515
+ 2023-03-25 18:29:18,150 44k INFO ====> Epoch: 360, cost 16.48 s
516
+ 2023-03-25 18:29:34,541 44k INFO ====> Epoch: 361, cost 16.39 s
517
+ 2023-03-25 18:29:51,314 44k INFO ====> Epoch: 362, cost 16.77 s
518
+ 2023-03-25 18:30:07,952 44k INFO ====> Epoch: 363, cost 16.64 s
519
+ 2023-03-25 18:30:19,472 44k INFO Train Epoch: 364 [64%]
520
+ 2023-03-25 18:30:19,473 44k INFO Losses: [2.204318046569824, 2.510396957397461, 13.009056091308594, 19.046606063842773, 0.6128262281417847], step: 12000, lr: 9.556363396329299e-05
521
+ 2023-03-25 18:30:23,964 44k INFO Saving model and optimizer state at iteration 364 to ./logs/44k/G_12000.pth
522
+ 2023-03-25 18:30:25,298 44k INFO Saving model and optimizer state at iteration 364 to ./logs/44k/D_12000.pth
523
+ 2023-03-25 18:30:30,855 44k INFO ====> Epoch: 364, cost 22.90 s
524
+ 2023-03-25 18:30:47,333 44k INFO ====> Epoch: 365, cost 16.48 s
525
+ 2023-03-25 18:31:03,833 44k INFO ====> Epoch: 366, cost 16.50 s
526
+ 2023-03-25 18:31:20,318 44k INFO ====> Epoch: 367, cost 16.49 s
527
+ 2023-03-25 18:31:36,481 44k INFO ====> Epoch: 368, cost 16.16 s
528
+ 2023-03-25 18:31:52,959 44k INFO ====> Epoch: 369, cost 16.48 s
529
+ 2023-03-25 18:32:05,768 44k INFO Train Epoch: 370 [70%]
530
+ 2023-03-25 18:32:05,769 44k INFO Losses: [2.8103220462799072, 1.9135441780090332, 4.667817115783691, 12.864496231079102, 0.5750285983085632], step: 12200, lr: 9.54919836318146e-05
531
+ 2023-03-25 18:32:10,376 44k INFO ====> Epoch: 370, cost 17.42 s
532
+ 2023-03-25 18:32:26,649 44k INFO ====> Epoch: 371, cost 16.27 s
533
+ 2023-03-25 18:32:43,415 44k INFO ====> Epoch: 372, cost 16.77 s
534
+ 2023-03-25 18:32:59,729 44k INFO ====> Epoch: 373, cost 16.31 s
535
+ 2023-03-25 18:33:16,345 44k INFO ====> Epoch: 374, cost 16.62 s
536
+ 2023-03-25 18:33:32,600 44k INFO ====> Epoch: 375, cost 16.26 s
537
+ 2023-03-25 18:33:45,666 44k INFO Train Epoch: 376 [76%]
538
+ 2023-03-25 18:33:45,667 44k INFO Losses: [2.4472713470458984, 2.2148959636688232, 10.443349838256836, 16.083932876586914, 0.6222093105316162], step: 12400, lr: 9.542038702129457e-05
539
+ 2023-03-25 18:33:49,459 44k INFO ====> Epoch: 376, cost 16.86 s
540
+ 2023-03-25 18:34:05,825 44k INFO ====> Epoch: 377, cost 16.37 s
541
+ 2023-03-25 18:34:22,051 44k INFO ====> Epoch: 378, cost 16.23 s
542
+ 2023-03-25 18:34:38,473 44k INFO ====> Epoch: 379, cost 16.42 s
543
+ 2023-03-25 18:34:54,719 44k INFO ====> Epoch: 380, cost 16.25 s
544
+ 2023-03-25 18:35:10,879 44k INFO ====> Epoch: 381, cost 16.16 s
545
+ 2023-03-25 18:35:24,611 44k INFO Train Epoch: 382 [82%]
546
+ 2023-03-25 18:35:24,612 44k INFO Losses: [2.351381778717041, 2.2640938758850098, 10.070740699768066, 19.411256790161133, 0.5809462666511536], step: 12600, lr: 9.534884409145477e-05
547
+ 2023-03-25 18:35:27,619 44k INFO ====> Epoch: 382, cost 16.74 s
548
+ 2023-03-25 18:35:44,312 44k INFO ====> Epoch: 383, cost 16.69 s
549
+ 2023-03-25 18:36:00,575 44k INFO ====> Epoch: 384, cost 16.26 s
550
+ 2023-03-25 18:36:17,367 44k INFO ====> Epoch: 385, cost 16.79 s
551
+ 2023-03-25 18:36:33,862 44k INFO ====> Epoch: 386, cost 16.49 s
552
+ 2023-03-25 18:36:50,147 44k INFO ====> Epoch: 387, cost 16.28 s
553
+ 2023-03-25 18:37:04,689 44k INFO Train Epoch: 388 [88%]
554
+ 2023-03-25 18:37:04,690 44k INFO Losses: [2.483032464981079, 2.0824134349823, 9.235447883605957, 15.548537254333496, 0.5608479976654053], step: 12800, lr: 9.527735480204728e-05
555
+ 2023-03-25 18:37:09,240 44k INFO Saving model and optimizer state at iteration 388 to ./logs/44k/G_12800.pth
556
+ 2023-03-25 18:37:10,581 44k INFO Saving model and optimizer state at iteration 388 to ./logs/44k/D_12800.pth
557
+ 2023-03-25 18:37:12,943 44k INFO ====> Epoch: 388, cost 22.80 s
558
+ 2023-03-25 18:37:29,278 44k INFO ====> Epoch: 389, cost 16.33 s
559
+ 2023-03-25 18:37:45,699 44k INFO ====> Epoch: 390, cost 16.42 s
560
+ 2023-03-25 18:38:02,084 44k INFO ====> Epoch: 391, cost 16.39 s
561
+ 2023-03-25 18:38:18,298 44k INFO ====> Epoch: 392, cost 16.21 s
562
+ 2023-03-25 18:38:34,621 44k INFO ====> Epoch: 393, cost 16.32 s
563
+ 2023-03-25 18:38:50,007 44k INFO Train Epoch: 394 [94%]
564
+ 2023-03-25 18:38:50,008 44k INFO Losses: [2.5371432304382324, 1.9721053838729858, 9.028430938720703, 12.427369117736816, 0.916906476020813], step: 13000, lr: 9.520591911285433e-05
565
+ 2023-03-25 18:38:51,490 44k INFO ====> Epoch: 394, cost 16.87 s
566
+ 2023-03-25 18:39:08,072 44k INFO ====> Epoch: 395, cost 16.58 s
567
+ 2023-03-25 18:39:24,382 44k INFO ====> Epoch: 396, cost 16.31 s
568
+ 2023-03-25 18:39:41,864 44k INFO ====> Epoch: 397, cost 17.48 s
569
+ 2023-03-25 18:39:58,320 44k INFO ====> Epoch: 398, cost 16.46 s
570
+ 2023-03-25 18:40:14,427 44k INFO ====> Epoch: 399, cost 16.11 s
571
+ 2023-03-25 18:40:31,773 44k INFO ====> Epoch: 400, cost 17.35 s
572
+ 2023-03-25 18:40:35,493 44k INFO Train Epoch: 401 [0%]
573
+ 2023-03-25 18:40:35,495 44k INFO Losses: [2.464968204498291, 2.3817968368530273, 5.272903919219971, 11.98283863067627, 0.8528004288673401], step: 13200, lr: 9.512264516656537e-05
574
+ 2023-03-25 18:40:48,882 44k INFO ====> Epoch: 401, cost 17.11 s
575
+ 2023-03-25 18:41:05,273 44k INFO ====> Epoch: 402, cost 16.39 s
576
+ 2023-03-25 18:41:21,778 44k INFO ====> Epoch: 403, cost 16.51 s
577
+ 2023-03-25 18:41:38,055 44k INFO ====> Epoch: 404, cost 16.28 s
578
+ 2023-03-25 18:41:54,305 44k INFO ====> Epoch: 405, cost 16.25 s
579
+ 2023-03-25 18:42:11,211 44k INFO ====> Epoch: 406, cost 16.91 s
580
+ 2023-03-25 18:42:15,794 44k INFO Train Epoch: 407 [6%]
581
+ 2023-03-25 18:42:15,795 44k INFO Losses: [2.748605728149414, 2.0154476165771484, 8.186812400817871, 10.141654014587402, -0.014795398339629173], step: 13400, lr: 9.505132547334502e-05
582
+ 2023-03-25 18:42:28,420 44k INFO ====> Epoch: 407, cost 17.21 s
583
+ 2023-03-25 18:42:45,205 44k INFO ====> Epoch: 408, cost 16.79 s
584
+ 2023-03-25 18:43:01,444 44k INFO ====> Epoch: 409, cost 16.24 s
585
+ 2023-03-25 18:43:17,926 44k INFO ====> Epoch: 410, cost 16.48 s
586
+ 2023-03-25 18:43:34,340 44k INFO ====> Epoch: 411, cost 16.41 s
587
+ 2023-03-25 18:43:50,913 44k INFO ====> Epoch: 412, cost 16.57 s
588
+ 2023-03-25 18:43:56,056 44k INFO Train Epoch: 413 [12%]
589
+ 2023-03-25 18:43:56,057 44k INFO Losses: [2.4909560680389404, 2.0936198234558105, 6.849160671234131, 12.217107772827148, 0.795183539390564], step: 13600, lr: 9.498005925318179e-05
590
+ 2023-03-25 18:44:00,475 44k INFO Saving model and optimizer state at iteration 413 to ./logs/44k/G_13600.pth
591
+ 2023-03-25 18:44:01,790 44k INFO Saving model and optimizer state at iteration 413 to ./logs/44k/D_13600.pth
592
+ 2023-03-25 18:44:13,600 44k INFO ====> Epoch: 413, cost 22.69 s
so-vits-svc-4.0/README.md ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ 推理项目地址
2
+ Inference project address
3
+
4
+ https://github.com/zwa73/so-vits-svc-fork
5
+ https://github.com/svc-develop-team/so-vits-svc