v2/ar_base.pth → DiT_seed_v2_uvit_facodec_small_wavenet_bigvgan_pruned.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:42e2afabf97e5d856580af1f81a9705423b4fb21bb9475cece9925b62355cc5c
3
- size 358846893
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4b52780fd069dd897309a2def2bd7c1664ca36b56f47e9f17ebf7a52b711e8c
3
+ size 440806506
v2/cfm_small.pth → DiT_seed_v2_uvit_facodec_small_wavenet_f0_bigvgan_pruned.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:73ae18615dc8cb01710086629973198226dedc63ebeaab0aee17863d767b3474
3
- size 353349694
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afa945f80fa1dd63410626df265b2f4d9774719c4be5f0d80ace0cf2cc093a5e
3
+ size 437658856
DiT_seed_v2_uvit_facodec_small_wavenet_f0_bigvgan_pruned_44k_L1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61807b267b8438326d635edbe6f7d374282826698cf82405cabd7082d450719d
3
+ size 442351868
DiT_seed_v2_uvit_facodec_small_wavenet_f0_bigvgan_pruned_44k_L2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:404aa78081d75fd3652c68453874a0aefbc6177f75655adcfef2628c2c337b79
3
+ size 442352798
DiT_step_298000_seed_uvit_facodec_small_wavenet_pruned.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfacb2331a9cdb0fe513f1ad1a75cee25285bae9744b1c96d58232ef7ef806e5
3
+ size 440805788
config_dit_mel_seed_facodec_small_wavenet.yml ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ log_dir: "./runs/run_dit_mel_seed_facodec_small"
2
+ save_freq: 1
3
+ log_interval: 10
4
+ save_interval: 1000
5
+ device: "cuda"
6
+ epochs: 1000 # number of epochs for first stage training (pre-training)
7
+ batch_size: 2
8
+ batch_length: 100 # maximum duration of audio in a batch (in seconds)
9
+ max_len: 80 # maximum number of frames
10
+ pretrained_model: ""
11
+ pretrained_encoder: ""
12
+ load_only_params: False # set to true if do not want to load epoch numbers and optimizer parameters
13
+
14
+ F0_path: "modules/JDC/bst.t7"
15
+
16
+ data_params:
17
+ train_data: "./data/train.txt"
18
+ val_data: "./data/val.txt"
19
+ root_path: "./data/"
20
+
21
+ preprocess_params:
22
+ sr: 22050
23
+ spect_params:
24
+ n_fft: 1024
25
+ win_length: 1024
26
+ hop_length: 256
27
+ n_mels: 80
28
+
29
+ model_params:
30
+ dit_type: "DiT" # uDiT or DiT
31
+ reg_loss_type: "l1" # l1 or l2
32
+
33
+ speech_tokenizer:
34
+ type: 'facodec'
35
+ path: "speech_tokenizer_v1.onnx"
36
+
37
+ style_encoder:
38
+ dim: 192
39
+ campplus_path: "campplus_cn_common.bin"
40
+
41
+ DAC:
42
+ encoder_dim: 64
43
+ encoder_rates: [2, 5, 5, 6]
44
+ decoder_dim: 1536
45
+ decoder_rates: [ 6, 5, 5, 2 ]
46
+ sr: 24000
47
+
48
+ length_regulator:
49
+ channels: 512
50
+ is_discrete: true
51
+ content_codebook_size: 1024
52
+ in_frame_rate: 80
53
+ out_frame_rate: 80
54
+ sampling_ratios: [1, 1, 1, 1]
55
+ token_dropout_prob: 0.3 # probability of performing token dropout
56
+ token_dropout_range: 1.0 # maximum percentage of tokens to drop out
57
+ n_codebooks: 3
58
+ quantizer_dropout: 0.5
59
+ f0_condition: false
60
+ n_f0_bins: 512
61
+
62
+ DiT:
63
+ hidden_dim: 512
64
+ num_heads: 8
65
+ depth: 13
66
+ class_dropout_prob: 0.1
67
+ block_size: 8192
68
+ in_channels: 80
69
+ style_condition: true
70
+ final_layer_type: 'wavenet'
71
+ target: 'mel' # mel or codec
72
+ content_dim: 512
73
+ content_codebook_size: 1024
74
+ content_type: 'discrete'
75
+ f0_condition: true
76
+ n_f0_bins: 512
77
+ content_codebooks: 1
78
+ is_causal: false
79
+ long_skip_connection: true
80
+ zero_prompt_speech_token: false # for prompt component, do not input corresponding speech token
81
+ time_as_token: false
82
+ style_as_token: false
83
+ uvit_skip_connection: true
84
+ add_resblock_in_transformer: false
85
+
86
+ wavenet:
87
+ hidden_dim: 512
88
+ num_layers: 8
89
+ kernel_size: 5
90
+ dilation_rate: 1
91
+ p_dropout: 0.2
92
+ style_condition: true
93
+
94
+ loss_params:
95
+ base_lr: 0.0001
96
+ lambda_mel: 45
97
+ lambda_kl: 1.0
config_dit_mel_seed_facodec_small_wavenet_f0.yml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ log_dir: ""
2
+ save_freq: 1
3
+ log_interval: 10
4
+ save_interval: 1000
5
+ device: "cuda"
6
+ epochs: 1000 # number of epochs for first stage training (pre-training)
7
+ batch_size: 2
8
+ batch_length: 100 # maximum duration of audio in a batch (in seconds)
9
+ max_len: 80 # maximum number of frames
10
+ pretrained_model: ""
11
+ pretrained_encoder: ""
12
+ load_only_params: False # set to true if do not want to load epoch numbers and optimizer parameters
13
+
14
+ preprocess_params:
15
+ sr: 22050
16
+ spect_params:
17
+ n_fft: 1024
18
+ win_length: 1024
19
+ hop_length: 256
20
+ n_mels: 80
21
+
22
+ model_params:
23
+ dit_type: "DiT" # uDiT or DiT
24
+ reg_loss_type: "l1" # l1 or l2
25
+
26
+ speech_tokenizer:
27
+ type: 'facodec'
28
+ path: "speech_tokenizer_v1.onnx"
29
+
30
+ cosyvoice:
31
+ path: "../CosyVoice/pretrained_models/CosyVoice-300M"
32
+
33
+ style_encoder:
34
+ dim: 192
35
+ campplus_path: "campplus_cn_common.bin"
36
+
37
+ DAC:
38
+ encoder_dim: 64
39
+ encoder_rates: [2, 5, 5, 6]
40
+ decoder_dim: 1536
41
+ decoder_rates: [ 6, 5, 5, 2 ]
42
+ sr: 24000
43
+
44
+ length_regulator:
45
+ channels: 512
46
+ is_discrete: true
47
+ content_codebook_size: 1024
48
+ in_frame_rate: 80
49
+ out_frame_rate: 80
50
+ sampling_ratios: [1, 1, 1, 1]
51
+ token_dropout_prob: 0.3 # probability of performing token dropout
52
+ token_dropout_range: 1.0 # maximum percentage of tokens to drop out
53
+ n_codebooks: 1
54
+ quantizer_dropout: 0.5
55
+ f0_condition: true
56
+ n_f0_bins: 512
57
+
58
+ DiT:
59
+ hidden_dim: 512
60
+ num_heads: 8
61
+ depth: 13
62
+ class_dropout_prob: 0.1
63
+ block_size: 8192
64
+ in_channels: 80
65
+ style_condition: true
66
+ final_layer_type: 'wavenet'
67
+ target: 'mel' # mel or codec
68
+ content_dim: 512
69
+ content_codebook_size: 1024
70
+ content_type: 'discrete'
71
+ f0_condition: true
72
+ n_f0_bins: 512
73
+ content_codebooks: 3
74
+ is_causal: false
75
+ long_skip_connection: true
76
+ zero_prompt_speech_token: false # for prompt component, do not input corresponding speech token
77
+ time_as_token: false
78
+ style_as_token: false
79
+ uvit_skip_connection: true
80
+ add_resblock_in_transformer: false
81
+
82
+ wavenet:
83
+ hidden_dim: 512
84
+ num_layers: 8
85
+ kernel_size: 5
86
+ dilation_rate: 1
87
+ p_dropout: 0.2
88
+ style_condition: true
89
+
90
+ loss_params:
91
+ base_lr: 0.0001
92
+ lambda_mel: 45
93
+ lambda_kl: 1.0
config_dit_mel_seed_facodec_small_wavenet_f0_44k.yml ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ log_dir: ""
2
+ save_freq: 1
3
+ log_interval: 10
4
+ save_interval: 1000
5
+ device: "cuda"
6
+ epochs: 1000 # number of epochs for first stage training (pre-training)
7
+ batch_size: 2
8
+ batch_length: 100 # maximum duration of audio in a batch (in seconds)
9
+ max_len: 80 # maximum number of frames
10
+ pretrained_model: ""
11
+ pretrained_encoder: ""
12
+ load_only_params: False # set to true if do not want to load epoch numbers and optimizer parameters
13
+
14
+ preprocess_params:
15
+ sr: 44100
16
+ spect_params:
17
+ n_fft: 2048
18
+ win_length: 2048
19
+ hop_length: 512
20
+ n_mels: 128
21
+ fmin: 0
22
+ fmax: "None"
23
+
24
+ model_params:
25
+ dit_type: "DiT" # uDiT or DiT
26
+ reg_loss_type: "l1" # l1 or l2
27
+
28
+ speech_tokenizer:
29
+ type: 'facodec'
30
+ path: "speech_tokenizer_v1.onnx"
31
+
32
+ cosyvoice:
33
+ path: "../CosyVoice/pretrained_models/CosyVoice-300M"
34
+
35
+ style_encoder:
36
+ dim: 192
37
+ campplus_path: "campplus_cn_common.bin"
38
+
39
+ DAC:
40
+ encoder_dim: 64
41
+ encoder_rates: [2, 5, 5, 6]
42
+ decoder_dim: 1536
43
+ decoder_rates: [ 6, 5, 5, 2 ]
44
+ sr: 24000
45
+
46
+ length_regulator:
47
+ channels: 512
48
+ is_discrete: true
49
+ content_codebook_size: 1024
50
+ in_frame_rate: 80
51
+ out_frame_rate: 80
52
+ sampling_ratios: [1, 1, 1, 1]
53
+ token_dropout_prob: 0.3 # probability of performing token dropout
54
+ token_dropout_range: 1.0 # maximum percentage of tokens to drop out
55
+ n_codebooks: 3
56
+ quantizer_dropout: 0.5
57
+ f0_condition: true
58
+ n_f0_bins: 512
59
+
60
+ DiT:
61
+ hidden_dim: 512
62
+ num_heads: 8
63
+ depth: 13
64
+ class_dropout_prob: 0.1
65
+ block_size: 8192
66
+ in_channels: 128
67
+ style_condition: true
68
+ final_layer_type: 'wavenet'
69
+ target: 'mel' # mel or codec
70
+ content_dim: 512
71
+ content_codebook_size: 1024
72
+ content_type: 'discrete'
73
+ f0_condition: true
74
+ n_f0_bins: 512
75
+ content_codebooks: 1
76
+ is_causal: false
77
+ long_skip_connection: true
78
+ zero_prompt_speech_token: false # for prompt component, do not input corresponding speech token
79
+ time_as_token: false
80
+ style_as_token: false
81
+ uvit_skip_connection: true
82
+ add_resblock_in_transformer: false
83
+
84
+ wavenet:
85
+ hidden_dim: 512
86
+ num_layers: 8
87
+ kernel_size: 5
88
+ dilation_rate: 1
89
+ p_dropout: 0.2
90
+ style_condition: true
91
+
92
+ loss_params:
93
+ base_lr: 0.0001
94
+ lambda_mel: 45
95
+ lambda_kl: 1.0