Politrees commited on
Commit
f590c23
·
verified ·
1 Parent(s): e3263dc

Adding new models (read description)

Browse files

- BandSplit Roformer | 4-stems FT by SYH99999
- MelBand Roformer | 4-stems FT Large v1 by SYH99999
- MelBand Roformer | 4-stems FT Large v2 by SYH99999
- MelBand Roformer | 4-stems Large v1 by Aname
- MelBand Roformer | 4-stems XL v1 by Aname

models/Roformer/BandSplit/BandSplit_Roformer_4stems_FT_by_SYH99999.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a64c61c5972fe3fdf331d98f06bb4d6e237147c83a4bc71fd4e00e254c9b7652
3
+ size 527245586
models/Roformer/BandSplit/config_BandSplit_Roformer_4stems_FT_by_SYH99999.yaml ADDED
@@ -0,0 +1,196 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ audio:
2
+ chunk_size: 485100
3
+ dim_f: 1024
4
+ dim_t: 801 # don't work (use in model)
5
+ hop_length: 441 # don't work (use in model)
6
+ n_fft: 2048
7
+ num_channels: 2
8
+ sample_rate: 44100
9
+ min_mean_abs: 0.000
10
+
11
+ model:
12
+ dim: 384
13
+ depth: 8
14
+ stereo: true
15
+ num_stems: 4
16
+ time_transformer_depth: 1
17
+ freq_transformer_depth: 1
18
+ linear_transformer_depth: 0
19
+ freqs_per_bands: !!python/tuple
20
+ - 2
21
+ - 2
22
+ - 2
23
+ - 2
24
+ - 2
25
+ - 2
26
+ - 2
27
+ - 2
28
+ - 2
29
+ - 2
30
+ - 2
31
+ - 2
32
+ - 2
33
+ - 2
34
+ - 2
35
+ - 2
36
+ - 2
37
+ - 2
38
+ - 2
39
+ - 2
40
+ - 2
41
+ - 2
42
+ - 2
43
+ - 2
44
+ - 4
45
+ - 4
46
+ - 4
47
+ - 4
48
+ - 4
49
+ - 4
50
+ - 4
51
+ - 4
52
+ - 4
53
+ - 4
54
+ - 4
55
+ - 4
56
+ - 12
57
+ - 12
58
+ - 12
59
+ - 12
60
+ - 12
61
+ - 12
62
+ - 12
63
+ - 12
64
+ - 24
65
+ - 24
66
+ - 24
67
+ - 24
68
+ - 24
69
+ - 24
70
+ - 24
71
+ - 24
72
+ - 48
73
+ - 48
74
+ - 48
75
+ - 48
76
+ - 48
77
+ - 48
78
+ - 48
79
+ - 48
80
+ - 128
81
+ - 129
82
+ dim_head: 64
83
+ heads: 8
84
+ attn_dropout: 0.1
85
+ ff_dropout: 0.1
86
+ flash_attn: true
87
+ dim_freqs_in: 1025
88
+ stft_n_fft: 2048
89
+ stft_hop_length: 441
90
+ stft_win_length: 2048
91
+ stft_normalized: false
92
+ mask_estimator_depth: 2
93
+ multi_stft_resolution_loss_weight: 1.0
94
+ multi_stft_resolutions_window_sizes: !!python/tuple
95
+ - 4096
96
+ - 2048
97
+ - 1024
98
+ - 512
99
+ - 256
100
+ multi_stft_hop_size: 147
101
+ multi_stft_normalized: False
102
+ mlp_expansion_factor: 2
103
+ use_torch_checkpoint: False # it allows to greatly reduce GPU memory consumption during training (not fully tested)
104
+ skip_connection: False # Enable skip connection between transformer blocks - can solve problem with gradients and probably faster training
105
+
106
+ training:
107
+ batch_size: 2
108
+ gradient_accumulation_steps: 1
109
+ grad_clip: 0
110
+ instruments: ['drums', 'bass', 'other', 'vocals']
111
+ patience: 3
112
+ reduce_factor: 0.95
113
+ target_instrument: null
114
+ num_epochs: 1000
115
+ num_steps: 1000
116
+ augmentation: false # enable augmentations by audiomentations and pedalboard
117
+ augmentation_type: simple1
118
+ use_mp3_compress: false # Deprecated
119
+ augmentation_mix: true # Mix several stems of the same type with some probability
120
+ augmentation_loudness: true # randomly change loudness of each stem
121
+ augmentation_loudness_type: 1 # Type 1 or 2
122
+ augmentation_loudness_min: 0.5
123
+ augmentation_loudness_max: 1.5
124
+ q: 0.95
125
+ coarse_loss_clip: true
126
+ ema_momentum: 0.999
127
+ # optimizer: prodigy
128
+ optimizer: adam
129
+ # lr: 1.0
130
+ lr: 1.0e-5
131
+ other_fix: false # it's needed for checking on multisong dataset if other is actually instrumental
132
+ use_amp: true # enable or disable usage of mixed precision (float16) - usually it must be true
133
+
134
+ augmentations:
135
+ enable: true # enable or disable all augmentations (to fast disable if needed)
136
+ loudness: true # randomly change loudness of each stem on the range (loudness_min; loudness_max)
137
+ loudness_min: 0.5
138
+ loudness_max: 1.5
139
+ mixup: true # mix several stems of same type with some probability (only works for dataset types: 1, 2, 3)
140
+ mixup_probs: !!python/tuple # 2 additional stems of the same type (1st with prob 0.2, 2nd with prob 0.02)
141
+ - 0.2
142
+ - 0.02
143
+ mixup_loudness_min: 0.5
144
+ mixup_loudness_max: 1.5
145
+
146
+ all:
147
+ channel_shuffle: 0.5 # Set 0 or lower to disable
148
+ random_inverse: 0.1 # inverse track (better lower probability)
149
+ random_polarity: 0.5 # polarity change (multiply waveform to -1)
150
+
151
+ vocals:
152
+ pitch_shift: 0.1
153
+ pitch_shift_min_semitones: -5
154
+ pitch_shift_max_semitones: 5
155
+ seven_band_parametric_eq: 0.1
156
+ seven_band_parametric_eq_min_gain_db: -9
157
+ seven_band_parametric_eq_max_gain_db: 9
158
+ tanh_distortion: 0.1
159
+ tanh_distortion_min: 0.1
160
+ tanh_distortion_max: 0.7
161
+ bass:
162
+ pitch_shift: 0.1
163
+ pitch_shift_min_semitones: -2
164
+ pitch_shift_max_semitones: 2
165
+ seven_band_parametric_eq: 0.1
166
+ seven_band_parametric_eq_min_gain_db: -3
167
+ seven_band_parametric_eq_max_gain_db: 6
168
+ tanh_distortion: 0.1
169
+ tanh_distortion_min: 0.1
170
+ tanh_distortion_max: 0.5
171
+ drums:
172
+ pitch_shift: 0.1
173
+ pitch_shift_min_semitones: -5
174
+ pitch_shift_max_semitones: 5
175
+ seven_band_parametric_eq: 0.1
176
+ seven_band_parametric_eq_min_gain_db: -9
177
+ seven_band_parametric_eq_max_gain_db: 9
178
+ tanh_distortion: 0.1
179
+ tanh_distortion_min: 0.1
180
+ tanh_distortion_max: 0.6
181
+ other:
182
+ pitch_shift: 0.1
183
+ pitch_shift_min_semitones: -4
184
+ pitch_shift_max_semitones: 4
185
+ gaussian_noise: 0.1
186
+ gaussian_noise_min_amplitude: 0.001
187
+ gaussian_noise_max_amplitude: 0.015
188
+ time_stretch: 0.1
189
+ time_stretch_min_rate: 0.8
190
+ time_stretch_max_rate: 1.25
191
+
192
+
193
+ inference:
194
+ batch_size: 2
195
+ dim_t: 2048
196
+ num_overlap: 4
models/Roformer/MelBand/MelBand_Roformer_4stems_FT_Large_v1_by_SYH99999.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:590358e6abe3255d87d7173ca2e4005de082f14cbb466ea4c5ffedbcc3964506
3
+ size 3764030874
models/Roformer/MelBand/MelBand_Roformer_4stems_FT_Large_v2_by_SYH99999.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:835ce9fa6651769c28ad5a2c87f9220b003a461d7d6c57ebe0898fff5479138c
3
+ size 3764030874
models/Roformer/MelBand/MelBand_Roformer_4stems_Large_v1_by_Aname.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2ca1ac95e203d2630994c3d4a0cebabd7160e70bb010adab794843340fb704f
3
+ size 3764030874
models/Roformer/MelBand/MelBand_Roformer_4stems_XL_v1_by_Aname.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca5f7c78543d84c369319f226b7337a6cc0b2e1a75d20409fc3ebde32686e94b
3
+ size 6411087858
models/Roformer/MelBand/config_MelBand_Roformer_4stems_FT_Large_by_SYH99999.yaml ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ audio:
2
+ chunk_size: 485100
3
+ dim_f: 1024
4
+ dim_t: 1101
5
+ hop_length: 882
6
+ n_fft: 2048
7
+ num_channels: 2
8
+ sample_rate: 44100
9
+ min_mean_abs: 0.000
10
+
11
+ model:
12
+ dim: 384
13
+ depth: 8
14
+ stereo: true
15
+ num_stems: 4
16
+ linear_transformer_depth: 0
17
+ time_transformer_depth: 1
18
+ freq_transformer_depth: 1
19
+ num_bands: 60
20
+ dim_head: 64
21
+ heads: 8
22
+ attn_dropout: 0.0
23
+ ff_dropout: 0.0
24
+ flash_attn: true
25
+ dim_freqs_in: 2049
26
+ sample_rate: 44100 # needed for mel filter bank from librosa
27
+ stft_n_fft: 4096
28
+ stft_hop_length: 882
29
+ stft_win_length: 4096
30
+ stft_normalized: False
31
+ mask_estimator_depth: 2
32
+ multi_stft_resolution_loss_weight: 1.0
33
+ multi_stft_resolutions_window_sizes: !!python/tuple
34
+ - 4096
35
+ - 2048
36
+ - 1024
37
+ - 512
38
+ - 256
39
+ multi_stft_hop_size: 147
40
+ multi_stft_normalized: False
41
+ mlp_expansion_factor: 4
42
+ use_torch_checkpoint: False # it allows to greatly reduce GPU memory consumption during training (not fully tested)
43
+ skip_connection: True # Enable skip connection between transformer blocks - can solve problem with gradients and probably faster training
44
+
45
+ training:
46
+ batch_size: 1
47
+ gradient_accumulation_steps: 1
48
+ grad_clip: 0
49
+ instruments: ['drums', 'bass', 'other', 'vocals']
50
+ lr: 1.0
51
+ patience: 3
52
+ reduce_factor: 0.95
53
+ target_instrument: null
54
+ num_epochs: 1000
55
+ num_steps: 1000
56
+ q: 0.95
57
+ coarse_loss_clip: false
58
+ ema_momentum: 0.999
59
+ optimizer: prodigy
60
+ read_metadata_procs: 8 # Number of processes to use during metadata reading for dataset. Can speed up metadata generation
61
+ normalize: false
62
+ other_fix: false # it's needed for checking on multisong dataset if other is actually instrumental
63
+ use_amp: true # enable or disable usage of mixed precision (float16) - usually it must be true
64
+
65
+ inference:
66
+ batch_size: 4
67
+ dim_t: 1101
68
+ num_overlap: 4
69
+ normalize: false
models/Roformer/MelBand/config_MelBand_Roformer_4stems_Large_v1_by_Aname.yaml ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ audio:
2
+ chunk_size: 661500
3
+ dim_f: 1024
4
+ dim_t: 1101
5
+ hop_length: 882
6
+ n_fft: 4096
7
+ num_channels: 2
8
+ sample_rate: 44100
9
+ min_mean_abs: 0.0001
10
+
11
+ model:
12
+ dim: 384
13
+ depth: 8
14
+ stereo: true
15
+ num_stems: 4
16
+ linear_transformer_depth: 0
17
+ time_transformer_depth: 1
18
+ freq_transformer_depth: 1
19
+ num_bands: 60
20
+ dim_head: 64
21
+ heads: 8
22
+ attn_dropout: 0.0
23
+ ff_dropout: 0.0
24
+ flash_attn: true
25
+ dim_freqs_in: 2049
26
+ sample_rate: 44100 # needed for mel filter bank from librosa
27
+ stft_n_fft: 4096
28
+ stft_hop_length: 882
29
+ stft_win_length: 4096
30
+ stft_normalized: False
31
+ mask_estimator_depth: 2
32
+ multi_stft_resolution_loss_weight: 1.0
33
+ multi_stft_resolutions_window_sizes: !!python/tuple
34
+ - 4096
35
+ - 2048
36
+ - 1024
37
+ - 512
38
+ - 256
39
+ multi_stft_hop_size: 147
40
+ multi_stft_normalized: False
41
+ mlp_expansion_factor: 4
42
+ use_torch_checkpoint: False # it allows to greatly reduce GPU memory consumption during training (not fully tested)
43
+ skip_connection: True # Enable skip connection between transformer blocks - can solve problem with gradients and probably faster training
44
+
45
+ training:
46
+ batch_size: 1
47
+ gradient_accumulation_steps: 4
48
+ grad_clip: 0
49
+ instruments: ['drums', 'bass', 'other', 'vocals']
50
+ lr: 2.0e-05
51
+ patience: 2
52
+ reduce_factor: 0.95
53
+ target_instrument: null
54
+ num_epochs: 1000
55
+ num_steps: 300
56
+ q: 0.95
57
+ coarse_loss_clip: false
58
+ ema_momentum: 0.999
59
+ optimizer: adamw
60
+ read_metadata_procs: 8 # Number of processes to use during metadata reading for dataset. Can speed up metadata generation
61
+ normalize: false
62
+ other_fix: false # it's needed for checking on multisong dataset if other is actually instrumental
63
+ use_amp: true # enable or disable usage of mixed precision (float16) - usually it must be true
64
+
65
+ augmentations:
66
+ enable: false # enable or disable all augmentations (to fast disable if needed)
67
+ loudness: false # randomly change loudness of each stem on the range (loudness_min; loudness_max)
68
+ loudness_min: 0.5
69
+ loudness_max: 1.5
70
+ mixup: true # mix several stems of same type with some probability (only works for dataset types: 1, 2, 3)
71
+ mixup_probs: !!python/tuple # 2 additional stems of the same type (1st with prob 0.2, 2nd with prob 0.02)
72
+ - 0.2
73
+ - 0.02
74
+ - 0.002
75
+ mixup_loudness_min: 0.5
76
+ mixup_loudness_max: 1.5
77
+
78
+ # apply mp3 compression to mixture only (emulate downloading mp3 from internet)
79
+ mp3_compression_on_mixture: 0.01
80
+ mp3_compression_on_mixture_bitrate_min: 32
81
+ mp3_compression_on_mixture_bitrate_max: 320
82
+ mp3_compression_on_mixture_backend: "lameenc"
83
+
84
+ all:
85
+ channel_shuffle: 0.5 # Set 0 or lower to disable
86
+ random_inverse: 0.01 # inverse track (better lower probability)
87
+ random_polarity: 0.5 # polarity change (multiply waveform to -1)
88
+
89
+ vocals:
90
+ pitch_shift: 1.0
91
+ pitch_shift_min_semitones: -12
92
+ pitch_shift_max_semitones: 12
93
+ seven_band_parametric_eq: 0.5
94
+ seven_band_parametric_eq_min_gain_db: -80
95
+ seven_band_parametric_eq_max_gain_db: 9
96
+ tanh_distortion: 0.5
97
+ tanh_distortion_min: 0.1
98
+ tanh_distortion_max: 1
99
+ time_stretch: 1.0
100
+ time_stretch_min_rate: 0.5
101
+ time_stretch_max_rate: 2
102
+ bass:
103
+ pitch_shift: 1.0
104
+ pitch_shift_min_semitones: -6
105
+ pitch_shift_max_semitones: 6
106
+ seven_band_parametric_eq: 0.4
107
+ seven_band_parametric_eq_min_gain_db: -32
108
+ seven_band_parametric_eq_max_gain_db: 6
109
+ tanh_distortion: 1.0
110
+ tanh_distortion_min: 0.1
111
+ tanh_distortion_max: 0.5
112
+ time_stretch: 1.0
113
+ time_stretch_min_rate: 0.5
114
+ time_stretch_max_rate: 1.5
115
+ drums:
116
+ pitch_shift: 0.1
117
+ pitch_shift_min_semitones: -6
118
+ pitch_shift_max_semitones: 6
119
+ seven_band_parametric_eq: 0.5
120
+ seven_band_parametric_eq_min_gain_db: -24
121
+ seven_band_parametric_eq_max_gain_db: 12
122
+ tanh_distortion: 0.3
123
+ tanh_distortion_min: 0.1
124
+ tanh_distortion_max: 0.6
125
+ time_stretch: 1.0
126
+ time_stretch_min_rate: 0.333
127
+ time_stretch_max_rate: 1.5
128
+ other:
129
+ pitch_shift: 1.0
130
+ pitch_shift_min_semitones: -12
131
+ pitch_shift_max_semitones: 12
132
+ gaussian_noise: 0.4
133
+ gaussian_noise_min_amplitude: 0.001
134
+ gaussian_noise_max_amplitude: 0.15
135
+ time_stretch: 0.01
136
+ time_stretch_min_rate: 0.25
137
+ time_stretch_max_rate: 1.5
138
+
139
+ inference:
140
+ batch_size: 1
141
+ dim_t: 256
142
+ num_overlap: 4
143
+ normalize: false
144
+
145
+ loss_multistft:
146
+ fft_sizes:
147
+ - 1024
148
+ - 2048
149
+ - 4096
150
+ hop_sizes:
151
+ - 147
152
+ - 256
153
+ - 512
154
+ win_lengths:
155
+ - 1024
156
+ - 2048
157
+ - 4096
158
+ window: "hann_window"
159
+ scale: "mel"
160
+ n_bins: 128
161
+ sample_rate: 44100
162
+ perceptual_weighting: true
163
+ w_sc: 16.0
164
+ w_log_mag: 16.0
165
+ w_lin_mag: 16.0
166
+ w_phs: 0.0
167
+ mag_distance: "L1"
models/Roformer/MelBand/config_MelBand_Roformer_4stems_XL_v1_by_Aname.yaml ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ audio:
2
+ chunk_size: 661500
3
+ dim_f: 1024
4
+ dim_t: 1101
5
+ hop_length: 882
6
+ n_fft: 4096
7
+ num_channels: 2
8
+ sample_rate: 44100
9
+ min_mean_abs: 0.0001
10
+
11
+ model:
12
+ dim: 512
13
+ depth: 12
14
+ stereo: true
15
+ num_stems: 4
16
+ linear_transformer_depth: 0
17
+ time_transformer_depth: 1
18
+ freq_transformer_depth: 1
19
+ num_bands: 60
20
+ dim_head: 64
21
+ heads: 8
22
+ attn_dropout: 0.0
23
+ ff_dropout: 0.0
24
+ flash_attn: true
25
+ dim_freqs_in: 2049
26
+ sample_rate: 44100 # needed for mel filter bank from librosa
27
+ stft_n_fft: 4096
28
+ stft_hop_length: 882
29
+ stft_win_length: 4096
30
+ stft_normalized: False
31
+ mask_estimator_depth: 2
32
+ multi_stft_resolution_loss_weight: 1.0
33
+ multi_stft_resolutions_window_sizes: !!python/tuple
34
+ - 4096
35
+ - 2048
36
+ - 1024
37
+ - 512
38
+ - 256
39
+ multi_stft_hop_size: 147
40
+ multi_stft_normalized: False
41
+ mlp_expansion_factor: 4
42
+ use_torch_checkpoint: False # it allows to greatly reduce GPU memory consumption during training (not fully tested)
43
+ skip_connection: True # Enable skip connection between transformer blocks - can solve problem with gradients and probably faster training
44
+
45
+ training:
46
+ batch_size: 1
47
+ gradient_accumulation_steps: 4
48
+ grad_clip: 0
49
+ instruments: ['drums', 'bass', 'other', 'vocals']
50
+ lr: 2.0e-05
51
+ patience: 2
52
+ reduce_factor: 0.95
53
+ target_instrument: null
54
+ num_epochs: 1000
55
+ num_steps: 300
56
+ q: 0.95
57
+ coarse_loss_clip: false
58
+ ema_momentum: 0.999
59
+ optimizer: adamw
60
+ read_metadata_procs: 8 # Number of processes to use during metadata reading for dataset. Can speed up metadata generation
61
+ normalize: false
62
+ other_fix: false # it's needed for checking on multisong dataset if other is actually instrumental
63
+ use_amp: true # enable or disable usage of mixed precision (float16) - usually it must be true
64
+
65
+ augmentations:
66
+ enable: false # enable or disable all augmentations (to fast disable if needed)
67
+ loudness: false # randomly change loudness of each stem on the range (loudness_min; loudness_max)
68
+ loudness_min: 0.5
69
+ loudness_max: 1.5
70
+ mixup: true # mix several stems of same type with some probability (only works for dataset types: 1, 2, 3)
71
+ mixup_probs: !!python/tuple # 2 additional stems of the same type (1st with prob 0.2, 2nd with prob 0.02)
72
+ - 0.2
73
+ - 0.02
74
+ - 0.002
75
+ mixup_loudness_min: 0.5
76
+ mixup_loudness_max: 1.5
77
+
78
+ # apply mp3 compression to mixture only (emulate downloading mp3 from internet)
79
+ mp3_compression_on_mixture: 0.01
80
+ mp3_compression_on_mixture_bitrate_min: 32
81
+ mp3_compression_on_mixture_bitrate_max: 320
82
+ mp3_compression_on_mixture_backend: "lameenc"
83
+
84
+ all:
85
+ channel_shuffle: 0.5 # Set 0 or lower to disable
86
+ random_inverse: 0.01 # inverse track (better lower probability)
87
+ random_polarity: 0.5 # polarity change (multiply waveform to -1)
88
+
89
+ vocals:
90
+ pitch_shift: 1.0
91
+ pitch_shift_min_semitones: -12
92
+ pitch_shift_max_semitones: 12
93
+ seven_band_parametric_eq: 0.5
94
+ seven_band_parametric_eq_min_gain_db: -80
95
+ seven_band_parametric_eq_max_gain_db: 9
96
+ tanh_distortion: 0.5
97
+ tanh_distortion_min: 0.1
98
+ tanh_distortion_max: 1
99
+ time_stretch: 1.0
100
+ time_stretch_min_rate: 0.5
101
+ time_stretch_max_rate: 2
102
+ bass:
103
+ pitch_shift: 1.0
104
+ pitch_shift_min_semitones: -6
105
+ pitch_shift_max_semitones: 6
106
+ seven_band_parametric_eq: 0.4
107
+ seven_band_parametric_eq_min_gain_db: -32
108
+ seven_band_parametric_eq_max_gain_db: 6
109
+ tanh_distortion: 1.0
110
+ tanh_distortion_min: 0.1
111
+ tanh_distortion_max: 0.5
112
+ time_stretch: 1.0
113
+ time_stretch_min_rate: 0.5
114
+ time_stretch_max_rate: 1.5
115
+ drums:
116
+ pitch_shift: 0.1
117
+ pitch_shift_min_semitones: -6
118
+ pitch_shift_max_semitones: 6
119
+ seven_band_parametric_eq: 0.5
120
+ seven_band_parametric_eq_min_gain_db: -24
121
+ seven_band_parametric_eq_max_gain_db: 12
122
+ tanh_distortion: 0.3
123
+ tanh_distortion_min: 0.1
124
+ tanh_distortion_max: 0.6
125
+ time_stretch: 1.0
126
+ time_stretch_min_rate: 0.333
127
+ time_stretch_max_rate: 1.5
128
+ other:
129
+ pitch_shift: 1.0
130
+ pitch_shift_min_semitones: -12
131
+ pitch_shift_max_semitones: 12
132
+ gaussian_noise: 0.4
133
+ gaussian_noise_min_amplitude: 0.001
134
+ gaussian_noise_max_amplitude: 0.15
135
+ time_stretch: 0.01
136
+ time_stretch_min_rate: 0.25
137
+ time_stretch_max_rate: 1.5
138
+
139
+ inference:
140
+ batch_size: 1
141
+ dim_t: 256
142
+ num_overlap: 4
143
+ normalize: false
144
+
145
+ loss_multistft:
146
+ fft_sizes:
147
+ - 1024
148
+ - 2048
149
+ - 4096
150
+ hop_sizes:
151
+ - 147
152
+ - 256
153
+ - 512
154
+ win_lengths:
155
+ - 1024
156
+ - 2048
157
+ - 4096
158
+ window: "hann_window"
159
+ scale: "mel"
160
+ n_bins: 128
161
+ sample_rate: 44100
162
+ perceptual_weighting: true
163
+ w_sc: 16.0
164
+ w_log_mag: 16.0
165
+ w_lin_mag: 16.0
166
+ w_phs: 0.0
167
+ mag_distance: "L1"