|
|
freeze_preset_selector = 5
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
net_g_mod = net_g.module if hasattr(net_g, 'module') else net_g
|
|
|
|
|
|
|
|
|
for param in net_g_mod.parameters():
|
|
|
param.requires_grad = True
|
|
|
|
|
|
active_freezing = False
|
|
|
|
|
|
if freeze_preset_selector == 0:
|
|
|
print("no layer freeze")
|
|
|
active_freezing = False
|
|
|
elif freeze_preset_selector == 1:
|
|
|
print("freeze: phone embeddings, first text enc attention layer, pos encoder pre-processing")
|
|
|
active_freezing = True
|
|
|
|
|
|
for param in net_g_mod.enc_p.emb_phone.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
|
|
|
for i, layer in enumerate(net_g_mod.enc_p.encoder.attn_layers):
|
|
|
if i < 1:
|
|
|
for param in layer.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
|
|
|
for param in net_g_mod.enc_q.pre.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
|
|
|
elif freeze_preset_selector == 2:
|
|
|
print("freeze: phone, first 2 text enc attention layers, pos encoder pre-processing & first few layers, and initial layers of decoder")
|
|
|
active_freezing = True
|
|
|
|
|
|
for param in net_g_mod.enc_p.emb_phone.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
|
|
|
for i, layer in enumerate(net_g_mod.enc_p.encoder.attn_layers):
|
|
|
if i < 2:
|
|
|
for param in layer.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
|
|
|
for param in net_g_mod.enc_q.pre.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
|
|
|
wavenet_module = net_g_mod.enc_q.enc
|
|
|
num_wavenet_layers_to_freeze = 2
|
|
|
|
|
|
for i, layer in enumerate(wavenet_module.in_layers):
|
|
|
if i < num_wavenet_layers_to_freeze:
|
|
|
for param in layer.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
for i, layer in enumerate(wavenet_module.res_skip_layers):
|
|
|
if i < num_wavenet_layers_to_freeze:
|
|
|
for param in layer.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
|
|
|
for i, upsample_layer in enumerate(net_g_mod.dec.ups):
|
|
|
if i < 1:
|
|
|
for param in upsample_layer.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
elif freeze_preset_selector == 3:
|
|
|
print("freezing only phone embeddings")
|
|
|
active_freezing = True
|
|
|
|
|
|
for param in net_g_mod.enc_p.emb_phone.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
|
|
|
elif freeze_preset_selector == 4:
|
|
|
print("freezing phone embeddings, all text enc main layers, pos encoder pre-processing, first 4 layers in pos encoder")
|
|
|
active_freezing = True
|
|
|
|
|
|
for param in net_g_mod.enc_p.emb_phone.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
|
|
|
for param in net_g_mod.enc_p.encoder.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
|
|
|
for param in net_g_mod.enc_q.pre.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
wavenet_module_p4 = net_g_mod.enc_q.enc
|
|
|
num_wavenet_layers_to_freeze_p4 = 4
|
|
|
for i, layer in enumerate(wavenet_module_p4.in_layers):
|
|
|
if i < num_wavenet_layers_to_freeze_p4:
|
|
|
for param in layer.parameters():
|
|
|
param.requires_grad = False
|
|
|
for i, layer in enumerate(wavenet_module_p4.res_skip_layers):
|
|
|
if i < num_wavenet_layers_to_freeze_p4:
|
|
|
for param in layer.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
elif freeze_preset_selector == 5:
|
|
|
print("freeze phone embedding, first 2 text enc attention layeer, pos encoder pre-processing, first 3 layers in pos encoder, decoder upsample block")
|
|
|
active_freezing = True
|
|
|
for param in net_g_mod.enc_p.emb_phone.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
for i, layer in enumerate(net_g_mod.enc_p.encoder.attn_layers):
|
|
|
if i < 2:
|
|
|
for param in layer.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
|
|
|
for param in net_g_mod.enc_q.pre.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
|
|
|
wavenet_module_p5 = net_g_mod.enc_q.enc
|
|
|
num_wavenet_layers_to_freeze_p5 = 3
|
|
|
for i, layer in enumerate(wavenet_module_p5.in_layers):
|
|
|
if i < num_wavenet_layers_to_freeze_p5:
|
|
|
for param in layer.parameters():
|
|
|
param.requires_grad = False
|
|
|
for i, layer in enumerate(wavenet_module_p5.res_skip_layers):
|
|
|
if i < num_wavenet_layers_to_freeze_p5:
|
|
|
for param in layer.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
for i, upsample_layer in enumerate(net_g_mod.dec.ups):
|
|
|
if i < 1:
|
|
|
for param in upsample_layer.parameters():
|
|
|
param.requires_grad = False
|
|
|
|
|
|
|
|
|
else:
|
|
|
raise ValueError(f"invalid preset")
|
|
|
|
|
|
if active_freezing:
|
|
|
total_params = 0
|
|
|
frozen_params = 0
|
|
|
for name, param in net_g_mod.named_parameters():
|
|
|
total_params += param.numel()
|
|
|
if not param.requires_grad:
|
|
|
frozen_params += param.numel()
|
|
|
print(f"Freezing applied (Preset {freeze_preset_selector}): {frozen_params:,}/{total_params:,} parameters frozen.")
|
|
|
else:
|
|
|
total_params = sum(p.numel() for p in net_g_mod.parameters())
|
|
|
print(f"no freezing applied") |