|
|
from typing import Any, Callable |
|
|
|
|
|
import lightning as L |
|
|
import torch |
|
|
import torch.nn.functional as F |
|
|
import wandb |
|
|
from lightning.pytorch.loggers import TensorBoardLogger, WandbLogger |
|
|
from matplotlib import pyplot as plt |
|
|
from torch import nn |
|
|
|
|
|
|
|
|
from fish_speech.models.vqgan.utils import ( |
|
|
avg_with_mask, |
|
|
plot_mel, |
|
|
sequence_mask |
|
|
) |
|
|
from fish_speech.third_party.vocoder_infer import BigVGAN |
|
|
|
|
|
class FlowDecoder(L.LightningModule): |
|
|
def __init__( |
|
|
self, |
|
|
optimizer: Callable, |
|
|
lr_scheduler: Callable, |
|
|
generator: nn.Module, |
|
|
hop_length: int = 320, |
|
|
sample_rate: int = 24000, |
|
|
mel_min: float = -1, |
|
|
mel_max: float = 1, |
|
|
lambda_gen: float = 1.0, |
|
|
lambda_dur: float = 1.0, |
|
|
): |
|
|
super().__init__() |
|
|
|
|
|
|
|
|
self.optimizer_builder = optimizer |
|
|
self.lr_scheduler_builder = lr_scheduler |
|
|
|
|
|
|
|
|
self.generator = generator |
|
|
|
|
|
|
|
|
self.mel_min = mel_min |
|
|
self.mel_max = mel_max |
|
|
|
|
|
|
|
|
self.hop_length = hop_length |
|
|
self.sampling_rate = sample_rate |
|
|
self.lambda_gen =lambda_gen |
|
|
self.lambda_dur = lambda_dur |
|
|
|
|
|
|
|
|
self.automatic_optimization = False |
|
|
|
|
|
self.vocoder = BigVGAN(next(self.generator.parameters()).device) |
|
|
|
|
|
|
|
|
def configure_optimizers(self): |
|
|
|
|
|
optimizer_generator = self.optimizer_builder(self.generator.parameters()) |
|
|
|
|
|
lr_scheduler_generator = self.lr_scheduler_builder(optimizer_generator) |
|
|
|
|
|
return ( |
|
|
{ |
|
|
"optimizer": optimizer_generator, |
|
|
"lr_scheduler": { |
|
|
"scheduler": lr_scheduler_generator, |
|
|
"interval": "step", |
|
|
"name": "optimizer/generator", |
|
|
}, |
|
|
} |
|
|
) |
|
|
|
|
|
def training_step(self, batch, batch_idx): |
|
|
optim_g = self.optimizers() |
|
|
|
|
|
mels, mel_lengths = batch["mels"], batch["mel_lengths"] |
|
|
hubert_codes, hubert_code_lengths = batch["hubert_codes"], batch["hubert_code_lengths"] |
|
|
style_codes, style_code_lengths = batch["style_codes"], batch["style_code_lengths"] |
|
|
durations = batch["durations"] |
|
|
|
|
|
with torch.no_grad(): |
|
|
gt_mels = mels |
|
|
|
|
|
gt_mels = (gt_mels - self.mel_min) / (self.mel_max - self.mel_min) |
|
|
|
|
|
ret = self.generator(gt_mels,mel_lengths,hubert_codes,hubert_code_lengths, |
|
|
style_codes,style_code_lengths,durations) |
|
|
|
|
|
gen_loss = ret['loss'] |
|
|
dur_loss = ret['dur_loss'] |
|
|
self.log( |
|
|
"train/generator/loss", |
|
|
gen_loss, |
|
|
on_step=True, |
|
|
on_epoch=False, |
|
|
prog_bar=True, |
|
|
logger=True, |
|
|
sync_dist=True, |
|
|
) |
|
|
self.log( |
|
|
"train/generator/dur_loss", |
|
|
dur_loss, |
|
|
on_step=True, |
|
|
on_epoch=False, |
|
|
prog_bar=True, |
|
|
logger=True, |
|
|
sync_dist=True, |
|
|
) |
|
|
|
|
|
loss = gen_loss * self.lambda_gen + dur_loss * self.lambda_dur |
|
|
|
|
|
|
|
|
optim_g.zero_grad() |
|
|
|
|
|
self.manual_backward(loss) |
|
|
self.clip_gradients( |
|
|
optim_g, gradient_clip_val=1000.0, gradient_clip_algorithm="norm" |
|
|
) |
|
|
optim_g.step() |
|
|
|
|
|
|
|
|
scheduler_g = self.lr_schedulers() |
|
|
scheduler_g.step() |
|
|
|
|
|
def validation_step(self, batch: Any, batch_idx: int): |
|
|
if batch_idx >= 2: |
|
|
return |
|
|
mels, mel_lengths = batch["mels"], batch["mel_lengths"] |
|
|
hubert_codes, hubert_code_lengths = batch["hubert_codes"], batch["hubert_code_lengths"] |
|
|
style_codes, style_code_lengths = batch["style_codes"], batch["style_code_lengths"] |
|
|
durations = batch["durations"] |
|
|
|
|
|
with torch.no_grad(): |
|
|
gt_mels = mels |
|
|
|
|
|
norm_gt_mels = (gt_mels - self.mel_min) / (self.mel_max - self.mel_min) |
|
|
mel_masks = torch.unsqueeze( |
|
|
sequence_mask(mel_lengths, norm_gt_mels.shape[2]), 1 |
|
|
).to(norm_gt_mels.dtype) |
|
|
|
|
|
refine_mels, dur_loss = self.generator.inference(norm_gt_mels,mel_lengths,hubert_codes,hubert_code_lengths, |
|
|
style_codes,style_code_lengths,durations) |
|
|
|
|
|
refine_mels = refine_mels*(self.mel_max - self.mel_min)+ self.mel_min |
|
|
|
|
|
min_mel_length = min(gt_mels.shape[-1], refine_mels.shape[-1]) |
|
|
gt_mels = gt_mels[:, :, :min_mel_length] |
|
|
refine_mels = refine_mels[:, :, :min_mel_length] |
|
|
|
|
|
refine_mel_loss = avg_with_mask( |
|
|
F.l1_loss(gt_mels, refine_mels, reduction="none"), mel_masks |
|
|
) |
|
|
|
|
|
self.log( |
|
|
"val/refine_mel_loss", |
|
|
refine_mel_loss, |
|
|
on_step=False, |
|
|
on_epoch=True, |
|
|
prog_bar=False, |
|
|
logger=True, |
|
|
sync_dist=True, |
|
|
) |
|
|
self.log( |
|
|
"val/dur_loss", |
|
|
dur_loss, |
|
|
on_step=False, |
|
|
on_epoch=True, |
|
|
prog_bar=False, |
|
|
logger=True, |
|
|
sync_dist=True, |
|
|
) |
|
|
|
|
|
|
|
|
if batch_idx != 0: |
|
|
return |
|
|
|
|
|
gt_audios = self.vocoder.spec2wav(gt_mels) |
|
|
gen_audios = self.vocoder.spec2wav(refine_mels) |
|
|
|
|
|
for idx, ( |
|
|
mel, |
|
|
refine_mel, |
|
|
mel_len, |
|
|
gt_audio, |
|
|
gen_audio |
|
|
) in enumerate( |
|
|
zip( |
|
|
gt_mels, |
|
|
refine_mels, |
|
|
mel_lengths, |
|
|
gt_audios.cpu().float(), |
|
|
gen_audios.cpu().float() |
|
|
) |
|
|
): |
|
|
if idx > 4: |
|
|
break |
|
|
audio_len = mel_len * self.hop_length |
|
|
|
|
|
image_mels = plot_mel( |
|
|
[ |
|
|
refine_mel[:, :mel_len], |
|
|
mel[:, :mel_len], |
|
|
], |
|
|
[ |
|
|
"Refine (Flow)", |
|
|
"Ground-Truth", |
|
|
], |
|
|
) |
|
|
|
|
|
if isinstance(self.logger, WandbLogger): |
|
|
self.logger.experiment.log( |
|
|
{ |
|
|
"reconstruction_mel": wandb.Image(image_mels, caption="mels"), |
|
|
"wavs": [ |
|
|
wandb.Audio( |
|
|
gt_audio[:audio_len], |
|
|
sample_rate=self.sampling_rate, |
|
|
caption="gt", |
|
|
), |
|
|
wandb.Audio( |
|
|
gen_audio[:audio_len], |
|
|
sample_rate=self.sampling_rate, |
|
|
caption="recon", |
|
|
), |
|
|
], |
|
|
}, |
|
|
) |
|
|
|
|
|
if isinstance(self.logger, TensorBoardLogger): |
|
|
self.logger.experiment.add_figure( |
|
|
f"sample-{idx}/mels", |
|
|
image_mels, |
|
|
global_step=self.global_step, |
|
|
) |
|
|
self.logger.experiment.add_audio( |
|
|
f"sample-{idx}/wavs/gt", |
|
|
gt_audio[:audio_len], |
|
|
self.global_step, |
|
|
sample_rate=self.sampling_rate, |
|
|
) |
|
|
self.logger.experiment.add_audio( |
|
|
f"sample-{idx}/wavs/gen", |
|
|
gen_audio[:audio_len], |
|
|
self.global_step, |
|
|
sample_rate=self.sampling_rate, |
|
|
) |
|
|
|
|
|
plt.close(image_mels) |
|
|
|
|
|
|
|
|
class FlowNewDecoder(L.LightningModule): |
|
|
def __init__( |
|
|
self, |
|
|
optimizer: Callable, |
|
|
lr_scheduler: Callable, |
|
|
generator: nn.Module, |
|
|
hop_length: int = 320, |
|
|
sample_rate: int = 24000, |
|
|
mel_min: float = -1, |
|
|
mel_max: float = 1, |
|
|
lambda_gen: float = 1.0, |
|
|
lambda_dur: float = 1.0, |
|
|
lambda_prior: float = 1.0, |
|
|
): |
|
|
super().__init__() |
|
|
|
|
|
|
|
|
self.optimizer_builder = optimizer |
|
|
self.lr_scheduler_builder = lr_scheduler |
|
|
|
|
|
|
|
|
self.generator = generator |
|
|
|
|
|
|
|
|
self.mel_min = mel_min |
|
|
self.mel_max = mel_max |
|
|
|
|
|
|
|
|
self.hop_length = hop_length |
|
|
self.sampling_rate = sample_rate |
|
|
self.lambda_gen =lambda_gen |
|
|
self.lambda_dur = lambda_dur |
|
|
self.lambda_prior = lambda_prior |
|
|
|
|
|
|
|
|
self.automatic_optimization = False |
|
|
|
|
|
self.vocoder = BigVGAN(next(self.generator.parameters()).device) |
|
|
|
|
|
|
|
|
def configure_optimizers(self): |
|
|
|
|
|
optimizer_generator = self.optimizer_builder(self.generator.parameters()) |
|
|
|
|
|
lr_scheduler_generator = self.lr_scheduler_builder(optimizer_generator) |
|
|
|
|
|
return ( |
|
|
{ |
|
|
"optimizer": optimizer_generator, |
|
|
"lr_scheduler": { |
|
|
"scheduler": lr_scheduler_generator, |
|
|
"interval": "step", |
|
|
"name": "optimizer/generator", |
|
|
}, |
|
|
} |
|
|
) |
|
|
|
|
|
def training_step(self, batch, batch_idx): |
|
|
optim_g = self.optimizers() |
|
|
|
|
|
mels, mel_lengths = batch["mels"], batch["mel_lengths"] |
|
|
hubert_codes, hubert_code_lengths = batch["hubert_codes"], batch["hubert_code_lengths"] |
|
|
style_codes, style_code_lengths = batch["style_codes"], batch["style_code_lengths"] |
|
|
durations = batch["durations"] |
|
|
|
|
|
with torch.no_grad(): |
|
|
gt_mels = mels |
|
|
|
|
|
gt_mels = (gt_mels - self.mel_min) / (self.mel_max - self.mel_min) |
|
|
|
|
|
ret = self.generator(gt_mels,mel_lengths,hubert_codes,hubert_code_lengths, |
|
|
style_codes,style_code_lengths,durations) |
|
|
|
|
|
gen_loss = ret['loss'] |
|
|
dur_loss = ret['dur_loss'] |
|
|
prior_loss = ret['prior_loss'] |
|
|
self.log( |
|
|
"train/generator/loss", |
|
|
gen_loss, |
|
|
on_step=True, |
|
|
on_epoch=False, |
|
|
prog_bar=True, |
|
|
logger=True, |
|
|
sync_dist=True, |
|
|
) |
|
|
self.log( |
|
|
"train/generator/dur_loss", |
|
|
dur_loss, |
|
|
on_step=True, |
|
|
on_epoch=False, |
|
|
prog_bar=True, |
|
|
logger=True, |
|
|
sync_dist=True, |
|
|
) |
|
|
self.log( |
|
|
"train/generator/prior_loss", |
|
|
prior_loss, |
|
|
on_step=True, |
|
|
on_epoch=False, |
|
|
prog_bar=True, |
|
|
logger=True, |
|
|
sync_dist=True, |
|
|
) |
|
|
|
|
|
loss = gen_loss * self.lambda_gen + dur_loss * self.lambda_dur + prior_loss * self.lambda_prior |
|
|
|
|
|
|
|
|
optim_g.zero_grad() |
|
|
|
|
|
self.manual_backward(loss) |
|
|
self.clip_gradients( |
|
|
optim_g, gradient_clip_val=1000.0, gradient_clip_algorithm="norm" |
|
|
) |
|
|
optim_g.step() |
|
|
|
|
|
|
|
|
scheduler_g = self.lr_schedulers() |
|
|
scheduler_g.step() |
|
|
|
|
|
def validation_step(self, batch: Any, batch_idx: int): |
|
|
mels, mel_lengths = batch["mels"], batch["mel_lengths"] |
|
|
hubert_codes, hubert_code_lengths = batch["hubert_codes"], batch["hubert_code_lengths"] |
|
|
style_codes, style_code_lengths = batch["style_codes"], batch["style_code_lengths"] |
|
|
durations = batch["durations"] |
|
|
|
|
|
with torch.no_grad(): |
|
|
gt_mels = mels |
|
|
|
|
|
norm_gt_mels = (gt_mels - self.mel_min) / (self.mel_max - self.mel_min) |
|
|
mel_masks = torch.unsqueeze( |
|
|
sequence_mask(mel_lengths, norm_gt_mels.shape[2]), 1 |
|
|
).to(norm_gt_mels.dtype) |
|
|
|
|
|
refine_mels, dur_loss, prior_loss = self.generator.inference(norm_gt_mels,mel_lengths,hubert_codes,hubert_code_lengths, |
|
|
style_codes,style_code_lengths,durations) |
|
|
|
|
|
refine_mels = refine_mels*(self.mel_max - self.mel_min)+ self.mel_min |
|
|
|
|
|
min_mel_length = min(gt_mels.shape[-1], refine_mels.shape[-1]) |
|
|
gt_mels = gt_mels[:, :, :min_mel_length] |
|
|
refine_mels = refine_mels[:, :, :min_mel_length] |
|
|
|
|
|
refine_mel_loss = avg_with_mask( |
|
|
F.l1_loss(gt_mels, refine_mels, reduction="none"), mel_masks |
|
|
) |
|
|
|
|
|
self.log( |
|
|
"val/refine_mel_loss", |
|
|
refine_mel_loss, |
|
|
on_step=False, |
|
|
on_epoch=True, |
|
|
prog_bar=False, |
|
|
logger=True, |
|
|
sync_dist=True, |
|
|
) |
|
|
self.log( |
|
|
"val/dur_loss", |
|
|
dur_loss, |
|
|
on_step=False, |
|
|
on_epoch=True, |
|
|
prog_bar=False, |
|
|
logger=True, |
|
|
sync_dist=True, |
|
|
) |
|
|
self.log( |
|
|
"val/prior_loss", |
|
|
prior_loss, |
|
|
on_step=False, |
|
|
on_epoch=True, |
|
|
prog_bar=False, |
|
|
logger=True, |
|
|
sync_dist=True, |
|
|
) |
|
|
|
|
|
|
|
|
if batch_idx != 0: |
|
|
return |
|
|
|
|
|
gt_audios = self.vocoder.spec2wav(gt_mels) |
|
|
gen_audios = self.vocoder.spec2wav(refine_mels) |
|
|
|
|
|
for idx, ( |
|
|
mel, |
|
|
refine_mel, |
|
|
mel_len, |
|
|
gt_audio, |
|
|
gen_audio |
|
|
) in enumerate( |
|
|
zip( |
|
|
gt_mels, |
|
|
refine_mels, |
|
|
mel_lengths, |
|
|
gt_audios.cpu().float(), |
|
|
gen_audios.cpu().float() |
|
|
) |
|
|
): |
|
|
if idx > 4: |
|
|
break |
|
|
audio_len = mel_len * self.hop_length |
|
|
|
|
|
image_mels = plot_mel( |
|
|
[ |
|
|
refine_mel[:, :mel_len], |
|
|
mel[:, :mel_len], |
|
|
], |
|
|
[ |
|
|
"Refine (Flow)", |
|
|
"Ground-Truth", |
|
|
], |
|
|
) |
|
|
|
|
|
if isinstance(self.logger, WandbLogger): |
|
|
self.logger.experiment.log( |
|
|
{ |
|
|
"reconstruction_mel": wandb.Image(image_mels, caption="mels"), |
|
|
"wavs": [ |
|
|
wandb.Audio( |
|
|
gt_audio[:audio_len], |
|
|
sample_rate=self.sampling_rate, |
|
|
caption="gt", |
|
|
), |
|
|
wandb.Audio( |
|
|
gen_audio[:audio_len], |
|
|
sample_rate=self.sampling_rate, |
|
|
caption="recon", |
|
|
), |
|
|
], |
|
|
}, |
|
|
) |
|
|
|
|
|
if isinstance(self.logger, TensorBoardLogger): |
|
|
self.logger.experiment.add_figure( |
|
|
f"sample-{idx}/mels", |
|
|
image_mels, |
|
|
global_step=self.global_step, |
|
|
) |
|
|
self.logger.experiment.add_audio( |
|
|
f"sample-{idx}/wavs/gt", |
|
|
gt_audio[:audio_len], |
|
|
self.global_step, |
|
|
sample_rate=self.sampling_rate, |
|
|
) |
|
|
self.logger.experiment.add_audio( |
|
|
f"sample-{idx}/wavs/gen", |
|
|
gen_audio[:audio_len], |
|
|
self.global_step, |
|
|
sample_rate=self.sampling_rate, |
|
|
) |
|
|
|
|
|
plt.close(image_mels) |
|
|
|
|
|
|
|
|
class FlowDecoder_Wo_Style(L.LightningModule): |
|
|
def __init__( |
|
|
self, |
|
|
optimizer: Callable, |
|
|
lr_scheduler: Callable, |
|
|
generator: nn.Module, |
|
|
hop_length: int = 320, |
|
|
sample_rate: int = 24000, |
|
|
mel_min: float = -1, |
|
|
mel_max: float = 1, |
|
|
lambda_gen: float = 1.0, |
|
|
lambda_dur: float = 1.0 |
|
|
): |
|
|
super().__init__() |
|
|
|
|
|
|
|
|
self.optimizer_builder = optimizer |
|
|
self.lr_scheduler_builder = lr_scheduler |
|
|
|
|
|
|
|
|
self.generator = generator |
|
|
|
|
|
|
|
|
self.mel_min = mel_min |
|
|
self.mel_max = mel_max |
|
|
|
|
|
|
|
|
self.hop_length = hop_length |
|
|
self.sampling_rate = sample_rate |
|
|
self.lambda_gen =lambda_gen |
|
|
self.lambda_dur = lambda_dur |
|
|
|
|
|
|
|
|
self.automatic_optimization = False |
|
|
|
|
|
self.vocoder = BigVGAN(next(self.generator.parameters()).device) |
|
|
|
|
|
|
|
|
def configure_optimizers(self): |
|
|
|
|
|
optimizer_generator = self.optimizer_builder(self.generator.parameters()) |
|
|
|
|
|
lr_scheduler_generator = self.lr_scheduler_builder(optimizer_generator) |
|
|
|
|
|
return ( |
|
|
{ |
|
|
"optimizer": optimizer_generator, |
|
|
"lr_scheduler": { |
|
|
"scheduler": lr_scheduler_generator, |
|
|
"interval": "step", |
|
|
"name": "optimizer/generator", |
|
|
}, |
|
|
} |
|
|
) |
|
|
|
|
|
def training_step(self, batch, batch_idx): |
|
|
optim_g = self.optimizers() |
|
|
|
|
|
mels, mel_lengths = batch["mels"], batch["mel_lengths"] |
|
|
hubert_codes, hubert_code_lengths = batch["hubert_codes"], batch["hubert_code_lengths"] |
|
|
durations = batch["durations"] |
|
|
|
|
|
with torch.no_grad(): |
|
|
gt_mels = mels |
|
|
|
|
|
gt_mels = (gt_mels - self.mel_min) / (self.mel_max - self.mel_min) |
|
|
|
|
|
ret = self.generator(gt_mels,mel_lengths,hubert_codes,hubert_code_lengths,durations) |
|
|
|
|
|
gen_loss = ret['loss'] |
|
|
dur_loss = ret['dur_loss'] |
|
|
self.log( |
|
|
"train/generator/loss", |
|
|
gen_loss, |
|
|
on_step=True, |
|
|
on_epoch=False, |
|
|
prog_bar=True, |
|
|
logger=True, |
|
|
sync_dist=True, |
|
|
) |
|
|
self.log( |
|
|
"train/generator/dur_loss", |
|
|
dur_loss, |
|
|
on_step=True, |
|
|
on_epoch=False, |
|
|
prog_bar=True, |
|
|
logger=True, |
|
|
sync_dist=True, |
|
|
) |
|
|
|
|
|
loss = gen_loss * self.lambda_gen + dur_loss * self.lambda_dur |
|
|
|
|
|
|
|
|
optim_g.zero_grad() |
|
|
|
|
|
self.manual_backward(loss) |
|
|
self.clip_gradients( |
|
|
optim_g, gradient_clip_val=1000.0, gradient_clip_algorithm="norm" |
|
|
) |
|
|
optim_g.step() |
|
|
|
|
|
|
|
|
scheduler_g = self.lr_schedulers() |
|
|
scheduler_g.step() |
|
|
|
|
|
def validation_step(self, batch: Any, batch_idx: int): |
|
|
if batch_idx >= 2: |
|
|
return |
|
|
mels, mel_lengths = batch["mels"], batch["mel_lengths"] |
|
|
hubert_codes, hubert_code_lengths = batch["hubert_codes"], batch["hubert_code_lengths"] |
|
|
durations = batch["durations"] |
|
|
|
|
|
with torch.no_grad(): |
|
|
gt_mels = mels |
|
|
|
|
|
norm_gt_mels = (gt_mels - self.mel_min) / (self.mel_max - self.mel_min) |
|
|
mel_masks = torch.unsqueeze( |
|
|
sequence_mask(mel_lengths, norm_gt_mels.shape[2]), 1 |
|
|
).to(norm_gt_mels.dtype) |
|
|
|
|
|
refine_mels, dur_loss = self.generator.inference(norm_gt_mels,mel_lengths,hubert_codes,hubert_code_lengths,durations) |
|
|
|
|
|
refine_mels = refine_mels*(self.mel_max - self.mel_min)+ self.mel_min |
|
|
|
|
|
min_mel_length = min(gt_mels.shape[-1], refine_mels.shape[-1]) |
|
|
gt_mels = gt_mels[:, :, :min_mel_length] |
|
|
refine_mels = refine_mels[:, :, :min_mel_length] |
|
|
|
|
|
refine_mel_loss = avg_with_mask( |
|
|
F.l1_loss(gt_mels, refine_mels, reduction="none"), mel_masks |
|
|
) |
|
|
|
|
|
self.log( |
|
|
"val/refine_mel_loss", |
|
|
refine_mel_loss, |
|
|
on_step=False, |
|
|
on_epoch=True, |
|
|
prog_bar=False, |
|
|
logger=True, |
|
|
sync_dist=True, |
|
|
) |
|
|
self.log( |
|
|
"val/dur_loss", |
|
|
dur_loss, |
|
|
on_step=False, |
|
|
on_epoch=True, |
|
|
prog_bar=False, |
|
|
logger=True, |
|
|
sync_dist=True, |
|
|
) |
|
|
|
|
|
|
|
|
if batch_idx != 0: |
|
|
return |
|
|
|
|
|
gt_audios = self.vocoder.spec2wav(gt_mels) |
|
|
gen_audios = self.vocoder.spec2wav(refine_mels) |
|
|
|
|
|
for idx, ( |
|
|
mel, |
|
|
refine_mel, |
|
|
mel_len, |
|
|
gt_audio, |
|
|
gen_audio |
|
|
) in enumerate( |
|
|
zip( |
|
|
gt_mels, |
|
|
refine_mels, |
|
|
mel_lengths, |
|
|
gt_audios.cpu().float(), |
|
|
gen_audios.cpu().float() |
|
|
) |
|
|
): |
|
|
if idx > 4: |
|
|
break |
|
|
audio_len = mel_len * self.hop_length |
|
|
|
|
|
image_mels = plot_mel( |
|
|
[ |
|
|
refine_mel[:, :mel_len], |
|
|
mel[:, :mel_len], |
|
|
], |
|
|
[ |
|
|
"Refine (Flow)", |
|
|
"Ground-Truth", |
|
|
], |
|
|
) |
|
|
|
|
|
if isinstance(self.logger, WandbLogger): |
|
|
self.logger.experiment.log( |
|
|
{ |
|
|
"reconstruction_mel": wandb.Image(image_mels, caption="mels"), |
|
|
"wavs": [ |
|
|
wandb.Audio( |
|
|
gt_audio[:audio_len], |
|
|
sample_rate=self.sampling_rate, |
|
|
caption="gt", |
|
|
), |
|
|
wandb.Audio( |
|
|
gen_audio[:audio_len], |
|
|
sample_rate=self.sampling_rate, |
|
|
caption="recon", |
|
|
), |
|
|
], |
|
|
}, |
|
|
) |
|
|
|
|
|
if isinstance(self.logger, TensorBoardLogger): |
|
|
self.logger.experiment.add_figure( |
|
|
f"sample-{idx}/mels", |
|
|
image_mels, |
|
|
global_step=self.global_step, |
|
|
) |
|
|
self.logger.experiment.add_audio( |
|
|
f"sample-{idx}/wavs/gt", |
|
|
gt_audio[:audio_len], |
|
|
self.global_step, |
|
|
sample_rate=self.sampling_rate, |
|
|
) |
|
|
self.logger.experiment.add_audio( |
|
|
f"sample-{idx}/wavs/gen", |
|
|
gen_audio[:audio_len], |
|
|
self.global_step, |
|
|
sample_rate=self.sampling_rate, |
|
|
) |
|
|
|
|
|
plt.close(image_mels) |