jlking's picture
Upload folder using huggingface_hub
7375975 verified
from typing import Any, Callable
import lightning as L
import torch
import torch.nn.functional as F
import wandb
from lightning.pytorch.loggers import TensorBoardLogger, WandbLogger
from matplotlib import pyplot as plt
from torch import nn
from fish_speech.models.vits_decoder.losses import (
discriminator_loss,
feature_loss,
generator_loss,
kl_loss,
)
from fish_speech.models.vqgan.utils import (
avg_with_mask,
plot_mel,
sequence_mask,
slice_segments,
)
class VITSDecoder(L.LightningModule):
def __init__(
self,
optimizer: Callable,
lr_scheduler: Callable,
generator: nn.Module,
discriminator: nn.Module,
mel_transform: nn.Module,
spec_transform: nn.Module,
hop_length: int = 512,
sample_rate: int = 44100,
freeze_discriminator: bool = False,
weight_mel: float = 45,
weight_kl: float = 0.1,
):
super().__init__()
# Model parameters
self.optimizer_builder = optimizer
self.lr_scheduler_builder = lr_scheduler
# Generator and discriminator
self.generator = generator
self.discriminator = discriminator
self.mel_transform = mel_transform
self.spec_transform = spec_transform
self.freeze_discriminator = freeze_discriminator
# Loss weights
self.weight_mel = weight_mel
self.weight_kl = weight_kl
# Other parameters
self.hop_length = hop_length
self.sampling_rate = sample_rate
# Disable automatic optimization
self.automatic_optimization = False
if self.freeze_discriminator:
for p in self.discriminator.parameters():
p.requires_grad = False
def configure_optimizers(self):
# Need two optimizers and two schedulers
optimizer_generator = self.optimizer_builder(self.generator.parameters())
optimizer_discriminator = self.optimizer_builder(
self.discriminator.parameters()
)
lr_scheduler_generator = self.lr_scheduler_builder(optimizer_generator)
lr_scheduler_discriminator = self.lr_scheduler_builder(optimizer_discriminator)
return (
{
"optimizer": optimizer_generator,
"lr_scheduler": {
"scheduler": lr_scheduler_generator,
"interval": "step",
"name": "optimizer/generator",
},
},
{
"optimizer": optimizer_discriminator,
"lr_scheduler": {
"scheduler": lr_scheduler_discriminator,
"interval": "step",
"name": "optimizer/discriminator",
},
},
)
def training_step(self, batch, batch_idx):
optim_g, optim_d = self.optimizers()
audios, audio_lengths = batch["audios"], batch["audio_lengths"]
texts, text_lengths = batch["texts"], batch["text_lengths"]
audios = audios.float()
audios = audios[:, None, :]
with torch.no_grad():
gt_mels = self.mel_transform(audios)
gt_specs = self.spec_transform(audios)
spec_lengths = audio_lengths // self.hop_length
spec_masks = torch.unsqueeze(
sequence_mask(spec_lengths, gt_mels.shape[2]), 1
).to(gt_mels.dtype)
(
fake_audios,
ids_slice,
y_mask,
(z, z_p, m_p, logs_p, m_q, logs_q),
) = self.generator(
audios,
audio_lengths,
gt_specs,
spec_lengths,
texts,
text_lengths,
)
gt_mels = slice_segments(gt_mels, ids_slice, self.generator.segment_size)
spec_masks = slice_segments(spec_masks, ids_slice, self.generator.segment_size)
audios = slice_segments(
audios,
ids_slice * self.hop_length,
self.generator.segment_size * self.hop_length,
)
fake_mels = self.mel_transform(fake_audios.squeeze(1))
assert (
audios.shape == fake_audios.shape
), f"{audios.shape} != {fake_audios.shape}"
# Discriminator
if self.freeze_discriminator is False:
y_d_hat_r, y_d_hat_g, _, _ = self.discriminator(
audios, fake_audios.detach()
)
with torch.autocast(device_type=audios.device.type, enabled=False):
loss_disc, _, _ = discriminator_loss(y_d_hat_r, y_d_hat_g)
self.log(
f"train/discriminator/loss",
loss_disc,
on_step=True,
on_epoch=False,
prog_bar=False,
logger=True,
sync_dist=True,
)
optim_d.zero_grad()
self.manual_backward(loss_disc)
self.clip_gradients(
optim_d, gradient_clip_val=1000.0, gradient_clip_algorithm="norm"
)
optim_d.step()
# Adv Loss
y_d_hat_r, y_d_hat_g, _, _ = self.discriminator(audios, fake_audios)
# Adversarial Loss
with torch.autocast(device_type=audios.device.type, enabled=False):
loss_adv, _ = generator_loss(y_d_hat_g)
self.log(
f"train/generator/adv",
loss_adv,
on_step=True,
on_epoch=False,
prog_bar=False,
logger=True,
sync_dist=True,
)
with torch.autocast(device_type=audios.device.type, enabled=False):
loss_fm = feature_loss(y_d_hat_r, y_d_hat_g)
self.log(
f"train/generator/adv_fm",
loss_fm,
on_step=True,
on_epoch=False,
prog_bar=False,
logger=True,
sync_dist=True,
)
with torch.autocast(device_type=audios.device.type, enabled=False):
loss_mel = avg_with_mask(
F.l1_loss(gt_mels, fake_mels, reduction="none"), spec_masks
)
self.log(
"train/generator/loss_mel",
loss_mel,
on_step=True,
on_epoch=False,
prog_bar=False,
logger=True,
sync_dist=True,
)
loss_kl = kl_loss(z_p, logs_q, m_p, logs_p, y_mask)
self.log(
"train/generator/loss_kl",
loss_kl,
on_step=True,
on_epoch=False,
prog_bar=False,
logger=True,
sync_dist=True,
)
loss = (
loss_mel * self.weight_mel + loss_kl * self.weight_kl + loss_adv + loss_fm
)
self.log(
"train/generator/loss",
loss,
on_step=True,
on_epoch=False,
prog_bar=True,
logger=True,
sync_dist=True,
)
# Backward
optim_g.zero_grad()
self.manual_backward(loss)
self.clip_gradients(
optim_g, gradient_clip_val=1000.0, gradient_clip_algorithm="norm"
)
optim_g.step()
# Manual LR Scheduler
scheduler_g, scheduler_d = self.lr_schedulers()
scheduler_g.step()
scheduler_d.step()
def validation_step(self, batch: Any, batch_idx: int):
audios, audio_lengths = batch["audios"], batch["audio_lengths"]
texts, text_lengths = batch["texts"], batch["text_lengths"]
audios = audios.float()
audios = audios[:, None, :]
gt_mels = self.mel_transform(audios)
gt_specs = self.spec_transform(audios)
spec_lengths = audio_lengths // self.hop_length
spec_masks = torch.unsqueeze(
sequence_mask(spec_lengths, gt_mels.shape[2]), 1
).to(gt_mels.dtype)
prior_audios = self.generator.infer(
audios, audio_lengths, gt_specs, spec_lengths, texts, text_lengths
)
posterior_audios = self.generator.infer_posterior(gt_specs, spec_lengths)
prior_mels = self.mel_transform(prior_audios.squeeze(1))
posterior_mels = self.mel_transform(posterior_audios.squeeze(1))
min_mel_length = min(
gt_mels.shape[-1], prior_mels.shape[-1], posterior_mels.shape[-1]
)
gt_mels = gt_mels[:, :, :min_mel_length]
prior_mels = prior_mels[:, :, :min_mel_length]
posterior_mels = posterior_mels[:, :, :min_mel_length]
prior_mel_loss = avg_with_mask(
F.l1_loss(gt_mels, prior_mels, reduction="none"), spec_masks
)
posterior_mel_loss = avg_with_mask(
F.l1_loss(gt_mels, posterior_mels, reduction="none"), spec_masks
)
self.log(
"val/prior_mel_loss",
prior_mel_loss,
on_step=False,
on_epoch=True,
prog_bar=False,
logger=True,
sync_dist=True,
)
self.log(
"val/posterior_mel_loss",
posterior_mel_loss,
on_step=False,
on_epoch=True,
prog_bar=False,
logger=True,
sync_dist=True,
)
# only log the first batch
if batch_idx != 0:
return
for idx, (
mel,
prior_mel,
posterior_mel,
audio,
prior_audio,
posterior_audio,
audio_len,
) in enumerate(
zip(
gt_mels,
prior_mels,
posterior_mels,
audios.detach().float(),
prior_audios.detach().float(),
posterior_audios.detach().float(),
audio_lengths,
)
):
mel_len = audio_len // self.hop_length
image_mels = plot_mel(
[
prior_mel[:, :mel_len],
posterior_mel[:, :mel_len],
mel[:, :mel_len],
],
[
"Prior (VQ)",
"Posterior (Reconstruction)",
"Ground-Truth",
],
)
if isinstance(self.logger, WandbLogger):
self.logger.experiment.log(
{
"reconstruction_mel": wandb.Image(image_mels, caption="mels"),
"wavs": [
wandb.Audio(
audio[0, :audio_len],
sample_rate=self.sampling_rate,
caption="gt",
),
wandb.Audio(
prior_audio[0, :audio_len],
sample_rate=self.sampling_rate,
caption="prior",
),
wandb.Audio(
posterior_audio[0, :audio_len],
sample_rate=self.sampling_rate,
caption="posterior",
),
],
},
)
if isinstance(self.logger, TensorBoardLogger):
self.logger.experiment.add_figure(
f"sample-{idx}/mels",
image_mels,
global_step=self.global_step,
)
self.logger.experiment.add_audio(
f"sample-{idx}/wavs/gt",
audio[0, :audio_len],
self.global_step,
sample_rate=self.sampling_rate,
)
self.logger.experiment.add_audio(
f"sample-{idx}/wavs/prior",
prior_audio[0, :audio_len],
self.global_step,
sample_rate=self.sampling_rate,
)
self.logger.experiment.add_audio(
f"sample-{idx}/wavs/posterior",
posterior_audio[0, :audio_len],
self.global_step,
sample_rate=self.sampling_rate,
)
plt.close(image_mels)