import math import torch from torch import nn from fish_speech.models.vqgan.modules.fsq import DownsampleFiniteScalarQuantize from fish_speech.models.vqgan.modules.wavenet import WaveNet from fish_speech.models.vqgan.utils import sequence_mask from fish_speech.utils.spectrogram import LogMelSpectrogram class VQEncoder(nn.Module): def __init__( self, ): super().__init__() self.encoder = WaveNet( input_channels=128, residual_channels=768, residual_layers=20, dilation_cycle=4, ) self.quantizer = DownsampleFiniteScalarQuantize( input_dim=768, n_codebooks=1, n_groups=2, levels=[8, 5, 5, 5] ) self.spec = LogMelSpectrogram( sample_rate=44100, n_fft=2048, win_length=2048, hop_length=512, n_mels=128, f_min=0.0, f_max=8000.0, ) self.eval() e = self.load_state_dict( torch.load("checkpoints/vq-gan-group-fsq-2x1024.pth", map_location="cpu"), strict=False, ) assert len(e.missing_keys) == 0, e.missing_keys assert all( k.startswith("decoder.") or k.startswith("quality_projection.") or k.startswith("discriminator.") for k in e.unexpected_keys ), e.unexpected_keys @torch.no_grad() def forward(self, audios, audio_lengths, sr=None): mel_spec = self.spec(audios, sample_rate=sr) if sr is not None: audio_lengths = audio_lengths * 44100 // sr mel_lengths = audio_lengths // self.spec.hop_length mel_masks = ( torch.arange(mel_spec.shape[2], device=mel_spec.device) < mel_lengths[:, None] ) mel_masks_float_conv = mel_masks[:, None, :].float() mels = mel_spec * mel_masks_float_conv # Encode encoded_features = self.encoder(mels) * mel_masks_float_conv encoded_features = self.quantizer(encoded_features).z * mel_masks_float_conv return encoded_features @torch.no_grad() def indicies_to_vq_features( self, indices, feature_lengths, ): factor = math.prod(self.quantizer.downsample_factor) mel_masks = sequence_mask(feature_lengths * factor, indices.shape[2] * factor) mel_masks_float_conv = mel_masks[:, None, :].float() z = self.quantizer.decode(indices) * mel_masks_float_conv return z @torch.no_grad() def encode(self, audios, audio_lengths, sr=None): audios = audios.float() mels = self.spec(audios, sample_rate=sr) mel_lengths = audio_lengths // self.spec.hop_length mel_masks = sequence_mask(mel_lengths, mels.shape[2]) mel_masks_float_conv = mel_masks[:, None, :].float() mels = mels * mel_masks_float_conv # Encode encoded_features = self.encoder(mels) * mel_masks_float_conv feature_lengths = mel_lengths // math.prod(self.quantizer.downsample_factor) return self.quantizer.encode(encoded_features), feature_lengths