| |
| |
| |
| |
|
|
| import json |
| import logging |
| from typing import Dict |
|
|
| import numpy as np |
| import torch |
| import torch.nn.functional as F |
| from torch import nn |
|
|
| from fairseq.data.audio.audio_utils import ( |
| TTSSpectrogram, |
| get_fourier_basis, |
| get_mel_filters, |
| get_window, |
| ) |
| from fairseq.data.audio.speech_to_text_dataset import S2TDataConfig |
| from fairseq.models import BaseFairseqModel, register_model |
| from agent.tts.codehifigan import CodeGenerator as CodeHiFiGANModel |
| from fairseq.models.text_to_speech.hifigan import Generator as HiFiGANModel |
| from fairseq.models.text_to_speech.hub_interface import VocoderHubInterface |
|
|
| logger = logging.getLogger(__name__) |
|
|
|
|
| @register_model("CodeHiFiGANVocoderWithDur") |
| class CodeHiFiGANVocoderWithDur(BaseFairseqModel): |
| def __init__( |
| self, checkpoint_path: str, model_cfg: Dict[str, str], fp16: bool = False |
| ) -> None: |
| super().__init__() |
| self.model = CodeHiFiGANModel(model_cfg) |
| if torch.cuda.is_available(): |
| state_dict = torch.load(checkpoint_path) |
| else: |
| state_dict = torch.load(checkpoint_path, map_location=torch.device("cpu")) |
| self.model.load_state_dict(state_dict["generator"]) |
| self.model.eval() |
| if fp16: |
| self.model.half() |
| self.model.remove_weight_norm() |
| logger.info(f"loaded CodeHiFiGAN checkpoint from {checkpoint_path}") |
|
|
| def forward(self, x: Dict[str, torch.Tensor], dur_prediction=False) -> torch.Tensor: |
| assert "code" in x |
| x["dur_prediction"] = dur_prediction |
|
|
| |
| mask = x["code"] >= 0 |
| x["code"] = x["code"][mask].unsqueeze(dim=0) |
| if "f0" in x: |
| f0_up_ratio = x["f0"].size(1) // x["code"].size(1) |
| mask = mask.unsqueeze(2).repeat(1, 1, f0_up_ratio).view(-1, x["f0"].size(1)) |
| x["f0"] = x["f0"][mask].unsqueeze(dim=0) |
| wav, dur = self.model(**x) |
| return wav.detach().squeeze(), dur |
|
|
| @classmethod |
| def from_data_cfg(cls, args, data_cfg): |
| vocoder_cfg = data_cfg.vocoder |
| assert vocoder_cfg is not None, "vocoder not specified in the data config" |
| with open(vocoder_cfg["config"]) as f: |
| model_cfg = json.load(f) |
| return cls(vocoder_cfg["checkpoint"], model_cfg, fp16=args.fp16) |
|
|
| @classmethod |
| def hub_models(cls): |
| base_url = "http://dl.fbaipublicfiles.com/fairseq/vocoder" |
| model_ids = [ |
| "unit_hifigan_mhubert_vp_en_es_fr_it3_400k_layer11_km1000_lj_dur", |
| "unit_hifigan_mhubert_vp_en_es_fr_it3_400k_layer11_km1000_es_css10_dur", |
| "unit_hifigan_HK_layer12.km2500_frame_TAT-TTS", |
| ] |
| return {i: f"{base_url}/{i}.tar.gz" for i in model_ids} |
|
|
| @classmethod |
| def from_pretrained( |
| cls, |
| model_name_or_path, |
| checkpoint_file="model.pt", |
| data_name_or_path=".", |
| config="config.json", |
| fp16: bool = False, |
| **kwargs, |
| ): |
| from fairseq import hub_utils |
|
|
| x = hub_utils.from_pretrained( |
| model_name_or_path, |
| checkpoint_file, |
| data_name_or_path, |
| archive_map=cls.hub_models(), |
| config_yaml=config, |
| fp16=fp16, |
| is_vocoder=True, |
| **kwargs, |
| ) |
|
|
| with open(f"{x['args']['data']}/{config}") as f: |
| vocoder_cfg = json.load(f) |
| assert len(x["args"]["model_path"]) == 1, "Too many vocoder models in the input" |
|
|
| vocoder = CodeHiFiGANVocoderWithDur(x["args"]["model_path"][0], vocoder_cfg) |
| return VocoderHubInterface(vocoder_cfg, vocoder) |
|
|