| |
| |
| |
| |
|
|
| import logging |
| from pathlib import Path |
| from typing import Any, Dict, List, Optional |
|
|
| import torch |
| from torch import Tensor |
|
|
| from fairseq import checkpoint_utils, utils |
| from fairseq.models import ( |
| FairseqEncoderDecoderModel, |
| FairseqEncoderModel, |
| FairseqLanguageModel, |
| register_model, |
| register_model_architecture, |
| ) |
| from fairseq.models.speech_to_speech.modules.ctc_decoder import CTCDecoder |
| from fairseq.models.speech_to_speech.modules.stacked_embedding import StackedEmbedding |
| from fairseq.models.speech_to_text import S2TTransformerEncoder |
| from fairseq.models.text_to_speech import TTSTransformerDecoder |
| from fairseq.models.transformer import Linear, TransformerModelBase |
|
|
| from uni_unity.modules.transformer_decoder import TransformerDecoder |
|
|
| logger = logging.getLogger(__name__) |
|
|
|
|
| class S2STransformerEncoder(S2TTransformerEncoder): |
| """Based on S2T transformer encoder, with support |
| to incorporate target speaker embedding.""" |
|
|
| def __init__(self, args): |
| super().__init__(args) |
|
|
| self.spk_emb_proj = None |
| if args.target_speaker_embed: |
| self.spk_emb_proj = Linear( |
| args.encoder_embed_dim + args.speaker_embed_dim, args.encoder_embed_dim |
| ) |
|
|
| def forward( |
| self, src_tokens, src_lengths, tgt_speaker=None, return_all_hiddens=False |
| ): |
| out = super().forward(src_tokens, src_lengths, return_all_hiddens) |
|
|
| if self.spk_emb_proj: |
| x = out["encoder_out"][0] |
| seq_len, bsz, _ = x.size() |
| tgt_speaker_emb = tgt_speaker.view(1, bsz, -1).expand(seq_len, bsz, -1) |
| x = self.spk_emb_proj(torch.cat([x, tgt_speaker_emb], dim=2)) |
| out["encoder_out"][0] = x |
|
|
| return out |
|
|
|
|
| class TransformerUnitDecoder(TransformerDecoder): |
| """Based on Transformer decoder, with support to decoding stacked units""" |
|
|
| def __init__( |
| self, |
| args, |
| dictionary, |
| embed_tokens, |
| no_encoder_attn=False, |
| output_projection=None, |
| ): |
| super().__init__( |
| args, dictionary, embed_tokens, no_encoder_attn, output_projection |
| ) |
| self.n_frames_per_step = args.n_frames_per_step |
|
|
| self.out_proj_n_frames = ( |
| Linear( |
| self.output_embed_dim, |
| self.output_embed_dim * self.n_frames_per_step, |
| bias=False, |
| ) |
| if self.n_frames_per_step > 1 |
| else None |
| ) |
|
|
| def forward( |
| self, |
| prev_output_tokens, |
| encoder_out: Optional[Dict[str, List[Tensor]]] = None, |
| incremental_state: Optional[Dict[str, Dict[str, Optional[Tensor]]]] = None, |
| features_only: bool = False, |
| full_context_alignment: bool = False, |
| alignment_layer: Optional[int] = None, |
| alignment_heads: Optional[int] = None, |
| src_lengths: Optional[Any] = None, |
| return_all_hiddens: bool = False, |
| streaming_config=None, |
| ): |
| """ |
| Args: |
| prev_output_tokens (LongTensor): previous decoder outputs of shape |
| `(batch, tgt_len)`, for teacher forcing |
| encoder_out (optional): output from the encoder, used for |
| encoder-side attention, should be of size T x B x C |
| incremental_state (dict): dictionary used for storing state during |
| :ref:`Incremental decoding` |
| features_only (bool, optional): only return features without |
| applying output layer (default: False). |
| full_context_alignment (bool, optional): don't apply |
| auto-regressive mask to self-attention (default: False). |
| |
| Returns: |
| tuple: |
| - the decoder's output of shape `(batch, tgt_len, vocab)` |
| - a dictionary with any model-specific outputs |
| """ |
|
|
| x, extra = self.extract_features( |
| prev_output_tokens, |
| encoder_out=encoder_out, |
| incremental_state=incremental_state, |
| full_context_alignment=full_context_alignment, |
| alignment_layer=alignment_layer, |
| alignment_heads=alignment_heads, |
| streaming_config=streaming_config, |
| ) |
|
|
| if not features_only: |
| bsz, seq_len, d = x.size() |
| if self.out_proj_n_frames: |
| x = self.out_proj_n_frames(x) |
| x = self.output_layer(x.view(bsz, seq_len, self.n_frames_per_step, d)) |
| x = x.view(bsz, seq_len * self.n_frames_per_step, -1) |
| if ( |
| incremental_state is None and self.n_frames_per_step > 1 |
| ): |
| x = x[ |
| :, : -(self.n_frames_per_step - 1), : |
| ] |
|
|
| return x, extra |
|
|
| def upgrade_state_dict_named(self, state_dict, name): |
| if self.n_frames_per_step > 1: |
| move_keys = [ |
| ( |
| f"{name}.project_in_dim.weight", |
| f"{name}.embed_tokens.project_in_dim.weight", |
| ) |
| ] |
| for from_k, to_k in move_keys: |
| if from_k in state_dict and to_k not in state_dict: |
| state_dict[to_k] = state_dict[from_k] |
| del state_dict[from_k] |
|
|
|
|
| class S2STransformerMultitaskModelBase(FairseqEncoderDecoderModel): |
| @classmethod |
| def build_encoder(cls, args): |
| encoder = S2STransformerEncoder(args) |
| pretraining_path = getattr(args, "load_pretrained_encoder_from", None) |
| if pretraining_path is not None: |
| if not Path(pretraining_path).exists(): |
| logger.warning( |
| f"skipped pretraining because {pretraining_path} does not exist" |
| ) |
| else: |
| encoder = checkpoint_utils.load_pretrained_component_from_model( |
| component=encoder, checkpoint=pretraining_path |
| ) |
| logger.info(f"loaded pretrained encoder from: {pretraining_path}") |
| return encoder |
|
|
| @classmethod |
| def build_multitask_decoder(cls, args, tgt_dict, in_dim): |
| decoder_args = args.decoder_args |
| decoder_args.encoder_embed_dim = in_dim |
| if args.decoder_type == "transformer": |
| base_multitask_text_transformer_decoder_arch(decoder_args) |
| task_decoder = TransformerDecoder( |
| decoder_args, |
| tgt_dict, |
| embed_tokens=TransformerModelBase.build_embedding( |
| decoder_args, |
| tgt_dict, |
| decoder_args.decoder_embed_dim, |
| ), |
| ) |
| elif args.decoder_type == "ctc": |
| task_decoder = CTCDecoder( |
| dictionary=tgt_dict, |
| in_dim=in_dim, |
| ) |
| else: |
| raise NotImplementedError( |
| "currently only support multitask decoder_type 'transformer', 'ctc'" |
| ) |
|
|
| return task_decoder |
|
|
| @classmethod |
| def build_model(cls, args, task): |
| encoder = cls.build_encoder(args) |
| decoder = ( |
| cls.build_decoder(args, task.target_dictionary) |
| if task.args.target_is_code |
| else cls.build_decoder(args) |
| ) |
| base_model = cls(encoder, decoder) |
|
|
| |
| base_model.multitask_decoders = {} |
| for task_name, task_obj in task.multitask_tasks.items(): |
| in_dim = ( |
| args.encoder_embed_dim |
| if task_obj.args.input_from == "encoder" |
| else args.decoder_embed_dim |
| ) |
| task_decoder = cls.build_multitask_decoder( |
| task_obj.args, task_obj.target_dictionary, in_dim |
| ) |
|
|
| setattr(base_model, f"{task_name}_decoder", task_decoder) |
| decoder_model_cls = ( |
| FairseqEncoderModel |
| if task_obj.args.decoder_type == "ctc" |
| else FairseqLanguageModel |
| ) |
| base_model.multitask_decoders[task_name] = decoder_model_cls( |
| getattr(base_model, f"{task_name}_decoder") |
| ) |
|
|
| return base_model |
|
|
| def forward_encoder(self, src_tokens, src_lengths, speaker=None, **kwargs): |
| return self.encoder( |
| src_tokens, src_lengths=src_lengths, tgt_speaker=speaker, **kwargs |
| ) |
|
|
|
|
| class S2UTTransformerModel(S2STransformerMultitaskModelBase): |
| """ |
| Direct speech-to-speech translation model with Transformer encoder + Transformer discrete unit decoder |
| https://arxiv.org/abs/2107.05604 |
| """ |
|
|
| @staticmethod |
| def add_args(parser): |
| |
| parser.add_argument( |
| "--conv-kernel-sizes", |
| type=str, |
| metavar="STR", |
| help="kernel sizes of Conv1d (s2t_transformer) subsampling layers", |
| ) |
| parser.add_argument( |
| "--conv-channels", |
| type=int, |
| metavar="N", |
| help="# of channels in Conv1d (s2t_transformer) subsampling layers", |
| ) |
| parser.add_argument( |
| "--conv-out-channels", |
| type=int, |
| metavar="N", |
| help="# of channels in Conv2d (convtransformer) subsampling layers", |
| ) |
| parser.add_argument( |
| "--conv-version", |
| type=str, |
| default="s2t_transformer", |
| choices=["s2t_transformer", "convtransformer"], |
| help="version of frontend convolutional layers", |
| ) |
| |
| parser.add_argument( |
| "--activation-fn", |
| type=str, |
| default="relu", |
| choices=utils.get_available_activation_fns(), |
| help="activation function to use", |
| ) |
| parser.add_argument( |
| "--dropout", type=float, metavar="D", help="dropout probability" |
| ) |
| parser.add_argument( |
| "--attention-dropout", |
| type=float, |
| metavar="D", |
| help="dropout probability for attention weights", |
| ) |
| parser.add_argument( |
| "--activation-dropout", |
| "--relu-dropout", |
| type=float, |
| metavar="D", |
| help="dropout probability after activation in FFN.", |
| ) |
| parser.add_argument( |
| "--encoder-embed-dim", |
| type=int, |
| metavar="N", |
| help="encoder embedding dimension", |
| ) |
| parser.add_argument( |
| "--encoder-ffn-embed-dim", |
| type=int, |
| metavar="N", |
| help="encoder embedding dimension for FFN", |
| ) |
| parser.add_argument( |
| "--encoder-layers", type=int, metavar="N", help="num encoder layers" |
| ) |
| parser.add_argument( |
| "--encoder-attention-heads", |
| type=int, |
| metavar="N", |
| help="num encoder attention heads", |
| ) |
| parser.add_argument( |
| "--encoder-normalize-before", |
| action="store_true", |
| help="apply layernorm before each encoder block", |
| ) |
| parser.add_argument( |
| "--decoder-embed-dim", |
| type=int, |
| metavar="N", |
| help="decoder embedding dimension", |
| ) |
| parser.add_argument( |
| "--decoder-ffn-embed-dim", |
| type=int, |
| metavar="N", |
| help="decoder embedding dimension for FFN", |
| ) |
| parser.add_argument( |
| "--decoder-layers", type=int, metavar="N", help="num decoder layers" |
| ) |
| parser.add_argument( |
| "--decoder-attention-heads", |
| type=int, |
| metavar="N", |
| help="num decoder attention heads", |
| ) |
| parser.add_argument( |
| "--decoder-normalize-before", |
| action="store_true", |
| help="apply layernorm before each decoder block", |
| ) |
| parser.add_argument( |
| "--share-decoder-input-output-embed", |
| action="store_true", |
| help="share decoder input and output embeddings", |
| ) |
| parser.add_argument( |
| "--layernorm-embedding", |
| action="store_true", |
| help="add layernorm to embedding", |
| ) |
| parser.add_argument( |
| "--no-scale-embedding", |
| action="store_true", |
| help="if True, dont scale embeddings", |
| ) |
| parser.add_argument( |
| "--load-pretrained-encoder-from", |
| type=str, |
| metavar="STR", |
| help="model to take encoder weights from (for initialization)", |
| ) |
| parser.add_argument( |
| "--encoder-freezing-updates", |
| type=int, |
| metavar="N", |
| help="freeze encoder for first N updates", |
| ) |
| |
| parser.add_argument( |
| "--speaker-embed-dim", |
| type=int, |
| metavar="N", |
| help="speaker embedding dimension", |
| ) |
|
|
| @classmethod |
| def build_decoder(cls, args, tgt_dict): |
| num_embeddings = len(tgt_dict) |
| padding_idx = tgt_dict.pad() |
| embed_tokens = StackedEmbedding( |
| num_embeddings, |
| args.decoder_embed_dim, |
| padding_idx, |
| num_stacked=args.n_frames_per_step, |
| ) |
|
|
| return TransformerUnitDecoder( |
| args, |
| tgt_dict, |
| embed_tokens, |
| ) |
|
|
| def forward( |
| self, |
| src_tokens, |
| src_lengths, |
| prev_output_tokens, |
| tgt_speaker=None, |
| return_all_hiddens=False, |
| ): |
| encoder_out = self.encoder( |
| src_tokens, |
| src_lengths=src_lengths, |
| tgt_speaker=tgt_speaker, |
| return_all_hiddens=return_all_hiddens, |
| ) |
| decoder_out = self.decoder( |
| prev_output_tokens, |
| encoder_out=encoder_out, |
| ) |
| if return_all_hiddens: |
| decoder_out[-1]["encoder_states"] = encoder_out["encoder_states"] |
| decoder_out[-1]["encoder_padding_mask"] = encoder_out[ |
| "encoder_padding_mask" |
| ] |
| return decoder_out |
|
|
|
|
| def base_multitask_text_transformer_decoder_arch(args): |
| args.dropout = getattr(args, "dropout", 0.3) |
| args.decoder_layerdrop = getattr(args, "decoder_layerdrop", 0.0) |
| args.share_decoder_input_output_embed = getattr( |
| args, "share_decoder_input_output_embed", True |
| ) |
| args.decoder_embed_dim = getattr(args, "decoder_embed_dim", 256) |
| args.decoder_output_dim = getattr( |
| args, "decoder_output_dim", args.decoder_embed_dim |
| ) |
| args.decoder_input_dim = getattr(args, "decoder_input_dim", args.decoder_embed_dim) |
|
|
| args.max_target_positions = getattr(args, "max_target_positions", 1024) |
| args.no_scale_embedding = getattr(args, "no_scale_embedding", False) |
|
|
| args.adaptive_input = getattr(args, "adaptive_input", False) |
| args.quant_noise_pq = getattr(args, "quant_noise_pq", 0) |
|
|
| args.decoder_learned_pos = getattr(args, "decoder_learned_pos", False) |
| args.no_token_positional_embeddings = getattr( |
| args, "no_token_positional_embeddings", False |
| ) |
|
|
| args.decoder_layers = getattr(args, "decoder_layers", 2) |
|
|
| args.adaptive_softmax_cutoff = getattr(args, "adaptive_softmax_cutoff", None) |
|
|
| |
| args.activation_dropout = getattr(args, "activation_dropout", args.dropout) |
| args.activation_fn = getattr(args, "activation_fn", "relu") |
| args.decoder_normalize_before = getattr(args, "decoder_normalize_before", True) |
| args.decoder_ffn_embed_dim = getattr(args, "decoder_ffn_embed_dim", 2048) |
|
|
| args.attention_dropout = getattr(args, "attention_dropout", args.dropout) |
| args.decoder_attention_heads = getattr(args, "decoder_attention_heads", 4) |
|
|
|
|
| def base_s2st_transformer_encoder_architecture(args): |
| args.encoder_freezing_updates = getattr(args, "encoder_freezing_updates", 0) |
|
|
| |
| args.input_channels = getattr(args, "input_channels", 1) |
| args.conv_kernel_sizes = getattr(args, "conv_kernel_sizes", "5,5") |
| args.conv_channels = getattr(args, "conv_channels", 1024) |
| args.conv_out_channels = getattr(args, "conv_out_channels", 256) |
| args.conv_version = getattr(args, "conv_version", "s2t_transformer") |
| |
| args.encoder_embed_dim = getattr(args, "encoder_embed_dim", 512) |
| args.encoder_ffn_embed_dim = getattr(args, "encoder_ffn_embed_dim", 2048) |
| args.encoder_layers = getattr(args, "encoder_layers", 12) |
| args.encoder_attention_heads = getattr(args, "encoder_attention_heads", 8) |
| args.encoder_normalize_before = getattr(args, "encoder_normalize_before", True) |
| args.no_scale_embedding = getattr(args, "no_scale_embedding", False) |
|
|
| args.dropout = getattr(args, "dropout", 0.1) |
| args.attention_dropout = getattr(args, "attention_dropout", args.dropout) |
| args.activation_dropout = getattr(args, "activation_dropout", args.dropout) |
| args.activation_fn = getattr(args, "activation_fn", "relu") |
|
|
| args.speaker_embed_dim = getattr(args, "speaker_embed_dim", 256) |
|
|