StreamSpeech / researches /ctc_unity /modules /transformer_encoder.py
fasdfsa's picture
init
901e06a
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
import torch
from fairseq import utils
import torch.nn as nn
import math
from fairseq.models import FairseqEncoder
from fairseq.modules import LayerNorm, PositionalEmbedding, FairseqDropout
from ctc_unity.modules.transformer_layer import TransformerEncoderLayer
class UniTransformerEncoderNoEmb(FairseqEncoder):
"""Transformer encoder without token embeddings."""
def __init__(self, args):
super().__init__(None)
self.layers = nn.ModuleList(
[TransformerEncoderLayer(args) for _ in range(args.encoder_layers)]
)
if args.encoder_normalize_before:
self.layer_norm = LayerNorm(args.encoder_embed_dim)
else:
self.layer_norm = None
self._future_mask = torch.empty(0)
self.unidirectional = getattr(args, "uni_encoder", False)
def forward(
self, x, encoder_padding_mask, return_all_hiddens=False, streaming_config=None
):
encoder_states = []
if streaming_config is None:
extra = {
"encoder_mask": (
self.buffered_future_mask(x) if self.unidirectional else None
)
}
else:
encoder_mask = None
if (
"encoder_mask" in streaming_config.keys()
and streaming_config["encoder_mask"] is not None
):
encoder_mask = streaming_config["encoder_mask"]
else:
encoder_mask = self.buffered_chunk_mask(
x, tgt_step=streaming_config["tgt_step"]
)
extra = {"encoder_mask": encoder_mask}
for layer in self.layers:
x = layer(x, encoder_padding_mask, extra=extra)
if return_all_hiddens:
encoder_states.append(x)
if self.layer_norm is not None:
x = self.layer_norm(x)
return {
"encoder_out": [x], # T x B x C
"encoder_padding_mask": (
[encoder_padding_mask]
if encoder_padding_mask is not None and encoder_padding_mask.any()
else []
), # B x T
"encoder_embedding": [], # B x T x C
"encoder_states": encoder_states, # List[T x B x C]
"src_tokens": [],
"src_lengths": [],
}
def buffered_future_mask(self, tensor):
dim = tensor.size(0)
# self._future_mask.device != tensor.device is not working in TorchScript. This is a workaround.
if (
self._future_mask.size(0) == 0
or (not self._future_mask.device == tensor.device)
or self._future_mask.size(0) < dim
):
self._future_mask = torch.triu(
utils.fill_with_neg_inf(torch.zeros([dim, dim])), 1
)
self._future_mask = self._future_mask.to(tensor)
return self._future_mask[:dim, :dim]
def buffered_chunk_mask(self, tensor, tgt_step):
dim = tensor.size(0)
idx = torch.arange(0, dim, device=tensor.device).unsqueeze(1)
idx = (idx // tgt_step + 1) * tgt_step
idx = idx.clamp(1, dim)
tmp = torch.arange(0, dim, device=tensor.device).unsqueeze(0).repeat(dim, 1)
chunk_mask = torch.where(
idx <= tmp, torch.tensor(float("-inf")), torch.tensor(0.0)
)
return chunk_mask[:dim, :dim]
def reorder_encoder_out(self, encoder_out, new_order):
new_encoder_out = (
[]
if len(encoder_out["encoder_out"]) == 0
else [x.index_select(1, new_order) for x in encoder_out["encoder_out"]]
)
new_encoder_padding_mask = (
[]
if len(encoder_out["encoder_padding_mask"]) == 0
else [
x.index_select(0, new_order)
for x in encoder_out["encoder_padding_mask"]
]
)
new_encoder_embedding = (
[]
if len(encoder_out["encoder_embedding"]) == 0
else [
x.index_select(0, new_order) for x in encoder_out["encoder_embedding"]
]
)
encoder_states = encoder_out["encoder_states"]
if len(encoder_states) > 0:
for idx, state in enumerate(encoder_states):
encoder_states[idx] = state.index_select(1, new_order)
return {
"encoder_out": new_encoder_out, # T x B x C
"encoder_padding_mask": new_encoder_padding_mask, # B x T
"encoder_embedding": new_encoder_embedding, # B x T x C
"encoder_states": encoder_states, # List[T x B x C]
"src_tokens": [], # B x T
"src_lengths": [], # B x 1
}
class UniTransformerEncoderWithEmb(FairseqEncoder):
"""Transformer encoder without token embeddings."""
def __init__(
self,
args,
dictionary,
embed_tokens,
):
super().__init__(dictionary)
self.dropout_module = FairseqDropout(args.dropout)
embed_dim = embed_tokens.embedding_dim
self.padding_idx = embed_tokens.padding_idx
self.max_source_positions = args.max_source_positions
self.embed_tokens = embed_tokens
self.embed_scale = 1.0 if args.no_scale_embedding else math.sqrt(embed_dim)
self.embed_positions = (
PositionalEmbedding(
args.max_source_positions,
embed_dim,
self.padding_idx,
learned=False,
)
if not args.no_token_positional_embeddings
else None
)
self.layers = nn.ModuleList(
[TransformerEncoderLayer(args) for _ in range(args.encoder_layers)]
)
if args.encoder_normalize_before:
self.layer_norm = LayerNorm(args.encoder_embed_dim)
else:
self.layer_norm = None
self._future_mask = torch.empty(0)
self.unidirectional = getattr(args, "uni_encoder", False)
def forward_embedding(self, src_tokens):
token_embedding = self.embed_tokens(src_tokens)
x = embed = self.embed_scale * token_embedding
if self.embed_positions is not None:
x = embed + self.embed_positions(src_tokens)
x = self.dropout_module(x)
return x, embed
def forward(
self,
src_tokens,
encoder_padding_mask,
return_all_hiddens=False,
streaming_config=None,
):
x, encoder_embedding = self.forward_embedding(src_tokens)
x = x.transpose(0, 1)
encoder_states = []
if streaming_config is None:
extra = {
"encoder_mask": (
self.buffered_future_mask(x) if self.unidirectional else None
)
}
else:
encoder_mask = None
if (
"encoder_mask" in streaming_config.keys()
and streaming_config["encoder_mask"] is not None
):
encoder_mask = streaming_config["encoder_mask"]
else:
encoder_mask = self.buffered_chunk_mask(
x, tgt_step=streaming_config["tgt_step"]
)
extra = {"encoder_mask": encoder_mask}
for layer in self.layers:
x = layer(x, encoder_padding_mask, extra=extra)
if return_all_hiddens:
encoder_states.append(x)
if self.layer_norm is not None:
x = self.layer_norm(x)
return {
"encoder_out": [x], # T x B x C
"encoder_padding_mask": (
[encoder_padding_mask]
if encoder_padding_mask is not None and encoder_padding_mask.any()
else []
), # B x T
"encoder_embedding": [], # B x T x C
"encoder_states": encoder_states, # List[T x B x C]
"src_tokens": [],
"src_lengths": [],
}
def buffered_future_mask(self, tensor):
dim = tensor.size(0)
# self._future_mask.device != tensor.device is not working in TorchScript. This is a workaround.
if (
self._future_mask.size(0) == 0
or (not self._future_mask.device == tensor.device)
or self._future_mask.size(0) < dim
):
self._future_mask = torch.triu(
utils.fill_with_neg_inf(torch.zeros([dim, dim])), 1
)
self._future_mask = self._future_mask.to(tensor)
return self._future_mask[:dim, :dim]
def buffered_chunk_mask(self, tensor, tgt_step):
dim = tensor.size(0)
idx = torch.arange(0, dim, device=tensor.device).unsqueeze(1)
idx = (idx // tgt_step + 1) * tgt_step
idx = idx.clamp(1, dim)
tmp = torch.arange(0, dim, device=tensor.device).unsqueeze(0).repeat(dim, 1)
chunk_mask = torch.where(
idx <= tmp, torch.tensor(float("-inf")), torch.tensor(0.0)
)
return chunk_mask[:dim, :dim]
def reorder_encoder_out(self, encoder_out, new_order):
new_encoder_out = (
[]
if len(encoder_out["encoder_out"]) == 0
else [x.index_select(1, new_order) for x in encoder_out["encoder_out"]]
)
new_encoder_padding_mask = (
[]
if len(encoder_out["encoder_padding_mask"]) == 0
else [
x.index_select(0, new_order)
for x in encoder_out["encoder_padding_mask"]
]
)
new_encoder_embedding = (
[]
if len(encoder_out["encoder_embedding"]) == 0
else [
x.index_select(0, new_order) for x in encoder_out["encoder_embedding"]
]
)
encoder_states = encoder_out["encoder_states"]
if len(encoder_states) > 0:
for idx, state in enumerate(encoder_states):
encoder_states[idx] = state.index_select(1, new_order)
return {
"encoder_out": new_encoder_out, # T x B x C
"encoder_padding_mask": new_encoder_padding_mask, # B x T
"encoder_embedding": new_encoder_embedding, # B x T x C
"encoder_states": encoder_states, # List[T x B x C]
"src_tokens": [], # B x T
"src_lengths": [], # B x 1
}