| import copy |
| import math |
| from typing import Optional, Any, Tuple |
|
|
| import torch |
| from torch import Tensor |
| import torch.nn.functional as F |
| from torch.nn.modules.module import Module |
| from torch.nn.modules.activation import MultiheadAttention |
| from torch.nn.modules.container import ModuleList |
| from torch.nn.init import xavier_uniform_ |
| from torch.nn.modules.dropout import Dropout |
| from torch.nn.modules.linear import Linear |
| from torch.nn.modules.normalization import LayerNorm |
| |
|
|
| from transformers.modeling_outputs import BaseModelOutputWithPooling |
| from transformers import PreTrainedModel |
| from transformers import BertForMaskedLM, BertForSequenceClassification |
|
|
| from .configuration_hibial import HiBiAlBertConfig |
|
|
| import warnings |
|
|
| |
| warnings.filterwarnings("ignore") |
|
|
|
|
| |
| def gen_encoder_ut_mask(src_seq, input_mask, utt_loc): |
| def _gen_mask_hierarchical(A, src_pad_mask): |
| |
| return ~(2 * A == (A + A.transpose(1, 2))).bool() |
|
|
| enc_mask_utt = _gen_mask_hierarchical(utt_loc.unsqueeze(1).expand(-1, src_seq.shape[1], -1), input_mask) |
| return enc_mask_utt |
|
|
|
|
| def _get_pe_inputs(src_seq, input_mask, utt_loc): |
| pe_utt_loc = torch.zeros(utt_loc.shape, device=utt_loc.device) |
| for i in range(1, utt_loc.shape[1]): |
| _logic = (utt_loc[:, i] == utt_loc[:, i - 1]).float() |
| pe_utt_loc[:, i] = pe_utt_loc[:, i - 1] + _logic - (1 - _logic) * pe_utt_loc[:, i - 1] |
| return pe_utt_loc |
|
|
|
|
| def _CLS_masks(src_seq, input_mask, utt_loc): |
| |
| pe_utt_loc = _get_pe_inputs(src_seq, input_mask, utt_loc) |
|
|
| |
| enc_mask_utt = gen_encoder_ut_mask(src_seq, input_mask, utt_loc) |
|
|
| |
| enc_mask_ct = ((pe_utt_loc + input_mask) != 0).unsqueeze(1).expand(-1, src_seq.shape[1], -1) |
|
|
| return pe_utt_loc, enc_mask_utt, enc_mask_ct |
|
|
|
|
| def get_hier_encoder_mask(src_seq, input_mask, utt_loc, type: str): |
| |
| |
| |
| utt_loc = utt_loc - 2 * input_mask * utt_loc |
|
|
| |
| assert type in ["hier", "cls", "full"] |
|
|
| if type == "hier": |
| raise Exception("Not used for BERT") |
| elif type == "cls": |
| return _CLS_masks(src_seq, input_mask, utt_loc) |
| elif type == "full": |
| raise Exception("Not used for BERT") |
|
|
| return None |
|
|
|
|
| def _get_clones(module, N): |
| return ModuleList([copy.deepcopy(module) for i in range(N)]) |
|
|
|
|
| def _get_activation_fn(activation): |
| if activation == "relu": |
| return F.relu |
| elif activation == "gelu": |
| return F.gelu |
|
|
| raise RuntimeError("activation should be relu/gelu, not {}".format(activation)) |
|
|
|
|
| class BiALiBi(Module): |
| def __init__(self, config): |
| super().__init__() |
| self.num_attention_heads = config.num_attention_heads |
|
|
| self.beta = torch.nn.Parameter(torch.empty(self.num_attention_heads), requires_grad=True) |
| self.gamma = torch.nn.Parameter(torch.empty(self.num_attention_heads), requires_grad=True) |
|
|
| torch.nn.init.normal_(self.beta, -2, 1) |
| torch.nn.init.normal_(self.gamma, -2, 1) |
|
|
| def get_rel_pos(self, seq_len: int): |
| memory = torch.arange(seq_len, dtype=torch.float32).to(self.gamma.device) |
| context = torch.arange(seq_len, dtype=torch.float32).unsqueeze(-1).to(self.gamma.device) |
| rel_pos = (memory - context).abs() |
|
|
| return rel_pos.unsqueeze(0).expand(self.num_attention_heads, -1, -1).clone() |
|
|
| def forward(self, seq_len: int): |
| rel_pos = self.get_rel_pos(seq_len) |
|
|
| |
| gamma_mask = torch.triu(torch.ones_like(rel_pos), diagonal=1) |
| gamma_mask *= -torch.sigmoid(self.gamma).view(-1, 1, 1) |
|
|
| beta_mask = torch.tril(torch.ones_like(rel_pos), diagonal=-1) |
| beta_mask *= -torch.sigmoid(self.beta).view(-1, 1, 1) |
|
|
| mask = beta_mask + gamma_mask |
|
|
| return rel_pos * mask |
|
|
|
|
| class TransformerEncoderLayer(Module): |
| r"""TransformerEncoderLayer is made up of self-attn and feedforward network. |
| This standard encoder layer is based on the paper "Attention Is All You Need". |
| Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, |
| Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in |
| Neural Information Processing Systems, pages 6000-6010. Users may modify or implement |
| in a different way during application. |
| Args: |
| d_model: the number of expected features in the input (required). |
| nhead: the number of heads in the multiheadattention models (required). |
| dim_feedforward: the dimension of the feedforward network model (default=2048). |
| dropout: the dropout value (default=0.1). |
| activation: the activation function of intermediate layer, relu or gelu (default=relu). |
| layer_norm_eps: the eps value in layer normalization components (default=1e-5). |
| Examples:: |
| >>> encoder_layer = nn.TransformerEncoderLayer(d_model=512, nhead=8) |
| >>> src = torch.rand(10, 32, 512) |
| >>> out = encoder_layer(src) |
| """ |
|
|
| def __init__(self, config): |
| super(TransformerEncoderLayer, self).__init__() |
|
|
| self.self_attn = MultiheadAttention(config.hidden_size, |
| config.num_attention_heads, |
| dropout=config.attention_probs_dropout_prob) |
| |
| self.linear1 = Linear(config.hidden_size, config.intermediate_size) |
| self.dropout = Dropout(config.hidden_dropout_prob) |
| self.linear2 = Linear(config.intermediate_size, config.hidden_size) |
|
|
| self.norm_first = config.norm_first |
| self.norm1 = LayerNorm(config.hidden_size, eps=config.layer_norm_eps) |
| self.norm2 = LayerNorm(config.hidden_size, eps=config.layer_norm_eps) |
| self.dropout1 = Dropout(config.hidden_dropout_prob) |
| self.dropout2 = Dropout(config.hidden_dropout_prob) |
|
|
| self.activation = _get_activation_fn(config.hidden_act) |
|
|
| def __setstate__(self, state): |
| if 'activation' not in state: |
| state['activation'] = F.relu |
| super(TransformerEncoderLayer, self).__setstate__(state) |
|
|
| def forward(self, src: Tensor, src_mask: Optional[Tensor] = None, |
| src_key_padding_mask: Optional[Tensor] = None) -> tuple[Tensor, Optional[Tensor]]: |
| r"""Pass the input through the encoder layer. |
| Args: |
| src: the sequence to the encoder layer (required). |
| src_mask: the mask for the src sequence (optional). |
| src_key_padding_mask: the mask for the src keys per batch (optional). |
| Shape: |
| see the docs in Transformer class. |
| """ |
|
|
| |
| |
|
|
| |
| if self.norm_first: |
|
|
| src = self.norm1(src) |
| src_attn = self.self_attn(src, src, src, attn_mask=src_mask, |
| key_padding_mask=src_key_padding_mask, average_attn_weights=False) |
| src = src + self.dropout1(src_attn[0]) |
| src = self.norm2(src) |
| src_ffn = self.linear2(self.dropout(self.activation(self.linear1(src)))) |
| src = src + self.dropout2(src_ffn) |
|
|
| else: |
| src_attn = self.self_attn(src, src, src, attn_mask=src_mask, |
| key_padding_mask=src_key_padding_mask, average_attn_weights=False) |
| src = src + self.dropout1(src_attn[0]) |
| src = self.norm1(src) |
| src_ffn = self.linear2(self.dropout(self.activation(self.linear1(src)))) |
| src = src + self.dropout2(src_ffn) |
| src = self.norm2(src) |
| return src, src_attn[1] |
|
|
|
|
| class HiBiAlBert(Module): |
| r"""A transformer model. User is able to modify the attributes as needed. The architecture |
| is based on the paper "Attention Is All You Need". Ashish Vaswani, Noam Shazeer, |
| Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and |
| Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information |
| Processing Systems, pages 6000-6010. Users can build the BERT(https://arxiv.org/abs/1810.04805) |
| model with corresponding parameters. |
| Args: |
| d_model: the number of expected features in the encoder/decoder inputs (default=512). |
| nhead: the number of heads in the multiheadattention models (default=8). |
| num_encoder_layers: the number of sub-encoder-layers in the encoder (default=6). |
| dim_feedforward: the dimension of the feedforward network model (default=2048). |
| dropout: the dropout value (default=0.1). |
| activation: the activation function of encoder/decoder intermediate layer, relu or gelu (default=relu). |
| custom_encoder: custom encoder (default=None). |
| custom_decoder: custom decoder (default=None). |
| layer_norm_eps: the eps value in layer normalization components (default=1e-5). |
| Examples:: |
| # >>> transformer_model = HIERTransformer(nhead=16, num_encoder_layers=12) |
| # >>> src = torch.rand((10, 32, 512)) |
| # >>> token_type_ids/utt_indices = torch.tensor([0, 0, 1, 1, 1, 2, 2, 3, 3, 3]) Represent each utterance to encode |
| # >>> out = transformer_model(src) |
| Note: A full example to apply nn.Transformer module for the word language model is available in |
| # https://github.com/pytorch/examples/tree/master/word_language_model |
| """ |
|
|
| def __init__(self, config) -> None: |
| super(HiBiAlBert, self).__init__() |
| self.config = config |
| |
|
|
| self.bialibi_utt = BiALiBi(config) |
| self.bialibi_ct = BiALiBi(config) |
|
|
| self.word_embeddings = torch.nn.Embedding(config.vocab_size, |
| config.hidden_size, |
| padding_idx=config.pad_token_id) |
|
|
| |
| self.enc_layers = _get_clones(TransformerEncoderLayer(config=config), |
| |
| |
| |
| |
| |
| |
| |
| |
| config.num_hidden_layers) |
| self.norm_e = LayerNorm(config.hidden_size, |
| eps=config.layer_norm_eps) |
|
|
| self._reset_parameters() |
| self.init_weights() |
|
|
| def init_weights(self) -> None: |
| initrange = 0.1 |
| self.word_embeddings.weight.data.uniform_(-initrange, initrange) |
|
|
| |
| def forward(self, input_ids: Tensor, |
| attention_mask: Optional[Tensor] = None, |
| token_type_ids: Optional[Tensor] = None, |
| ct_mask_type: str = "cls", |
| output_attentions: Optional[bool] = True, |
| memory_key_padding_mask: Optional[Tensor] = None, |
| **kwargs |
| ): |
| r"""Take in and process masked source/target sequences. |
| Args: |
| input_ids/src: the sequence to the encoder (required). |
| src_mask: the additive mask for the src sequence (optional). |
| |
| memory_mask: the additive mask for the encoder output (optional). |
| attention_mask/src_key_padding_mask: the ByteTensor mask for src keys per batch (optional). |
| |
| memory_key_padding_mask: the ByteTensor mask for memory keys per batch (optional). |
| Shape: |
| - input_ids/src: :math:`(S, N, E)`. |
| - src_mask: :math:`(S, S)`. |
| - memory_mask: :math:`(T, S)`. |
| - not(attention_mask)/src_key_padding_mask: :math:`(N, S)`. |
| - token_type_ids/utt_indices: :math:`(N, S)`. |
| - memory_key_padding_mask: :math:`(N, S)`. |
| Note: [src/memory]_mask ensures that position i is allowed to attend the unmasked |
| positions. If a ByteTensor is provided, the non-zero positions are not allowed to attend |
| while the zero positions will be unchanged. If a BoolTensor is provided, positions with ``True`` |
| are not allowed to attend while ``False`` values will be unchanged. If a FloatTensor |
| is provided, it will be added to the attention weight. |
| [src/memory]_key_padding_mask provides specified elements in the key to be ignored by |
| the attention. If a ByteTensor is provided, the non-zero positions will be ignored while the zero |
| positions will be unchanged. If a BoolTensor is provided, the positions with the |
| value of ``True`` will be ignored while the position with the value of ``False`` will be unchanged. |
| - output: :math:`(T, N, E)`. |
| Note: Due to the multi-head attention architecture in the transformer model, |
| the output sequence length of a transformer is same as the input sequence |
| (i.e. target) length of the decode. |
| where S is the source sequence length, T is the target sequence length, N is the |
| batch size, E is the feature number |
| Examples: |
| # >>> output = transformer_model(src, src_mask=src_mask) |
| """ |
| all_self_attentions = () if output_attentions else None |
| |
|
|
| if attention_mask is None: |
| |
| attention_mask = self.create_padding_mask(input_ids) |
| attention_mask = torch.tensor(attention_mask, dtype=torch.long) |
|
|
| if token_type_ids is None: |
| |
| token_type_ids = self.convert_input_ids_to_token_type_ids(input_ids) |
|
|
| src_key_padding_mask = torch.logical_not(attention_mask) |
| utt_indices = token_type_ids |
|
|
| _, enc_mask_utt, enc_mask_ct = get_hier_encoder_mask(input_ids, |
| src_key_padding_mask, |
| utt_indices, |
| type=ct_mask_type) |
| enc_mask_utt = enc_mask_utt.repeat(self.config.num_attention_heads, 1, 1) |
| enc_mask_ct = enc_mask_ct.repeat(self.config.num_attention_heads, 1, 1) |
|
|
| B, T = input_ids.shape |
| bialibi_utt_mask = self.bialibi_utt(T).repeat(B, 1, 1) |
| bialibi_ct_mask = self.bialibi_ct(T).repeat(B, 1, 1) |
|
|
| bialibi_utt_mask[enc_mask_utt.bool()] = float('-inf') |
| bialibi_ct_mask[enc_mask_ct.bool()] = float('-inf') |
|
|
| |
|
|
| |
| |
|
|
| enc_inp = self.word_embeddings(input_ids.transpose(0, 1)) |
|
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| |
| |
| for i, layer in enumerate(self.enc_layers): |
| if i % (2 + 1) < 2: |
| |
| |
| enc_inp, att_w = layer(enc_inp, |
| src_mask=bialibi_utt_mask) |
| else: |
| |
| |
| enc_inp, att_w = layer(enc_inp, |
| src_mask=bialibi_ct_mask) |
| if output_attentions: |
| all_self_attentions = all_self_attentions + (att_w,) |
|
|
| if self.norm_e is not None: |
| enc_inp = self.norm_e(enc_inp) |
|
|
| encoder_output = enc_inp.transpose(0, 1) |
| hidden_states = encoder_output |
|
|
| pooled_output = hidden_states[:, 0, :] |
| outputs = (hidden_states, pooled_output, all_self_attentions) |
|
|
| return outputs |
|
|
| def create_padding_mask(self, token_ids): |
| padding_mask = torch.ne(token_ids, self.config.pad_token_id).int() |
| return padding_mask |
|
|
| def generate_square_subsequent_mask(self, sz: int) -> Tensor: |
| r"""Generate a square mask for the sequence. The masked positions are filled with float('-inf'). |
| Unmasked positions are filled with float(0.0). |
| """ |
| mask = (torch.triu(torch.ones(sz, sz)) == 1).transpose(0, 1) |
| mask = mask.float().masked_fill(mask == 0, float('-inf')).masked_fill(mask == 1, float(0.0)) |
| return mask |
|
|
| def _reset_parameters(self): |
| r"""Initiate parameters in the transformer model.""" |
|
|
| for p in self.parameters(): |
| if p.dim() > 1: |
| xavier_uniform_(p) |
|
|
| def convert_input_ids_to_token_type_ids(self, input_ids): |
| token_type_ids = torch.zeros_like(input_ids) |
|
|
| for row, row_tensor in enumerate(input_ids): |
| sep_indices = torch.nonzero(row_tensor == self.config.sep_token_id) |
| prev_index = -1 |
| for type_id, index in enumerate(sep_indices): |
| token_type_ids[row, prev_index + 1:index + 1] = type_id |
| prev_index = index |
|
|
| return token_type_ids |
|
|
|
|
| class HiBiAlBertModel(PreTrainedModel): |
| config_class = HiBiAlBertConfig |
| base_model_prefix = "bert" |
|
|
| def __init__(self, config): |
| super().__init__(config) |
|
|
| self.model = HiBiAlBert(config) |
|
|
| def forward( |
| self, |
| input_ids: Optional[torch.Tensor] = None, |
| attention_mask: Optional[torch.Tensor] = None, |
| token_type_ids: Optional[torch.Tensor] = None, |
| position_ids: Optional[torch.Tensor] = None, |
| inputs_embeds: Optional[torch.Tensor] = None, |
| output_attentions: Optional[bool] = None, |
| output_hidden_states: Optional[bool] = None, |
| return_dict: Optional[bool] = None, |
| **kwargs |
| ): |
| outputs = self.model(input_ids=input_ids, |
| attention_mask=attention_mask, |
| token_type_ids=token_type_ids, |
| position_ids=position_ids, |
| inputs_embeds=inputs_embeds, |
| output_attentions=output_attentions, |
| output_hidden_states=output_hidden_states, |
| return_dict=return_dict) |
| if not return_dict: |
| return outputs |
|
|
| return BaseModelOutputWithPooling( |
| last_hidden_state=outputs[0], |
| pooler_output=outputs[1], |
| attentions=outputs[2]) |
|
|
| def get_input_embeddings(self): |
| return self.model.word_embeddings |
|
|
| def set_input_embeddings(self, value): |
| self.model.word_embeddings = value |
|
|
|
|
| class HiBiAlBertForMaskedLM(BertForMaskedLM): |
| config_class = HiBiAlBertConfig |
|
|
| def __init__(self, config): |
| super().__init__(config) |
| self.bert = HiBiAlBertModel(config) |
|
|
|
|
| class HiBiAlBertForSequenceClassification(BertForSequenceClassification): |
| config_class = HiBiAlBertConfig |
|
|
| def __init__(self, config): |
| super().__init__(config) |
| self.bert = HiBiAlBertModel(config) |
|
|