|
|
import torch.nn as nn |
|
|
from transformers import BertConfig, BertModel, BertTokenizer |
|
|
|
|
|
from modules.build import LANGUAGE_REGISTRY |
|
|
|
|
|
|
|
|
@LANGUAGE_REGISTRY.register() |
|
|
class BERTLanguageEncoder(nn.Module): |
|
|
def __init__(self, cfg, weights="bert-base-uncased", hidden_size=768, |
|
|
num_hidden_layers=4, num_attention_heads=12, type_vocab_size=2): |
|
|
super().__init__() |
|
|
self.tokenizer = BertTokenizer.from_pretrained( |
|
|
weights, do_lower_case=True |
|
|
) |
|
|
self.bert_config = BertConfig( |
|
|
hidden_size=hidden_size, |
|
|
num_hidden_layers=num_hidden_layers, |
|
|
num_attention_heads=num_attention_heads, |
|
|
type_vocab_size=type_vocab_size |
|
|
) |
|
|
self.model = BertModel.from_pretrained( |
|
|
weights, config=self.bert_config |
|
|
) |
|
|
|
|
|
def forward(self, txt_ids, txt_masks, **kwargs): |
|
|
return self.model(txt_ids, txt_masks).last_hidden_state |
|
|
|
|
|
|