|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import math |
|
|
from collections.abc import Sequence |
|
|
|
|
|
import torch |
|
|
from torch import nn |
|
|
|
|
|
from monai.config.type_definitions import PathLike |
|
|
from monai.utils import optional_import |
|
|
|
|
|
transformers = optional_import("transformers") |
|
|
load_tf_weights_in_bert = optional_import("transformers", name="load_tf_weights_in_bert")[0] |
|
|
cached_file = optional_import("transformers.utils", name="cached_file")[0] |
|
|
BertEmbeddings = optional_import("transformers.models.bert.modeling_bert", name="BertEmbeddings")[0] |
|
|
BertLayer = optional_import("transformers.models.bert.modeling_bert", name="BertLayer")[0] |
|
|
|
|
|
__all__ = ["BertPreTrainedModel", "BertAttention", "BertOutput", "BertMixedLayer", "Pooler", "MultiModal", "Transchex"] |
|
|
|
|
|
|
|
|
class BertPreTrainedModel(nn.Module): |
|
|
"""Module to load BERT pre-trained weights. |
|
|
Based on: |
|
|
LXMERT |
|
|
https://github.com/airsplay/lxmert |
|
|
BERT (pytorch-transformer) |
|
|
https://github.com/huggingface/transformers |
|
|
""" |
|
|
|
|
|
def __init__(self, *inputs, **kwargs) -> None: |
|
|
super().__init__() |
|
|
|
|
|
def init_bert_weights(self, module): |
|
|
if isinstance(module, (nn.Linear, nn.Embedding)): |
|
|
module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) |
|
|
elif isinstance(module, torch.nn.LayerNorm): |
|
|
module.bias.data.zero_() |
|
|
module.weight.data.fill_(1.0) |
|
|
if isinstance(module, nn.Linear) and module.bias is not None: |
|
|
module.bias.data.zero_() |
|
|
|
|
|
@classmethod |
|
|
def from_pretrained( |
|
|
cls, |
|
|
num_language_layers, |
|
|
num_vision_layers, |
|
|
num_mixed_layers, |
|
|
bert_config, |
|
|
state_dict=None, |
|
|
cache_dir=None, |
|
|
from_tf=False, |
|
|
path_or_repo_id="bert-base-uncased", |
|
|
filename="pytorch_model.bin", |
|
|
*inputs, |
|
|
**kwargs, |
|
|
): |
|
|
weights_path = cached_file(path_or_repo_id, filename, cache_dir=cache_dir) |
|
|
model = cls(num_language_layers, num_vision_layers, num_mixed_layers, bert_config, *inputs, **kwargs) |
|
|
if state_dict is None and not from_tf: |
|
|
state_dict = torch.load(weights_path, map_location="cpu" if not torch.cuda.is_available() else None) |
|
|
if from_tf: |
|
|
return load_tf_weights_in_bert(model, weights_path) |
|
|
old_keys = [] |
|
|
new_keys = [] |
|
|
for key in state_dict.keys(): |
|
|
new_key = None |
|
|
if "gamma" in key: |
|
|
new_key = key.replace("gamma", "weight") |
|
|
if "beta" in key: |
|
|
new_key = key.replace("beta", "bias") |
|
|
if new_key: |
|
|
old_keys.append(key) |
|
|
new_keys.append(new_key) |
|
|
for old_key, new_key in zip(old_keys, new_keys): |
|
|
state_dict[new_key] = state_dict.pop(old_key) |
|
|
missing_keys: list = [] |
|
|
unexpected_keys: list = [] |
|
|
error_msgs: list = [] |
|
|
metadata = getattr(state_dict, "_metadata", None) |
|
|
state_dict = state_dict.copy() |
|
|
if metadata is not None: |
|
|
state_dict._metadata = metadata |
|
|
|
|
|
def load(module, prefix=""): |
|
|
local_metadata = {} if metadata is None else metadata.get(prefix[:-1], {}) |
|
|
module._load_from_state_dict( |
|
|
state_dict, prefix, local_metadata, True, missing_keys, unexpected_keys, error_msgs |
|
|
) |
|
|
for name, child in module._modules.items(): |
|
|
if child is not None: |
|
|
load(child, prefix + name + ".") |
|
|
|
|
|
start_prefix = "" |
|
|
if not hasattr(model, "bert") and any(s.startswith("bert.") for s in state_dict.keys()): |
|
|
start_prefix = "bert." |
|
|
load(model, prefix=start_prefix) |
|
|
return model |
|
|
|
|
|
|
|
|
class BertAttention(nn.Module): |
|
|
"""BERT attention layer. |
|
|
Based on: BERT (pytorch-transformer) |
|
|
https://github.com/huggingface/transformers |
|
|
""" |
|
|
|
|
|
def __init__(self, config) -> None: |
|
|
super().__init__() |
|
|
self.num_attention_heads = config.num_attention_heads |
|
|
self.attention_head_size = int(config.hidden_size / config.num_attention_heads) |
|
|
self.all_head_size = self.num_attention_heads * self.attention_head_size |
|
|
self.query = nn.Linear(config.hidden_size, self.all_head_size) |
|
|
self.key = nn.Linear(config.hidden_size, self.all_head_size) |
|
|
self.value = nn.Linear(config.hidden_size, self.all_head_size) |
|
|
self.dropout = nn.Dropout(config.attention_probs_dropout_prob) |
|
|
|
|
|
def transpose_for_scores(self, x): |
|
|
new_x_shape = x.size()[:-1] + (self.num_attention_heads, self.attention_head_size) |
|
|
x = x.view(*new_x_shape) |
|
|
return x.permute(0, 2, 1, 3) |
|
|
|
|
|
def forward(self, hidden_states, context): |
|
|
mixed_query_layer = self.query(hidden_states) |
|
|
mixed_key_layer = self.key(context) |
|
|
mixed_value_layer = self.value(context) |
|
|
query_layer = self.transpose_for_scores(mixed_query_layer) |
|
|
key_layer = self.transpose_for_scores(mixed_key_layer) |
|
|
value_layer = self.transpose_for_scores(mixed_value_layer) |
|
|
attention_scores = torch.matmul(query_layer, key_layer.transpose(-1, -2)) |
|
|
attention_scores = attention_scores / math.sqrt(self.attention_head_size) |
|
|
attention_probs = self.dropout(nn.Softmax(dim=-1)(attention_scores)) |
|
|
context_layer = torch.matmul(attention_probs, value_layer) |
|
|
context_layer = context_layer.permute(0, 2, 1, 3).contiguous() |
|
|
new_context_layer_shape = context_layer.size()[:-2] + (self.all_head_size,) |
|
|
context_layer = context_layer.view(*new_context_layer_shape) |
|
|
return context_layer |
|
|
|
|
|
|
|
|
class BertOutput(nn.Module): |
|
|
"""BERT output layer. |
|
|
Based on: BERT (pytorch-transformer) |
|
|
https://github.com/huggingface/transformers |
|
|
""" |
|
|
|
|
|
def __init__(self, config) -> None: |
|
|
super().__init__() |
|
|
self.dense = nn.Linear(config.hidden_size, config.hidden_size) |
|
|
self.LayerNorm = torch.nn.LayerNorm(config.hidden_size, eps=1e-12) |
|
|
self.dropout = nn.Dropout(config.hidden_dropout_prob) |
|
|
|
|
|
def forward(self, hidden_states, input_tensor): |
|
|
hidden_states = self.dense(hidden_states) |
|
|
hidden_states = self.dropout(hidden_states) |
|
|
hidden_states = self.LayerNorm(hidden_states + input_tensor) |
|
|
return hidden_states |
|
|
|
|
|
|
|
|
class BertMixedLayer(nn.Module): |
|
|
"""BERT cross attention layer. |
|
|
Based on: BERT (pytorch-transformer) |
|
|
https://github.com/huggingface/transformers |
|
|
""" |
|
|
|
|
|
def __init__(self, config) -> None: |
|
|
super().__init__() |
|
|
self.att_x = BertAttention(config) |
|
|
self.output_x = BertOutput(config) |
|
|
self.att_y = BertAttention(config) |
|
|
self.output_y = BertOutput(config) |
|
|
|
|
|
def forward(self, x, y): |
|
|
output_x = self.att_x(x, y) |
|
|
output_y = self.att_y(y, x) |
|
|
return self.output_x(output_x, x), self.output_y(output_y, y) |
|
|
|
|
|
|
|
|
class Pooler(nn.Module): |
|
|
"""BERT pooler layer. |
|
|
Based on: BERT (pytorch-transformer) |
|
|
https://github.com/huggingface/transformers |
|
|
""" |
|
|
|
|
|
def __init__(self, hidden_size) -> None: |
|
|
super().__init__() |
|
|
self.dense = nn.Linear(hidden_size, hidden_size) |
|
|
self.activation = nn.Tanh() |
|
|
|
|
|
def forward(self, hidden_states): |
|
|
first_token_tensor = hidden_states[:, 0] |
|
|
pooled_output = self.dense(first_token_tensor) |
|
|
pooled_output = self.activation(pooled_output) |
|
|
return pooled_output |
|
|
|
|
|
|
|
|
class MultiModal(BertPreTrainedModel): |
|
|
""" |
|
|
Multimodal Transformers From Pretrained BERT Weights" |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, num_language_layers: int, num_vision_layers: int, num_mixed_layers: int, bert_config: dict |
|
|
) -> None: |
|
|
""" |
|
|
Args: |
|
|
num_language_layers: number of language transformer layers. |
|
|
num_vision_layers: number of vision transformer layers. |
|
|
bert_config: configuration for bert language transformer encoder. |
|
|
|
|
|
""" |
|
|
super().__init__() |
|
|
self.config = type("obj", (object,), bert_config) |
|
|
self.embeddings = BertEmbeddings(self.config) |
|
|
self.language_encoder = nn.ModuleList([BertLayer(self.config) for _ in range(num_language_layers)]) |
|
|
self.vision_encoder = nn.ModuleList([BertLayer(self.config) for _ in range(num_vision_layers)]) |
|
|
self.mixed_encoder = nn.ModuleList([BertMixedLayer(self.config) for _ in range(num_mixed_layers)]) |
|
|
self.apply(self.init_bert_weights) |
|
|
|
|
|
def forward(self, input_ids, token_type_ids=None, vision_feats=None, attention_mask=None): |
|
|
language_features = self.embeddings(input_ids, token_type_ids) |
|
|
for layer in self.vision_encoder: |
|
|
vision_feats = layer(vision_feats, None)[0] |
|
|
for layer in self.language_encoder: |
|
|
language_features = layer(language_features, attention_mask)[0] |
|
|
for layer in self.mixed_encoder: |
|
|
language_features, vision_feats = layer(language_features, vision_feats) |
|
|
return language_features, vision_feats |
|
|
|
|
|
|
|
|
class Transchex(torch.nn.Module): |
|
|
""" |
|
|
TransChex based on: "Hatamizadeh et al.,TransCheX: Self-Supervised Pretraining of Vision-Language |
|
|
Transformers for Chest X-ray Analysis" |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
in_channels: int, |
|
|
img_size: Sequence[int] | int, |
|
|
patch_size: int | tuple[int, int], |
|
|
num_classes: int, |
|
|
num_language_layers: int, |
|
|
num_vision_layers: int, |
|
|
num_mixed_layers: int, |
|
|
hidden_size: int = 768, |
|
|
drop_out: float = 0.0, |
|
|
attention_probs_dropout_prob: float = 0.1, |
|
|
gradient_checkpointing: bool = False, |
|
|
hidden_act: str = "gelu", |
|
|
hidden_dropout_prob: float = 0.1, |
|
|
initializer_range: float = 0.02, |
|
|
intermediate_size: int = 3072, |
|
|
layer_norm_eps: float = 1e-12, |
|
|
max_position_embeddings: int = 512, |
|
|
model_type: str = "bert", |
|
|
num_attention_heads: int = 12, |
|
|
num_hidden_layers: int = 12, |
|
|
pad_token_id: int = 0, |
|
|
position_embedding_type: str = "absolute", |
|
|
transformers_version: str = "4.10.2", |
|
|
type_vocab_size: int = 2, |
|
|
use_cache: bool = True, |
|
|
vocab_size: int = 30522, |
|
|
chunk_size_feed_forward: int = 0, |
|
|
is_decoder: bool = False, |
|
|
add_cross_attention: bool = False, |
|
|
path_or_repo_id: str | PathLike = "bert-base-uncased", |
|
|
filename: str = "pytorch_model.bin", |
|
|
) -> None: |
|
|
""" |
|
|
Args: |
|
|
in_channels: dimension of input channels. |
|
|
img_size: dimension of input image. |
|
|
patch_size: dimension of patch size. |
|
|
num_classes: number of classes if classification is used. |
|
|
num_language_layers: number of language transformer layers. |
|
|
num_vision_layers: number of vision transformer layers. |
|
|
num_mixed_layers: number of mixed transformer layers. |
|
|
drop_out: fraction of the input units to drop. |
|
|
path_or_repo_id: This can be either: |
|
|
- a string, the *model id* of a model repo on huggingface.co. |
|
|
- a path to a *directory* potentially containing the file. |
|
|
filename: The name of the file to locate in `path_or_repo`. |
|
|
|
|
|
The other parameters are part of the `bert_config` to `MultiModal.from_pretrained`. |
|
|
|
|
|
Examples: |
|
|
|
|
|
.. code-block:: python |
|
|
|
|
|
# for 3-channel with image size of (224,224), patch size of (32,32), 3 classes, 2 language layers, |
|
|
# 2 vision layers, 2 mixed modality layers and dropout of 0.2 in the classification head |
|
|
net = Transchex(in_channels=3, |
|
|
img_size=(224, 224), |
|
|
num_classes=3, |
|
|
num_language_layers=2, |
|
|
num_vision_layers=2, |
|
|
num_mixed_layers=2, |
|
|
drop_out=0.2) |
|
|
|
|
|
""" |
|
|
super().__init__() |
|
|
bert_config = { |
|
|
"attention_probs_dropout_prob": attention_probs_dropout_prob, |
|
|
"classifier_dropout": None, |
|
|
"gradient_checkpointing": gradient_checkpointing, |
|
|
"hidden_act": hidden_act, |
|
|
"hidden_dropout_prob": hidden_dropout_prob, |
|
|
"hidden_size": hidden_size, |
|
|
"initializer_range": initializer_range, |
|
|
"intermediate_size": intermediate_size, |
|
|
"layer_norm_eps": layer_norm_eps, |
|
|
"max_position_embeddings": max_position_embeddings, |
|
|
"model_type": model_type, |
|
|
"num_attention_heads": num_attention_heads, |
|
|
"num_hidden_layers": num_hidden_layers, |
|
|
"pad_token_id": pad_token_id, |
|
|
"position_embedding_type": position_embedding_type, |
|
|
"transformers_version": transformers_version, |
|
|
"type_vocab_size": type_vocab_size, |
|
|
"use_cache": use_cache, |
|
|
"vocab_size": vocab_size, |
|
|
"chunk_size_feed_forward": chunk_size_feed_forward, |
|
|
"is_decoder": is_decoder, |
|
|
"add_cross_attention": add_cross_attention, |
|
|
} |
|
|
if not (0 <= drop_out <= 1): |
|
|
raise ValueError("dropout_rate should be between 0 and 1.") |
|
|
|
|
|
if (img_size[0] % patch_size[0] != 0) or (img_size[1] % patch_size[1] != 0): |
|
|
raise ValueError("img_size should be divisible by patch_size.") |
|
|
|
|
|
self.multimodal = MultiModal.from_pretrained( |
|
|
num_language_layers=num_language_layers, |
|
|
num_vision_layers=num_vision_layers, |
|
|
num_mixed_layers=num_mixed_layers, |
|
|
bert_config=bert_config, |
|
|
path_or_repo_id=path_or_repo_id, |
|
|
filename=filename, |
|
|
) |
|
|
|
|
|
self.patch_size = patch_size |
|
|
self.num_patches = (img_size[0] // self.patch_size[0]) * (img_size[1] // self.patch_size[1]) |
|
|
self.vision_proj = nn.Conv2d( |
|
|
in_channels=in_channels, out_channels=hidden_size, kernel_size=self.patch_size, stride=self.patch_size |
|
|
) |
|
|
self.norm_vision_pos = nn.LayerNorm(hidden_size) |
|
|
self.pos_embed_vis = nn.Parameter(torch.zeros(1, self.num_patches, hidden_size)) |
|
|
self.pooler = Pooler(hidden_size=hidden_size) |
|
|
self.drop = torch.nn.Dropout(drop_out) |
|
|
self.cls_head = torch.nn.Linear(hidden_size, num_classes) |
|
|
|
|
|
def forward(self, input_ids, token_type_ids=None, vision_feats=None): |
|
|
attention_mask = torch.ones_like(input_ids).unsqueeze(1).unsqueeze(2) |
|
|
attention_mask = attention_mask.to(dtype=next(self.parameters()).dtype) |
|
|
attention_mask = (1.0 - attention_mask) * -10000.0 |
|
|
vision_feats = self.vision_proj(vision_feats).flatten(2).transpose(1, 2) |
|
|
vision_feats = self.norm_vision_pos(vision_feats) |
|
|
vision_feats = vision_feats + self.pos_embed_vis |
|
|
hidden_state_lang, hidden_state_vis = self.multimodal( |
|
|
input_ids=input_ids, token_type_ids=token_type_ids, vision_feats=vision_feats, attention_mask=attention_mask |
|
|
) |
|
|
pooled_features = self.pooler(hidden_state_lang) |
|
|
logits = self.cls_head(self.drop(pooled_features)) |
|
|
return logits |
|
|
|