Spaces:
Runtime error
Runtime error
| import math | |
| import torch | |
| from torch import nn, Tensor | |
| from torch.nn import functional as F | |
| from typing import Optional | |
| class ShortCut_CrossAttention(nn.Module): | |
| def __init__(self, d_model, nhead, panoptic_on = False): | |
| super().__init__() | |
| self.multihead_attn = nn.MultiheadAttention(d_model, nhead, dropout=0.0) | |
| self.norm = nn.LayerNorm(d_model) | |
| self.activation = F.relu | |
| self._reset_parameters() | |
| self.MLP = nn.Linear(d_model, d_model) | |
| self.panoptic_on = panoptic_on | |
| if panoptic_on: | |
| nn.init.constant(self.MLP.weight, 0.0) | |
| nn.init.constant(self.MLP.bias, 0.0) | |
| def _reset_parameters(self): | |
| for p in self.parameters(): | |
| if p.dim() > 1: | |
| nn.init.xavier_uniform_(p) | |
| def with_pos_embed(self, tensor, pos: Optional[Tensor]): | |
| return tensor if pos is None else tensor + pos | |
| def forward(self, tgt, memory, | |
| memory_mask: Optional[Tensor] = None, | |
| memory_key_padding_mask: Optional[Tensor] = None, | |
| pos: Optional[Tensor] = None, | |
| query_pos: Optional[Tensor] = None): | |
| tgt2 = self.multihead_attn(query=self.with_pos_embed(tgt, query_pos), | |
| key=self.with_pos_embed(memory, pos), | |
| value=memory, attn_mask=memory_mask, | |
| key_padding_mask=memory_key_padding_mask)[0] | |
| if self.panoptic_on: | |
| tgt = tgt + self.norm(self.MLP(tgt2)) | |
| else: | |
| tgt = self.norm(tgt + self.MLP(tgt2)) | |
| return tgt | |
| class ContentDependentTransfer(nn.Module): | |
| def __init__(self, d_model, nhead, panoptic_on): | |
| super().__init__() | |
| self.pe_layer = PositionEmbeddingSine(d_model//2, normalize=True) | |
| self.cross_atten = ShortCut_CrossAttention(d_model = d_model, nhead = nhead, panoptic_on = panoptic_on) | |
| def visual_prediction_forward_convnext(self, x): | |
| batch, channel, h, w = x.shape | |
| x = x.reshape(batch*h*w, channel).unsqueeze(-1).unsqueeze(-1) # fake 2D input | |
| x = self.truck_head(x) | |
| x = self.head(x) | |
| return x.reshape(batch, h, w, x.shape[-1]).permute(0,3,1,2) # B x num_queries x 640 | |
| def forward(self, img_feat, text_classifier, ): | |
| text_classifier = text_classifier.unsqueeze(0).repeat(img_feat.shape[0],1,1) | |
| pos = self.pe_layer(img_feat, None).flatten(2).permute(2, 0, 1) # hw * b * c | |
| img_feat = img_feat.flatten(2).permute(2, 0, 1) # hw * b * c | |
| bias = self.cross_atten(text_classifier.permute(1, 0, 2), img_feat, memory_mask=None, memory_key_padding_mask=None, pos=pos, query_pos=None) | |
| return bias.permute(1, 0, 2) | |
| class PositionEmbeddingSine(nn.Module): | |
| """ | |
| This is a more standard version of the position embedding, very similar to the one | |
| used by the Attention is all you need paper, generalized to work on images. | |
| """ | |
| def __init__(self, num_pos_feats=64, temperature=10000, normalize=False, scale=None): | |
| super().__init__() | |
| self.num_pos_feats = num_pos_feats | |
| self.temperature = temperature | |
| self.normalize = normalize | |
| if scale is not None and normalize is False: | |
| raise ValueError("normalize should be True if scale is passed") | |
| if scale is None: | |
| scale = 2 * math.pi | |
| self.scale = scale | |
| def forward(self, x, mask=None): | |
| if mask is None: | |
| mask = torch.zeros((x.size(0), x.size(2), x.size(3)), device=x.device, dtype=torch.bool) | |
| not_mask = ~mask | |
| y_embed = not_mask.cumsum(1, dtype=torch.float32) | |
| x_embed = not_mask.cumsum(2, dtype=torch.float32) | |
| if self.normalize: | |
| eps = 1e-6 | |
| y_embed = y_embed / (y_embed[:, -1:, :] + eps) * self.scale | |
| x_embed = x_embed / (x_embed[:, :, -1:] + eps) * self.scale | |
| dim_t = torch.arange(self.num_pos_feats, dtype=torch.float32, device=x.device) | |
| dim_t = self.temperature ** (2 * (dim_t // 2) / self.num_pos_feats) | |
| pos_x = x_embed[:, :, :, None] / dim_t | |
| pos_y = y_embed[:, :, :, None] / dim_t | |
| pos_x = torch.stack( | |
| (pos_x[:, :, :, 0::2].sin(), pos_x[:, :, :, 1::2].cos()), dim=4 | |
| ).flatten(3) | |
| pos_y = torch.stack( | |
| (pos_y[:, :, :, 0::2].sin(), pos_y[:, :, :, 1::2].cos()), dim=4 | |
| ).flatten(3) | |
| pos = torch.cat((pos_y, pos_x), dim=3).permute(0, 3, 1, 2) | |
| return pos | |
| def __repr__(self, _repr_indent=4): | |
| head = "Positional encoding " + self.__class__.__name__ | |
| body = [ | |
| "num_pos_feats: {}".format(self.num_pos_feats), | |
| "temperature: {}".format(self.temperature), | |
| "normalize: {}".format(self.normalize), | |
| "scale: {}".format(self.scale), | |
| ] | |
| # _repr_indent = 4 | |
| lines = [head] + [" " * _repr_indent + line for line in body] | |
| return "\n".join(lines) | |