constanceCM's picture
upload
8aa674c
import math
import torch
import torch.nn as nn
from .mlp import MLP
from ...builder import TRANSFORMERS
@TRANSFORMERS.register_module()
class SharedHead(nn.Module):
def __init__(self, embed_dim, num_classes, num_layers=3):
super().__init__()
# define classification head and box head
self.class_embed = nn.Linear(
embed_dim, num_classes
) # here we do not use +1, since we are using sigmoid focal loss
self.bbox_embed = MLP(input_dim=embed_dim, hidden_dim=embed_dim, output_dim=2, num_layers=num_layers)
def forward(self, hidden_states):
outputs_class = self.class_embed(hidden_states)
outputs_coord = self.bbox_embed(hidden_states).sigmoid()
return outputs_class, outputs_coord