AnyBox / project /grasp_box /submodules /SAM6D /model /fine_point_matching.py
savoji's picture
Add files using upload-large-folder tool
b24c748 verified
import torch
import torch.nn as nn
import torch.nn.functional as F
from transformer import SparseToDenseTransformer
from model_utils import compute_feature_similarity, compute_fine_Rt
from loss_utils import compute_correspondence_loss
from pointnet2_utils import QueryAndGroup
from pytorch_utils import SharedMLP, Conv1d
class FinePointMatching(nn.Module):
def __init__(self, cfg, return_feat=False):
super(FinePointMatching, self).__init__()
self.cfg = cfg
self.return_feat = return_feat
self.nblock = self.cfg.nblock
self.in_proj = nn.Linear(cfg.input_dim, cfg.hidden_dim)
self.out_proj = nn.Linear(cfg.hidden_dim, cfg.out_dim)
self.bg_token = nn.Parameter(torch.randn(1, 1, cfg.hidden_dim) * .02)
self.PE = PositionalEncoding(cfg.hidden_dim, r1=cfg.pe_radius1, r2=cfg.pe_radius2)
self.transformers = []
for _ in range(self.nblock):
self.transformers.append(SparseToDenseTransformer(
cfg.hidden_dim,
num_heads=4,
sparse_blocks=['self', 'cross'],
dropout=None,
activation_fn='ReLU',
focusing_factor=cfg.focusing_factor,
with_bg_token=True,
replace_bg_token=True
))
self.transformers = nn.ModuleList(self.transformers)
def forward(self, p1, f1, geo1, fps_idx1, p2, f2, geo2, fps_idx2, radius, end_points):
B = p1.size(0)
init_R = end_points['init_R']
init_t = end_points['init_t']
p1_ = (p1 - init_t.unsqueeze(1)) @ init_R
f1 = self.in_proj(f1) + self.PE(p1_)
f1 = torch.cat([self.bg_token.repeat(B,1,1), f1], dim=1) # adding bg
f2 = self.in_proj(f2) + self.PE(p2)
f2 = torch.cat([self.bg_token.repeat(B,1,1), f2], dim=1) # adding bg
atten_list = []
for idx in range(self.nblock):
f1, f2 = self.transformers[idx](f1, geo1, fps_idx1, f2, geo2, fps_idx2)
if self.training or idx==self.nblock-1:
atten_list.append(compute_feature_similarity(
self.out_proj(f1),
self.out_proj(f2),
self.cfg.sim_type,
self.cfg.temp,
self.cfg.normalize_feat
))
if self.training:
gt_R = end_points['rotation_label']
gt_t = end_points['translation_label'] / (radius.reshape(-1, 1)+1e-6)
end_points = compute_correspondence_loss(
end_points, atten_list, p1, p2, gt_R, gt_t,
dis_thres=self.cfg.loss_dis_thres,
loss_str='fine'
)
else:
pred_R, pred_t, pred_pose_score, points = compute_fine_Rt(
atten_list[-1], p1, p2, radius, end_points,
end_points['model'] / (radius.reshape(-1, 1, 1) + 1e-6),
)
end_points['pred_R'] = pred_R
end_points['pred_t'] = pred_t * (radius.reshape(-1, 1)+1e-6)
end_points['pred_pose_score'] = pred_pose_score
if torch.equal(pred_t, torch.tensor([0.,0.,0.], device = 'cuda').reshape(1,3)):
end_points['pred_pose_score'] = None
if self.return_feat:
return end_points, self.out_proj(f1), self.out_proj(f2)
else:
return end_points, points
class PositionalEncoding(nn.Module):
def __init__(self, out_dim, r1=0.1, r2=0.2, nsample1=32, nsample2=64, use_xyz=True, bn=True):
super(PositionalEncoding, self).__init__()
self.group1 = QueryAndGroup(r1, nsample1, use_xyz=use_xyz)
self.group2 = QueryAndGroup(r2, nsample2, use_xyz=use_xyz)
input_dim = 6 if use_xyz else 3
self.mlp1 = SharedMLP([input_dim, 32, 64, 128], bn=bn)
self.mlp2 = SharedMLP([input_dim, 32, 64, 128], bn=bn)
self.mlp3 = Conv1d(256, out_dim, 1, activation=None, bn=None)
def forward(self, pts1, pts2=None):
if pts2 is None:
pts2 = pts1
# scale1
feat1 = self.group1(
pts1.contiguous(), pts2.contiguous(), pts1.transpose(1,2).contiguous()
)
feat1 = self.mlp1(feat1)
feat1 = F.max_pool2d(
feat1, kernel_size=[1, feat1.size(3)]
)
# scale2
feat2 = self.group2(
pts1.contiguous(), pts2.contiguous(), pts1.transpose(1,2).contiguous()
)
feat2 = self.mlp2(feat2)
feat2 = F.max_pool2d(
feat2, kernel_size=[1, feat2.size(3)]
)
feat = torch.cat([feat1, feat2], dim=1).squeeze(-1)
feat = self.mlp3(feat).transpose(1,2)
return feat