| | |
| | import torch |
| | from torch import nn |
| | from torch.autograd import Function |
| | from torch.autograd.function import once_differentiable |
| | from torch.nn.modules.utils import _pair |
| |
|
| |
|
| | class _ROIAlignRotated(Function): |
| | @staticmethod |
| | def forward(ctx, input, roi, output_size, spatial_scale, sampling_ratio): |
| | ctx.save_for_backward(roi) |
| | ctx.output_size = _pair(output_size) |
| | ctx.spatial_scale = spatial_scale |
| | ctx.sampling_ratio = sampling_ratio |
| | ctx.input_shape = input.size() |
| | output = torch.ops.detectron2.roi_align_rotated_forward( |
| | input, roi, spatial_scale, output_size[0], output_size[1], sampling_ratio |
| | ) |
| | return output |
| |
|
| | @staticmethod |
| | @once_differentiable |
| | def backward(ctx, grad_output): |
| | (rois,) = ctx.saved_tensors |
| | output_size = ctx.output_size |
| | spatial_scale = ctx.spatial_scale |
| | sampling_ratio = ctx.sampling_ratio |
| | bs, ch, h, w = ctx.input_shape |
| | grad_input = torch.ops.detectron2.roi_align_rotated_backward( |
| | grad_output, |
| | rois, |
| | spatial_scale, |
| | output_size[0], |
| | output_size[1], |
| | bs, |
| | ch, |
| | h, |
| | w, |
| | sampling_ratio, |
| | ) |
| | return grad_input, None, None, None, None, None |
| |
|
| |
|
| | roi_align_rotated = _ROIAlignRotated.apply |
| |
|
| |
|
| | class ROIAlignRotated(nn.Module): |
| | def __init__(self, output_size, spatial_scale, sampling_ratio): |
| | """ |
| | Args: |
| | output_size (tuple): h, w |
| | spatial_scale (float): scale the input boxes by this number |
| | sampling_ratio (int): number of inputs samples to take for each output |
| | sample. 0 to take samples densely. |
| | |
| | Note: |
| | ROIAlignRotated supports continuous coordinate by default: |
| | Given a continuous coordinate c, its two neighboring pixel indices (in our |
| | pixel model) are computed by floor(c - 0.5) and ceil(c - 0.5). For example, |
| | c=1.3 has pixel neighbors with discrete indices [0] and [1] (which are sampled |
| | from the underlying signal at continuous coordinates 0.5 and 1.5). |
| | """ |
| | super(ROIAlignRotated, self).__init__() |
| | self.output_size = output_size |
| | self.spatial_scale = spatial_scale |
| | self.sampling_ratio = sampling_ratio |
| |
|
| | def forward(self, input, rois): |
| | """ |
| | Args: |
| | input: NCHW images |
| | rois: Bx6 boxes. First column is the index into N. |
| | The other 5 columns are (x_ctr, y_ctr, width, height, angle_degrees). |
| | """ |
| | assert rois.dim() == 2 and rois.size(1) == 6 |
| | orig_dtype = input.dtype |
| | if orig_dtype == torch.float16: |
| | input = input.float() |
| | rois = rois.float() |
| | output_size = _pair(self.output_size) |
| |
|
| | |
| | |
| | if torch.jit.is_scripting() or torch.jit.is_tracing(): |
| | return torch.ops.detectron2.roi_align_rotated_forward( |
| | input, rois, self.spatial_scale, output_size[0], output_size[1], self.sampling_ratio |
| | ).to(dtype=orig_dtype) |
| |
|
| | return roi_align_rotated( |
| | input, rois, self.output_size, self.spatial_scale, self.sampling_ratio |
| | ).to(dtype=orig_dtype) |
| |
|
| | def __repr__(self): |
| | tmpstr = self.__class__.__name__ + "(" |
| | tmpstr += "output_size=" + str(self.output_size) |
| | tmpstr += ", spatial_scale=" + str(self.spatial_scale) |
| | tmpstr += ", sampling_ratio=" + str(self.sampling_ratio) |
| | tmpstr += ")" |
| | return tmpstr |
| |
|