| import torch |
| from torch.autograd import Function |
|
|
| from ..utils import ext_loader |
|
|
| ext_module = ext_loader.load_ext( |
| '_ext', ['gather_points_forward', 'gather_points_backward']) |
|
|
|
|
| class GatherPoints(Function): |
| """Gather points with given index.""" |
|
|
| @staticmethod |
| def forward(ctx, features: torch.Tensor, |
| indices: torch.Tensor) -> torch.Tensor: |
| """ |
| Args: |
| features (Tensor): (B, C, N) features to gather. |
| indices (Tensor): (B, M) where M is the number of points. |
| |
| Returns: |
| Tensor: (B, C, M) where M is the number of points. |
| """ |
| assert features.is_contiguous() |
| assert indices.is_contiguous() |
|
|
| B, npoint = indices.size() |
| _, C, N = features.size() |
| output = torch.cuda.FloatTensor(B, C, npoint) |
|
|
| ext_module.gather_points_forward( |
| features, indices, output, b=B, c=C, n=N, npoints=npoint) |
|
|
| ctx.for_backwards = (indices, C, N) |
| if torch.__version__ != 'parrots': |
| ctx.mark_non_differentiable(indices) |
| return output |
|
|
| @staticmethod |
| def backward(ctx, grad_out): |
| idx, C, N = ctx.for_backwards |
| B, npoint = idx.size() |
|
|
| grad_features = torch.cuda.FloatTensor(B, C, N).zero_() |
| grad_out_data = grad_out.data.contiguous() |
| ext_module.gather_points_backward( |
| grad_out_data, |
| idx, |
| grad_features.data, |
| b=B, |
| c=C, |
| n=N, |
| npoints=npoint) |
| return grad_features, None |
|
|
|
|
| gather_points = GatherPoints.apply |
|
|