|
|
| """
|
| Wrappers around on some nn functions, mainly to support empty tensors.
|
|
|
| Ideally, add support directly in PyTorch to empty tensors in those functions.
|
|
|
| These can be removed once https://github.com/pytorch/pytorch/issues/12013
|
| is implemented
|
| """
|
|
|
| import warnings
|
| from typing import List, Optional
|
| import torch
|
| from torch.nn import functional as F
|
|
|
| from annotator.oneformer.detectron2.utils.env import TORCH_VERSION
|
|
|
|
|
| def shapes_to_tensor(x: List[int], device: Optional[torch.device] = None) -> torch.Tensor:
|
| """
|
| Turn a list of integer scalars or integer Tensor scalars into a vector,
|
| in a way that's both traceable and scriptable.
|
|
|
| In tracing, `x` should be a list of scalar Tensor, so the output can trace to the inputs.
|
| In scripting or eager, `x` should be a list of int.
|
| """
|
| if torch.jit.is_scripting():
|
| return torch.as_tensor(x, device=device)
|
| if torch.jit.is_tracing():
|
| assert all(
|
| [isinstance(t, torch.Tensor) for t in x]
|
| ), "Shape should be tensor during tracing!"
|
|
|
| ret = torch.stack(x)
|
| if ret.device != device:
|
| ret = ret.to(device=device)
|
| return ret
|
| return torch.as_tensor(x, device=device)
|
|
|
|
|
| def check_if_dynamo_compiling():
|
| if TORCH_VERSION >= (1, 14):
|
| from torch._dynamo import is_compiling
|
|
|
| return is_compiling()
|
| else:
|
| return False
|
|
|
|
|
| def cat(tensors: List[torch.Tensor], dim: int = 0):
|
| """
|
| Efficient version of torch.cat that avoids a copy if there is only a single element in a list
|
| """
|
| assert isinstance(tensors, (list, tuple))
|
| if len(tensors) == 1:
|
| return tensors[0]
|
| return torch.cat(tensors, dim)
|
|
|
|
|
| def empty_input_loss_func_wrapper(loss_func):
|
| def wrapped_loss_func(input, target, *, reduction="mean", **kwargs):
|
| """
|
| Same as `loss_func`, but returns 0 (instead of nan) for empty inputs.
|
| """
|
| if target.numel() == 0 and reduction == "mean":
|
| return input.sum() * 0.0
|
| return loss_func(input, target, reduction=reduction, **kwargs)
|
|
|
| return wrapped_loss_func
|
|
|
|
|
| cross_entropy = empty_input_loss_func_wrapper(F.cross_entropy)
|
|
|
|
|
| class _NewEmptyTensorOp(torch.autograd.Function):
|
| @staticmethod
|
| def forward(ctx, x, new_shape):
|
| ctx.shape = x.shape
|
| return x.new_empty(new_shape)
|
|
|
| @staticmethod
|
| def backward(ctx, grad):
|
| shape = ctx.shape
|
| return _NewEmptyTensorOp.apply(grad, shape), None
|
|
|
|
|
| class Conv2d(torch.nn.Conv2d):
|
| """
|
| A wrapper around :class:`torch.nn.Conv2d` to support empty inputs and more features.
|
| """
|
|
|
| def __init__(self, *args, **kwargs):
|
| """
|
| Extra keyword arguments supported in addition to those in `torch.nn.Conv2d`:
|
|
|
| Args:
|
| norm (nn.Module, optional): a normalization layer
|
| activation (callable(Tensor) -> Tensor): a callable activation function
|
|
|
| It assumes that norm layer is used before activation.
|
| """
|
| norm = kwargs.pop("norm", None)
|
| activation = kwargs.pop("activation", None)
|
| super().__init__(*args, **kwargs)
|
|
|
| self.norm = norm
|
| self.activation = activation
|
|
|
| def forward(self, x):
|
|
|
|
|
|
|
|
|
|
|
|
|
| if not torch.jit.is_scripting():
|
|
|
| is_dynamo_compiling = check_if_dynamo_compiling()
|
| if not is_dynamo_compiling:
|
| with warnings.catch_warnings(record=True):
|
| if x.numel() == 0 and self.training:
|
|
|
| assert not isinstance(
|
| self.norm, torch.nn.SyncBatchNorm
|
| ), "SyncBatchNorm does not support empty inputs!"
|
|
|
| x = F.conv2d(
|
| x, self.weight, self.bias, self.stride, self.padding, self.dilation, self.groups
|
| )
|
| if self.norm is not None:
|
| x = self.norm(x)
|
| if self.activation is not None:
|
| x = self.activation(x)
|
| return x
|
|
|
|
|
| ConvTranspose2d = torch.nn.ConvTranspose2d
|
| BatchNorm2d = torch.nn.BatchNorm2d
|
| interpolate = F.interpolate
|
| Linear = torch.nn.Linear
|
|
|
|
|
| def nonzero_tuple(x):
|
| """
|
| A 'as_tuple=True' version of torch.nonzero to support torchscript.
|
| because of https://github.com/pytorch/pytorch/issues/38718
|
| """
|
| if torch.jit.is_scripting():
|
| if x.dim() == 0:
|
| return x.unsqueeze(0).nonzero().unbind(1)
|
| return x.nonzero().unbind(1)
|
| else:
|
| return x.nonzero(as_tuple=True)
|
|
|
|
|
| @torch.jit.script_if_tracing
|
| def move_device_like(src: torch.Tensor, dst: torch.Tensor) -> torch.Tensor:
|
| """
|
| Tracing friendly way to cast tensor to another tensor's device. Device will be treated
|
| as constant during tracing, scripting the casting process as whole can workaround this issue.
|
| """
|
| return src.to(dst.device)
|
|
|