|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
from collections.abc import Sequence |
|
|
|
|
|
import numpy as np |
|
|
import torch |
|
|
|
|
|
__all__ = ["same_padding", "stride_minus_kernel_padding", "calculate_out_shape", "gaussian_1d", "polyval"] |
|
|
|
|
|
|
|
|
def same_padding(kernel_size: Sequence[int] | int, dilation: Sequence[int] | int = 1) -> tuple[int, ...] | int: |
|
|
""" |
|
|
Return the padding value needed to ensure a convolution using the given kernel size produces an output of the same |
|
|
shape as the input for a stride of 1, otherwise ensure a shape of the input divided by the stride rounded down. |
|
|
|
|
|
Raises: |
|
|
NotImplementedError: When ``np.any((kernel_size - 1) * dilation % 2 == 1)``. |
|
|
|
|
|
""" |
|
|
|
|
|
kernel_size_np = np.atleast_1d(kernel_size) |
|
|
dilation_np = np.atleast_1d(dilation) |
|
|
|
|
|
if np.any((kernel_size_np - 1) * dilation % 2 == 1): |
|
|
raise NotImplementedError( |
|
|
f"Same padding not available for kernel_size={kernel_size_np} and dilation={dilation_np}." |
|
|
) |
|
|
|
|
|
padding_np = (kernel_size_np - 1) / 2 * dilation_np |
|
|
padding = tuple(int(p) for p in padding_np) |
|
|
|
|
|
return padding if len(padding) > 1 else padding[0] |
|
|
|
|
|
|
|
|
def stride_minus_kernel_padding(kernel_size: Sequence[int] | int, stride: Sequence[int] | int) -> tuple[int, ...] | int: |
|
|
kernel_size_np = np.atleast_1d(kernel_size) |
|
|
stride_np = np.atleast_1d(stride) |
|
|
|
|
|
out_padding_np = stride_np - kernel_size_np |
|
|
out_padding = tuple(int(p) for p in out_padding_np) |
|
|
|
|
|
return out_padding if len(out_padding) > 1 else out_padding[0] |
|
|
|
|
|
|
|
|
def calculate_out_shape( |
|
|
in_shape: Sequence[int] | int | np.ndarray, |
|
|
kernel_size: Sequence[int] | int, |
|
|
stride: Sequence[int] | int, |
|
|
padding: Sequence[int] | int, |
|
|
) -> tuple[int, ...] | int: |
|
|
""" |
|
|
Calculate the output tensor shape when applying a convolution to a tensor of shape `inShape` with kernel size |
|
|
`kernel_size`, stride value `stride`, and input padding value `padding`. All arguments can be scalars or multiple |
|
|
values, return value is a scalar if all inputs are scalars. |
|
|
""" |
|
|
in_shape_np = np.atleast_1d(in_shape) |
|
|
kernel_size_np = np.atleast_1d(kernel_size) |
|
|
stride_np = np.atleast_1d(stride) |
|
|
padding_np = np.atleast_1d(padding) |
|
|
|
|
|
out_shape_np = ((in_shape_np - kernel_size_np + padding_np + padding_np) // stride_np) + 1 |
|
|
out_shape = tuple(int(s) for s in out_shape_np) |
|
|
|
|
|
return out_shape |
|
|
|
|
|
|
|
|
def gaussian_1d( |
|
|
sigma: torch.Tensor, truncated: float = 4.0, approx: str = "erf", normalize: bool = False |
|
|
) -> torch.Tensor: |
|
|
""" |
|
|
one dimensional Gaussian kernel. |
|
|
|
|
|
Args: |
|
|
sigma: std of the kernel |
|
|
truncated: tail length |
|
|
approx: discrete Gaussian kernel type, available options are "erf", "sampled", and "scalespace". |
|
|
|
|
|
- ``erf`` approximation interpolates the error function; |
|
|
- ``sampled`` uses a sampled Gaussian kernel; |
|
|
- ``scalespace`` corresponds to |
|
|
https://en.wikipedia.org/wiki/Scale_space_implementation#The_discrete_Gaussian_kernel |
|
|
based on the modified Bessel functions. |
|
|
|
|
|
normalize: whether to normalize the kernel with `kernel.sum()`. |
|
|
|
|
|
Raises: |
|
|
ValueError: When ``truncated`` is non-positive. |
|
|
|
|
|
Returns: |
|
|
1D torch tensor |
|
|
|
|
|
""" |
|
|
sigma = torch.as_tensor(sigma, dtype=torch.float, device=sigma.device if isinstance(sigma, torch.Tensor) else None) |
|
|
device = sigma.device |
|
|
if truncated <= 0.0: |
|
|
raise ValueError(f"truncated must be positive, got {truncated}.") |
|
|
tail = int(max(float(sigma) * truncated, 0.5) + 0.5) |
|
|
if approx.lower() == "erf": |
|
|
x = torch.arange(-tail, tail + 1, dtype=torch.float, device=device) |
|
|
t = 0.70710678 / torch.abs(sigma) |
|
|
out = 0.5 * ((t * (x + 0.5)).erf() - (t * (x - 0.5)).erf()) |
|
|
out = out.clamp(min=0) |
|
|
elif approx.lower() == "sampled": |
|
|
x = torch.arange(-tail, tail + 1, dtype=torch.float, device=sigma.device) |
|
|
out = torch.exp(-0.5 / (sigma * sigma) * x**2) |
|
|
if not normalize: |
|
|
out = out / (2.5066282 * sigma) |
|
|
elif approx.lower() == "scalespace": |
|
|
sigma2 = sigma * sigma |
|
|
out_pos: list[torch.Tensor | None] = [None] * (tail + 1) |
|
|
out_pos[0] = _modified_bessel_0(sigma2) |
|
|
out_pos[1] = _modified_bessel_1(sigma2) |
|
|
for k in range(2, len(out_pos)): |
|
|
out_pos[k] = _modified_bessel_i(k, sigma2) |
|
|
out = out_pos[:0:-1] |
|
|
out.extend(out_pos) |
|
|
out = torch.stack(out) * torch.exp(-sigma2) |
|
|
else: |
|
|
raise NotImplementedError(f"Unsupported option: approx='{approx}'.") |
|
|
return out / out.sum() if normalize else out |
|
|
|
|
|
|
|
|
def polyval(coef, x) -> torch.Tensor: |
|
|
""" |
|
|
Evaluates the polynomial defined by `coef` at `x`. |
|
|
|
|
|
For a 1D sequence of coef (length n), evaluate:: |
|
|
|
|
|
y = coef[n-1] + x * (coef[n-2] + ... + x * (coef[1] + x * coef[0])) |
|
|
|
|
|
Args: |
|
|
coef: a sequence of floats representing the coefficients of the polynomial |
|
|
x: float or a sequence of floats representing the variable of the polynomial |
|
|
|
|
|
Returns: |
|
|
1D torch tensor |
|
|
""" |
|
|
device = x.device if isinstance(x, torch.Tensor) else None |
|
|
coef = torch.as_tensor(coef, dtype=torch.float, device=device) |
|
|
if coef.ndim == 0 or (len(coef) < 1): |
|
|
return torch.zeros(x.shape) |
|
|
x = torch.as_tensor(x, dtype=torch.float, device=device) |
|
|
ans = coef[0] |
|
|
for c in coef[1:]: |
|
|
ans = ans * x + c |
|
|
return ans |
|
|
|
|
|
|
|
|
def _modified_bessel_0(x: torch.Tensor) -> torch.Tensor: |
|
|
x = torch.as_tensor(x, dtype=torch.float, device=x.device if isinstance(x, torch.Tensor) else None) |
|
|
if torch.abs(x) < 3.75: |
|
|
y = x * x / 14.0625 |
|
|
return polyval([0.45813e-2, 0.360768e-1, 0.2659732, 1.2067492, 3.0899424, 3.5156229, 1.0], y) |
|
|
ax = torch.abs(x) |
|
|
y = 3.75 / ax |
|
|
_coef = [ |
|
|
0.392377e-2, |
|
|
-0.1647633e-1, |
|
|
0.2635537e-1, |
|
|
-0.2057706e-1, |
|
|
0.916281e-2, |
|
|
-0.157565e-2, |
|
|
0.225319e-2, |
|
|
0.1328592e-1, |
|
|
0.39894228, |
|
|
] |
|
|
return polyval(_coef, y) * torch.exp(ax) / torch.sqrt(ax) |
|
|
|
|
|
|
|
|
def _modified_bessel_1(x: torch.Tensor) -> torch.Tensor: |
|
|
x = torch.as_tensor(x, dtype=torch.float, device=x.device if isinstance(x, torch.Tensor) else None) |
|
|
if torch.abs(x) < 3.75: |
|
|
y = x * x / 14.0625 |
|
|
_coef = [0.32411e-3, 0.301532e-2, 0.2658733e-1, 0.15084934, 0.51498869, 0.87890594, 0.5] |
|
|
return torch.abs(x) * polyval(_coef, y) |
|
|
ax = torch.abs(x) |
|
|
y = 3.75 / ax |
|
|
_coef = [ |
|
|
-0.420059e-2, |
|
|
0.1787654e-1, |
|
|
-0.2895312e-1, |
|
|
0.2282967e-1, |
|
|
-0.1031555e-1, |
|
|
0.163801e-2, |
|
|
-0.362018e-2, |
|
|
-0.3988024e-1, |
|
|
0.39894228, |
|
|
] |
|
|
ans = polyval(_coef, y) * torch.exp(ax) / torch.sqrt(ax) |
|
|
return -ans if x < 0.0 else ans |
|
|
|
|
|
|
|
|
def _modified_bessel_i(n: int, x: torch.Tensor) -> torch.Tensor: |
|
|
if n < 2: |
|
|
raise ValueError(f"n must be greater than 1, got n={n}.") |
|
|
x = torch.as_tensor(x, dtype=torch.float, device=x.device if isinstance(x, torch.Tensor) else None) |
|
|
if x == 0.0: |
|
|
return x |
|
|
device = x.device |
|
|
tox = 2.0 / torch.abs(x) |
|
|
ans, bip, bi = torch.tensor(0.0, device=device), torch.tensor(0.0, device=device), torch.tensor(1.0, device=device) |
|
|
m = int(2 * (n + np.floor(np.sqrt(40.0 * n)))) |
|
|
for j in range(m, 0, -1): |
|
|
bim = bip + float(j) * tox * bi |
|
|
bip = bi |
|
|
bi = bim |
|
|
if abs(bi) > 1.0e10: |
|
|
ans = ans * 1.0e-10 |
|
|
bi = bi * 1.0e-10 |
|
|
bip = bip * 1.0e-10 |
|
|
if j == n: |
|
|
ans = bip |
|
|
ans = ans * _modified_bessel_0(x) / bi |
|
|
return -ans if x < 0.0 and (n % 2) == 1 else ans |
|
|
|