| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | from typing import Optional, Sequence, Union |
| |
|
| | import torch |
| | import torch.nn as nn |
| |
|
| | from monai.networks.layers.factories import Conv, Pad, Pool |
| | from monai.networks.utils import icnr_init, pixelshuffle |
| | from monai.utils import UpsampleMode, ensure_tuple_rep |
| |
|
| |
|
| | class UpSample(nn.Module): |
| | """ |
| | Upsample with either kernel 1 conv + interpolation or transposed conv. |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | dimensions: int, |
| | in_channels: int, |
| | out_channels: Optional[int] = None, |
| | scale_factor: Union[Sequence[float], float] = 2, |
| | with_conv: bool = False, |
| | mode: Union[UpsampleMode, str] = UpsampleMode.LINEAR, |
| | align_corners: Optional[bool] = True, |
| | ) -> None: |
| | """ |
| | Args: |
| | dimensions: number of spatial dimensions of the input image. |
| | in_channels: number of channels of the input image. |
| | out_channels: number of channels of the output image. Defaults to `in_channels`. |
| | scale_factor: multiplier for spatial size. Has to match input size if it is a tuple. Defaults to 2. |
| | with_conv: whether to use a transposed convolution for upsampling. Defaults to False. |
| | mode: {``"nearest"``, ``"linear"``, ``"bilinear"``, ``"bicubic"``, ``"trilinear"``} |
| | If ends with ``"linear"`` will use ``spatial dims`` to determine the correct interpolation. |
| | This corresponds to linear, bilinear, trilinear for 1D, 2D, and 3D respectively. |
| | The interpolation mode. Defaults to ``"linear"``. |
| | See also: https://pytorch.org/docs/stable/nn.html#upsample |
| | align_corners: set the align_corners parameter of `torch.nn.Upsample`. Defaults to True. |
| | """ |
| | super().__init__() |
| | scale_factor_ = ensure_tuple_rep(scale_factor, dimensions) |
| | if not out_channels: |
| | out_channels = in_channels |
| | if not with_conv: |
| | mode = UpsampleMode(mode) |
| | linear_mode = [UpsampleMode.LINEAR, UpsampleMode.BILINEAR, UpsampleMode.TRILINEAR] |
| | if mode in linear_mode: |
| | mode = linear_mode[dimensions - 1] |
| | self.upsample = nn.Sequential( |
| | Conv[Conv.CONV, dimensions](in_channels=in_channels, out_channels=out_channels, kernel_size=1), |
| | nn.Upsample(scale_factor=scale_factor_, mode=mode.value, align_corners=align_corners), |
| | ) |
| | else: |
| | self.upsample = Conv[Conv.CONVTRANS, dimensions]( |
| | in_channels=in_channels, out_channels=out_channels, kernel_size=scale_factor_, stride=scale_factor_ |
| | ) |
| |
|
| | def forward(self, x: torch.Tensor) -> torch.Tensor: |
| | """ |
| | Args: |
| | x: Tensor in shape (batch, channel, spatial_1[, spatial_2, ...). |
| | """ |
| | return torch.as_tensor(self.upsample(x)) |
| |
|
| |
|
| | class SubpixelUpsample(nn.Module): |
| | """ |
| | Upsample via using a subpixel CNN. This module supports 1D, 2D and 3D input images. |
| | The module is consisted with two parts. First of all, a convolutional layer is employed |
| | to increase the number of channels into: ``in_channels * (scale_factor ** dimensions)``. |
| | Secondly, a pixel shuffle manipulation is utilized to aggregates the feature maps from |
| | low resolution space and build the super resolution space. |
| | The first part of the module is not fixed, a sequential layers can be used to replace the |
| | default single layer. |
| | |
| | See: Shi et al., 2016, "Real-Time Single Image and Video Super-Resolution |
| | Using a nEfficient Sub-Pixel Convolutional Neural Network." |
| | |
| | See: Aitken et al., 2017, "Checkerboard artifact free sub-pixel convolution". |
| | |
| | The idea comes from: |
| | https://arxiv.org/abs/1609.05158 |
| | |
| | The pixel shuffle mechanism refers to: |
| | https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/PixelShuffle.cpp |
| | and: |
| | https://github.com/pytorch/pytorch/pull/6340/files |
| | |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | dimensions: int, |
| | in_channels: int, |
| | scale_factor: int = 2, |
| | conv_block: Optional[nn.Module] = None, |
| | apply_pad_pool: bool = True, |
| | ) -> None: |
| | """ |
| | Args: |
| | dimensions: number of spatial dimensions of the input image. |
| | in_channels: number of channels of the input image. |
| | scale_factor: multiplier for spatial size. Defaults to 2. |
| | conv_block: a conv block to extract feature maps before upsampling. Defaults to None. |
| | When ``conv_block is None``, one reserved conv layer will be utilized. |
| | apply_pad_pool: if True the upsampled tensor is padded then average pooling is applied with a kernel the |
| | size of `scale_factor` with a stride of 1. This implements the nearest neighbour resize convolution |
| | component of subpixel convolutions described in Aitken et al. |
| | """ |
| | super().__init__() |
| |
|
| | if scale_factor <= 0: |
| | raise ValueError(f"The `scale_factor` multiplier must be an integer greater than 0, got {scale_factor}.") |
| |
|
| | self.dimensions = dimensions |
| | self.scale_factor = scale_factor |
| |
|
| | if conv_block is None: |
| | conv_out_channels = in_channels * (scale_factor ** dimensions) |
| | self.conv_block = Conv[Conv.CONV, dimensions]( |
| | in_channels=in_channels, |
| | out_channels=conv_out_channels, |
| | kernel_size=3, |
| | stride=1, |
| | padding=1, |
| | ) |
| |
|
| | icnr_init(self.conv_block, self.scale_factor) |
| | else: |
| | self.conv_block = conv_block |
| |
|
| | self.pad_pool: nn.Module = nn.Identity() |
| |
|
| | if apply_pad_pool: |
| | pool_type = Pool[Pool.AVG, self.dimensions] |
| | pad_type = Pad[Pad.CONSTANTPAD, self.dimensions] |
| |
|
| | self.pad_pool = nn.Sequential( |
| | pad_type(padding=(self.scale_factor - 1, 0) * self.dimensions, value=0.0), |
| | pool_type(kernel_size=self.scale_factor, stride=1), |
| | ) |
| |
|
| | def forward(self, x: torch.Tensor) -> torch.Tensor: |
| | """ |
| | Args: |
| | x: Tensor in shape (batch, channel, spatial_1[, spatial_2, ...). |
| | """ |
| | x = self.conv_block(x) |
| | x = pixelshuffle(x, self.dimensions, self.scale_factor) |
| | x = self.pad_pool(x) |
| | return x |
| |
|