|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
from collections.abc import Sequence |
|
|
|
|
|
import numpy as np |
|
|
import torch |
|
|
import torch.nn as nn |
|
|
|
|
|
from monai.networks.blocks.convolutions import Convolution |
|
|
from monai.networks.layers.factories import Act, Norm |
|
|
from monai.networks.layers.utils import get_act_layer, get_norm_layer |
|
|
|
|
|
|
|
|
class UnetResBlock(nn.Module): |
|
|
""" |
|
|
A skip-connection based module that can be used for DynUNet, based on: |
|
|
`Automated Design of Deep Learning Methods for Biomedical Image Segmentation <https://arxiv.org/abs/1904.08128>`_. |
|
|
`nnU-Net: Self-adapting Framework for U-Net-Based Medical Image Segmentation <https://arxiv.org/abs/1809.10486>`_. |
|
|
|
|
|
Args: |
|
|
spatial_dims: number of spatial dimensions. |
|
|
in_channels: number of input channels. |
|
|
out_channels: number of output channels. |
|
|
kernel_size: convolution kernel size. |
|
|
stride: convolution stride. |
|
|
norm_name: feature normalization type and arguments. |
|
|
act_name: activation layer type and arguments. |
|
|
dropout: dropout probability. |
|
|
|
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
spatial_dims: int, |
|
|
in_channels: int, |
|
|
out_channels: int, |
|
|
kernel_size: Sequence[int] | int, |
|
|
stride: Sequence[int] | int, |
|
|
norm_name: tuple | str, |
|
|
act_name: tuple | str = ("leakyrelu", {"inplace": True, "negative_slope": 0.01}), |
|
|
dropout: tuple | str | float | None = None, |
|
|
): |
|
|
super().__init__() |
|
|
self.conv1 = get_conv_layer( |
|
|
spatial_dims, |
|
|
in_channels, |
|
|
out_channels, |
|
|
kernel_size=kernel_size, |
|
|
stride=stride, |
|
|
dropout=dropout, |
|
|
act=None, |
|
|
norm=None, |
|
|
conv_only=False, |
|
|
) |
|
|
self.conv2 = get_conv_layer( |
|
|
spatial_dims, |
|
|
out_channels, |
|
|
out_channels, |
|
|
kernel_size=kernel_size, |
|
|
stride=1, |
|
|
dropout=dropout, |
|
|
act=None, |
|
|
norm=None, |
|
|
conv_only=False, |
|
|
) |
|
|
self.lrelu = get_act_layer(name=act_name) |
|
|
self.norm1 = get_norm_layer(name=norm_name, spatial_dims=spatial_dims, channels=out_channels) |
|
|
self.norm2 = get_norm_layer(name=norm_name, spatial_dims=spatial_dims, channels=out_channels) |
|
|
self.downsample = in_channels != out_channels |
|
|
stride_np = np.atleast_1d(stride) |
|
|
if not np.all(stride_np == 1): |
|
|
self.downsample = True |
|
|
if self.downsample: |
|
|
self.conv3 = get_conv_layer( |
|
|
spatial_dims, |
|
|
in_channels, |
|
|
out_channels, |
|
|
kernel_size=1, |
|
|
stride=stride, |
|
|
dropout=dropout, |
|
|
act=None, |
|
|
norm=None, |
|
|
conv_only=False, |
|
|
) |
|
|
self.norm3 = get_norm_layer(name=norm_name, spatial_dims=spatial_dims, channels=out_channels) |
|
|
|
|
|
def forward(self, inp): |
|
|
residual = inp |
|
|
out = self.conv1(inp) |
|
|
out = self.norm1(out) |
|
|
out = self.lrelu(out) |
|
|
out = self.conv2(out) |
|
|
out = self.norm2(out) |
|
|
if hasattr(self, "conv3"): |
|
|
residual = self.conv3(residual) |
|
|
if hasattr(self, "norm3"): |
|
|
residual = self.norm3(residual) |
|
|
out += residual |
|
|
out = self.lrelu(out) |
|
|
return out |
|
|
|
|
|
|
|
|
class UnetBasicBlock(nn.Module): |
|
|
""" |
|
|
A CNN module that can be used for DynUNet, based on: |
|
|
`Automated Design of Deep Learning Methods for Biomedical Image Segmentation <https://arxiv.org/abs/1904.08128>`_. |
|
|
`nnU-Net: Self-adapting Framework for U-Net-Based Medical Image Segmentation <https://arxiv.org/abs/1809.10486>`_. |
|
|
|
|
|
Args: |
|
|
spatial_dims: number of spatial dimensions. |
|
|
in_channels: number of input channels. |
|
|
out_channels: number of output channels. |
|
|
kernel_size: convolution kernel size. |
|
|
stride: convolution stride. |
|
|
norm_name: feature normalization type and arguments. |
|
|
act_name: activation layer type and arguments. |
|
|
dropout: dropout probability. |
|
|
|
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
spatial_dims: int, |
|
|
in_channels: int, |
|
|
out_channels: int, |
|
|
kernel_size: Sequence[int] | int, |
|
|
stride: Sequence[int] | int, |
|
|
norm_name: tuple | str, |
|
|
act_name: tuple | str = ("leakyrelu", {"inplace": True, "negative_slope": 0.01}), |
|
|
dropout: tuple | str | float | None = None, |
|
|
): |
|
|
super().__init__() |
|
|
self.conv1 = get_conv_layer( |
|
|
spatial_dims, |
|
|
in_channels, |
|
|
out_channels, |
|
|
kernel_size=kernel_size, |
|
|
stride=stride, |
|
|
dropout=dropout, |
|
|
act=None, |
|
|
norm=None, |
|
|
conv_only=False, |
|
|
) |
|
|
self.conv2 = get_conv_layer( |
|
|
spatial_dims, |
|
|
out_channels, |
|
|
out_channels, |
|
|
kernel_size=kernel_size, |
|
|
stride=1, |
|
|
dropout=dropout, |
|
|
act=None, |
|
|
norm=None, |
|
|
conv_only=False, |
|
|
) |
|
|
self.lrelu = get_act_layer(name=act_name) |
|
|
self.norm1 = get_norm_layer(name=norm_name, spatial_dims=spatial_dims, channels=out_channels) |
|
|
self.norm2 = get_norm_layer(name=norm_name, spatial_dims=spatial_dims, channels=out_channels) |
|
|
|
|
|
def forward(self, inp): |
|
|
out = self.conv1(inp) |
|
|
out = self.norm1(out) |
|
|
out = self.lrelu(out) |
|
|
out = self.conv2(out) |
|
|
out = self.norm2(out) |
|
|
out = self.lrelu(out) |
|
|
return out |
|
|
|
|
|
|
|
|
class UnetUpBlock(nn.Module): |
|
|
""" |
|
|
An upsampling module that can be used for DynUNet, based on: |
|
|
`Automated Design of Deep Learning Methods for Biomedical Image Segmentation <https://arxiv.org/abs/1904.08128>`_. |
|
|
`nnU-Net: Self-adapting Framework for U-Net-Based Medical Image Segmentation <https://arxiv.org/abs/1809.10486>`_. |
|
|
|
|
|
Args: |
|
|
spatial_dims: number of spatial dimensions. |
|
|
in_channels: number of input channels. |
|
|
out_channels: number of output channels. |
|
|
kernel_size: convolution kernel size. |
|
|
stride: convolution stride. |
|
|
upsample_kernel_size: convolution kernel size for transposed convolution layers. |
|
|
norm_name: feature normalization type and arguments. |
|
|
act_name: activation layer type and arguments. |
|
|
dropout: dropout probability. |
|
|
trans_bias: transposed convolution bias. |
|
|
|
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
spatial_dims: int, |
|
|
in_channels: int, |
|
|
out_channels: int, |
|
|
kernel_size: Sequence[int] | int, |
|
|
stride: Sequence[int] | int, |
|
|
upsample_kernel_size: Sequence[int] | int, |
|
|
norm_name: tuple | str, |
|
|
act_name: tuple | str = ("leakyrelu", {"inplace": True, "negative_slope": 0.01}), |
|
|
dropout: tuple | str | float | None = None, |
|
|
trans_bias: bool = False, |
|
|
): |
|
|
super().__init__() |
|
|
upsample_stride = upsample_kernel_size |
|
|
self.transp_conv = get_conv_layer( |
|
|
spatial_dims, |
|
|
in_channels, |
|
|
out_channels, |
|
|
kernel_size=upsample_kernel_size, |
|
|
stride=upsample_stride, |
|
|
dropout=dropout, |
|
|
bias=trans_bias, |
|
|
act=None, |
|
|
norm=None, |
|
|
conv_only=False, |
|
|
is_transposed=True, |
|
|
) |
|
|
self.conv_block = UnetBasicBlock( |
|
|
spatial_dims, |
|
|
out_channels + out_channels, |
|
|
out_channels, |
|
|
kernel_size=kernel_size, |
|
|
stride=1, |
|
|
dropout=dropout, |
|
|
norm_name=norm_name, |
|
|
act_name=act_name, |
|
|
) |
|
|
|
|
|
def forward(self, inp, skip): |
|
|
|
|
|
out = self.transp_conv(inp) |
|
|
out = torch.cat((out, skip), dim=1) |
|
|
out = self.conv_block(out) |
|
|
return out |
|
|
|
|
|
|
|
|
class UnetOutBlock(nn.Module): |
|
|
|
|
|
def __init__( |
|
|
self, spatial_dims: int, in_channels: int, out_channels: int, dropout: tuple | str | float | None = None |
|
|
): |
|
|
super().__init__() |
|
|
self.conv = get_conv_layer( |
|
|
spatial_dims, |
|
|
in_channels, |
|
|
out_channels, |
|
|
kernel_size=1, |
|
|
stride=1, |
|
|
dropout=dropout, |
|
|
bias=True, |
|
|
act=None, |
|
|
norm=None, |
|
|
conv_only=False, |
|
|
) |
|
|
|
|
|
def forward(self, inp): |
|
|
return self.conv(inp) |
|
|
|
|
|
|
|
|
def get_conv_layer( |
|
|
spatial_dims: int, |
|
|
in_channels: int, |
|
|
out_channels: int, |
|
|
kernel_size: Sequence[int] | int = 3, |
|
|
stride: Sequence[int] | int = 1, |
|
|
act: tuple | str | None = Act.PRELU, |
|
|
norm: tuple | str | None = Norm.INSTANCE, |
|
|
dropout: tuple | str | float | None = None, |
|
|
bias: bool = False, |
|
|
conv_only: bool = True, |
|
|
is_transposed: bool = False, |
|
|
): |
|
|
padding = get_padding(kernel_size, stride) |
|
|
output_padding = None |
|
|
if is_transposed: |
|
|
output_padding = get_output_padding(kernel_size, stride, padding) |
|
|
return Convolution( |
|
|
spatial_dims, |
|
|
in_channels, |
|
|
out_channels, |
|
|
strides=stride, |
|
|
kernel_size=kernel_size, |
|
|
act=act, |
|
|
norm=norm, |
|
|
dropout=dropout, |
|
|
bias=bias, |
|
|
conv_only=conv_only, |
|
|
is_transposed=is_transposed, |
|
|
padding=padding, |
|
|
output_padding=output_padding, |
|
|
) |
|
|
|
|
|
|
|
|
def get_padding(kernel_size: Sequence[int] | int, stride: Sequence[int] | int) -> tuple[int, ...] | int: |
|
|
kernel_size_np = np.atleast_1d(kernel_size) |
|
|
stride_np = np.atleast_1d(stride) |
|
|
padding_np = (kernel_size_np - stride_np + 1) / 2 |
|
|
if np.min(padding_np) < 0: |
|
|
raise AssertionError("padding value should not be negative, please change the kernel size and/or stride.") |
|
|
padding = tuple(int(p) for p in padding_np) |
|
|
|
|
|
return padding if len(padding) > 1 else padding[0] |
|
|
|
|
|
|
|
|
def get_output_padding( |
|
|
kernel_size: Sequence[int] | int, stride: Sequence[int] | int, padding: Sequence[int] | int |
|
|
) -> tuple[int, ...] | int: |
|
|
kernel_size_np = np.atleast_1d(kernel_size) |
|
|
stride_np = np.atleast_1d(stride) |
|
|
padding_np = np.atleast_1d(padding) |
|
|
|
|
|
out_padding_np = 2 * padding_np + stride_np - kernel_size_np |
|
|
if np.min(out_padding_np) < 0: |
|
|
raise AssertionError("out_padding value should not be negative, please change the kernel size and/or stride.") |
|
|
out_padding = tuple(int(p) for p in out_padding_np) |
|
|
|
|
|
return out_padding if len(out_padding) > 1 else out_padding[0] |
|
|
|