|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
from collections.abc import Sequence |
|
|
|
|
|
import torch |
|
|
import torch.nn as nn |
|
|
|
|
|
from monai.networks.blocks import ADN, Convolution |
|
|
from monai.networks.layers.simplelayers import ChannelPad |
|
|
from monai.utils import ChannelMatching |
|
|
|
|
|
__all__ = ["HighResBlock", "HighResNet"] |
|
|
|
|
|
DEFAULT_LAYER_PARAMS_3D = ( |
|
|
|
|
|
{"name": "conv_0", "n_features": 16, "kernel_size": 3}, |
|
|
|
|
|
{"name": "res_1", "n_features": 16, "kernels": (3, 3), "repeat": 3}, |
|
|
{"name": "res_2", "n_features": 32, "kernels": (3, 3), "repeat": 3}, |
|
|
{"name": "res_3", "n_features": 64, "kernels": (3, 3), "repeat": 3}, |
|
|
|
|
|
{"name": "conv_1", "n_features": 80, "kernel_size": 1}, |
|
|
{"name": "conv_2", "kernel_size": 1}, |
|
|
) |
|
|
|
|
|
|
|
|
class HighResBlock(nn.Module): |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
spatial_dims: int, |
|
|
in_channels: int, |
|
|
out_channels: int, |
|
|
kernels: Sequence[int] = (3, 3), |
|
|
dilation: Sequence[int] | int = 1, |
|
|
norm_type: tuple | str = ("batch", {"affine": True}), |
|
|
acti_type: tuple | str = ("relu", {"inplace": True}), |
|
|
bias: bool = False, |
|
|
channel_matching: ChannelMatching | str = ChannelMatching.PAD, |
|
|
) -> None: |
|
|
""" |
|
|
Args: |
|
|
spatial_dims: number of spatial dimensions of the input image. |
|
|
in_channels: number of input channels. |
|
|
out_channels: number of output channels. |
|
|
kernels: each integer k in `kernels` corresponds to a convolution layer with kernel size k. |
|
|
dilation: spacing between kernel elements. |
|
|
norm_type: feature normalization type and arguments. |
|
|
Defaults to ``("batch", {"affine": True})``. |
|
|
acti_type: {``"relu"``, ``"prelu"``, ``"relu6"``} |
|
|
Non-linear activation using ReLU or PReLU. Defaults to ``"relu"``. |
|
|
bias: whether to have a bias term in convolution blocks. Defaults to False. |
|
|
According to `Performance Tuning Guide <https://pytorch.org/tutorials/recipes/recipes/tuning_guide.html>`_, |
|
|
if a conv layer is directly followed by a batch norm layer, bias should be False. |
|
|
channel_matching: {``"pad"``, ``"project"``} |
|
|
Specifies handling residual branch and conv branch channel mismatches. Defaults to ``"pad"``. |
|
|
|
|
|
- ``"pad"``: with zero padding. |
|
|
- ``"project"``: with a trainable conv with kernel size one. |
|
|
|
|
|
Raises: |
|
|
ValueError: When ``channel_matching=pad`` and ``in_channels > out_channels``. Incompatible values. |
|
|
|
|
|
""" |
|
|
super().__init__() |
|
|
self.chn_pad = ChannelPad( |
|
|
spatial_dims=spatial_dims, in_channels=in_channels, out_channels=out_channels, mode=channel_matching |
|
|
) |
|
|
|
|
|
layers = nn.ModuleList() |
|
|
_in_chns, _out_chns = in_channels, out_channels |
|
|
|
|
|
for kernel_size in kernels: |
|
|
layers.append( |
|
|
ADN(ordering="NA", in_channels=_in_chns, act=acti_type, norm=norm_type, norm_dim=spatial_dims) |
|
|
) |
|
|
layers.append( |
|
|
Convolution( |
|
|
spatial_dims=spatial_dims, |
|
|
in_channels=_in_chns, |
|
|
out_channels=_out_chns, |
|
|
kernel_size=kernel_size, |
|
|
dilation=dilation, |
|
|
bias=bias, |
|
|
conv_only=True, |
|
|
) |
|
|
) |
|
|
_in_chns = _out_chns |
|
|
|
|
|
self.layers = nn.Sequential(*layers) |
|
|
|
|
|
def forward(self, x: torch.Tensor) -> torch.Tensor: |
|
|
x_conv: torch.Tensor = self.layers(x) |
|
|
return x_conv + torch.as_tensor(self.chn_pad(x)) |
|
|
|
|
|
|
|
|
class HighResNet(nn.Module): |
|
|
""" |
|
|
Reimplementation of highres3dnet based on |
|
|
Li et al., "On the compactness, efficiency, and representation of 3D |
|
|
convolutional networks: Brain parcellation as a pretext task", IPMI '17 |
|
|
|
|
|
Adapted from: |
|
|
https://github.com/NifTK/NiftyNet/blob/v0.6.0/niftynet/network/highres3dnet.py |
|
|
https://github.com/fepegar/highresnet |
|
|
|
|
|
Args: |
|
|
spatial_dims: number of spatial dimensions of the input image. |
|
|
in_channels: number of input channels. |
|
|
out_channels: number of output channels. |
|
|
norm_type: feature normalization type and arguments. |
|
|
Defaults to ``("batch", {"affine": True})``. |
|
|
acti_type: activation type and arguments. |
|
|
Defaults to ``("relu", {"inplace": True})``. |
|
|
dropout_prob: probability of the feature map to be zeroed |
|
|
(only applies to the penultimate conv layer). |
|
|
bias: whether to have a bias term in convolution blocks. Defaults to False. |
|
|
According to `Performance Tuning Guide <https://pytorch.org/tutorials/recipes/recipes/tuning_guide.html>`_, |
|
|
if a conv layer is directly followed by a batch norm layer, bias should be False. |
|
|
layer_params: specifying key parameters of each layer/block. |
|
|
channel_matching: {``"pad"``, ``"project"``} |
|
|
Specifies handling residual branch and conv branch channel mismatches. Defaults to ``"pad"``. |
|
|
|
|
|
- ``"pad"``: with zero padding. |
|
|
- ``"project"``: with a trainable conv with kernel size one. |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
spatial_dims: int = 3, |
|
|
in_channels: int = 1, |
|
|
out_channels: int = 1, |
|
|
norm_type: str | tuple = ("batch", {"affine": True}), |
|
|
acti_type: str | tuple = ("relu", {"inplace": True}), |
|
|
dropout_prob: tuple | str | float | None = 0.0, |
|
|
bias: bool = False, |
|
|
layer_params: Sequence[dict] = DEFAULT_LAYER_PARAMS_3D, |
|
|
channel_matching: ChannelMatching | str = ChannelMatching.PAD, |
|
|
) -> None: |
|
|
super().__init__() |
|
|
blocks = nn.ModuleList() |
|
|
|
|
|
|
|
|
params = layer_params[0] |
|
|
_in_chns, _out_chns = in_channels, params["n_features"] |
|
|
blocks.append( |
|
|
Convolution( |
|
|
spatial_dims=spatial_dims, |
|
|
in_channels=_in_chns, |
|
|
out_channels=_out_chns, |
|
|
kernel_size=params["kernel_size"], |
|
|
adn_ordering="NA", |
|
|
act=acti_type, |
|
|
norm=norm_type, |
|
|
bias=bias, |
|
|
) |
|
|
) |
|
|
|
|
|
|
|
|
for idx, params in enumerate(layer_params[1:-2]): |
|
|
_in_chns, _out_chns = _out_chns, params["n_features"] |
|
|
_dilation = 2**idx |
|
|
for _ in range(params["repeat"]): |
|
|
blocks.append( |
|
|
HighResBlock( |
|
|
spatial_dims=spatial_dims, |
|
|
in_channels=_in_chns, |
|
|
out_channels=_out_chns, |
|
|
kernels=params["kernels"], |
|
|
dilation=_dilation, |
|
|
norm_type=norm_type, |
|
|
acti_type=acti_type, |
|
|
bias=bias, |
|
|
channel_matching=channel_matching, |
|
|
) |
|
|
) |
|
|
_in_chns = _out_chns |
|
|
|
|
|
|
|
|
params = layer_params[-2] |
|
|
_in_chns, _out_chns = _out_chns, params["n_features"] |
|
|
blocks.append( |
|
|
Convolution( |
|
|
spatial_dims=spatial_dims, |
|
|
in_channels=_in_chns, |
|
|
out_channels=_out_chns, |
|
|
kernel_size=params["kernel_size"], |
|
|
adn_ordering="NAD", |
|
|
act=acti_type, |
|
|
norm=norm_type, |
|
|
bias=bias, |
|
|
dropout=dropout_prob, |
|
|
) |
|
|
) |
|
|
|
|
|
params = layer_params[-1] |
|
|
_in_chns = _out_chns |
|
|
blocks.append( |
|
|
Convolution( |
|
|
spatial_dims=spatial_dims, |
|
|
in_channels=_in_chns, |
|
|
out_channels=out_channels, |
|
|
kernel_size=params["kernel_size"], |
|
|
adn_ordering="NAD", |
|
|
act=acti_type, |
|
|
norm=norm_type, |
|
|
bias=bias, |
|
|
dropout=dropout_prob, |
|
|
) |
|
|
) |
|
|
|
|
|
self.blocks = nn.Sequential(*blocks) |
|
|
|
|
|
def forward(self, x: torch.Tensor) -> torch.Tensor: |
|
|
return torch.as_tensor(self.blocks(x)) |
|
|
|