|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import torch |
|
|
import torch.nn as nn |
|
|
import torch.nn.functional as F |
|
|
|
|
|
from monai.networks.blocks.convolutions import Convolution |
|
|
from monai.networks.blocks.upsample import UpSample |
|
|
from monai.networks.layers.factories import Act, Conv, Norm |
|
|
from monai.utils import optional_import |
|
|
|
|
|
models, _ = optional_import("torchvision", name="models") |
|
|
|
|
|
|
|
|
class GCN(nn.Module): |
|
|
""" |
|
|
The Global Convolutional Network module using large 1D |
|
|
Kx1 and 1xK kernels to represent 2D kernels. |
|
|
""" |
|
|
|
|
|
def __init__(self, inplanes: int, planes: int, ks: int = 7): |
|
|
""" |
|
|
Args: |
|
|
inplanes: number of input channels. |
|
|
planes: number of output channels. |
|
|
ks: kernel size for one dimension. Defaults to 7. |
|
|
""" |
|
|
super().__init__() |
|
|
|
|
|
conv2d_type: type[nn.Conv2d] = Conv[Conv.CONV, 2] |
|
|
self.conv_l1 = conv2d_type(in_channels=inplanes, out_channels=planes, kernel_size=(ks, 1), padding=(ks // 2, 0)) |
|
|
self.conv_l2 = conv2d_type(in_channels=planes, out_channels=planes, kernel_size=(1, ks), padding=(0, ks // 2)) |
|
|
self.conv_r1 = conv2d_type(in_channels=inplanes, out_channels=planes, kernel_size=(1, ks), padding=(0, ks // 2)) |
|
|
self.conv_r2 = conv2d_type(in_channels=planes, out_channels=planes, kernel_size=(ks, 1), padding=(ks // 2, 0)) |
|
|
|
|
|
def forward(self, x: torch.Tensor) -> torch.Tensor: |
|
|
""" |
|
|
Args: |
|
|
x: in shape (batch, inplanes, spatial_1, spatial_2). |
|
|
""" |
|
|
x_l = self.conv_l1(x) |
|
|
x_l = self.conv_l2(x_l) |
|
|
x_r = self.conv_r1(x) |
|
|
x_r = self.conv_r2(x_r) |
|
|
x = x_l + x_r |
|
|
return x |
|
|
|
|
|
|
|
|
class Refine(nn.Module): |
|
|
""" |
|
|
Simple residual block to refine the details of the activation maps. |
|
|
""" |
|
|
|
|
|
def __init__(self, planes: int): |
|
|
""" |
|
|
Args: |
|
|
planes: number of input channels. |
|
|
""" |
|
|
super().__init__() |
|
|
|
|
|
relu_type: type[nn.ReLU] = Act[Act.RELU] |
|
|
conv2d_type: type[nn.Conv2d] = Conv[Conv.CONV, 2] |
|
|
norm2d_type: type[nn.BatchNorm2d] = Norm[Norm.BATCH, 2] |
|
|
|
|
|
self.bn = norm2d_type(num_features=planes) |
|
|
self.relu = relu_type(inplace=True) |
|
|
self.conv1 = conv2d_type(in_channels=planes, out_channels=planes, kernel_size=3, padding=1) |
|
|
self.conv2 = conv2d_type(in_channels=planes, out_channels=planes, kernel_size=3, padding=1) |
|
|
|
|
|
def forward(self, x: torch.Tensor) -> torch.Tensor: |
|
|
""" |
|
|
Args: |
|
|
x: in shape (batch, planes, spatial_1, spatial_2). |
|
|
""" |
|
|
residual = x |
|
|
x = self.bn(x) |
|
|
x = self.relu(x) |
|
|
x = self.conv1(x) |
|
|
x = self.bn(x) |
|
|
x = self.relu(x) |
|
|
x = self.conv2(x) |
|
|
|
|
|
return residual + x |
|
|
|
|
|
|
|
|
class FCN(nn.Module): |
|
|
""" |
|
|
2D FCN network with 3 input channels. The small decoder is built |
|
|
with the GCN and Refine modules. |
|
|
The code is adapted from `lsqshr's official 2D code <https://github.com/lsqshr/AH-Net/blob/master/net2d.py>`_. |
|
|
|
|
|
Args: |
|
|
out_channels: number of output channels. Defaults to 1. |
|
|
upsample_mode: [``"transpose"``, ``"bilinear"``] |
|
|
The mode of upsampling manipulations. |
|
|
Using the second mode cannot guarantee the model's reproducibility. Defaults to ``bilinear``. |
|
|
|
|
|
- ``transpose``, uses transposed convolution layers. |
|
|
- ``bilinear``, uses bilinear interpolation. |
|
|
|
|
|
pretrained: If True, returns a model pre-trained on ImageNet |
|
|
progress: If True, displays a progress bar of the download to stderr. |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, out_channels: int = 1, upsample_mode: str = "bilinear", pretrained: bool = True, progress: bool = True |
|
|
): |
|
|
super().__init__() |
|
|
|
|
|
conv2d_type: type[nn.Conv2d] = Conv[Conv.CONV, 2] |
|
|
|
|
|
self.upsample_mode = upsample_mode |
|
|
self.conv2d_type = conv2d_type |
|
|
self.out_channels = out_channels |
|
|
resnet = models.resnet50(pretrained=pretrained, progress=progress) |
|
|
|
|
|
self.conv1 = resnet.conv1 |
|
|
self.bn0 = resnet.bn1 |
|
|
self.relu = resnet.relu |
|
|
self.maxpool = resnet.maxpool |
|
|
|
|
|
self.layer1 = resnet.layer1 |
|
|
self.layer2 = resnet.layer2 |
|
|
self.layer3 = resnet.layer3 |
|
|
self.layer4 = resnet.layer4 |
|
|
|
|
|
self.gcn1 = GCN(2048, self.out_channels) |
|
|
self.gcn2 = GCN(1024, self.out_channels) |
|
|
self.gcn3 = GCN(512, self.out_channels) |
|
|
self.gcn4 = GCN(64, self.out_channels) |
|
|
self.gcn5 = GCN(64, self.out_channels) |
|
|
|
|
|
self.refine1 = Refine(self.out_channels) |
|
|
self.refine2 = Refine(self.out_channels) |
|
|
self.refine3 = Refine(self.out_channels) |
|
|
self.refine4 = Refine(self.out_channels) |
|
|
self.refine5 = Refine(self.out_channels) |
|
|
self.refine6 = Refine(self.out_channels) |
|
|
self.refine7 = Refine(self.out_channels) |
|
|
self.refine8 = Refine(self.out_channels) |
|
|
self.refine9 = Refine(self.out_channels) |
|
|
self.refine10 = Refine(self.out_channels) |
|
|
self.transformer = self.conv2d_type(in_channels=256, out_channels=64, kernel_size=1) |
|
|
|
|
|
if self.upsample_mode == "transpose": |
|
|
self.up_conv = UpSample(spatial_dims=2, in_channels=self.out_channels, scale_factor=2, mode="deconv") |
|
|
|
|
|
def forward(self, x: torch.Tensor): |
|
|
""" |
|
|
Args: |
|
|
x: in shape (batch, 3, spatial_1, spatial_2). |
|
|
""" |
|
|
org_input = x |
|
|
x = self.conv1(x) |
|
|
x = self.bn0(x) |
|
|
x = self.relu(x) |
|
|
conv_x = x |
|
|
x = self.maxpool(x) |
|
|
pool_x = x |
|
|
|
|
|
fm1 = self.layer1(x) |
|
|
fm2 = self.layer2(fm1) |
|
|
fm3 = self.layer3(fm2) |
|
|
fm4 = self.layer4(fm3) |
|
|
|
|
|
gcfm1 = self.refine1(self.gcn1(fm4)) |
|
|
gcfm2 = self.refine2(self.gcn2(fm3)) |
|
|
gcfm3 = self.refine3(self.gcn3(fm2)) |
|
|
gcfm4 = self.refine4(self.gcn4(pool_x)) |
|
|
gcfm5 = self.refine5(self.gcn5(conv_x)) |
|
|
|
|
|
if self.upsample_mode == "transpose": |
|
|
fs1 = self.refine6(self.up_conv(gcfm1) + gcfm2) |
|
|
fs2 = self.refine7(self.up_conv(fs1) + gcfm3) |
|
|
fs3 = self.refine8(self.up_conv(fs2) + gcfm4) |
|
|
fs4 = self.refine9(self.up_conv(fs3) + gcfm5) |
|
|
return self.refine10(self.up_conv(fs4)) |
|
|
fs1 = self.refine6(F.interpolate(gcfm1, fm3.size()[2:], mode=self.upsample_mode, align_corners=True) + gcfm2) |
|
|
fs2 = self.refine7(F.interpolate(fs1, fm2.size()[2:], mode=self.upsample_mode, align_corners=True) + gcfm3) |
|
|
fs3 = self.refine8(F.interpolate(fs2, pool_x.size()[2:], mode=self.upsample_mode, align_corners=True) + gcfm4) |
|
|
fs4 = self.refine9(F.interpolate(fs3, conv_x.size()[2:], mode=self.upsample_mode, align_corners=True) + gcfm5) |
|
|
return self.refine10(F.interpolate(fs4, org_input.size()[2:], mode=self.upsample_mode, align_corners=True)) |
|
|
|
|
|
|
|
|
class MCFCN(FCN): |
|
|
""" |
|
|
The multi-channel version of the 2D FCN module. |
|
|
Adds a projection layer to take arbitrary number of inputs. |
|
|
|
|
|
Args: |
|
|
in_channels: number of input channels. Defaults to 3. |
|
|
out_channels: number of output channels. Defaults to 1. |
|
|
upsample_mode: [``"transpose"``, ``"bilinear"``] |
|
|
The mode of upsampling manipulations. |
|
|
Using the second mode cannot guarantee the model's reproducibility. Defaults to ``bilinear``. |
|
|
|
|
|
- ``transpose``, uses transposed convolution layers. |
|
|
- ``bilinear``, uses bilinear interpolate. |
|
|
pretrained: If True, returns a model pre-trained on ImageNet |
|
|
progress: If True, displays a progress bar of the download to stderr. |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
in_channels: int = 3, |
|
|
out_channels: int = 1, |
|
|
upsample_mode: str = "bilinear", |
|
|
pretrained: bool = True, |
|
|
progress: bool = True, |
|
|
): |
|
|
super().__init__( |
|
|
out_channels=out_channels, upsample_mode=upsample_mode, pretrained=pretrained, progress=progress |
|
|
) |
|
|
|
|
|
self.init_proj = Convolution( |
|
|
spatial_dims=2, |
|
|
in_channels=in_channels, |
|
|
out_channels=3, |
|
|
kernel_size=1, |
|
|
act=("relu", {"inplace": True}), |
|
|
norm=Norm.BATCH, |
|
|
bias=False, |
|
|
) |
|
|
|
|
|
def forward(self, x: torch.Tensor): |
|
|
""" |
|
|
Args: |
|
|
x: in shape (batch, in_channels, spatial_1, spatial_2). |
|
|
""" |
|
|
x = self.init_proj(x) |
|
|
return super().forward(x) |
|
|
|