| | |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | import torch.nn as nn |
| | import pytorch_lightning as pl |
| |
|
| |
|
| | class BaseNetwork(pl.LightningModule): |
| |
|
| | def __init__(self): |
| | super(BaseNetwork, self).__init__() |
| |
|
| | def init_weights(self, init_type='xavier', gain=0.02): |
| | ''' |
| | initializes network's weights |
| | init_type: normal | xavier | kaiming | orthogonal |
| | https://github.com/junyanz/pytorch-CycleGAN-and-pix2pix/blob/9451e70673400885567d08a9e97ade2524c700d0/models/networks.py#L39 |
| | ''' |
| |
|
| | def init_func(m): |
| | classname = m.__class__.__name__ |
| | if hasattr(m, 'weight') and (classname.find('Conv') != -1 |
| | or classname.find('Linear') != -1): |
| | if init_type == 'normal': |
| | nn.init.normal_(m.weight.data, 0.0, gain) |
| | elif init_type == 'xavier': |
| | nn.init.xavier_normal_(m.weight.data, gain=gain) |
| | elif init_type == 'kaiming': |
| | nn.init.kaiming_normal_(m.weight.data, a=0, mode='fan_in') |
| | elif init_type == 'orthogonal': |
| | nn.init.orthogonal_(m.weight.data, gain=gain) |
| |
|
| | if hasattr(m, 'bias') and m.bias is not None: |
| | nn.init.constant_(m.bias.data, 0.0) |
| |
|
| | elif classname.find('BatchNorm2d') != -1: |
| | nn.init.normal_(m.weight.data, 1.0, gain) |
| | nn.init.constant_(m.bias.data, 0.0) |
| |
|
| | self.apply(init_func) |
| |
|
| |
|
| | class Residual3D(BaseNetwork): |
| |
|
| | def __init__(self, numIn, numOut): |
| | super(Residual3D, self).__init__() |
| | self.numIn = numIn |
| | self.numOut = numOut |
| | self.with_bias = True |
| | |
| | self.bn = nn.BatchNorm3d(self.numIn) |
| | self.relu = nn.ReLU(inplace=True) |
| | self.conv1 = nn.Conv3d(self.numIn, |
| | self.numOut, |
| | bias=self.with_bias, |
| | kernel_size=3, |
| | stride=1, |
| | padding=2, |
| | dilation=2) |
| | |
| | self.bn1 = nn.BatchNorm3d(self.numOut) |
| | self.conv2 = nn.Conv3d(self.numOut, |
| | self.numOut, |
| | bias=self.with_bias, |
| | kernel_size=3, |
| | stride=1, |
| | padding=1) |
| | |
| | self.bn2 = nn.BatchNorm3d(self.numOut) |
| | self.conv3 = nn.Conv3d(self.numOut, |
| | self.numOut, |
| | bias=self.with_bias, |
| | kernel_size=3, |
| | stride=1, |
| | padding=1) |
| |
|
| | if self.numIn != self.numOut: |
| | self.conv4 = nn.Conv3d(self.numIn, |
| | self.numOut, |
| | bias=self.with_bias, |
| | kernel_size=1) |
| | self.init_weights() |
| |
|
| | def forward(self, x): |
| | residual = x |
| | |
| | |
| | out = self.conv1(x) |
| | out = self.bn1(out) |
| | out = self.relu(out) |
| | out = self.conv2(out) |
| | out = self.bn2(out) |
| | |
| | |
| |
|
| | if self.numIn != self.numOut: |
| | residual = self.conv4(x) |
| |
|
| | return out + residual |
| |
|
| |
|
| | class VolumeEncoder(BaseNetwork): |
| | """CycleGan Encoder""" |
| |
|
| | def __init__(self, num_in=3, num_out=32, num_stacks=2): |
| | super(VolumeEncoder, self).__init__() |
| | self.num_in = num_in |
| | self.num_out = num_out |
| | self.num_inter = 8 |
| | self.num_stacks = num_stacks |
| | self.with_bias = True |
| |
|
| | self.relu = nn.ReLU(inplace=True) |
| | self.conv1 = nn.Conv3d(self.num_in, |
| | self.num_inter, |
| | bias=self.with_bias, |
| | kernel_size=5, |
| | stride=2, |
| | padding=4, |
| | dilation=2) |
| | |
| | self.bn1 = nn.BatchNorm3d(self.num_inter) |
| | self.conv2 = nn.Conv3d(self.num_inter, |
| | self.num_out, |
| | bias=self.with_bias, |
| | kernel_size=5, |
| | stride=2, |
| | padding=4, |
| | dilation=2) |
| | |
| | self.bn2 = nn.BatchNorm3d(self.num_out) |
| |
|
| | self.conv_out1 = nn.Conv3d(self.num_out, |
| | self.num_out, |
| | bias=self.with_bias, |
| | kernel_size=3, |
| | stride=1, |
| | padding=1, |
| | dilation=1) |
| | self.conv_out2 = nn.Conv3d(self.num_out, |
| | self.num_out, |
| | bias=self.with_bias, |
| | kernel_size=3, |
| | stride=1, |
| | padding=1, |
| | dilation=1) |
| |
|
| | for idx in range(self.num_stacks): |
| | self.add_module("res" + str(idx), |
| | Residual3D(self.num_out, self.num_out)) |
| |
|
| | self.init_weights() |
| |
|
| | def forward(self, x, intermediate_output=True): |
| | out = self.conv1(x) |
| | out = self.bn1(out) |
| | out = self.relu(out) |
| |
|
| | out = self.conv2(out) |
| | out = self.bn2(out) |
| | out = self.relu(out) |
| |
|
| | out_lst = [] |
| | for idx in range(self.num_stacks): |
| | out = self._modules["res" + str(idx)](out) |
| | out_lst.append(out) |
| |
|
| | if intermediate_output: |
| | return out_lst |
| | else: |
| | return [out_lst[-1]] |
| |
|