| | import torch.nn as nn |
| |
|
| | def make_linear_layers(feat_dims, relu_final=True, use_bn=False): |
| | layers = [] |
| | for i in range(len(feat_dims)-1): |
| | layers.append(nn.Linear(feat_dims[i], feat_dims[i+1])) |
| |
|
| | |
| | if i < len(feat_dims)-2 or (i == len(feat_dims)-2 and relu_final): |
| | if use_bn: |
| | layers.append(nn.BatchNorm1d(feat_dims[i+1])) |
| | layers.append(nn.ReLU(inplace=True)) |
| |
|
| | return nn.Sequential(*layers) |
| |
|
| | def make_conv_layers(feat_dims, kernel=3, stride=1, padding=1, bnrelu_final=True): |
| | layers = [] |
| | for i in range(len(feat_dims)-1): |
| | layers.append( |
| | nn.Conv2d( |
| | in_channels=feat_dims[i], |
| | out_channels=feat_dims[i+1], |
| | kernel_size=kernel, |
| | stride=stride, |
| | padding=padding |
| | )) |
| | |
| | if i < len(feat_dims)-2 or (i == len(feat_dims)-2 and bnrelu_final): |
| | layers.append(nn.BatchNorm2d(feat_dims[i+1])) |
| | layers.append(nn.ReLU(inplace=True)) |
| |
|
| | return nn.Sequential(*layers) |
| |
|
| | def make_deconv_layers(feat_dims, bnrelu_final=True): |
| | layers = [] |
| | for i in range(len(feat_dims)-1): |
| | layers.append( |
| | nn.ConvTranspose2d( |
| | in_channels=feat_dims[i], |
| | out_channels=feat_dims[i+1], |
| | kernel_size=4, |
| | stride=2, |
| | padding=1, |
| | output_padding=0, |
| | bias=False)) |
| |
|
| | |
| | if i < len(feat_dims)-2 or (i == len(feat_dims)-2 and bnrelu_final): |
| | layers.append(nn.BatchNorm2d(feat_dims[i+1])) |
| | layers.append(nn.ReLU(inplace=True)) |
| |
|
| | return nn.Sequential(*layers) |
| |
|
| |
|