| | import torch |
| | import torch.nn as nn |
| | import torch.nn.functional as F |
| | import torchvision.models as M |
| | import math |
| | from torch import Tensor |
| | from torch.nn import Parameter |
| |
|
| | '''https://github.com/orashi/AlacGAN/blob/master/models/standard.py''' |
| |
|
| | def l2normalize(v, eps=1e-12): |
| | return v / (v.norm() + eps) |
| |
|
| |
|
| | class SpectralNorm(nn.Module): |
| | def __init__(self, module, name='weight', power_iterations=1): |
| | super(SpectralNorm, self).__init__() |
| | self.module = module |
| | self.name = name |
| | self.power_iterations = power_iterations |
| | if not self._made_params(): |
| | self._make_params() |
| |
|
| | def _update_u_v(self): |
| | u = getattr(self.module, self.name + "_u") |
| | v = getattr(self.module, self.name + "_v") |
| | w = getattr(self.module, self.name + "_bar") |
| |
|
| | height = w.data.shape[0] |
| | for _ in range(self.power_iterations): |
| | v.data = l2normalize(torch.mv(torch.t(w.view(height,-1).data), u.data)) |
| | u.data = l2normalize(torch.mv(w.view(height,-1).data, v.data)) |
| |
|
| | |
| | sigma = u.dot(w.view(height, -1).mv(v)) |
| | setattr(self.module, self.name, w / sigma.expand_as(w)) |
| |
|
| | def _made_params(self): |
| | try: |
| | u = getattr(self.module, self.name + "_u") |
| | v = getattr(self.module, self.name + "_v") |
| | w = getattr(self.module, self.name + "_bar") |
| | return True |
| | except AttributeError: |
| | return False |
| |
|
| |
|
| | def _make_params(self): |
| | w = getattr(self.module, self.name) |
| | height = w.data.shape[0] |
| | width = w.view(height, -1).data.shape[1] |
| |
|
| | u = Parameter(w.data.new(height).normal_(0, 1), requires_grad=False) |
| | v = Parameter(w.data.new(width).normal_(0, 1), requires_grad=False) |
| | u.data = l2normalize(u.data) |
| | v.data = l2normalize(v.data) |
| | w_bar = Parameter(w.data) |
| |
|
| | del self.module._parameters[self.name] |
| |
|
| | self.module.register_parameter(self.name + "_u", u) |
| | self.module.register_parameter(self.name + "_v", v) |
| | self.module.register_parameter(self.name + "_bar", w_bar) |
| |
|
| |
|
| | def forward(self, *args): |
| | self._update_u_v() |
| | return self.module.forward(*args) |
| |
|
| | class Selayer(nn.Module): |
| | def __init__(self, inplanes): |
| | super(Selayer, self).__init__() |
| | self.global_avgpool = nn.AdaptiveAvgPool2d(1) |
| | self.conv1 = nn.Conv2d(inplanes, inplanes // 16, kernel_size=1, stride=1) |
| | self.conv2 = nn.Conv2d(inplanes // 16, inplanes, kernel_size=1, stride=1) |
| | self.relu = nn.ReLU(inplace=True) |
| | self.sigmoid = nn.Sigmoid() |
| |
|
| | def forward(self, x): |
| | out = self.global_avgpool(x) |
| | out = self.conv1(out) |
| | out = self.relu(out) |
| | out = self.conv2(out) |
| | out = self.sigmoid(out) |
| |
|
| | return x * out |
| | |
| | class SelayerSpectr(nn.Module): |
| | def __init__(self, inplanes): |
| | super(SelayerSpectr, self).__init__() |
| | self.global_avgpool = nn.AdaptiveAvgPool2d(1) |
| | self.conv1 = SpectralNorm(nn.Conv2d(inplanes, inplanes // 16, kernel_size=1, stride=1)) |
| | self.conv2 = SpectralNorm(nn.Conv2d(inplanes // 16, inplanes, kernel_size=1, stride=1)) |
| | self.relu = nn.ReLU(inplace=True) |
| | self.sigmoid = nn.Sigmoid() |
| |
|
| | def forward(self, x): |
| | out = self.global_avgpool(x) |
| | out = self.conv1(out) |
| | out = self.relu(out) |
| | out = self.conv2(out) |
| | out = self.sigmoid(out) |
| |
|
| | return x * out |
| |
|
| | class ResNeXtBottleneck(nn.Module): |
| | def __init__(self, in_channels=256, out_channels=256, stride=1, cardinality=32, dilate=1): |
| | super(ResNeXtBottleneck, self).__init__() |
| | D = out_channels // 2 |
| | self.out_channels = out_channels |
| | self.conv_reduce = nn.Conv2d(in_channels, D, kernel_size=1, stride=1, padding=0, bias=False) |
| | self.conv_conv = nn.Conv2d(D, D, kernel_size=2 + stride, stride=stride, padding=dilate, dilation=dilate, |
| | groups=cardinality, |
| | bias=False) |
| | self.conv_expand = nn.Conv2d(D, out_channels, kernel_size=1, stride=1, padding=0, bias=False) |
| | self.shortcut = nn.Sequential() |
| | if stride != 1: |
| | self.shortcut.add_module('shortcut', |
| | nn.AvgPool2d(2, stride=2)) |
| | |
| | self.selayer = Selayer(out_channels) |
| |
|
| | def forward(self, x): |
| | bottleneck = self.conv_reduce.forward(x) |
| | bottleneck = F.leaky_relu(bottleneck, 0.2, True) |
| | bottleneck = self.conv_conv.forward(bottleneck) |
| | bottleneck = F.leaky_relu(bottleneck, 0.2, True) |
| | bottleneck = self.conv_expand.forward(bottleneck) |
| | bottleneck = self.selayer(bottleneck) |
| | |
| | x = self.shortcut.forward(x) |
| | return x + bottleneck |
| | |
| | class SpectrResNeXtBottleneck(nn.Module): |
| | def __init__(self, in_channels=256, out_channels=256, stride=1, cardinality=32, dilate=1): |
| | super(SpectrResNeXtBottleneck, self).__init__() |
| | D = out_channels // 2 |
| | self.out_channels = out_channels |
| | self.conv_reduce = SpectralNorm(nn.Conv2d(in_channels, D, kernel_size=1, stride=1, padding=0, bias=False)) |
| | self.conv_conv = SpectralNorm(nn.Conv2d(D, D, kernel_size=2 + stride, stride=stride, padding=dilate, dilation=dilate, |
| | groups=cardinality, |
| | bias=False)) |
| | self.conv_expand = SpectralNorm(nn.Conv2d(D, out_channels, kernel_size=1, stride=1, padding=0, bias=False)) |
| | self.shortcut = nn.Sequential() |
| | if stride != 1: |
| | self.shortcut.add_module('shortcut', |
| | nn.AvgPool2d(2, stride=2)) |
| | |
| | self.selayer = SelayerSpectr(out_channels) |
| |
|
| | def forward(self, x): |
| | bottleneck = self.conv_reduce.forward(x) |
| | bottleneck = F.leaky_relu(bottleneck, 0.2, True) |
| | bottleneck = self.conv_conv.forward(bottleneck) |
| | bottleneck = F.leaky_relu(bottleneck, 0.2, True) |
| | bottleneck = self.conv_expand.forward(bottleneck) |
| | bottleneck = self.selayer(bottleneck) |
| | |
| | x = self.shortcut.forward(x) |
| | return x + bottleneck |
| | |
| | class FeatureConv(nn.Module): |
| | def __init__(self, input_dim=512, output_dim=512): |
| | super(FeatureConv, self).__init__() |
| |
|
| | no_bn = True |
| | |
| | seq = [] |
| | seq.append(nn.Conv2d(input_dim, output_dim, kernel_size=3, stride=1, padding=1, bias=False)) |
| | if not no_bn: seq.append(nn.BatchNorm2d(output_dim)) |
| | seq.append(nn.ReLU(inplace=True)) |
| | seq.append(nn.Conv2d(output_dim, output_dim, kernel_size=3, stride=2, padding=1, bias=False)) |
| | if not no_bn: seq.append(nn.BatchNorm2d(output_dim)) |
| | seq.append(nn.ReLU(inplace=True)) |
| | seq.append(nn.Conv2d(output_dim, output_dim, kernel_size=3, stride=1, padding=1, bias=False)) |
| | seq.append(nn.ReLU(inplace=True)) |
| |
|
| | self.network = nn.Sequential(*seq) |
| |
|
| | def forward(self, x): |
| | return self.network(x) |
| | |
| | class Generator(nn.Module): |
| | def __init__(self, ngf=64): |
| | super(Generator, self).__init__() |
| | |
| | self.feature_conv = FeatureConv() |
| |
|
| | self.to0 = self._make_encoder_block_first(6, 32) |
| | self.to1 = self._make_encoder_block(32, 64) |
| | self.to2 = self._make_encoder_block(64, 128) |
| | self.to3 = self._make_encoder_block(128, 256) |
| | self.to4 = self._make_encoder_block(256, 512) |
| | |
| | self.deconv_for_decoder = nn.Sequential( |
| | nn.ConvTranspose2d(256, 128, 3, stride=2, padding=1, output_padding=1), |
| | nn.LeakyReLU(0.2), |
| | nn.ConvTranspose2d(128, 64, 3, stride=2, padding=1, output_padding=1), |
| | nn.LeakyReLU(0.2), |
| | nn.ConvTranspose2d(64, 32, 3, stride=2, padding=1, output_padding=1), |
| | nn.LeakyReLU(0.2), |
| | nn.ConvTranspose2d(32, 3, 3, stride=1, padding=1, output_padding=0), |
| | nn.Tanh(), |
| | ) |
| |
|
| | tunnel4 = nn.Sequential(*[ResNeXtBottleneck(ngf * 8, ngf * 8, cardinality=32, dilate=1) for _ in range(20)]) |
| |
|
| | self.tunnel4 = nn.Sequential(nn.Conv2d(ngf * 8 + 512, ngf * 8, kernel_size=3, stride=1, padding=1), |
| | nn.LeakyReLU(0.2, True), |
| | tunnel4, |
| | nn.Conv2d(ngf * 8, ngf * 4 * 4, kernel_size=3, stride=1, padding=1), |
| | nn.PixelShuffle(2), |
| | nn.LeakyReLU(0.2, True) |
| | ) |
| |
|
| | depth = 2 |
| | tunnel = [ResNeXtBottleneck(ngf * 4, ngf * 4, cardinality=32, dilate=1) for _ in range(depth)] |
| | tunnel += [ResNeXtBottleneck(ngf * 4, ngf * 4, cardinality=32, dilate=2) for _ in range(depth)] |
| | tunnel += [ResNeXtBottleneck(ngf * 4, ngf * 4, cardinality=32, dilate=4) for _ in range(depth)] |
| | tunnel += [ResNeXtBottleneck(ngf * 4, ngf * 4, cardinality=32, dilate=2), |
| | ResNeXtBottleneck(ngf * 4, ngf * 4, cardinality=32, dilate=1)] |
| | tunnel3 = nn.Sequential(*tunnel) |
| |
|
| | self.tunnel3 = nn.Sequential(nn.Conv2d(ngf * 8, ngf * 4, kernel_size=3, stride=1, padding=1), |
| | nn.LeakyReLU(0.2, True), |
| | tunnel3, |
| | nn.Conv2d(ngf * 4, ngf * 2 * 4, kernel_size=3, stride=1, padding=1), |
| | nn.PixelShuffle(2), |
| | nn.LeakyReLU(0.2, True) |
| | ) |
| |
|
| | tunnel = [ResNeXtBottleneck(ngf * 2, ngf * 2, cardinality=32, dilate=1) for _ in range(depth)] |
| | tunnel += [ResNeXtBottleneck(ngf * 2, ngf * 2, cardinality=32, dilate=2) for _ in range(depth)] |
| | tunnel += [ResNeXtBottleneck(ngf * 2, ngf * 2, cardinality=32, dilate=4) for _ in range(depth)] |
| | tunnel += [ResNeXtBottleneck(ngf * 2, ngf * 2, cardinality=32, dilate=2), |
| | ResNeXtBottleneck(ngf * 2, ngf * 2, cardinality=32, dilate=1)] |
| | tunnel2 = nn.Sequential(*tunnel) |
| |
|
| | self.tunnel2 = nn.Sequential(nn.Conv2d(ngf * 4, ngf * 2, kernel_size=3, stride=1, padding=1), |
| | nn.LeakyReLU(0.2, True), |
| | tunnel2, |
| | nn.Conv2d(ngf * 2, ngf * 4, kernel_size=3, stride=1, padding=1), |
| | nn.PixelShuffle(2), |
| | nn.LeakyReLU(0.2, True) |
| | ) |
| |
|
| | tunnel = [ResNeXtBottleneck(ngf, ngf, cardinality=16, dilate=1)] |
| | tunnel += [ResNeXtBottleneck(ngf, ngf, cardinality=16, dilate=2)] |
| | tunnel += [ResNeXtBottleneck(ngf, ngf, cardinality=16, dilate=4)] |
| | tunnel += [ResNeXtBottleneck(ngf, ngf, cardinality=16, dilate=2), |
| | ResNeXtBottleneck(ngf, ngf, cardinality=16, dilate=1)] |
| | tunnel1 = nn.Sequential(*tunnel) |
| |
|
| | self.tunnel1 = nn.Sequential(nn.Conv2d(ngf * 2, ngf, kernel_size=3, stride=1, padding=1), |
| | nn.LeakyReLU(0.2, True), |
| | tunnel1, |
| | nn.Conv2d(ngf, ngf * 2, kernel_size=3, stride=1, padding=1), |
| | nn.PixelShuffle(2), |
| | nn.LeakyReLU(0.2, True) |
| | ) |
| |
|
| | self.exit = nn.Conv2d(ngf, 3, kernel_size=3, stride=1, padding=1) |
| |
|
| | |
| | def _make_encoder_block(self, inplanes, planes): |
| | return nn.Sequential( |
| | nn.Conv2d(inplanes, planes, 3, 2, 1), |
| | nn.LeakyReLU(0.2), |
| | nn.Conv2d(planes, planes, 3, 1, 1), |
| | nn.LeakyReLU(0.2), |
| | ) |
| |
|
| | def _make_encoder_block_first(self, inplanes, planes): |
| | return nn.Sequential( |
| | nn.Conv2d(inplanes, planes, 3, 1, 1), |
| | nn.LeakyReLU(0.2), |
| | nn.Conv2d(planes, planes, 3, 1, 1), |
| | nn.LeakyReLU(0.2), |
| | ) |
| | |
| | def forward(self, sketch, sketch_feat): |
| |
|
| | x0 = self.to0(sketch) |
| | x1 = self.to1(x0) |
| | x2 = self.to2(x1) |
| | x3 = self.to3(x2) |
| | x4 = self.to4(x3) |
| |
|
| | sketch_feat = self.feature_conv(sketch_feat) |
| | |
| | out = self.tunnel4(torch.cat([x4, sketch_feat], 1)) |
| | |
| | |
| | |
| | |
| | x = self.tunnel3(torch.cat([out, x3], 1)) |
| | x = self.tunnel2(torch.cat([x, x2], 1)) |
| | x = self.tunnel1(torch.cat([x, x1], 1)) |
| | x = torch.tanh(self.exit(torch.cat([x, x0], 1))) |
| | |
| | decoder_output = self.deconv_for_decoder(out) |
| |
|
| | return x, decoder_output |
| | ''' |
| | class Colorizer(nn.Module): |
| | def __init__(self, extractor_path = 'model/model.pth'): |
| | super(Colorizer, self).__init__() |
| | |
| | self.generator = Generator() |
| | self.extractor = se_resnext_half(dump_path=extractor_path, num_classes=370, input_channels=1) |
| | |
| | def extractor_eval(self): |
| | for param in self.extractor.parameters(): |
| | param.requires_grad = False |
| | |
| | def extractor_train(self): |
| | for param in extractor.parameters(): |
| | param.requires_grad = True |
| | |
| | def forward(self, x, extractor_grad = False): |
| | |
| | if extractor_grad: |
| | features = self.extractor(x[:, 0:1]) |
| | else: |
| | with torch.no_grad(): |
| | features = self.extractor(x[:, 0:1]).detach() |
| | |
| | fake, guide = self.generator(x, features) |
| | |
| | return fake, guide |
| | ''' |
| |
|
| | class Colorizer(nn.Module): |
| | def __init__(self, generator_model, extractor_model): |
| | super(Colorizer, self).__init__() |
| | |
| | self.generator = generator_model |
| | self.extractor = extractor_model |
| | |
| | def load_generator_weights(self, gen_weights): |
| | self.generator.load_state_dict(gen_weights) |
| | |
| | def load_extractor_weights(self, ext_weights): |
| | self.extractor.load_state_dict(ext_weights) |
| | |
| | def extractor_eval(self): |
| | for param in self.extractor.parameters(): |
| | param.requires_grad = False |
| | self.extractor.eval() |
| | |
| | def extractor_train(self): |
| | for param in extractor.parameters(): |
| | param.requires_grad = True |
| | self.extractor.train() |
| | |
| | def forward(self, x, extractor_grad = False): |
| | |
| | if extractor_grad: |
| | features = self.extractor(x[:, 0:1]) |
| | else: |
| | with torch.no_grad(): |
| | features = self.extractor(x[:, 0:1]).detach() |
| |
|
| | fake, guide = self.generator(x, features) |
| |
|
| | return fake, guide |
| |
|
| | class Discriminator(nn.Module): |
| | def __init__(self, ndf=64): |
| | super(Discriminator, self).__init__() |
| |
|
| | self.feed = nn.Sequential(SpectralNorm(nn.Conv2d(3, 64, 3, 1, 1)), |
| | nn.LeakyReLU(0.2, True), |
| | SpectralNorm(nn.Conv2d(64, 64, 3, 2, 0)), |
| | nn.LeakyReLU(0.2, True), |
| | |
| | |
| | |
| | |
| | SpectrResNeXtBottleneck(ndf, ndf, cardinality=8, dilate=1), |
| | SpectrResNeXtBottleneck(ndf, ndf, cardinality=8, dilate=1, stride=2), |
| | SpectralNorm(nn.Conv2d(ndf, ndf * 2, kernel_size=1, stride=1, padding=0, bias=False)), |
| | nn.LeakyReLU(0.2, True), |
| |
|
| | SpectrResNeXtBottleneck(ndf * 2, ndf * 2, cardinality=8, dilate=1), |
| | SpectrResNeXtBottleneck(ndf * 2, ndf * 2, cardinality=8, dilate=1, stride=2), |
| | SpectralNorm(nn.Conv2d(ndf * 2, ndf * 4, kernel_size=1, stride=1, padding=0, bias=False)), |
| | nn.LeakyReLU(0.2, True), |
| |
|
| | SpectrResNeXtBottleneck(ndf * 4, ndf * 4, cardinality=8, dilate=1), |
| | SpectrResNeXtBottleneck(ndf * 4, ndf * 4, cardinality=8, dilate=1, stride=2), |
| | SpectralNorm(nn.Conv2d(ndf * 4, ndf * 8, kernel_size=1, stride=1, padding=1, bias=False)), |
| | nn.LeakyReLU(0.2, True), |
| | SpectrResNeXtBottleneck(ndf * 8, ndf * 8, cardinality=8, dilate=1), |
| | SpectrResNeXtBottleneck(ndf * 8, ndf * 8, cardinality=8, dilate=1, stride=2), |
| | SpectrResNeXtBottleneck(ndf * 8, ndf * 8, cardinality=8, dilate=1), |
| | SpectrResNeXtBottleneck(ndf * 8, ndf * 8, cardinality=8, dilate=1), |
| | nn.AdaptiveAvgPool2d((1, 1)) |
| | ) |
| |
|
| | self.out = nn.Linear(512, 1) |
| |
|
| | def forward(self, color): |
| | x = self.feed(color) |
| | |
| | out = self.out(x.view(color.size(0), -1)) |
| | return out |
| | |
| | class Content(nn.Module): |
| | def __init__(self, path): |
| | super(Content, self).__init__() |
| | vgg16 = M.vgg16() |
| | vgg16.load_state_dict(torch.load(path)) |
| | vgg16.features = nn.Sequential( |
| | *list(vgg16.features.children())[:9] |
| | ) |
| | self.model = vgg16.features |
| | self.register_buffer('mean', torch.FloatTensor([0.485 - 0.5, 0.456 - 0.5, 0.406 - 0.5]).view(1, 3, 1, 1)) |
| | self.register_buffer('std', torch.FloatTensor([0.229, 0.224, 0.225]).view(1, 3, 1, 1)) |
| |
|
| | def forward(self, images): |
| | return self.model((images.mul(0.5) - self.mean) / self.std) |