LARRES / utilpack /mim_modules.py
Staty's picture
Upload 50 files
2b21abc verified
import torch
import torch.nn as nn
class MIMBlock(nn.Module):
def __init__(self, in_channel, num_hidden, height, width, filter_size, stride, layer_norm):
super(MIMBlock, self).__init__()
self.convlstm_c = None
self.num_hidden = num_hidden
self.padding = filter_size // 2
self._forget_bias = 1.0
self.ct_weight = nn.Parameter(torch.zeros(num_hidden*2, height, width))
self.oc_weight = nn.Parameter(torch.zeros(num_hidden, height, width))
if layer_norm:
self.conv_t_cc = nn.Sequential(
nn.Conv2d(in_channel, num_hidden * 3, kernel_size=filter_size,
stride=stride, padding=self.padding, bias=False),
nn.LayerNorm([num_hidden * 3, height, width])
)
self.conv_s_cc = nn.Sequential(
nn.Conv2d(num_hidden, num_hidden * 4, kernel_size=filter_size,
stride=stride, padding=self.padding, bias=False),
nn.LayerNorm([num_hidden * 4, height, width])
)
self.conv_x_cc = nn.Sequential(
nn.Conv2d(num_hidden, num_hidden * 4, kernel_size=filter_size,
stride=stride, padding=self.padding, bias=False),
nn.LayerNorm([num_hidden * 4, height, width])
)
self.conv_h_concat = nn.Sequential(
nn.Conv2d(num_hidden, num_hidden * 4, kernel_size=filter_size,
stride=stride, padding=self.padding, bias=False),
nn.LayerNorm([num_hidden * 4, height, width])
)
self.conv_x_concat = nn.Sequential(
nn.Conv2d(num_hidden, num_hidden * 4, kernel_size=filter_size,
stride=stride, padding=self.padding, bias=False),
nn.LayerNorm([num_hidden * 4, height, width])
)
else:
self.conv_t_cc = nn.Sequential(
nn.Conv2d(in_channel, num_hidden * 3, kernel_size=filter_size,
stride=stride, padding=self.padding, bias=False),
)
self.conv_s_cc = nn.Sequential(
nn.Conv2d(num_hidden, num_hidden * 4, kernel_size=filter_size,
stride=stride, padding=self.padding, bias=False),
)
self.conv_x_cc = nn.Sequential(
nn.Conv2d(num_hidden, num_hidden * 4, kernel_size=filter_size,
stride=stride, padding=self.padding, bias=False),
)
self.conv_h_concat = nn.Sequential(
nn.Conv2d(num_hidden, num_hidden * 4, kernel_size=filter_size,
stride=stride, padding=self.padding, bias=False),
)
self.conv_x_concat = nn.Sequential(
nn.Conv2d(num_hidden, num_hidden * 4, kernel_size=filter_size,
stride=stride, padding=self.padding, bias=False),
)
self.conv_last = nn.Conv2d(num_hidden * 2, num_hidden, kernel_size=1,
stride=1, padding=0, bias=False)
def _init_state(self, inputs):
return torch.zeros_like(inputs)
def MIMS(self, x, h_t, c_t):
if h_t is None:
h_t = self._init_state(x)
if c_t is None:
c_t = self._init_state(x)
h_concat = self.conv_h_concat(h_t)
i_h, g_h, f_h, o_h = torch.split(h_concat, self.num_hidden, dim=1)
ct_activation = torch.mul(c_t.repeat(1,2,1,1), self.ct_weight)
i_c, f_c = torch.split(ct_activation, self.num_hidden, dim=1)
i_ = i_h + i_c
f_ = f_h + f_c
g_ = g_h
o_ = o_h
if x != None:
x_concat = self.conv_x_concat(x)
i_x, g_x, f_x, o_x = torch.split(x_concat, self.num_hidden, dim=1)
i_ = i_ + i_x
f_ = f_ + f_x
g_ = g_ + g_x
o_ = o_ + o_x
i_ = torch.sigmoid(i_)
f_ = torch.sigmoid(f_ + self._forget_bias)
c_new = f_ * c_t + i_ * torch.tanh(g_)
o_c = torch.mul(c_new, self.oc_weight)
h_new = torch.sigmoid(o_ + o_c) * torch.tanh(c_new)
return h_new, c_new
def forward(self, x, diff_h, h, c, m):
h = self._init_state(x) if h is None else h
c = self._init_state(x) if c is None else c
m = self._init_state(x) if m is None else m
diff_h = self._init_state(x) if diff_h is None else diff_h
t_cc = self.conv_t_cc(h)
s_cc = self.conv_s_cc(m)
x_cc = self.conv_x_cc(x)
i_s, g_s, f_s, o_s = torch.split(s_cc, self.num_hidden, dim=1)
i_t, g_t, o_t = torch.split(t_cc, self.num_hidden, dim=1)
i_x, g_x, f_x, o_x = torch.split(x_cc, self.num_hidden, dim=1)
i = torch.sigmoid(i_x + i_t)
i_ = torch.sigmoid(i_x + i_s)
g = torch.tanh(g_x + g_t)
g_ = torch.tanh(g_x + g_s)
f_ = torch.sigmoid(f_x + f_s + self._forget_bias)
o = torch.sigmoid(o_x + o_t + o_s)
new_m = f_ * m + i_ * g_
c, self.convlstm_c = self.MIMS(diff_h, c, self.convlstm_c \
if self.convlstm_c is None else self.convlstm_c.detach())
new_c = c + i * g
cell = torch.cat((new_c, new_m), 1)
new_h = o * torch.tanh(self.conv_last(cell))
return new_h, new_c, new_m
class MIMN(nn.Module):
def __init__(self, in_channel, num_hidden, height, width, filter_size, stride, layer_norm):
super(MIMN, self).__init__()
self.num_hidden = num_hidden
self.padding = filter_size // 2
self._forget_bias = 1.0
self.ct_weight = nn.Parameter(torch.zeros(num_hidden*2, height, width))
self.oc_weight = nn.Parameter(torch.zeros(num_hidden, height, width))
if layer_norm:
self.conv_h_concat = nn.Sequential(
nn.Conv2d(in_channel, num_hidden * 4, kernel_size=filter_size,
stride=stride, padding=self.padding, bias=False),
nn.LayerNorm([num_hidden * 4, height, width])
)
self.conv_x_concat = nn.Sequential(
nn.Conv2d(in_channel, num_hidden * 4, kernel_size=filter_size,
stride=stride, padding=self.padding, bias=False),
nn.LayerNorm([num_hidden * 4, height, width])
)
else:
self.conv_h_concat = nn.Sequential(
nn.Conv2d(in_channel, num_hidden * 4, kernel_size=filter_size,
stride=stride, padding=self.padding, bias=False),
)
self.conv_x_concat = nn.Sequential(
nn.Conv2d(in_channel, num_hidden * 4, kernel_size=filter_size,
stride=stride, padding=self.padding, bias=False),
)
self.conv_last = nn.Conv2d(num_hidden * 2, num_hidden, kernel_size=1,
stride=1, padding=0, bias=False)
def _init_state(self, inputs):
return torch.zeros_like(inputs)
def forward(self, x, h_t, c_t):
if h_t is None:
h_t = self._init_state(x)
if c_t is None:
c_t = self._init_state(x)
h_concat = self.conv_h_concat(h_t)
i_h, g_h, f_h, o_h = torch.split(h_concat, self.num_hidden, dim=1)
ct_activation = torch.mul(c_t.repeat(1,2,1,1), self.ct_weight)
i_c, f_c = torch.split(ct_activation, self.num_hidden, dim=1)
i_ = i_h + i_c
f_ = f_h + f_c
g_ = g_h
o_ = o_h
if x != None:
x_concat = self.conv_x_concat(x)
i_x, g_x, f_x, o_x = torch.split(x_concat, self.num_hidden, dim=1)
i_ = i_ + i_x
f_ = f_ + f_x
g_ = g_ + g_x
o_ = o_ + o_x
i_ = torch.sigmoid(i_)
f_ = torch.sigmoid(f_ + self._forget_bias)
c_new = f_ * c_t + i_ * torch.tanh(g_)
o_c = torch.mul(c_new, self.oc_weight)
h_new = torch.sigmoid(o_ + o_c) * torch.tanh(c_new)
return h_new, c_new