| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | import torch.nn as nn |
| | from utils.utils import instantiate_from_config |
| |
|
| |
|
| | def disabled_train(self, mode=True): |
| | """Overwrite model.train with this function to make sure train/eval mode |
| | does not change anymore.""" |
| | return self |
| |
|
| | def zero_module(module): |
| | """ |
| | Zero out the parameters of a module and return it. |
| | """ |
| | for p in module.parameters(): |
| | p.detach().zero_() |
| | return module |
| |
|
| | def scale_module(module, scale): |
| | """ |
| | Scale the parameters of a module and return it. |
| | """ |
| | for p in module.parameters(): |
| | p.detach().mul_(scale) |
| | return module |
| |
|
| |
|
| | def conv_nd(dims, *args, **kwargs): |
| | """ |
| | Create a 1D, 2D, or 3D convolution module. |
| | """ |
| | if dims == 1: |
| | return nn.Conv1d(*args, **kwargs) |
| | elif dims == 2: |
| | return nn.Conv2d(*args, **kwargs) |
| | elif dims == 3: |
| | return nn.Conv3d(*args, **kwargs) |
| | raise ValueError(f"unsupported dimensions: {dims}") |
| |
|
| |
|
| | def linear(*args, **kwargs): |
| | """ |
| | Create a linear module. |
| | """ |
| | return nn.Linear(*args, **kwargs) |
| |
|
| |
|
| | def avg_pool_nd(dims, *args, **kwargs): |
| | """ |
| | Create a 1D, 2D, or 3D average pooling module. |
| | """ |
| | if dims == 1: |
| | return nn.AvgPool1d(*args, **kwargs) |
| | elif dims == 2: |
| | return nn.AvgPool2d(*args, **kwargs) |
| | elif dims == 3: |
| | return nn.AvgPool3d(*args, **kwargs) |
| | raise ValueError(f"unsupported dimensions: {dims}") |
| |
|
| |
|
| | def nonlinearity(type='silu'): |
| | if type == 'silu': |
| | return nn.SiLU() |
| | elif type == 'leaky_relu': |
| | return nn.LeakyReLU() |
| |
|
| |
|
| | class GroupNormSpecific(nn.GroupNorm): |
| | def forward(self, x): |
| | return super().forward(x.float()).type(x.dtype) |
| |
|
| |
|
| | def normalization(channels, num_groups=32): |
| | """ |
| | Make a standard normalization layer. |
| | :param channels: number of input channels. |
| | :return: an nn.Module for normalization. |
| | """ |
| | return GroupNormSpecific(num_groups, channels) |
| |
|
| |
|
| | class HybridConditioner(nn.Module): |
| |
|
| | def __init__(self, c_concat_config, c_crossattn_config): |
| | super().__init__() |
| | self.concat_conditioner = instantiate_from_config(c_concat_config) |
| | self.crossattn_conditioner = instantiate_from_config(c_crossattn_config) |
| |
|
| | def forward(self, c_concat, c_crossattn): |
| | c_concat = self.concat_conditioner(c_concat) |
| | c_crossattn = self.crossattn_conditioner(c_crossattn) |
| | return {'c_concat': [c_concat], 'c_crossattn': [c_crossattn]} |