""" Contains torch Modules that correspond to basic network building blocks, like MLP, RNN, and CNN backbones. """ import math import abc import numpy as np import textwrap from collections import OrderedDict import torch import torch.nn as nn import torch.nn.functional as F from torchvision import transforms from torchvision import models as vision_models from torchvision import transforms import robomimic.utils.tensor_utils as TensorUtils CONV_ACTIVATIONS = { "relu": nn.ReLU, "None": None, None: None, } def rnn_args_from_config(rnn_config): """ Takes a Config object corresponding to RNN settings (for example `config.algo.rnn` in BCConfig) and extracts rnn kwargs for instantiating rnn networks. """ return dict( rnn_hidden_dim=rnn_config.hidden_dim, rnn_num_layers=rnn_config.num_layers, rnn_type=rnn_config.rnn_type, rnn_kwargs=dict(rnn_config.kwargs), ) def transformer_args_from_config(transformer_config): """ Takes a Config object corresponding to Transformer settings (for example `config.algo.transformer` in BCConfig) and extracts transformer kwargs for instantiating transformer networks. """ transformer_args = dict( transformer_context_length=transformer_config.context_length, transformer_embed_dim=transformer_config.embed_dim, transformer_num_heads=transformer_config.num_heads, transformer_emb_dropout=transformer_config.emb_dropout, transformer_attn_dropout=transformer_config.attn_dropout, transformer_block_output_dropout=transformer_config.block_output_dropout, transformer_sinusoidal_embedding=transformer_config.sinusoidal_embedding, transformer_activation=transformer_config.activation, transformer_nn_parameter_for_timesteps=transformer_config.nn_parameter_for_timesteps, ) if "num_layers" in transformer_config: transformer_args["transformer_num_layers"] = transformer_config.num_layers return transformer_args class Module(torch.nn.Module): """ Base class for networks. The only difference from torch.nn.Module is that it requires implementing @output_shape. """ @abc.abstractmethod def output_shape(self, input_shape=None): """ Function to compute output shape from inputs to this module. Args: input_shape (iterable of int): shape of input. Does not include batch dimension. Some modules may not need this argument, if their output does not depend on the size of the input, or if they assume fixed size input. Returns: out_shape ([int]): list of integers corresponding to output shape """ raise NotImplementedError class Sequential(torch.nn.Sequential, Module): """ Compose multiple Modules together (defined above). """ def __init__(self, *args, has_output_shape = True): """ Args: has_output_shape (bool, optional): indicates whether output_shape can be called on the Sequential module. torch.nn modules do not have an output_shape, but Modules (defined above) do. Defaults to True. """ for arg in args: if has_output_shape: assert isinstance(arg, Module) else: assert isinstance(arg, nn.Module) torch.nn.Sequential.__init__(self, *args) self.fixed = False self.has_output_shape = has_output_shape def output_shape(self, input_shape=None): """ Function to compute output shape from inputs to this module. Args: input_shape (iterable of int): shape of input. Does not include batch dimension. Some modules may not need this argument, if their output does not depend on the size of the input, or if they assume fixed size input. Returns: out_shape ([int]): list of integers corresponding to output shape """ if not self.has_output_shape: raise NotImplementedError("Output shape is not defined for this module") out_shape = input_shape for module in self: out_shape = module.output_shape(out_shape) return out_shape def freeze(self): self.fixed = True def train(self, mode): if self.fixed: super().train(False) else: super().train(mode) class Parameter(Module): """ A class that is a thin wrapper around a torch.nn.Parameter to make for easy saving and optimization. """ def __init__(self, init_tensor): """ Args: init_tensor (torch.Tensor): initial tensor """ super(Parameter, self).__init__() self.param = torch.nn.Parameter(init_tensor) def output_shape(self, input_shape=None): """ Function to compute output shape from inputs to this module. Args: input_shape (iterable of int): shape of input. Does not include batch dimension. Some modules may not need this argument, if their output does not depend on the size of the input, or if they assume fixed size input. Returns: out_shape ([int]): list of integers corresponding to output shape """ return list(self.param.shape) def forward(self, inputs=None): """ Forward call just returns the parameter tensor. """ return self.param class Unsqueeze(Module): """ Trivial class that unsqueezes the input. Useful for including in a nn.Sequential network """ def __init__(self, dim): super(Unsqueeze, self).__init__() self.dim = dim def output_shape(self, input_shape=None): assert input_shape is not None return input_shape + [1] if self.dim == -1 else input_shape[:self.dim + 1] + [1] + input_shape[self.dim + 1:] def forward(self, x): return x.unsqueeze(dim=self.dim) class Squeeze(Module): """ Trivial class that squeezes the input. Useful for including in a nn.Sequential network """ def __init__(self, dim): super(Squeeze, self).__init__() self.dim = dim def output_shape(self, input_shape=None): assert input_shape is not None return input_shape[:self.dim] + input_shape[self.dim+1:] if input_shape[self.dim] == 1 else input_shape def forward(self, x): return x.squeeze(dim=self.dim) class MLP(Module): """ Base class for simple Multi-Layer Perceptrons. """ def __init__( self, input_dim, output_dim, layer_dims=(), layer_func=nn.Linear, layer_func_kwargs=None, activation=nn.ReLU, dropouts=None, normalization=False, output_activation=None, ): """ Args: input_dim (int): dimension of inputs output_dim (int): dimension of outputs layer_dims ([int]): sequence of integers for the hidden layers sizes layer_func: mapping per layer - defaults to Linear layer_func_kwargs (dict): kwargs for @layer_func activation: non-linearity per layer - defaults to ReLU dropouts ([float]): if not None, adds dropout layers with the corresponding probabilities after every layer. Must be same size as @layer_dims. normalization (bool): if True, apply layer normalization after each layer output_activation: if provided, applies the provided non-linearity to the output layer """ super(MLP, self).__init__() layers = [] dim = input_dim if layer_func_kwargs is None: layer_func_kwargs = dict() if dropouts is not None: assert(len(dropouts) == len(layer_dims)) for i, l in enumerate(layer_dims): layers.append(layer_func(dim, l, **layer_func_kwargs)) if normalization: layers.append(nn.LayerNorm(l)) layers.append(activation()) if dropouts is not None and dropouts[i] > 0.: layers.append(nn.Dropout(dropouts[i])) dim = l layers.append(layer_func(dim, output_dim)) if output_activation is not None: layers.append(output_activation()) self._layer_func = layer_func self.nets = layers self._model = nn.Sequential(*layers) self._layer_dims = layer_dims self._input_dim = input_dim self._output_dim = output_dim self._dropouts = dropouts self._act = activation self._output_act = output_activation def output_shape(self, input_shape=None): """ Function to compute output shape from inputs to this module. Args: input_shape (iterable of int): shape of input. Does not include batch dimension. Some modules may not need this argument, if their output does not depend on the size of the input, or if they assume fixed size input. Returns: out_shape ([int]): list of integers corresponding to output shape """ return [self._output_dim] def forward(self, inputs): """ Forward pass. """ return self._model(inputs) def __repr__(self): """Pretty print network.""" header = str(self.__class__.__name__) act = None if self._act is None else self._act.__name__ output_act = None if self._output_act is None else self._output_act.__name__ indent = ' ' * 4 msg = "input_dim={}\noutput_dim={}\nlayer_dims={}\nlayer_func={}\ndropout={}\nact={}\noutput_act={}".format( self._input_dim, self._output_dim, self._layer_dims, self._layer_func.__name__, self._dropouts, act, output_act ) msg = textwrap.indent(msg, indent) msg = header + '(\n' + msg + '\n)' return msg class RNN_Base(Module): """ A wrapper class for a multi-step RNN and a per-step network. """ def __init__( self, input_dim, rnn_hidden_dim, rnn_num_layers, rnn_type="LSTM", # [LSTM, GRU] rnn_kwargs=None, per_step_net=None, ): """ Args: input_dim (int): dimension of inputs rnn_hidden_dim (int): RNN hidden dimension rnn_num_layers (int): number of RNN layers rnn_type (str): [LSTM, GRU] rnn_kwargs (dict): kwargs for the torch.nn.LSTM / GRU per_step_net: a network that runs per time step on top of the RNN output """ super(RNN_Base, self).__init__() self.per_step_net = per_step_net if per_step_net is not None: assert isinstance(per_step_net, Module), "RNN_Base: per_step_net is not instance of Module" assert rnn_type in ["LSTM", "GRU"] rnn_cls = nn.LSTM if rnn_type == "LSTM" else nn.GRU rnn_kwargs = rnn_kwargs if rnn_kwargs is not None else {} rnn_is_bidirectional = rnn_kwargs.get("bidirectional", False) self.nets = rnn_cls( input_size=input_dim, hidden_size=rnn_hidden_dim, num_layers=rnn_num_layers, batch_first=True, **rnn_kwargs, ) self._hidden_dim = rnn_hidden_dim self._num_layers = rnn_num_layers self._rnn_type = rnn_type self._num_directions = int(rnn_is_bidirectional) + 1 # 2 if bidirectional, 1 otherwise @property def rnn_type(self): return self._rnn_type def get_rnn_init_state(self, batch_size, device): """ Get a default RNN state (zeros) Args: batch_size (int): batch size dimension device: device the hidden state should be sent to. Returns: hidden_state (torch.Tensor or tuple): returns hidden state tensor or tuple of hidden state tensors depending on the RNN type """ h_0 = torch.zeros(self._num_layers * self._num_directions, batch_size, self._hidden_dim).to(device) if self._rnn_type == "LSTM": c_0 = torch.zeros(self._num_layers * self._num_directions, batch_size, self._hidden_dim).to(device) return h_0, c_0 else: return h_0 def output_shape(self, input_shape): """ Function to compute output shape from inputs to this module. Args: input_shape (iterable of int): shape of input. Does not include batch dimension. Some modules may not need this argument, if their output does not depend on the size of the input, or if they assume fixed size input. Returns: out_shape ([int]): list of integers corresponding to output shape """ # infer time dimension from input shape and add to per_step_net output shape if self.per_step_net is not None: out = self.per_step_net.output_shape(input_shape[1:]) if isinstance(out, dict): out = {k: [input_shape[0]] + out[k] for k in out} else: out = [input_shape[0]] + out else: out = [input_shape[0], self._num_layers * self._hidden_dim] return out def forward(self, inputs, rnn_init_state=None, return_state=False): """ Forward a sequence of inputs through the RNN and the per-step network. Args: inputs (torch.Tensor): tensor input of shape [B, T, D], where D is the RNN input size rnn_init_state: rnn hidden state, initialize to zero state if set to None return_state (bool): whether to return hidden state Returns: outputs: outputs of the per_step_net rnn_state: return rnn state at the end if return_state is set to True """ assert inputs.ndimension() == 3 # [B, T, D] batch_size, seq_length, inp_dim = inputs.shape if rnn_init_state is None: rnn_init_state = self.get_rnn_init_state(batch_size, device=inputs.device) outputs, rnn_state = self.nets(inputs, rnn_init_state) if self.per_step_net is not None: outputs = TensorUtils.time_distributed(outputs, self.per_step_net) if return_state: return outputs, rnn_state else: return outputs def forward_step(self, inputs, rnn_state): """ Forward a single step input through the RNN and per-step network, and return the new hidden state. Args: inputs (torch.Tensor): tensor input of shape [B, D], where D is the RNN input size rnn_state: rnn hidden state, initialize to zero state if set to None Returns: outputs: outputs of the per_step_net rnn_state: return the new rnn state """ assert inputs.ndimension() == 2 inputs = TensorUtils.to_sequence(inputs) outputs, rnn_state = self.forward( inputs, rnn_init_state=rnn_state, return_state=True, ) return outputs[:, 0], rnn_state """ ================================================ Visual Backbone Networks ================================================ """ class ConvBase(Module): """ Base class for ConvNets. """ def __init__(self): super(ConvBase, self).__init__() # dirty hack - re-implement to pass the buck onto subclasses from ABC parent def output_shape(self, input_shape): """ Function to compute output shape from inputs to this module. Args: input_shape (iterable of int): shape of input. Does not include batch dimension. Some modules may not need this argument, if their output does not depend on the size of the input, or if they assume fixed size input. Returns: out_shape ([int]): list of integers corresponding to output shape """ raise NotImplementedError def forward(self, inputs): x = self.nets(inputs) if list(self.output_shape(list(inputs.shape)[1:])) != list(x.shape)[1:]: raise ValueError('Size mismatch: expect size %s, but got size %s' % ( str(self.output_shape(list(inputs.shape)[1:])), str(list(x.shape)[1:])) ) return x class ResNet18Conv(ConvBase): """ A ResNet18 block that can be used to process input images. """ def __init__( self, input_channel=3, pretrained=False, input_coord_conv=False, ): """ Args: input_channel (int): number of input channels for input images to the network. If not equal to 3, modifies first conv layer in ResNet to handle the number of input channels. pretrained (bool): if True, load pretrained weights for all ResNet layers. input_coord_conv (bool): if True, use a coordinate convolution for the first layer (a convolution where input channels are modified to encode spatial pixel location) """ super(ResNet18Conv, self).__init__() net = vision_models.resnet18(pretrained=pretrained) if input_coord_conv: net.conv1 = CoordConv2d(input_channel, 64, kernel_size=7, stride=2, padding=3, bias=False) elif input_channel != 3: net.conv1 = nn.Conv2d(input_channel, 64, kernel_size=7, stride=2, padding=3, bias=False) # cut the last fc layer self._input_coord_conv = input_coord_conv self._input_channel = input_channel self.nets = torch.nn.Sequential(*(list(net.children())[:-2])) def output_shape(self, input_shape): """ Function to compute output shape from inputs to this module. Args: input_shape (iterable of int): shape of input. Does not include batch dimension. Some modules may not need this argument, if their output does not depend on the size of the input, or if they assume fixed size input. Returns: out_shape ([int]): list of integers corresponding to output shape """ assert(len(input_shape) == 3) out_h = int(math.ceil(input_shape[1] / 32.)) out_w = int(math.ceil(input_shape[2] / 32.)) return [512, out_h, out_w] def __repr__(self): """Pretty print network.""" header = '{}'.format(str(self.__class__.__name__)) return header + '(input_channel={}, input_coord_conv={})'.format(self._input_channel, self._input_coord_conv) class R3MConv(ConvBase): """ Base class for ConvNets pretrained with R3M (https://arxiv.org/abs/2203.12601) """ def __init__( self, input_channel=3, r3m_model_class='resnet18', freeze=True, ): """ Using R3M pretrained observation encoder network proposed by https://arxiv.org/abs/2203.12601 Args: input_channel (int): number of input channels for input images to the network. If not equal to 3, modifies first conv layer in ResNet to handle the number of input channels. r3m_model_class (str): select one of the r3m pretrained model "resnet18", "resnet34" or "resnet50" freeze (bool): if True, use a frozen R3M pretrained model. """ super(R3MConv, self).__init__() try: from r3m import load_r3m except ImportError: print("WARNING: could not load r3m library! Please follow https://github.com/facebookresearch/r3m to install R3M") net = load_r3m(r3m_model_class) assert input_channel == 3 # R3M only support input image with channel size 3 assert r3m_model_class in ["resnet18", "resnet34", "resnet50"] # make sure the selected r3m model do exist # cut the last fc layer self._input_channel = input_channel self._r3m_model_class = r3m_model_class self._freeze = freeze self._input_coord_conv = False self._pretrained = True preprocess = nn.Sequential( transforms.Resize(256), transforms.CenterCrop(224), transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]), ) self.nets = Sequential(*([preprocess] + list(net.module.convnet.children())), has_output_shape = False) if freeze: self.nets.freeze() self.weight_sum = np.sum([param.cpu().data.numpy().sum() for param in self.nets.parameters()]) if freeze: for param in self.nets.parameters(): param.requires_grad = False self.nets.eval() def output_shape(self, input_shape): """ Function to compute output shape from inputs to this module. Args: input_shape (iterable of int): shape of input. Does not include batch dimension. Some modules may not need this argument, if their output does not depend on the size of the input, or if they assume fixed size input. Returns: out_shape ([int]): list of integers corresponding to output shape """ assert(len(input_shape) == 3) if self._r3m_model_class == 'resnet50': out_dim = 2048 else: out_dim = 512 return [out_dim, 1, 1] def __repr__(self): """Pretty print network.""" header = '{}'.format(str(self.__class__.__name__)) return header + '(input_channel={}, input_coord_conv={}, pretrained={}, freeze={})'.format(self._input_channel, self._input_coord_conv, self._pretrained, self._freeze) class MVPConv(ConvBase): """ Base class for ConvNets pretrained with MVP (https://arxiv.org/abs/2203.06173) """ def __init__( self, input_channel=3, mvp_model_class='vitb-mae-egosoup', freeze=True, ): """ Using MVP pretrained observation encoder network proposed by https://arxiv.org/abs/2203.06173 Args: input_channel (int): number of input channels for input images to the network. If not equal to 3, modifies first conv layer in ResNet to handle the number of input channels. mvp_model_class (str): select one of the mvp pretrained model "vits-mae-hoi", "vits-mae-in", "vits-sup-in", "vitb-mae-egosoup" or "vitl-256-mae-egosoup" freeze (bool): if True, use a frozen MVP pretrained model. """ super(MVPConv, self).__init__() try: import mvp except ImportError: print("WARNING: could not load mvp library! Please follow https://github.com/ir413/mvp to install MVP.") self.nets = mvp.load(mvp_model_class) if freeze: self.nets.freeze() assert input_channel == 3 # MVP only support input image with channel size 3 assert mvp_model_class in ["vits-mae-hoi", "vits-mae-in", "vits-sup-in", "vitb-mae-egosoup", "vitl-256-mae-egosoup"] # make sure the selected r3m model do exist self._input_channel = input_channel self._freeze = freeze self._mvp_model_class = mvp_model_class self._input_coord_conv = False self._pretrained = True if '256' in mvp_model_class: input_img_size = 256 else: input_img_size = 224 self.preprocess = nn.Sequential( transforms.Resize(input_img_size) ) def forward(self, inputs): x = self.preprocess(inputs) x = self.nets(x) if list(self.output_shape(list(inputs.shape)[1:])) != list(x.shape)[1:]: raise ValueError('Size mismatch: expect size %s, but got size %s' % ( str(self.output_shape(list(inputs.shape)[1:])), str(list(x.shape)[1:])) ) return x def output_shape(self, input_shape): """ Function to compute output shape from inputs to this module. Args: input_shape (iterable of int): shape of input. Does not include batch dimension. Some modules may not need this argument, if their output does not depend on the size of the input, or if they assume fixed size input. Returns: out_shape ([int]): list of integers corresponding to output shape """ assert(len(input_shape) == 3) if 'vitb' in self._mvp_model_class: output_shape = [768] elif 'vitl' in self._mvp_model_class: output_shape = [1024] else: output_shape = [384] return output_shape def __repr__(self): """Pretty print network.""" header = '{}'.format(str(self.__class__.__name__)) return header + '(input_channel={}, input_coord_conv={}, pretrained={}, freeze={})'.format(self._input_channel, self._input_coord_conv, self._pretrained, self._freeze) class CoordConv2d(nn.Conv2d, Module): """ 2D Coordinate Convolution Source: An Intriguing Failing of Convolutional Neural Networks and the CoordConv Solution https://arxiv.org/abs/1807.03247 (e.g. adds 2 channels per input feature map corresponding to (x, y) location on map) """ def __init__( self, in_channels, out_channels, kernel_size, stride=1, padding=0, dilation=1, groups=1, bias=True, padding_mode='zeros', coord_encoding='position', ): """ Args: in_channels: number of channels of the input tensor [C, H, W] out_channels: number of output channels of the layer kernel_size: convolution kernel size stride: conv stride padding: conv padding dilation: conv dilation groups: conv groups bias: conv bias padding_mode: conv padding mode coord_encoding: type of coordinate encoding. currently only 'position' is implemented """ assert(coord_encoding in ['position']) self.coord_encoding = coord_encoding if coord_encoding == 'position': in_channels += 2 # two extra channel for positional encoding self._position_enc = None # position encoding else: raise Exception("CoordConv2d: coord encoding {} not implemented".format(self.coord_encoding)) nn.Conv2d.__init__( self, in_channels=in_channels, out_channels=out_channels, kernel_size=kernel_size, stride=stride, padding=padding, dilation=dilation, groups=groups, bias=bias, padding_mode=padding_mode ) def output_shape(self, input_shape): """ Function to compute output shape from inputs to this module. Args: input_shape (iterable of int): shape of input. Does not include batch dimension. Some modules may not need this argument, if their output does not depend on the size of the input, or if they assume fixed size input. Returns: out_shape ([int]): list of integers corresponding to output shape """ # adds 2 to channel dimension return [input_shape[0] + 2] + input_shape[1:] def forward(self, input): b, c, h, w = input.shape if self.coord_encoding == 'position': if self._position_enc is None: pos_y, pos_x = torch.meshgrid(torch.arange(h), torch.arange(w)) pos_y = pos_y.float().to(input.device) / float(h) pos_x = pos_x.float().to(input.device) / float(w) self._position_enc = torch.stack((pos_y, pos_x)).unsqueeze(0) pos_enc = self._position_enc.expand(b, -1, -1, -1) input = torch.cat((input, pos_enc), dim=1) return super(CoordConv2d, self).forward(input) class ShallowConv(ConvBase): """ A shallow convolutional encoder from https://rll.berkeley.edu/dsae/dsae.pdf """ def __init__(self, input_channel=3, output_channel=32): super(ShallowConv, self).__init__() self._input_channel = input_channel self._output_channel = output_channel self.nets = nn.Sequential( torch.nn.Conv2d(input_channel, 64, kernel_size=7, stride=2, padding=3), torch.nn.ReLU(), torch.nn.Conv2d(64, 32, kernel_size=1, stride=1, padding=0), torch.nn.ReLU(), torch.nn.Conv2d(32, 32, kernel_size=3, stride=1, padding=1), torch.nn.ReLU(), torch.nn.Conv2d(32, 32, kernel_size=3, stride=1, padding=1), ) def output_shape(self, input_shape): """ Function to compute output shape from inputs to this module. Args: input_shape (iterable of int): shape of input. Does not include batch dimension. Some modules may not need this argument, if their output does not depend on the size of the input, or if they assume fixed size input. Returns: out_shape ([int]): list of integers corresponding to output shape """ assert(len(input_shape) == 3) assert(input_shape[0] == self._input_channel) out_h = int(math.floor(input_shape[1] / 2.)) out_w = int(math.floor(input_shape[2] / 2.)) return [self._output_channel, out_h, out_w] class Conv1dBase(Module): """ Base class for stacked Conv1d layers. Args: input_channel (int): Number of channels for inputs to this network activation (None or str): Per-layer activation to use. Defaults to "relu". Valid options are currently {relu, None} for no activation out_channels (list of int): Output channel size for each sequential Conv1d layer kernel_size (list of int): Kernel sizes for each sequential Conv1d layer stride (list of int): Stride sizes for each sequential Conv1d layer conv_kwargs (dict): additional nn.Conv1D args to use, in list form, where the ith element corresponds to the argument to be passed to the ith Conv1D layer. See https://pytorch.org/docs/stable/generated/torch.nn.Conv1d.html for specific possible arguments. """ def __init__( self, input_channel=1, activation="relu", out_channels=(32, 64, 64), kernel_size=(8, 4, 2), stride=(4, 2, 1), **conv_kwargs, ): super(Conv1dBase, self).__init__() # Get activation requested activation = CONV_ACTIVATIONS[activation] # Add layer kwargs conv_kwargs["out_channels"] = out_channels conv_kwargs["kernel_size"] = kernel_size conv_kwargs["stride"] = stride # Generate network self.n_layers = len(out_channels) layers = OrderedDict() for i in range(self.n_layers): layer_kwargs = {k: v[i] for k, v in conv_kwargs.items()} layers[f'conv{i}'] = nn.Conv1d( in_channels=input_channel, **layer_kwargs, ) if activation is not None: layers[f'act{i}'] = activation() input_channel = layer_kwargs["out_channels"] # Store network self.nets = nn.Sequential(layers) def output_shape(self, input_shape): """ Function to compute output shape from inputs to this module. Args: input_shape (iterable of int): shape of input. Does not include batch dimension. Some modules may not need this argument, if their output does not depend on the size of the input, or if they assume fixed size input. Returns: out_shape ([int]): list of integers corresponding to output shape """ channels, length = input_shape for i in range(self.n_layers): net = getattr(self.nets, f"conv{i}") channels = net.out_channels length = int((length + 2 * net.padding[0] - net.dilation[0] * (net.kernel_size[0] - 1) - 1) / net.stride[0]) + 1 return [channels, length] def forward(self, inputs): x = self.nets(inputs) if list(self.output_shape(list(inputs.shape)[1:])) != list(x.shape)[1:]: raise ValueError('Size mismatch: expect size %s, but got size %s' % ( str(self.output_shape(list(inputs.shape)[1:])), str(list(x.shape)[1:])) ) return x """ ================================================ Pooling Networks ================================================ """ class SpatialSoftmax(ConvBase): """ Spatial Softmax Layer. Based on Deep Spatial Autoencoders for Visuomotor Learning by Finn et al. https://rll.berkeley.edu/dsae/dsae.pdf """ def __init__( self, input_shape, num_kp=32, temperature=1., learnable_temperature=False, output_variance=False, noise_std=0.0, ): """ Args: input_shape (list): shape of the input feature (C, H, W) num_kp (int): number of keypoints (None for not using spatialsoftmax) temperature (float): temperature term for the softmax. learnable_temperature (bool): whether to learn the temperature output_variance (bool): treat attention as a distribution, and compute second-order statistics to return noise_std (float): add random spatial noise to the predicted keypoints """ super(SpatialSoftmax, self).__init__() assert len(input_shape) == 3 self._in_c, self._in_h, self._in_w = input_shape # (C, H, W) if num_kp is not None: self.nets = torch.nn.Conv2d(self._in_c, num_kp, kernel_size=1) self._num_kp = num_kp else: self.nets = None self._num_kp = self._in_c self.learnable_temperature = learnable_temperature self.output_variance = output_variance self.noise_std = noise_std if self.learnable_temperature: # temperature will be learned temperature = torch.nn.Parameter(torch.ones(1) * temperature, requires_grad=True) self.register_parameter('temperature', temperature) else: # temperature held constant after initialization temperature = torch.nn.Parameter(torch.ones(1) * temperature, requires_grad=False) self.register_buffer('temperature', temperature) pos_x, pos_y = np.meshgrid( np.linspace(-1., 1., self._in_w), np.linspace(-1., 1., self._in_h) ) pos_x = torch.from_numpy(pos_x.reshape(1, self._in_h * self._in_w)).float() pos_y = torch.from_numpy(pos_y.reshape(1, self._in_h * self._in_w)).float() self.register_buffer('pos_x', pos_x) self.register_buffer('pos_y', pos_y) self.kps = None def __repr__(self): """Pretty print network.""" header = format(str(self.__class__.__name__)) return header + '(num_kp={}, temperature={}, noise={})'.format( self._num_kp, self.temperature.item(), self.noise_std) def output_shape(self, input_shape): """ Function to compute output shape from inputs to this module. Args: input_shape (iterable of int): shape of input. Does not include batch dimension. Some modules may not need this argument, if their output does not depend on the size of the input, or if they assume fixed size input. Returns: out_shape ([int]): list of integers corresponding to output shape """ assert(len(input_shape) == 3) assert(input_shape[0] == self._in_c) return [self._num_kp, 2] def forward(self, feature): """ Forward pass through spatial softmax layer. For each keypoint, a 2D spatial probability distribution is created using a softmax, where the support is the pixel locations. This distribution is used to compute the expected value of the pixel location, which becomes a keypoint of dimension 2. K such keypoints are created. Returns: out (torch.Tensor or tuple): mean keypoints of shape [B, K, 2], and possibly keypoint variance of shape [B, K, 2, 2] corresponding to the covariance under the 2D spatial softmax distribution """ assert(feature.shape[1] == self._in_c) assert(feature.shape[2] == self._in_h) assert(feature.shape[3] == self._in_w) if self.nets is not None: feature = self.nets(feature) # [B, K, H, W] -> [B * K, H * W] where K is number of keypoints feature = feature.reshape(-1, self._in_h * self._in_w) # 2d softmax normalization attention = F.softmax(feature / self.temperature, dim=-1) # [1, H * W] x [B * K, H * W] -> [B * K, 1] for spatial coordinate mean in x and y dimensions expected_x = torch.sum(self.pos_x * attention, dim=1, keepdim=True) expected_y = torch.sum(self.pos_y * attention, dim=1, keepdim=True) # stack to [B * K, 2] expected_xy = torch.cat([expected_x, expected_y], 1) # reshape to [B, K, 2] feature_keypoints = expected_xy.view(-1, self._num_kp, 2) if self.training: noise = torch.randn_like(feature_keypoints) * self.noise_std feature_keypoints += noise if self.output_variance: # treat attention as a distribution, and compute second-order statistics to return expected_xx = torch.sum(self.pos_x * self.pos_x * attention, dim=1, keepdim=True) expected_yy = torch.sum(self.pos_y * self.pos_y * attention, dim=1, keepdim=True) expected_xy = torch.sum(self.pos_x * self.pos_y * attention, dim=1, keepdim=True) var_x = expected_xx - expected_x * expected_x var_y = expected_yy - expected_y * expected_y var_xy = expected_xy - expected_x * expected_y # stack to [B * K, 4] and then reshape to [B, K, 2, 2] where last 2 dims are covariance matrix feature_covar = torch.cat([var_x, var_xy, var_xy, var_y], 1).reshape(-1, self._num_kp, 2, 2) feature_keypoints = (feature_keypoints, feature_covar) if isinstance(feature_keypoints, tuple): self.kps = (feature_keypoints[0].detach(), feature_keypoints[1].detach()) else: self.kps = feature_keypoints.detach() return feature_keypoints class SpatialMeanPool(Module): """ Module that averages inputs across all spatial dimensions (dimension 2 and after), leaving only the batch and channel dimensions. """ def __init__(self, input_shape): super(SpatialMeanPool, self).__init__() assert len(input_shape) == 3 # [C, H, W] self.in_shape = input_shape def output_shape(self, input_shape=None): """ Function to compute output shape from inputs to this module. Args: input_shape (iterable of int): shape of input. Does not include batch dimension. Some modules may not need this argument, if their output does not depend on the size of the input, or if they assume fixed size input. Returns: out_shape ([int]): list of integers corresponding to output shape """ return list(self.in_shape[:1]) # [C, H, W] -> [C] def forward(self, inputs): """Forward pass - average across all dimensions except batch and channel.""" return TensorUtils.flatten(inputs, begin_axis=2).mean(dim=2) class FeatureAggregator(Module): """ Helpful class for aggregating features across a dimension. This is useful in practice when training models that break an input image up into several patches since features can be extraced per-patch using the same encoder and then aggregated using this module. """ def __init__(self, dim=1, agg_type="avg"): super(FeatureAggregator, self).__init__() self.dim = dim self.agg_type = agg_type def set_weight(self, w): assert self.agg_type == "w_avg" self.agg_weight = w def clear_weight(self): assert self.agg_type == "w_avg" self.agg_weight = None def output_shape(self, input_shape): """ Function to compute output shape from inputs to this module. Args: input_shape (iterable of int): shape of input. Does not include batch dimension. Some modules may not need this argument, if their output does not depend on the size of the input, or if they assume fixed size input. Returns: out_shape ([int]): list of integers corresponding to output shape """ # aggregates on @self.dim, so it is removed from the output shape return list(input_shape[:self.dim]) + list(input_shape[self.dim+1:]) def forward(self, x): """Forward pooling pass.""" if self.agg_type == "avg": # mean-pooling return torch.mean(x, dim=1) if self.agg_type == "w_avg": # weighted mean-pooling return torch.sum(x * self.agg_weight, dim=1) raise Exception("unexpected agg type: {}".forward(self.agg_type))