import torch import torch.nn as nn from einops import rearrange from typing import List class ConvBlock(nn.Module): def __init__(self, hidden_size, kernel_size, activation): super(ConvBlock, self).__init__() self.conv = nn.Conv1d( in_channels=hidden_size, out_channels=hidden_size, kernel_size=kernel_size, padding=(kernel_size - 1) // 2, ) self.norm = nn.LayerNorm(hidden_size) self.activation = getattr(nn, activation)() self.dropout = nn.Dropout(0.1) def forward(self, x): x = self.activation(x) x = self.dropout(x) x = self.conv(x) x = rearrange(x, "B D T -> B T D") x = self.norm(x) x = rearrange(x, "B T D -> B D T") return x class ConvStack(nn.Module): def __init__(self, hidden_size, n_blocks, kernel_size, activation): super(ConvStack, self).__init__() blocks = [] for i in range(n_blocks): blocks += [ ConvBlock( hidden_size=hidden_size, kernel_size=kernel_size, activation=activation, ) ] self.blocks = nn.Sequential(*blocks) def forward(self, x): return self.blocks(x) class ResidualBlockStack(nn.Module): def __init__(self, hidden_size, n_stacks, n_blocks, kernel_size, activation): super(ResidualBlockStack, self).__init__() self.conv_stacks = [] for i in range(n_stacks): self.conv_stacks += [ ConvStack( hidden_size=hidden_size, n_blocks=n_blocks, kernel_size=kernel_size, activation=activation, ) ] self.conv_stacks = nn.Sequential(*self.conv_stacks) def forward(self, x): for conv_stack in self.conv_stacks: x = x + conv_stack(x) return x class ConvNet(nn.Module): def __init__( self, in_channels: int, out_channels: int, hidden_size: int, n_stacks: int, n_blocks: int, kernel_size: int, activation: str, last_layer_avg_pooling: bool = False, ): super(ConvNet, self).__init__() self.first_layer = nn.Conv1d( in_channels=in_channels, out_channels=hidden_size, kernel_size=kernel_size, stride=1, padding=(kernel_size - 1) // 2, ) self.conv_stack = ResidualBlockStack( hidden_size=hidden_size, n_stacks=n_stacks, n_blocks=n_blocks, kernel_size=kernel_size, activation=activation, ) if last_layer_avg_pooling: self.last_layer = nn.AdaptiveAvgPool1d(1) else: self.last_layer = nn.Conv1d( in_channels=hidden_size, out_channels=out_channels, kernel_size=kernel_size, stride=1, padding=(kernel_size - 1) // 2, ) def forward(self, x): x = self.first_layer(x) x = self.conv_stack(x) x = self.last_layer(x) return x class ConvNetDoubleLayer(nn.Module): def __init__( self, hidden_size: int, n_stacks: int, n_blocks: int, middle_layer: nn.Module, kernel_size: int, activation: str, ): super(ConvNetDoubleLayer, self).__init__() self.conv_stack1 = ResidualBlockStack( hidden_size=hidden_size, n_stacks=n_stacks, n_blocks=n_blocks, kernel_size=kernel_size, activation=activation, ) self.middle_layer = middle_layer self.conv_stack2 = ResidualBlockStack( hidden_size=hidden_size, n_stacks=n_stacks, n_blocks=n_blocks, kernel_size=kernel_size, activation=activation, ) def forward(self, x): x = self.conv_stack1(x) x = self.middle_layer(x) x = self.conv_stack2(x) return x class ConvNetDouble(nn.Module): def __init__( self, in_channels: int, out_channels: int, hidden_size: int, n_layers: int, n_stacks: int, n_blocks: int, middle_layer: nn.Module, kernel_size: int, activation: str, ): super(ConvNetDouble, self).__init__() self.first_layer = first_conv = nn.Conv1d( in_channels=in_channels, out_channels=hidden_size, kernel_size=kernel_size, stride=1, padding=(kernel_size - 1) // 2, ) self.layers = [] for i in range(n_layers): self.layers += [ ConvNetDoubleLayer( hidden_size=hidden_size, n_stacks=n_stacks, n_blocks=n_blocks, middle_layer=middle_layer, kernel_size=kernel_size, activation=activation, ) ] self.layers = nn.Sequential(*self.layers) self.last_layer = nn.Conv1d( in_channels=hidden_size, out_channels=out_channels, kernel_size=kernel_size, stride=1, padding=(kernel_size - 1) // 2, ) def forward(self, x): x = self.first_layer(x) x_out = self.layers[0](x) for layer in self.layers[1:]: x_out = x_out + layer(x) x = self.last_layer(x_out) return x def test(): x = torch.rand(2, 128, 240) convnet = ConvNet( in_channels=128, out_channels=128, hidden_size=128, n_stacks=2, n_blocks=2, kernel_size=3, activation="ReLU", ) y = convnet(x) print(y.shape) convnet = ConvNetDouble( in_channels=128, out_channels=128, hidden_size=128, n_layers=2, n_stacks=2, n_blocks=2, middle_layer=nn.MaxPool1d( kernel_size=8, stride=8, ), kernel_size=3, activation="ReLU", ) y = convnet(x) print(y.shape)