| import torch |
| import torch.nn.functional as F |
| from torch import nn |
|
|
|
|
| class PositionalEncoding(nn.Module): |
| def __init__(self, dim, max_len=5000): |
| super().__init__() |
| self.max_len = max_len |
| pe = torch.zeros(max_len, dim) |
| position = torch.arange(0, max_len, dtype=torch.float).unsqueeze(1) |
| div_term = torch.exp(torch.arange(0, dim, 2).float() * (-torch.log(torch.tensor(10000.0)) / dim)) |
| pe[:, 0::2] = torch.sin(position * div_term) |
| pe[:, 1::2] = torch.cos(position * div_term) |
| pe = pe.unsqueeze(0).transpose(0, 1) |
| self.register_buffer('pe', pe) |
|
|
| def forward(self, x): |
| |
| assert x.size(1) <= self.max_len |
| return x + self.pe[:x.size(0), :] |
|
|
| class TransformerBlock(nn.Module): |
| def __init__(self, dim, n_heads): |
| super().__init__() |
| self.attn = nn.MultiheadAttention(dim, n_heads) |
| self.ff = nn.Sequential( |
| nn.Linear(dim, 2*dim), |
| nn.ReLU(), |
| nn.Linear(2*dim, dim) |
| ) |
| self.norm1 = nn.LayerNorm(dim) |
| self.norm2 = nn.LayerNorm(dim) |
|
|
| def forward(self, x): |
| norm_x = self.norm1(x) |
| attn_output, _ = self.attn(norm_x, norm_x, norm_x) |
| x = x + attn_output |
| x = self.norm2(x) |
| ff_output = self.ff(x) |
| x = x + ff_output |
| return x |
|
|
| class TransformerModel(nn.Module): |
| def __init__(self, text_vocab, audio_vocab, dim, n_blocks, n_heads = 8): |
| super().__init__() |
| self.text_embed = nn.Embedding(text_vocab, dim) |
| self.audio_embed = nn.Embedding(audio_vocab, dim) |
| |
| self.blocks = nn.ModuleList([ |
| TransformerBlock(dim, n_heads) |
| for _ in range(n_blocks) |
| ]) |
| |
| self.pos_encoder = PositionalEncoding(dim) |
| self.lmhead = nn.Linear(dim, audio_vocab) |
| self.norm_in = nn.LayerNorm(dim) |
| self.norm_out = nn.LayerNorm(dim) |
|
|
| def forward(self, text, text_attention_mask, audio): |
| |
| if text is None and audio is not None: |
| audio = self.audio_embed(audio) |
| x = audio |
| elif text is not None and audio is not None: |
| text = self.text_embed(text) |
| audio = self.audio_embed(audio) |
|
|
| |
| text_list = [] |
| audio_list = [] |
| |
| for i in range(text.shape[0]): |
| masked_text = text[i][text_attention_mask[i] == 1] |
| text_list.append(masked_text) |
| audio_segment = audio[i] |
| audio_list.append(audio_segment) |
| |
| tensor_list = [] |
| max_len = max([len(x) + len(y) for x, y in zip(text_list, audio_list)]) |
| |
| for i in range(len(text_list)): |
| combine_tensor = torch.cat((text_list[i], audio_list[i]), dim=0) |
| if combine_tensor.shape[0] < max_len: |
| padding = audio_list[i][-1:].expand(max_len - combine_tensor.shape[0], -1) |
| combine_tensor = torch.cat((combine_tensor, padding), dim=0) |
| tensor_list.append(combine_tensor) |
| |
| x = torch.stack(tensor_list, dim=0) |
| |
| elif text is not None and audio is None: |
| x = self.text_embed(text) |
| |
| x = self.pos_encoder(x) |
| |
| x = self.norm_in(x) |
| for block in self.blocks: |
| x = block(x) |
| |
| return self.lmhead(self.norm_out(x)) |
| |
| def generate(self, audio, text, MAX_LENGTH, device, temperature=1.0): |
| tokens = [] |
| for i in range(MAX_LENGTH): |
| text_out = self.text_embed(text) |
| if audio is None: |
| x = text_out |
| else: |
| audio_out = self.audio_embed(audio) |
| x = torch.cat([text_out, audio_out], dim=1) |
|
|
| x = self.pos_encoder(x) |
| x = self.norm_in(x) |
| for block in self.blocks: |
| x = block(x) |
| x = self.lmhead(self.norm_out(x)) |
| |
| last_vector = x[:, -1, :] |
| probabilities = F.softmax(last_vector / temperature, dim=-1) |
| token_id = torch.argmax(probabilities, dim=-1) |
| print(token_id.item()) |
| |
| if token_id.item() == 8192: |
| print("ending with pad") |
| return torch.tensor(tokens).unsqueeze(0).to(device) |
|
|
| tokens.append(token_id.item()) |
| audio = torch.tensor(tokens).unsqueeze(0).to(device) |
|
|
| return torch.tensor(tokens).unsqueeze(0).to(device) |