|
|
from __future__ import annotations |
|
|
|
|
|
import torch |
|
|
from torch import nn, cat, is_tensor |
|
|
import torch.nn.functional as F |
|
|
from torch.nn import Module, ModuleList |
|
|
from torch.utils._pytree import tree_map |
|
|
|
|
|
from einops.layers.torch import Rearrange |
|
|
|
|
|
from rotary_embedding_torch import RotaryEmbedding |
|
|
|
|
|
|
|
|
|
|
|
def exists(v): |
|
|
return v is not None |
|
|
|
|
|
|
|
|
|
|
|
class NestedAttention(Module): |
|
|
def __init__( |
|
|
self, |
|
|
dim, |
|
|
*, |
|
|
dim_head = 64, |
|
|
heads = 8, |
|
|
prenorm = True, |
|
|
keys_rmsnorm = True |
|
|
): |
|
|
super().__init__() |
|
|
|
|
|
self.norm = nn.RMSNorm(dim) if prenorm else nn.Identity() |
|
|
|
|
|
dim_inner = dim_head * heads |
|
|
self.to_queries = nn.Linear(dim, dim_inner, bias = False) |
|
|
|
|
|
|
|
|
|
|
|
self.rotary_embed = RotaryEmbedding(dim_head) |
|
|
|
|
|
self.to_keys = nn.Linear(dim, dim_inner * 3, bias = False) |
|
|
self.to_values = nn.Linear(dim, dim_inner * 3, bias = False) |
|
|
|
|
|
self.key_norms = ModuleList([nn.RMSNorm(dim_head) for _ in range(3)]) |
|
|
self.nested_key_norm = nn.RMSNorm(dim_head) |
|
|
|
|
|
self.split_heads = Rearrange('b n (h d) -> b h n d', h = heads) |
|
|
self.merge_heads = Rearrange('b h n d -> b n (h d)') |
|
|
|
|
|
self.to_out = nn.Linear(dim_inner, dim, bias = False) |
|
|
|
|
|
def forward( |
|
|
self, |
|
|
tokens, |
|
|
cache = None, |
|
|
return_kv_cache = False |
|
|
): |
|
|
batch, seq_len, device = *tokens.shape[:2], tokens.device |
|
|
|
|
|
tokens = self.norm(tokens) |
|
|
|
|
|
queries = self.to_queries(tokens) |
|
|
|
|
|
keys = self.to_keys(tokens).chunk(3, dim = -1) |
|
|
values = self.to_values(tokens).chunk(3, dim = -1) |
|
|
|
|
|
|
|
|
|
|
|
queries, keys, values = tree_map(self.split_heads, (queries, keys, values)) |
|
|
|
|
|
|
|
|
|
|
|
keys = [norm(k) for norm, k in zip(self.key_norms, keys)] |
|
|
|
|
|
|
|
|
|
|
|
if exists(cache): |
|
|
(cache_keys, cache_values), (cache_nested_keys, cache_nested_values) = cache |
|
|
|
|
|
keys = [cat(args, dim = -2) for args in zip(cache_keys, keys)] |
|
|
values = [cat(args, dim = -2) for args in zip(cache_values, values)] |
|
|
|
|
|
|
|
|
|
|
|
def attend(q, k, v): |
|
|
q, k = self.rotary_embed.rotate_queries_with_cached_keys(q, k) |
|
|
|
|
|
return F.scaled_dot_product_attention(q, k, v, is_causal = True) |
|
|
|
|
|
|
|
|
|
|
|
nq, nk, nv = [attend(queries, key, value) for key, value in zip(keys, values)] |
|
|
|
|
|
nk = self.nested_key_norm(nk) |
|
|
|
|
|
if exists(cache): |
|
|
nk = cat((cache_nested_keys, nk), dim = -2) |
|
|
nv = cat((cache_nested_values, nv), dim = -2) |
|
|
|
|
|
out = attend(nq, nk, nv) |
|
|
|
|
|
|
|
|
|
|
|
out = self.merge_heads(out) |
|
|
|
|
|
out = self.to_out(out) |
|
|
|
|
|
if not return_kv_cache: |
|
|
return out |
|
|
|
|
|
return out, ((keys, values), (nk, nv)) |
|
|
|
|
|
if __name__ == '__main__': |
|
|
|
|
|
nested_attn = NestedAttention(512) |
|
|
|
|
|
tokens = torch.randn(1, 1024, 512) |
|
|
|
|
|
out1, cache = nested_attn(tokens) |
|
|
out2, cache = nested_attn(tokens[:, -1:], cache = cache) |
|
|
|
|
|
assert out1.shape == tokens.shape |
|
|
assert out2.shape == (1, 1, 512) |
|
|
|