titans_NPC / tests /test_titans.py
ChipYTY's picture
Update full code snapshot (exclude data and model checkpoints)
4196369 verified
from contextlib import contextmanager
import torch
from torch import nn
import pytest
from titans_pytorch import NeuralMemory
from titans_pytorch.mac_transformer import flex_attention, SegmentedAttention, MemoryAsContextTransformer
# functions
def exists(v):
return v is not None
def diff(x, y):
return (x - y).abs().amax()
@contextmanager
def torch_default_dtype(dtype):
prev_dtype = torch.get_default_dtype()
torch.set_default_dtype(dtype)
yield
torch.set_default_dtype(prev_dtype)
# main test
@pytest.mark.parametrize('seq_len', (32, 512, 77))
@pytest.mark.parametrize('silu', (False, True))
@pytest.mark.parametrize('chunk_size, attn_pool_chunks', ((64, True), (64, False), (1, False)))
@pytest.mark.parametrize('momentum', (False, True))
@pytest.mark.parametrize('qk_rmsnorm', (False, True))
@pytest.mark.parametrize('heads', (1, 4))
@pytest.mark.parametrize('max_grad_norm', (None, 2.))
@pytest.mark.parametrize('num_kv_per_token', (1, 2))
@pytest.mark.parametrize('per_parameter_lr_modulation', (False, True))
@pytest.mark.parametrize('per_head_learned_parameters', (False, True))
@pytest.mark.parametrize('test_store_mask', (False, True))
def test_titans(
seq_len,
silu,
attn_pool_chunks,
chunk_size,
momentum,
qk_rmsnorm,
heads,
max_grad_norm,
num_kv_per_token,
per_parameter_lr_modulation,
per_head_learned_parameters,
test_store_mask
):
mem = NeuralMemory(
dim = 16,
chunk_size = chunk_size,
activation = nn.SiLU() if silu else None,
attn_pool_chunks = attn_pool_chunks,
max_grad_norm = max_grad_norm,
num_kv_per_token = num_kv_per_token,
momentum = momentum,
qk_rmsnorm = qk_rmsnorm,
heads = heads,
per_parameter_lr_modulation = per_parameter_lr_modulation,
per_head_learned_parameters = per_head_learned_parameters
)
seq = torch.randn(2, seq_len, 16)
store_mask = None
if test_store_mask:
store_mask = torch.randint(0, 2, (2, seq_len)).bool()
retrieved, _ = mem(seq, store_mask = store_mask)
assert seq.shape == retrieved.shape
def test_return_surprises():
mem = NeuralMemory(
dim = 384,
chunk_size = 2,
dim_head = 64,
heads = 4,
)
seq = torch.randn(4, 64, 384)
_, _, (surprises, adaptive_lr) = mem(seq, return_surprises = True)
assert all([t.shape == (4, 4, 64) for t in (surprises, adaptive_lr)])
@pytest.mark.parametrize('learned_momentum_combine', (False, True))
@pytest.mark.parametrize('learned_combine_include_zeroth', (False, True))
def test_titans_second_order_momentum(
learned_momentum_combine,
learned_combine_include_zeroth
):
mem = NeuralMemory(
dim = 384,
dim_head = 64,
heads = 2,
chunk_size = 1,
batch_size = 2,
momentum_order = 2,
learned_momentum_combine = learned_momentum_combine,
learned_combine_include_zeroth = learned_combine_include_zeroth
)
seq = torch.randn(2, 5, 384)
parallel_retrieved, state = mem(seq)
assert seq.shape == parallel_retrieved.shape
def test_titans_attn_memory():
from titans_pytorch.memory_models import MemoryAttention
mem = NeuralMemory(
dim = 16,
chunk_size = 64,
model = MemoryAttention(
dim = 16
)
)
seq = torch.randn(2, 1024, 16)
retrieved, _ = mem(seq)
assert seq.shape == retrieved.shape
def test_swiglu_ff_memory():
from titans_pytorch.memory_models import MemorySwiGluMLP
mem = NeuralMemory(
dim = 16,
chunk_size = 2,
mem_model_norm_add_residual = False,
model = MemorySwiGluMLP(
dim = 16,
depth = 2
)
)
seq = torch.randn(2, 64, 16)
retrieved, _ = mem(seq)
assert seq.shape == retrieved.shape
@pytest.mark.parametrize('gated_transition', (True, False))
def test_neural_mem_chaining_chunks(
gated_transition
):
mem = NeuralMemory(
dim = 16,
dim_head = 16,
heads = 2,
chunk_size = 16,
gated_transition = gated_transition
)
seq = torch.randn(2, 48, 16)
parallel_retrieved, state = mem(seq)
seq_first, seq_second, seq_third = seq.split(16, dim = 1)
first_retrieved, state = mem(seq_first)
second_retrieved, state = mem(seq_second, state = state)
third_retrieved, state = mem(seq_third, state = state)
assert torch.allclose(parallel_retrieved, torch.cat((first_retrieved, second_retrieved, third_retrieved), dim = 1), atol = 1e-5)
def test_neural_mem_chaining_with_weight_residual():
mem = NeuralMemory(
dim = 16,
dim_head = 16,
heads = 2,
chunk_size = 64
)
mem2 = NeuralMemory(
dim = 16,
dim_head = 16,
heads = 2,
chunk_size = 64,
accept_weight_residual = True
)
seq = torch.randn(2, 256, 16)
seq, state = mem(seq)
parallel_retrieved, _ = mem2(seq, prev_weights = state.updates)
seq_first, seq_second = seq[:, :128], seq[:, 128:]
first_retrieved, state1 = mem2(seq_first, prev_weights = state.updates)
second_retrieved, state2 = mem2(seq_second, state = state1, prev_weights = state.updates)
assert torch.allclose(parallel_retrieved, torch.cat((first_retrieved, second_retrieved), dim = 1), atol = 1e-5)
def test_neural_mem_chaining_with_batch_size():
mem = NeuralMemory(
dim = 16,
dim_head = 16,
heads = 2,
chunk_size = 16,
batch_size = 64
)
seq = torch.randn(2, 112, 16)
parallel_retrieved, state = mem(seq)
seq_first, seq_second, seq_third = seq[:, :16], seq[:, 16:64], seq[:, 64:]
first_retrieved, state = mem(seq_first)
second_retrieved, state = mem(seq_second, state = state)
third_retrieved, state = mem(seq_third, state = state)
parallel_part_retrieved = torch.cat((first_retrieved, second_retrieved, third_retrieved), dim = 1)
assert torch.allclose(parallel_retrieved, parallel_part_retrieved, atol = 1e-5)
@pytest.mark.parametrize('seq_len', (1023, 17))
@pytest.mark.parametrize('num_persist_mem_tokens', (0, 16))
@pytest.mark.parametrize('num_longterm_mem_tokens', (0, 16))
@pytest.mark.parametrize('neural_mem_gate_attn_output', (False, True))
@pytest.mark.parametrize('neural_mem_segment_len', (8, 16))
@pytest.mark.parametrize('neural_mem_weight_residual', (False, True))
@pytest.mark.parametrize('neural_mem_batch_size', (None, 64))
@pytest.mark.parametrize('neural_mem_qkv_receives_diff_views', (False, True))
@pytest.mark.parametrize('neural_mem_momentum', (False, True))
def test_mac(
seq_len,
num_persist_mem_tokens,
num_longterm_mem_tokens,
neural_mem_gate_attn_output,
neural_mem_segment_len,
neural_mem_weight_residual,
neural_mem_batch_size,
neural_mem_qkv_receives_diff_views,
neural_mem_momentum
):
transformer = MemoryAsContextTransformer(
num_tokens = 256,
dim = 16,
depth = 2,
num_persist_mem_tokens = num_persist_mem_tokens,
num_longterm_mem_tokens = num_longterm_mem_tokens,
segment_len = 128,
neural_mem_gate_attn_output = neural_mem_gate_attn_output,
neural_memory_segment_len = neural_mem_segment_len,
neural_memory_batch_size = neural_mem_batch_size,
neural_memory_qkv_receives_diff_views = neural_mem_qkv_receives_diff_views,
neural_mem_weight_residual = neural_mem_weight_residual,
neural_memory_kwargs = dict(
momentum = neural_mem_momentum
)
)
x = torch.randint(0, 256, (1, seq_len))
logits = transformer(x)
assert logits.shape == (1, seq_len, 256)
@pytest.mark.parametrize('sliding', (False, True))
@pytest.mark.parametrize('mem_layers', ((), None))
@pytest.mark.parametrize('longterm_mems', (0, 4, 16))
@pytest.mark.parametrize('prompt_len', (4, 16))
@torch_default_dtype(torch.float64)
def test_mac_sampling(
sliding,
mem_layers,
longterm_mems,
prompt_len
):
transformer = MemoryAsContextTransformer(
num_tokens = 256,
dim = 16,
depth = 4,
segment_len = 32,
num_persist_mem_tokens = 4,
num_longterm_mem_tokens = longterm_mems,
sliding_window_attn = sliding,
neural_memory_layers = mem_layers,
neural_mem_gate_attn_output = False
)
ids = torch.randint(0, 256, (1, 1023))
# after much training
prompt = ids[:, :prompt_len]
sampled = transformer.sample(prompt, 53, use_cache = False, temperature = 0.)
sampled_with_cache = transformer.sample(prompt, 53, use_cache = True, temperature = 0.)
assert torch.allclose(sampled, sampled_with_cache)
@pytest.mark.parametrize('seq_len', (2, 64, 256))
@pytest.mark.parametrize('prompt_len', (0, 65))
@pytest.mark.parametrize('mem_chunk_size', (2, 32, 64))
@pytest.mark.parametrize('gated_transition', (False, True))
@torch_default_dtype(torch.float64)
def test_neural_mem_inference(
seq_len,
prompt_len,
mem_chunk_size,
gated_transition
):
mem = NeuralMemory(
dim = 16,
chunk_size = mem_chunk_size,
gated_transition = gated_transition
)
seq = torch.randn(2, seq_len, 16)
parallel_retrieved, _ = mem(seq)
assert seq.shape == parallel_retrieved.shape
state = None
sequential_retrieved = []
# test initial parallel prompt
test_parallel_prompt = prompt_len > 0 and prompt_len < seq_len
if test_parallel_prompt:
prompt, seq = seq[:, :prompt_len], seq[:, prompt_len:]
retrieved_prompt, state = mem(prompt)
sequential_retrieved.append(retrieved_prompt)
# sequential inference
for token in seq.unbind(dim = 1):
one_retrieved, state = mem.forward(
token,
state = state,
)
sequential_retrieved.append(one_retrieved)
sequential_retrieved = torch.cat(sequential_retrieved, dim = -2)
assert torch.allclose(parallel_retrieved, sequential_retrieved, atol = 1e-6)
@pytest.mark.parametrize('seq_len', (1023, 17))
@pytest.mark.parametrize('sliding', (True, False))
def test_flex(
seq_len,
sliding
):
if not (torch.cuda.is_available() and exists(flex_attention)):
pytest.skip()
attn = SegmentedAttention(
dim = 16,
segment_len = 32,
num_persist_mem_tokens = 1,
num_longterm_mem_tokens = 1,
use_flex_attn = True,
sliding = sliding
).cuda()
seq = torch.randn(1, seq_len, 16).cuda()
out_flex, _ = attn(seq)
out_non_flex, _ = attn(seq, disable_flex_attn = True)
assert torch.allclose(out_flex, out_non_flex, atol = 1e-5)
@pytest.mark.parametrize('use_accelerated', (True, False))
def test_assoc_scan(
use_accelerated
):
from titans_pytorch.neural_memory import AssocScan
if use_accelerated and not torch.cuda.is_available():
pytest.skip()
scan = AssocScan(use_accelerated = use_accelerated)
seq_len = 128
mid_point = seq_len // 2
gates = torch.randn(2, seq_len, 16).sigmoid()
inputs = torch.randn(2, seq_len, 16)
if use_accelerated:
gates = gates.cuda()
inputs = inputs.cuda()
output = scan(gates, inputs)
gates1, gates2 = gates[:, :mid_point], gates[:, mid_point:]
inputs1, inputs2 = inputs[:, :mid_point], inputs[:, mid_point:]
first_half = scan(gates1, inputs1)
second_half = scan(gates2, inputs2, prev = first_half[:, -1])
assert second_half.shape == inputs2.shape
assert torch.allclose(output[:, -1], second_half[:, -1], atol = 1e-5)
def test_mem_state_detach():
from titans_pytorch.neural_memory import mem_state_detach
mem = NeuralMemory(
dim = 384,
chunk_size = 2,
qk_rmsnorm = True,
dim_head = 64,
heads = 4,
)
seq = torch.randn(4, 64, 384)
state = None
for _ in range(2):
parallel_retrieved, state = mem(seq, state = state)
state = mem_state_detach(state)
parallel_retrieved.sum().backward()