|
|
""" |
|
|
Qwen + Titans 流式处理超长序列 |
|
|
|
|
|
核心思想: |
|
|
- Qwen 作为 Core(短期处理器),每次只处理一个 chunk |
|
|
- Titans NeuralMemory 作为长期记忆,跨 chunk 保持状态 |
|
|
- 虽然 Core 窗口有限(如 4k/8k),但整体能处理任意长度的上下文 |
|
|
|
|
|
处理流程: |
|
|
┌─────────────────────────────────────────────────────────────────────┐ |
|
|
│ 超长文档 (1M tokens) │ |
|
|
│ [chunk_0] [chunk_1] [chunk_2] ... [chunk_n-1] [chunk_n + question] │ |
|
|
└─────────────────────────────────────────────────────────────────────┘ |
|
|
│ │ │ │ │ |
|
|
▼ ▼ ▼ ▼ ▼ |
|
|
┌─────┐ ┌─────┐ ┌─────┐ ┌─────┐ ┌─────┐ |
|
|
│Qwen │ │Qwen │ │Qwen │ ... │Qwen │ │Qwen │ |
|
|
│Core │ │Core │ │Core │ │Core │ │Core │ |
|
|
└──┬──┘ └──┬──┘ └──┬──┘ └──┬──┘ └──┬──┘ |
|
|
│ │ │ │ │ |
|
|
▼ ▼ ▼ ▼ ▼ |
|
|
┌─────────────────────────────────────────────────────────┐ |
|
|
│ Titans Long-term Memory │ |
|
|
│ M_0 ──write──> M_1 ──write──> M_2 ... M_n-1 ──read──> │ |
|
|
│ │ |
|
|
│ 存储:关键事实、实体关系、重要信息 │ |
|
|
│ 检索:回答问题时取回相关记忆 │ |
|
|
└─────────────────────────────────────────────────────────┘ |
|
|
""" |
|
|
|
|
|
import torch |
|
|
import torch.nn as nn |
|
|
import torch.nn.functional as F |
|
|
from torch import Tensor |
|
|
from typing import Optional, List, Dict, Any, Tuple |
|
|
from dataclasses import dataclass |
|
|
from tqdm import tqdm |
|
|
import math |
|
|
|
|
|
from einops import rearrange, repeat |
|
|
|
|
|
|
|
|
from titans_pytorch import NeuralMemory, MemoryMLP |
|
|
from titans_pytorch.neural_memory import NeuralMemState |
|
|
|
|
|
|
|
|
@dataclass |
|
|
class StreamingConfig: |
|
|
"""流式处理配置""" |
|
|
chunk_size: int = 4096 |
|
|
memory_chunk_size: int = 64 |
|
|
memory_batch_size: int = 128 |
|
|
num_memory_tokens: int = 16 |
|
|
overlap_size: int = 128 |
|
|
|
|
|
|
|
|
class TitansLongTermMemory(nn.Module): |
|
|
""" |
|
|
Titans 长期记忆模块 |
|
|
|
|
|
负责: |
|
|
1. 将 chunk 的信息写入长期记忆 |
|
|
2. 从长期记忆中检索相关信息 |
|
|
3. 生成 memory tokens 注入到 Core 中 |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
hidden_size: int, |
|
|
chunk_size: int = 64, |
|
|
batch_size: int = 128, |
|
|
dim_head: int = 64, |
|
|
num_heads: int = None, |
|
|
memory_depth: int = 2, |
|
|
): |
|
|
super().__init__() |
|
|
|
|
|
self.hidden_size = hidden_size |
|
|
num_heads = num_heads or (hidden_size // dim_head) |
|
|
|
|
|
|
|
|
memory_model = MemoryMLP( |
|
|
dim=dim_head, |
|
|
depth=memory_depth, |
|
|
expansion_factor=2.0 |
|
|
) |
|
|
|
|
|
|
|
|
self.neural_memory = NeuralMemory( |
|
|
dim=hidden_size, |
|
|
chunk_size=chunk_size, |
|
|
batch_size=batch_size, |
|
|
dim_head=dim_head, |
|
|
heads=num_heads, |
|
|
model=memory_model, |
|
|
momentum=True, |
|
|
momentum_order=1, |
|
|
qk_rmsnorm=True, |
|
|
pre_rmsnorm=True, |
|
|
default_step_transform_max_lr=0.1, |
|
|
) |
|
|
|
|
|
|
|
|
|
|
|
self.memory_query_tokens = nn.Parameter( |
|
|
torch.randn(1, 16, hidden_size) * 0.02 |
|
|
) |
|
|
|
|
|
|
|
|
self.memory_proj = nn.Sequential( |
|
|
nn.LayerNorm(hidden_size), |
|
|
nn.Linear(hidden_size, hidden_size), |
|
|
nn.GELU(), |
|
|
nn.Linear(hidden_size, hidden_size), |
|
|
) |
|
|
|
|
|
def write( |
|
|
self, |
|
|
hidden_states: Tensor, |
|
|
state: Optional[NeuralMemState] = None, |
|
|
) -> Tuple[Tensor, NeuralMemState]: |
|
|
""" |
|
|
将当前 chunk 的信息写入长期记忆 |
|
|
|
|
|
Args: |
|
|
hidden_states: 当前 chunk 经过 Qwen 处理后的隐藏状态 |
|
|
state: 之前的记忆状态 |
|
|
|
|
|
Returns: |
|
|
retrieved: 从记忆中检索到的信息 |
|
|
next_state: 更新后的记忆状态 |
|
|
""" |
|
|
|
|
|
retrieved, next_state = self.neural_memory( |
|
|
hidden_states, |
|
|
state=state |
|
|
) |
|
|
|
|
|
return retrieved, next_state |
|
|
|
|
|
def read( |
|
|
self, |
|
|
batch_size: int, |
|
|
state: NeuralMemState, |
|
|
num_tokens: int = 16, |
|
|
) -> Tensor: |
|
|
""" |
|
|
从长期记忆中读取信息,生成 memory tokens |
|
|
|
|
|
这些 tokens 会被注入到 Core 的输入中 |
|
|
""" |
|
|
|
|
|
queries = repeat( |
|
|
self.memory_query_tokens[:, :num_tokens], |
|
|
'1 n d -> b n d', |
|
|
b=batch_size |
|
|
) |
|
|
|
|
|
|
|
|
retrieved, _ = self.neural_memory( |
|
|
queries, |
|
|
state=state |
|
|
) |
|
|
|
|
|
|
|
|
memory_tokens = self.memory_proj(retrieved) |
|
|
|
|
|
return memory_tokens |
|
|
|
|
|
|
|
|
class QwenTitansStreaming(nn.Module): |
|
|
""" |
|
|
Qwen + Titans 流式处理模型 |
|
|
|
|
|
能够处理任意长度的序列,通过: |
|
|
1. 将序列分成 chunks |
|
|
2. 每个 chunk 用 Qwen Core 处理 |
|
|
3. 用 Titans 长期记忆跨 chunk 传递信息 |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
qwen_model, |
|
|
config: StreamingConfig = None, |
|
|
): |
|
|
super().__init__() |
|
|
|
|
|
self.qwen = qwen_model |
|
|
self.config = config or StreamingConfig() |
|
|
self.hidden_size = qwen_model.config.hidden_size |
|
|
|
|
|
|
|
|
self.long_term_memory = TitansLongTermMemory( |
|
|
hidden_size=self.hidden_size, |
|
|
chunk_size=self.config.memory_chunk_size, |
|
|
batch_size=self.config.memory_batch_size, |
|
|
) |
|
|
|
|
|
|
|
|
self.memory_gate = nn.Sequential( |
|
|
nn.Linear(self.hidden_size * 2, self.hidden_size), |
|
|
nn.Sigmoid() |
|
|
) |
|
|
|
|
|
print(f"[QwenTitansStreaming] 初始化完成:") |
|
|
print(f" - Chunk size: {self.config.chunk_size}") |
|
|
print(f" - Memory chunk size: {self.config.memory_chunk_size}") |
|
|
print(f" - Memory batch size: {self.config.memory_batch_size}") |
|
|
print(f" - Overlap size: {self.config.overlap_size}") |
|
|
|
|
|
def _split_into_chunks( |
|
|
self, |
|
|
input_ids: Tensor, |
|
|
chunk_size: int, |
|
|
overlap: int = 0, |
|
|
) -> List[Tensor]: |
|
|
""" |
|
|
将输入序列分成 chunks |
|
|
|
|
|
Args: |
|
|
input_ids: [batch, seq_len] |
|
|
chunk_size: 每个 chunk 的大小 |
|
|
overlap: chunk 之间的重叠 |
|
|
|
|
|
Returns: |
|
|
List of chunks, each [batch, chunk_size] |
|
|
""" |
|
|
batch_size, seq_len = input_ids.shape |
|
|
chunks = [] |
|
|
|
|
|
stride = chunk_size - overlap |
|
|
|
|
|
for start in range(0, seq_len, stride): |
|
|
end = min(start + chunk_size, seq_len) |
|
|
chunk = input_ids[:, start:end] |
|
|
|
|
|
|
|
|
if chunk.shape[1] < chunk_size: |
|
|
pad_len = chunk_size - chunk.shape[1] |
|
|
chunk = F.pad(chunk, (0, pad_len), value=0) |
|
|
|
|
|
chunks.append(chunk) |
|
|
|
|
|
if end >= seq_len: |
|
|
break |
|
|
|
|
|
return chunks |
|
|
|
|
|
def process_document( |
|
|
self, |
|
|
input_ids: Tensor, |
|
|
attention_mask: Optional[Tensor] = None, |
|
|
return_all_hidden_states: bool = False, |
|
|
show_progress: bool = True, |
|
|
) -> Dict[str, Any]: |
|
|
""" |
|
|
流式处理整个文档 |
|
|
|
|
|
这是核心方法: |
|
|
1. 将文档分成 chunks |
|
|
2. 逐个 chunk 处理 |
|
|
3. 每个 chunk 后更新长期记忆 |
|
|
|
|
|
Args: |
|
|
input_ids: [batch, seq_len] - 可以是任意长度! |
|
|
attention_mask: [batch, seq_len] |
|
|
|
|
|
Returns: |
|
|
包含最终隐藏状态、记忆状态等的字典 |
|
|
""" |
|
|
batch_size, total_seq_len = input_ids.shape |
|
|
device = input_ids.device |
|
|
|
|
|
|
|
|
chunks = self._split_into_chunks( |
|
|
input_ids, |
|
|
self.config.chunk_size, |
|
|
self.config.overlap_size |
|
|
) |
|
|
|
|
|
num_chunks = len(chunks) |
|
|
print(f"[process_document] 总长度: {total_seq_len}, 分成 {num_chunks} 个 chunks") |
|
|
|
|
|
|
|
|
memory_state = None |
|
|
all_hidden_states = [] |
|
|
|
|
|
|
|
|
iterator = tqdm(enumerate(chunks), total=num_chunks, desc="Processing chunks") \ |
|
|
if show_progress else enumerate(chunks) |
|
|
|
|
|
for chunk_idx, chunk_ids in iterator: |
|
|
|
|
|
|
|
|
|
|
|
memory_tokens = None |
|
|
if memory_state is not None and chunk_idx > 0: |
|
|
memory_tokens = self.long_term_memory.read( |
|
|
batch_size=batch_size, |
|
|
state=memory_state, |
|
|
num_tokens=self.config.num_memory_tokens |
|
|
) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
chunk_hidden = self._process_chunk_with_memory( |
|
|
chunk_ids, |
|
|
memory_tokens=memory_tokens, |
|
|
) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
_, memory_state = self.long_term_memory.write( |
|
|
chunk_hidden, |
|
|
state=memory_state |
|
|
) |
|
|
|
|
|
if return_all_hidden_states: |
|
|
all_hidden_states.append(chunk_hidden) |
|
|
|
|
|
|
|
|
result = { |
|
|
'last_hidden_states': chunk_hidden, |
|
|
'memory_state': memory_state, |
|
|
'num_chunks_processed': num_chunks, |
|
|
} |
|
|
|
|
|
if return_all_hidden_states: |
|
|
result['all_hidden_states'] = all_hidden_states |
|
|
|
|
|
return result |
|
|
|
|
|
def _process_chunk_with_memory( |
|
|
self, |
|
|
chunk_ids: Tensor, |
|
|
memory_tokens: Optional[Tensor] = None, |
|
|
) -> Tensor: |
|
|
""" |
|
|
处理单个 chunk,可选地注入 memory tokens |
|
|
|
|
|
Args: |
|
|
chunk_ids: [batch, chunk_size] |
|
|
memory_tokens: [batch, num_mem_tokens, hidden] - 从长期记忆检索的 |
|
|
|
|
|
Returns: |
|
|
hidden_states: [batch, chunk_size, hidden] |
|
|
""" |
|
|
batch_size = chunk_ids.shape[0] |
|
|
|
|
|
|
|
|
if hasattr(self.qwen.model, 'embed_tokens'): |
|
|
hidden_states = self.qwen.model.embed_tokens(chunk_ids) |
|
|
else: |
|
|
hidden_states = self.qwen.get_input_embeddings()(chunk_ids) |
|
|
|
|
|
|
|
|
if memory_tokens is not None: |
|
|
|
|
|
hidden_states = torch.cat([memory_tokens, hidden_states], dim=1) |
|
|
|
|
|
|
|
|
for layer in self.qwen.model.layers: |
|
|
layer_output = layer(hidden_states, attention_mask=None) |
|
|
if isinstance(layer_output, tuple): |
|
|
hidden_states = layer_output[0] |
|
|
else: |
|
|
hidden_states = layer_output |
|
|
|
|
|
|
|
|
hidden_states = self.qwen.model.norm(hidden_states) |
|
|
|
|
|
|
|
|
if memory_tokens is not None: |
|
|
num_mem = memory_tokens.shape[1] |
|
|
hidden_states = hidden_states[:, num_mem:] |
|
|
|
|
|
return hidden_states |
|
|
|
|
|
def generate_answer( |
|
|
self, |
|
|
question_ids: Tensor, |
|
|
memory_state: NeuralMemState, |
|
|
max_new_tokens: int = 100, |
|
|
temperature: float = 0.7, |
|
|
) -> Tensor: |
|
|
""" |
|
|
基于长期记忆生成答案 |
|
|
|
|
|
关键:虽然 Core 只看到问题,但它能从长期记忆中 |
|
|
检索到之前 1M tokens 中的相关事实! |
|
|
|
|
|
Args: |
|
|
question_ids: [batch, question_len] - 问题的 token ids |
|
|
memory_state: 处理完整个文档后的记忆状态 |
|
|
max_new_tokens: 最大生成长度 |
|
|
|
|
|
Returns: |
|
|
generated_ids: [batch, question_len + generated_len] |
|
|
""" |
|
|
batch_size = question_ids.shape[0] |
|
|
generated = question_ids.clone() |
|
|
|
|
|
for _ in range(max_new_tokens): |
|
|
|
|
|
memory_tokens = self.long_term_memory.read( |
|
|
batch_size=batch_size, |
|
|
state=memory_state, |
|
|
num_tokens=self.config.num_memory_tokens |
|
|
) |
|
|
|
|
|
|
|
|
hidden = self._process_chunk_with_memory( |
|
|
generated, |
|
|
memory_tokens=memory_tokens |
|
|
) |
|
|
|
|
|
|
|
|
logits = self.qwen.lm_head(hidden[:, -1:, :]) |
|
|
|
|
|
if temperature > 0: |
|
|
probs = F.softmax(logits.squeeze(1) / temperature, dim=-1) |
|
|
next_token = torch.multinomial(probs, num_samples=1) |
|
|
else: |
|
|
next_token = logits.squeeze(1).argmax(dim=-1, keepdim=True) |
|
|
|
|
|
generated = torch.cat([generated, next_token], dim=1) |
|
|
|
|
|
|
|
|
if hasattr(self.qwen.config, 'eos_token_id'): |
|
|
if (next_token == self.qwen.config.eos_token_id).all(): |
|
|
break |
|
|
|
|
|
return generated |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def babilong_style_example(): |
|
|
""" |
|
|
演示如何用 Qwen + Titans 处理 BABILong 风格的超长序列任务 |
|
|
|
|
|
BABILong 任务结构: |
|
|
- 前面是很长的背景文档(可能 1M tokens) |
|
|
- 最后是一个问题 |
|
|
- 需要从文档中找到相关事实来回答 |
|
|
""" |
|
|
|
|
|
print("=" * 70) |
|
|
print("Qwen + Titans 流式处理超长序列示例") |
|
|
print("=" * 70) |
|
|
|
|
|
try: |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
|
|
|
|
|
print("\n[1] 加载 Qwen 模型...") |
|
|
model_name = "Qwen/Qwen2-0.5B" |
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True) |
|
|
qwen_model = AutoModelForCausalLM.from_pretrained( |
|
|
model_name, |
|
|
torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32, |
|
|
device_map="auto" if torch.cuda.is_available() else None, |
|
|
trust_remote_code=True |
|
|
) |
|
|
|
|
|
|
|
|
print("\n[2] 创建 QwenTitansStreaming 模型...") |
|
|
config = StreamingConfig( |
|
|
chunk_size=2048, |
|
|
memory_chunk_size=64, |
|
|
memory_batch_size=128, |
|
|
num_memory_tokens=16, |
|
|
overlap_size=64, |
|
|
) |
|
|
|
|
|
model = QwenTitansStreaming(qwen_model, config) |
|
|
|
|
|
if torch.cuda.is_available(): |
|
|
model = model.cuda() |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
print("\n[3] 模拟 BABILong 任务...") |
|
|
|
|
|
|
|
|
long_document = """ |
|
|
这是一个关于人工智能发展历史的长篇文档。 |
|
|
|
|
|
第一章:早期发展 |
|
|
人工智能的概念最早可以追溯到 1950 年代。1956 年的达特茅斯会议 |
|
|
被认为是人工智能作为一门学科正式诞生的标志。 |
|
|
|
|
|
[这里假设有很多很多内容...] |
|
|
|
|
|
重要事实:达特茅斯会议在 1956 年举行。 |
|
|
|
|
|
[更多内容...] |
|
|
|
|
|
第五十章:现代发展 |
|
|
2022 年,大型语言模型取得了突破性进展。 |
|
|
|
|
|
重要事实:GPT-4 在 2023 年发布。 |
|
|
|
|
|
[更多内容...] |
|
|
""" |
|
|
|
|
|
|
|
|
|
|
|
very_long_document = long_document * 100 |
|
|
|
|
|
question = "\n问题:达特茅斯会议是在哪一年举行的?" |
|
|
full_input = very_long_document + question |
|
|
|
|
|
|
|
|
print(f" 文档长度(字符): {len(full_input)}") |
|
|
inputs = tokenizer(full_input, return_tensors="pt") |
|
|
input_ids = inputs.input_ids |
|
|
print(f" 文档长度(tokens): {input_ids.shape[1]}") |
|
|
|
|
|
device = next(model.parameters()).device |
|
|
input_ids = input_ids.to(device) |
|
|
|
|
|
|
|
|
print("\n[4] 流式处理文档...") |
|
|
with torch.no_grad(): |
|
|
result = model.process_document( |
|
|
input_ids, |
|
|
show_progress=True |
|
|
) |
|
|
|
|
|
print(f"\n 处理完成!") |
|
|
print(f" - 处理了 {result['num_chunks_processed']} 个 chunks") |
|
|
print(f" - 记忆状态 seq_index: {result['memory_state'].seq_index}") |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
except ImportError as e: |
|
|
print(f"\n需要安装依赖: pip install transformers") |
|
|
print(f"错误: {e}") |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
print("\n" + "=" * 70) |
|
|
print("[独立测试] Titans 长期记忆模块") |
|
|
print("=" * 70) |
|
|
|
|
|
device = 'cuda' if torch.cuda.is_available() else 'cpu' |
|
|
|
|
|
|
|
|
ltm = TitansLongTermMemory( |
|
|
hidden_size=512, |
|
|
chunk_size=64, |
|
|
batch_size=128, |
|
|
dim_head=64, |
|
|
num_heads=8, |
|
|
).to(device) |
|
|
|
|
|
|
|
|
batch_size = 2 |
|
|
chunk_size = 256 |
|
|
hidden_dim = 512 |
|
|
num_chunks = 5 |
|
|
|
|
|
print(f"\n模拟处理 {num_chunks} 个 chunks:") |
|
|
|
|
|
memory_state = None |
|
|
for i in range(num_chunks): |
|
|
|
|
|
chunk_hidden = torch.randn(batch_size, chunk_size, hidden_dim).to(device) |
|
|
|
|
|
|
|
|
retrieved, memory_state = ltm.write(chunk_hidden, state=memory_state) |
|
|
|
|
|
print(f" Chunk {i}: 写入完成, seq_index = {memory_state.seq_index}") |
|
|
|
|
|
|
|
|
print(f"\n从长期记忆中读取:") |
|
|
memory_tokens = ltm.read(batch_size, memory_state, num_tokens=16) |
|
|
print(f" Memory tokens 形状: {memory_tokens.shape}") |
|
|
|
|
|
print("\n" + "=" * 70) |
|
|
print("完成!") |
|
|
print("=" * 70) |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
babilong_style_example() |
|
|
|