File size: 26,712 Bytes
4196369
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
"""
MAC (Memory-as-Context) 结构集成到 Qwen 的详细实现

=== MAC 结构核心原理 ===

1. 将长序列分成多个 segment(如每 128 个 token 一个 segment)
2. 在每个 segment 的【开头】插入 longterm_mem_tokens(如 16 个)
3. 这些 memory tokens 会参与 attention 计算
4. 使用 NeuralMemory 模块来动态更新这些 memory tokens 的内容

原始序列: [t1, t2, t3, ..., t128, t129, ..., t256, ...]

MAC 序列:  [M1..M16, t1...t128, M1..M16, t129...t256, ...]
            ↑                    ↑
         memory tokens       memory tokens

=== Qwen2 架构 ===

Qwen2DecoderLayer:
├── input_layernorm (RMSNorm)
├── self_attn (Qwen2Attention/Qwen2FlashAttention2)
│   ├── q_proj, k_proj, v_proj
│   ├── RoPE (rotary positional embedding)
│   └── o_proj
├── post_attention_layernorm (RMSNorm)
└── mlp (Qwen2MLP: gate_proj, up_proj, down_proj with SiLU)

我们需要在特定层添加 NeuralMemory 模块
"""

import torch
import torch.nn as nn
import torch.nn.functional as F
from torch import Tensor
from typing import Optional, Tuple, List, Dict, Any
from copy import deepcopy
from functools import partial

from einops import rearrange, repeat, pack, unpack

# Titans 组件
from titans_pytorch import NeuralMemory, MemoryMLP
from titans_pytorch.neural_memory import NeuralMemState


# ============================================================================
# 辅助函数
# ============================================================================

def exists(v):
    return v is not None

def default(v, d):
    return v if exists(v) else d

def divisible_by(num, den):
    return (num % den) == 0

def round_up_multiple(seq, mult):
    return ((seq + mult - 1) // mult) * mult


# ============================================================================
# MAC 风格的 Qwen 实现
# ============================================================================

class QwenMACTransformer(nn.Module):
    """
    将 MAC 结构应用到 Qwen 模型的完整实现
    
    架构图:
    
    Input IDs


    ┌─────────────────┐
    │  Token Embed    │
    └────────┬────────┘


    ┌─────────────────────────────────────────┐
    │  插入 Memory Tokens 到每个 Segment 开头   │
    │  [M1..Mn, t1..t_seg, M1..Mn, ...]       │
    └────────┬────────────────────────────────┘


    ╔═════════════════════════════════════════╗
    ║         Qwen Decoder Layer 1            ║
    ║  ┌────────────────────────────────┐     ║
    ║  │ RMSNorm → Self-Attention → Add │     ║
    ║  └────────────────────────────────┘     ║
    ║  ┌────────────────────────────────┐     ║
    ║  │ RMSNorm → MLP → Add            │     ║
    ║  └────────────────────────────────┘     ║
    ╚═════════════════════════════════════════╝


    ╔═════════════════════════════════════════╗
    ║    Qwen Decoder Layer 2 (with Memory)   ║
    ║  ┌────────────────────────────────┐     ║
    ║  │ RMSNorm → Self-Attention → Add │     ║
    ║  └────────────────────────────────┘     ║
    ║  ┌──────────────────────────────────┐   ║
    ║  │  ★ NeuralMemory 记忆增强 ★        │   ║
    ║  │  retrieved = mem(hidden_states)  │   ║
    ║  │  hidden += gate * retrieved      │   ║
    ║  └──────────────────────────────────┘   ║
    ║  ┌────────────────────────────────┐     ║
    ║  │ RMSNorm → MLP → Add            │     ║
    ║  └────────────────────────────────┘     ║
    ╚═════════════════════════════════════════╝


          ... 更多层 ...


    ┌─────────────────┐
    │  Final RMSNorm  │
    └────────┬────────┘


    ┌─────────────────┐
    │    LM Head      │
    └────────┬────────┘


        Logits
    """
    
    def __init__(
        self,
        qwen_model,
        # === Segment 配置 ===
        segment_len: int = 128,              # 每个 segment 的长度
        num_longterm_mem_tokens: int = 16,   # 每个 segment 开头的 memory token 数量
        num_persist_mem_tokens: int = 4,     # 全局持久 memory token 数量
        
        # === NeuralMemory 配置 ===
        neural_memory_layers: Tuple[int, ...] = (2, 4, 6),  # 哪些层使用记忆
        memory_chunk_size: int = 64,         # 记忆模块的 chunk 大小
        memory_batch_size: int = 128,        # 记忆更新的批次大小
        memory_depth: int = 2,               # 记忆 MLP 的深度
        
        # === 其他配置 ===
        dim_head: int = 64,
        num_heads: int = None,               # 默认从模型配置读取
        use_momentum: bool = True,
        gate_memory_output: bool = False,    # 是否用记忆门控 attention 输出
    ):
        super().__init__()
        
        # 保存原始 Qwen 模型
        self.qwen = qwen_model
        self.config = qwen_model.config
        
        # 获取模型维度
        self.hidden_size = self.config.hidden_size
        self.num_layers = self.config.num_hidden_layers
        num_heads = default(num_heads, self.hidden_size // dim_head)
        
        # Segment 配置
        self.segment_len = segment_len
        self.num_longterm_mem_tokens = num_longterm_mem_tokens
        self.num_persist_mem_tokens = num_persist_mem_tokens
        self.total_segment_len = segment_len + num_longterm_mem_tokens
        
        # =====================================================================
        # Memory Tokens (这是 MAC 的核心!)
        # =====================================================================
        
        # 持久记忆 tokens - 放在序列最前面,所有 segment 共享
        # 用于存储全局上下文信息
        self.persist_mem_tokens = nn.Parameter(
            torch.randn(num_persist_mem_tokens, self.hidden_size) * 0.02
        )
        
        # 长期记忆 tokens - 插入到每个 segment 的开头
        # 这些 token 会被 NeuralMemory 动态更新
        self.longterm_mem_tokens = nn.Parameter(
            torch.randn(num_longterm_mem_tokens, self.hidden_size) * 0.02
        )
        
        # =====================================================================
        # NeuralMemory 模块
        # =====================================================================
        
        self.neural_memory_layers = neural_memory_layers
        self.gate_memory_output = gate_memory_output
        
        # 为每个指定层创建 NeuralMemory
        self.neural_memories = nn.ModuleDict()
        self.memory_projections = nn.ModuleDict()  # 投影层
        self.memory_gates = nn.ModuleDict()        # 门控层
        
        # 创建记忆网络模板
        memory_model_template = MemoryMLP(
            dim=dim_head,
            depth=memory_depth,
            expansion_factor=2.0
        )
        
        for layer_idx in neural_memory_layers:
            layer_key = str(layer_idx)
            
            # NeuralMemory 模块
            self.neural_memories[layer_key] = NeuralMemory(
                dim=self.hidden_size,
                chunk_size=memory_chunk_size,
                batch_size=memory_batch_size,
                dim_head=dim_head,
                heads=num_heads,
                model=deepcopy(memory_model_template),
                momentum=use_momentum,
                momentum_order=1,
                qk_rmsnorm=True,
                pre_rmsnorm=True,
                default_step_transform_max_lr=0.1,
            )
            
            # 门控层 - 控制记忆的影响程度
            self.memory_gates[layer_key] = nn.Sequential(
                nn.Linear(self.hidden_size, self.hidden_size),
                nn.Sigmoid()
            )
        
        print(f"[QwenMAC] 初始化完成:")
        print(f"  - 隐藏层大小: {self.hidden_size}")
        print(f"  - 层数: {self.num_layers}")
        print(f"  - Segment 长度: {segment_len}")
        print(f"  - Longterm Memory Tokens: {num_longterm_mem_tokens}")
        print(f"  - Persist Memory Tokens: {num_persist_mem_tokens}")
        print(f"  - 记忆层: {neural_memory_layers}")
    
    def _insert_memory_tokens(
        self,
        hidden_states: Tensor,  # [batch, seq_len, hidden]
        batch_size: int,
        seq_len: int,
    ) -> Tuple[Tensor, int]:
        """
        在序列中插入 memory tokens
        
        输入: [batch, seq_len, hidden]
        输出: [batch, new_seq_len, hidden]
        
        处理流程:
        原始: [t1, t2, ..., t128, t129, ..., t256]
        
        1. 分成 segments:
           Seg1: [t1, ..., t128]
           Seg2: [t129, ..., t256]
        
        2. 每个 segment 前插入 longterm_mem:
           Seg1: [M1, ..., M16, t1, ..., t128]
           Seg2: [M1, ..., M16, t129, ..., t256]
        
        3. 合并 + 前置 persist_mem:
           [P1, ..., P4, M1..M16, t1..t128, M1..M16, t129..t256]
        """
        segment_len = self.segment_len
        num_longterm = self.num_longterm_mem_tokens
        num_persist = self.num_persist_mem_tokens
        
        # 计算需要多少个 segment
        num_segments = (seq_len + segment_len - 1) // segment_len
        
        # Padding 到 segment_len 的整数倍
        padded_len = num_segments * segment_len
        if seq_len < padded_len:
            padding = padded_len - seq_len
            hidden_states = F.pad(hidden_states, (0, 0, 0, padding))
        
        # 重塑为 segments: [batch, num_segments, segment_len, hidden]
        hidden_states = rearrange(
            hidden_states,
            'b (s n) d -> b s n d',
            s=num_segments,
            n=segment_len
        )
        
        # 扩展 longterm memory tokens: [batch, num_segments, num_longterm, hidden]
        longterm_mem = repeat(
            self.longterm_mem_tokens,
            'n d -> b s n d',
            b=batch_size,
            s=num_segments
        )
        
        # 在每个 segment 前插入 memory tokens
        # [batch, num_segments, num_longterm + segment_len, hidden]
        hidden_states = torch.cat([longterm_mem, hidden_states], dim=2)
        
        # 展平 segments: [batch, num_segments * (num_longterm + segment_len), hidden]
        hidden_states = rearrange(hidden_states, 'b s n d -> b (s n) d')
        
        # 添加持久记忆 tokens 在最前面
        persist_mem = repeat(
            self.persist_mem_tokens,
            'n d -> b n d',
            b=batch_size
        )
        hidden_states = torch.cat([persist_mem, hidden_states], dim=1)
        
        new_seq_len = hidden_states.shape[1]
        
        return hidden_states, new_seq_len
    
    def _remove_memory_tokens(
        self,
        hidden_states: Tensor,
        original_seq_len: int,
    ) -> Tensor:
        """
        从输出中移除 memory tokens,恢复原始序列长度
        """
        segment_len = self.segment_len
        num_longterm = self.num_longterm_mem_tokens
        num_persist = self.num_persist_mem_tokens
        total_segment_len = segment_len + num_longterm
        
        batch_size = hidden_states.shape[0]
        
        # 移除 persist tokens
        hidden_states = hidden_states[:, num_persist:]
        
        # 计算 segments
        num_segments = (original_seq_len + segment_len - 1) // segment_len
        
        # 重塑为 segments
        hidden_states = rearrange(
            hidden_states,
            'b (s n) d -> b s n d',
            s=num_segments,
            n=total_segment_len
        )
        
        # 移除每个 segment 开头的 memory tokens
        hidden_states = hidden_states[:, :, num_longterm:, :]
        
        # 展平并截取原始长度
        hidden_states = rearrange(hidden_states, 'b s n d -> b (s n) d')
        hidden_states = hidden_states[:, :original_seq_len, :]
        
        return hidden_states
    
    def _create_mac_attention_mask(
        self,
        seq_len_with_mem: int,
        device: torch.device,
        dtype: torch.dtype,
    ) -> Tensor:
        """
        创建 MAC 风格的 attention mask
        
        MAC mask 的特点:
        1. Persist memory tokens 对所有位置可见
        2. 每个 segment 内部是 causal 的
        3. Memory tokens 可以 attend 到之前的 segment
        
        这是一个简化版本,完整版需要考虑更多细节
        """
        # 创建基础 causal mask
        mask = torch.ones(seq_len_with_mem, seq_len_with_mem, device=device, dtype=dtype)
        mask = torch.tril(mask)
        
        # Persist memory 对所有位置可见
        num_persist = self.num_persist_mem_tokens
        mask[:, :num_persist] = 1.0
        
        return mask
    
    def forward(
        self,
        input_ids: Tensor,
        attention_mask: Optional[Tensor] = None,
        position_ids: Optional[Tensor] = None,
        memory_states: Optional[Dict[str, NeuralMemState]] = None,
        return_memory_states: bool = True,
        **kwargs
    ) -> Dict[str, Any]:
        """
        前向传播
        
        Args:
            input_ids: [batch, seq_len]
            attention_mask: [batch, seq_len]
            memory_states: 各层的记忆状态(用于增量推理)
            
        Returns:
            dict with 'logits', 'hidden_states', 'memory_states'
        """
        batch_size, seq_len = input_ids.shape
        device = input_ids.device
        
        # =====================================================================
        # Step 1: Token Embedding
        # =====================================================================
        if hasattr(self.qwen.model, 'embed_tokens'):
            hidden_states = self.qwen.model.embed_tokens(input_ids)
        else:
            hidden_states = self.qwen.get_input_embeddings()(input_ids)
        
        # =====================================================================
        # Step 2: 插入 Memory Tokens
        # =====================================================================
        hidden_states, seq_len_with_mem = self._insert_memory_tokens(
            hidden_states, batch_size, seq_len
        )
        
        # =====================================================================
        # Step 3: 创建 Attention Mask (简化版)
        # =====================================================================
        # 注意: 完整实现需要更复杂的 mask 处理
        # 这里用 None 让模型使用默认 causal mask
        mac_attention_mask = None
        
        # =====================================================================
        # Step 4: 逐层处理
        # =====================================================================
        if memory_states is None:
            memory_states = {}
        next_memory_states = {}
        
        # 遍历 Qwen 的所有层
        for layer_idx, layer in enumerate(self.qwen.model.layers):
            layer_key = str(layer_idx)
            
            # -----------------------------------------------------------------
            # 4.1 标准的 Qwen Decoder Layer 前向传播
            # -----------------------------------------------------------------
            # Qwen2DecoderLayer.forward() 的简化版本
            
            residual = hidden_states
            
            # Input LayerNorm
            hidden_states = layer.input_layernorm(hidden_states)
            
            # Self Attention
            # 注意: 这里简化了 attention 的调用,实际可能需要更多参数
            attn_output = layer.self_attn(
                hidden_states=hidden_states,
                attention_mask=mac_attention_mask,
                position_ids=None,  # 会自动生成
            )
            
            # 处理不同返回格式
            if isinstance(attn_output, tuple):
                attn_output = attn_output[0]
            
            hidden_states = residual + attn_output
            
            # -----------------------------------------------------------------
            # 4.2 NeuralMemory 记忆增强 (仅在指定层)
            # -----------------------------------------------------------------
            if layer_key in self.neural_memories:
                neural_mem = self.neural_memories[layer_key]
                gate_fn = self.memory_gates[layer_key]
                
                # 获取该层的记忆状态
                mem_state = memory_states.get(layer_key)
                
                # 记忆检索和更新
                retrieved, next_mem_state = neural_mem(
                    hidden_states,
                    state=mem_state
                )
                
                # 门控融合
                gate = gate_fn(hidden_states)
                
                if self.gate_memory_output:
                    # 方式1: 用检索到的记忆门控后续输出
                    hidden_states = hidden_states * (1 + gate * retrieved.sigmoid())
                else:
                    # 方式2: 直接加上门控后的记忆(更常用)
                    hidden_states = hidden_states + gate * retrieved
                
                # 保存记忆状态
                next_memory_states[layer_key] = next_mem_state
            
            # -----------------------------------------------------------------
            # 4.3 Feed Forward Network
            # -----------------------------------------------------------------
            residual = hidden_states
            hidden_states = layer.post_attention_layernorm(hidden_states)
            hidden_states = layer.mlp(hidden_states)
            hidden_states = residual + hidden_states
        
        # =====================================================================
        # Step 5: Final LayerNorm
        # =====================================================================
        hidden_states = self.qwen.model.norm(hidden_states)
        
        # =====================================================================
        # Step 6: 移除 Memory Tokens
        # =====================================================================
        hidden_states = self._remove_memory_tokens(hidden_states, seq_len)
        
        # =====================================================================
        # Step 7: LM Head
        # =====================================================================
        logits = self.qwen.lm_head(hidden_states)
        
        # =====================================================================
        # 返回结果
        # =====================================================================
        result = {
            'logits': logits,
            'hidden_states': hidden_states,
        }
        
        if return_memory_states:
            result['memory_states'] = next_memory_states
            
        return result
    
    def generate(
        self,
        input_ids: Tensor,
        max_new_tokens: int = 100,
        temperature: float = 1.0,
        top_p: float = 0.9,
        memory_states: Optional[Dict] = None,
        **kwargs
    ) -> Tensor:
        """
        简单的生成函数
        """
        generated = input_ids.clone()
        
        for _ in range(max_new_tokens):
            outputs = self.forward(
                generated,
                memory_states=memory_states,
                return_memory_states=True
            )
            
            logits = outputs['logits'][:, -1, :]
            memory_states = outputs['memory_states']
            
            # 采样
            if temperature > 0:
                probs = F.softmax(logits / temperature, dim=-1)
                next_token = torch.multinomial(probs, num_samples=1)
            else:
                next_token = logits.argmax(dim=-1, keepdim=True)
            
            generated = torch.cat([generated, next_token], dim=-1)
            
            # 检查 EOS
            if hasattr(self.config, 'eos_token_id'):
                if (next_token == self.config.eos_token_id).all():
                    break
        
        return generated


# ============================================================================
# 使用示例
# ============================================================================

def main():
    """
    完整的使用示例
    """
    print("=" * 70)
    print("MAC (Memory-as-Context) 集成到 Qwen 的示例")
    print("=" * 70)
    
    # -------------------------------------------------------------------------
    # 方式 1: 使用 Hugging Face 的 Qwen 模型
    # -------------------------------------------------------------------------
    
    try:
        from transformers import AutoModelForCausalLM, AutoTokenizer
        
        print("\n[1] 加载 Qwen 模型...")
        model_name = "Qwen/Qwen2-0.5B"
        
        tokenizer = AutoTokenizer.from_pretrained(
            model_name,
            trust_remote_code=True
        )
        
        qwen_model = AutoModelForCausalLM.from_pretrained(
            model_name,
            torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
            device_map="auto" if torch.cuda.is_available() else None,
            trust_remote_code=True
        )
        
        print(f"  模型配置:")
        print(f"  - hidden_size: {qwen_model.config.hidden_size}")
        print(f"  - num_layers: {qwen_model.config.num_hidden_layers}")
        print(f"  - num_heads: {qwen_model.config.num_attention_heads}")
        
        # 创建 MAC 版本
        print("\n[2] 创建 QwenMAC 模型...")
        mac_model = QwenMACTransformer(
            qwen_model=qwen_model,
            segment_len=64,              # 每 64 个 token 一个 segment
            num_longterm_mem_tokens=8,   # 每个 segment 8 个 memory token
            num_persist_mem_tokens=4,    # 4 个全局 memory token
            neural_memory_layers=(1, 3, 5),  # 在第 1, 3, 5 层添加记忆
            memory_chunk_size=32,
            memory_batch_size=64,
        )
        
        if torch.cuda.is_available():
            mac_model = mac_model.cuda()
        
        # 测试前向传播
        print("\n[3] 测试前向传播...")
        test_text = "人工智能正在改变世界,它可以"
        inputs = tokenizer(test_text, return_tensors="pt")
        
        device = next(mac_model.parameters()).device
        input_ids = inputs.input_ids.to(device)
        
        with torch.no_grad():
            outputs = mac_model(input_ids)
            
        print(f"  输入形状: {input_ids.shape}")
        print(f"  输出 logits 形状: {outputs['logits'].shape}")
        print(f"  记忆状态数量: {len(outputs['memory_states'])}")
        
        # 测试生成
        print("\n[4] 测试文本生成...")
        with torch.no_grad():
            generated = mac_model.generate(
                input_ids,
                max_new_tokens=50,
                temperature=0.7
            )
        
        generated_text = tokenizer.decode(generated[0], skip_special_tokens=True)
        print(f"  生成文本: {generated_text}")
        
    except ImportError as e:
        print(f"\n注意: 需要安装 transformers")
        print(f"pip install transformers")
        print(f"错误: {e}")
    
    # -------------------------------------------------------------------------
    # 方式 2: 独立测试 NeuralMemory 组件
    # -------------------------------------------------------------------------
    
    print("\n" + "=" * 70)
    print("[独立测试] NeuralMemory 组件")
    print("=" * 70)
    
    device = 'cuda' if torch.cuda.is_available() else 'cpu'
    
    # 创建 NeuralMemory
    mem = NeuralMemory(
        dim=512,           # 隐藏维度
        chunk_size=32,     # 分块大小
        batch_size=64,     # 批次大小
        dim_head=64,       # 每个头的维度
        heads=8,           # 头数
        model=MemoryMLP(dim=64, depth=2),
        momentum=True,
        qk_rmsnorm=True,
    ).to(device)
    
    # 模拟输入
    batch_size = 2
    seq_len = 256
    hidden_dim = 512
    
    x = torch.randn(batch_size, seq_len, hidden_dim).to(device)
    
    print(f"\n输入形状: {x.shape}")
    
    # 第一次前向传播
    retrieved, state = mem(x)
    print(f"检索输出形状: {retrieved.shape}")
    print(f"记忆状态 seq_index: {state.seq_index}")
    
    # 第二次前向传播(传入之前的状态)
    x2 = torch.randn(batch_size, seq_len, hidden_dim).to(device)
    retrieved2, state2 = mem(x2, state=state)
    print(f"第二次检索输出形状: {retrieved2.shape}")
    print(f"更新后 seq_index: {state2.seq_index}")
    
    print("\n" + "=" * 70)
    print("完成!")
    print("=" * 70)


if __name__ == "__main__":
    main()