File size: 2,244 Bytes
8709e9b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
import torch
import numpy as np
from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
from datasets import load_dataset
import gc
import os

# MODEL SELECTION
selected_model = "qwen-moe-a2.7b" # Change this
LOAD_PATH = "outputs/qwen-moe-a2.7b_20251009_084529"  # Change this

input_file = os.path.join(LOAD_PATH, 'tokenized_input.npz')
metadata_file = os.path.join(LOAD_PATH, 'metadata.txt')
attention_file = os.path.join(LOAD_PATH, 'attention_matrices_multihead.npz')
routing_file = os.path.join(LOAD_PATH, 'routing_matrices.npz')

# METADATA
print("--- Metadata ---")
with open(metadata_file, 'r') as f:
    print(f.read())

# INPUT DATA
print("\n--- Input data ---")

input_data = np.load(input_file)
print(f"File size: {os.path.getsize(input_file) / (1024**2):.2f} MB")

input_ids = input_data['input_ids']
print(f"\nTokenized Input:")
print(f"Shape: {input_ids.shape}")
print(f"First 20 token IDs: {input_ids[:20]}")


input_text = str(input_data['input_text'])
print(f"\nOriginal text length: {len(input_text)} characters")
print(f"First 200 characters: {input_text[:200]}...")

# ATTENTION MATRICES
print("\n--- Attention matrices ---")
attn_data = np.load(attention_file, allow_pickle=True)
print(f"File size: {os.path.getsize(attention_file) / (1024**2):.2f} MB")
print(f"Available layers: {len(attn_data.files)}")
print(f"Layer names: {attn_data.files}")

# Print info for first layer
layer_0 = attn_data['layer_0']
print(f"\nLayer 0 attention matrix:")
print(f"Shape: {layer_0.shape} [num_heads, seq_len, seq_len]")
print(f"Number of heads: {layer_0.shape[0]}")
print(f"Min value: {layer_0.min():.6f}")
print(f"Max value: {layer_0.max():.6f}")
print(f"Mean value: {layer_0.mean():.6f}")

print(f"\n  Head 0 attention sample (first 5x5 tokens):")
print(layer_0[0, :5, :5])


# ROUTING MATRICES
print("\n--- Routing matrices ---")
routing_data = np.load(routing_file)
print(f"File size: {os.path.getsize(routing_file) / (1024**2):.2f} MB")
print(f"Available layers: {len(routing_data.files)}")
print(f"Layer names: {routing_data.files}")
layer_0 = routing_data['layer_0']

print(f"Number of experts: {len(layer_0[0])}")

print(f"\nRouting logits sample (first 5 tokens, all experts):")
print(layer_0[:5, :])