AKG2 commited on
Commit
e6fea7a
·
verified ·
1 Parent(s): fc7e8d9

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +38 -0
README.md ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - sparse-autoencoder
4
+ - llava
5
+ - clip
6
+ - machine-unlearning
7
+ - vision-language-model
8
+ ---
9
+
10
+ # Micro-SAE for LLaVA 1.5 7B Vision Tower
11
+
12
+ A Sparse Autoencoder trained on CLIP ViT-L/14 patch activations (Layer 23)
13
+ extracted from `llava-hf/llava-1.5-7b-hf`.
14
+
15
+ ## Architecture
16
+ - **Input dim**: 1024 (CLIP ViT-L/14 hidden size)
17
+ - **Dictionary size**: 4096 (4× expansion)
18
+ - **Activation**: ReLU + L1
19
+
20
+ ## Usage
21
+
22
+ ```python
23
+ import torch
24
+ import json
25
+
26
+ # Load config
27
+ with open("config.json") as f:
28
+ config = json.load(f)
29
+
30
+ # Recreate and load SAE
31
+ sae = SparseAutoencoder(config["input_dim"], config["dict_size"])
32
+ sae.load_state_dict(torch.load("micro_sae_1024d.pt"))
33
+ ```
34
+
35
+ ## Training Data
36
+ - 20,000 COCO images
37
+ - 500 zebra images (concept-specific)
38
+ - 500 fire-truck images (concept-specific)