davidoneil commited on
Commit
6a35891
·
verified ·
1 Parent(s): 6208ced

Upload folder using huggingface_hub

Browse files
config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"sae": {"activation": "topk", "expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": false}, "batch_size": 4, "grad_acc_steps": 1, "micro_acc_steps": 1, "loss_fn": "fvu", "optimizer": "signum", "lr": null, "lr_warmup_steps": 1000, "k_decay_steps": 0, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "exclude_tokens": [], "hookpoints": ["layers.12", "layers.16", "layers.20", "layers.24"], "init_seeds": [0], "layers": [12, 16, 20, 24], "layer_stride": 1, "distribute_modules": true, "save_every": 1000, "save_best": false, "finetune": null, "log_to_wandb": false, "run_name": null, "wandb_log_frequency": 1, "save_dir": "train-sae/artifacts/sparsify_sae_llama32_3b_instruct_personas", "model": "meta-llama/Llama-3.2-3B-Instruct", "dataset": "json", "split": "train", "ctx_len": 2048, "hf_token": null, "revision": null, "load_in_8bit": false, "max_examples": null, "resume": false, "text_column": "content", "shuffle_seed": 42, "data_preprocessing_num_proc": 112, "data_args": "data_files=/workspace/ACL-Paper-SAE/train-sae/data/unified_persona_corpus.jsonl"}
layers.12/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": false, "d_in": 3072}
layers.12/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7221981ed88e2af17b62b7efa50e1dfbd4cb5254971783eacfa5ecbe8d0c4544
3
+ size 604090704
layers.16/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": false, "d_in": 3072}
layers.16/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ef1a8b1b6dc23d4d4b7c1550fb9c8d7e6c0fd19426aee12d32b4b11c538088b
3
+ size 604090704
layers.20/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": false, "d_in": 3072}
layers.20/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f8f75be9611b297a22bab07225ae51f61fa9b7fe551fffac78586759173752d
3
+ size 604090704
layers.24/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": false, "d_in": 3072}
layers.24/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce8e9d4a4f4a9908ab2ca71a2d2323e92f4d3939dfeed55cbaa4e929c32fa648
3
+ size 604090704