pellement99 commited on
Commit
e5cb284
·
verified ·
1 Parent(s): fb0566f

Upload folder using huggingface_hub

Browse files
batch_topk_expansion_factor16_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"sae": {"expansion_factor": 16, "normalize_decoder": true, "num_latents": 0, "k": 32, "batch_topk": true, "sample_topk": false, "input_unit_norm": false, "multi_topk": false}, "dataset_path": ["/w/nobackup/385/scratch-space/expires-2025-Nov-12/pellement/8499/SAeUron/activations/laion-coco-aesthetic/full_pipeline/output"], "effective_batch_size": 4096, "num_workers": 4, "persistent_workers": true, "prefetch_factor": 2, "grad_acc_steps": 1, "micro_acc_steps": 1, "lr": 0.0004, "lr_scheduler": "linear", "lr_warmup_steps": 0, "auxk_alpha": 0.03125, "dead_feature_threshold": 10000000, "feature_sampling_window": 100, "hookpoints": ["unet.up_blocks.1.attentions.1"], "distribute_modules": false, "save_every": 5000, "log_to_wandb": true, "run_name": "batch_topk_expansion_factor16_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline", "wandb_log_frequency": 4000, "wandb_project": "sae_stable-diffusion-v1-4", "mixed_precision": "no", "max_examples": null, "seed": 42, "device": "cuda", "num_epochs": 5}
batch_topk_expansion_factor16_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/lr_scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1877b546e2492e13adaa56d455d87d7554d7cdceb9d4f9b8291ae80db7154dce
3
+ size 1012
batch_topk_expansion_factor16_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e11742d675306bde86ae95d340315a223c983bad8f56ee6f540d4f9a2eaa2bb
3
+ size 419608786
batch_topk_expansion_factor16_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/state.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c907a823c994f7bb266922b8e0f6b4a850da4b9a40056fc767f9005ed2cddf82
3
+ size 165010
batch_topk_expansion_factor16_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/unet.up_blocks.1.attentions.1/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 16, "normalize_decoder": true, "num_latents": 0, "k": 32, "batch_topk": true, "sample_topk": false, "input_unit_norm": false, "multi_topk": false, "d_in": 1280}
batch_topk_expansion_factor16_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/unet.up_blocks.1.attentions.1/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b671273e0176ac43b31ad55eced6ce7869908e929e4c607f06aab320884e397
3
+ size 209802576