pellement99 commited on
Commit
3f9aa94
·
verified ·
1 Parent(s): 46d52e6

Upload folder using huggingface_hub

Browse files
Files changed (24) hide show
  1. sae_stable-diffusion-v1-4-cat/nov30_cat_patch_topk_expansion_factor8_k64_multi_topkFalse_auxk_alpha0.002_full_pipeline/config.json +1 -0
  2. sae_stable-diffusion-v1-4-cat/nov30_cat_patch_topk_expansion_factor8_k64_multi_topkFalse_auxk_alpha0.002_full_pipeline/lr_scheduler.pt +3 -0
  3. sae_stable-diffusion-v1-4-cat/nov30_cat_patch_topk_expansion_factor8_k64_multi_topkFalse_auxk_alpha0.002_full_pipeline/optimizer.pt +3 -0
  4. sae_stable-diffusion-v1-4-cat/nov30_cat_patch_topk_expansion_factor8_k64_multi_topkFalse_auxk_alpha0.002_full_pipeline/state.pt +3 -0
  5. sae_stable-diffusion-v1-4-cat/nov30_cat_patch_topk_expansion_factor8_k64_multi_topkFalse_auxk_alpha0.002_full_pipeline/unet.up_blocks.1.attentions.1/cfg.json +1 -0
  6. sae_stable-diffusion-v1-4-cat/nov30_cat_patch_topk_expansion_factor8_k64_multi_topkFalse_auxk_alpha0.002_full_pipeline/unet.up_blocks.1.attentions.1/sae.safetensors +3 -0
  7. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor4_k64_multi_topkFalse_auxk_alpha0.015_full_pipeline/config.json +1 -0
  8. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor4_k64_multi_topkFalse_auxk_alpha0.015_full_pipeline/lr_scheduler.pt +3 -0
  9. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor4_k64_multi_topkFalse_auxk_alpha0.015_full_pipeline/optimizer.pt +3 -0
  10. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor4_k64_multi_topkFalse_auxk_alpha0.015_full_pipeline/state.pt +3 -0
  11. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor4_k64_multi_topkFalse_auxk_alpha0.015_full_pipeline/unet.up_blocks.1.attentions.1/cfg.json +1 -0
  12. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor4_k64_multi_topkFalse_auxk_alpha0.015_full_pipeline/unet.up_blocks.1.attentions.1/sae.safetensors +3 -0
  13. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor6_k64_multi_topkFalse_auxk_alpha0.005_full_pipeline/config.json +1 -0
  14. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor6_k64_multi_topkFalse_auxk_alpha0.005_full_pipeline/lr_scheduler.pt +3 -0
  15. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor6_k64_multi_topkFalse_auxk_alpha0.005_full_pipeline/optimizer.pt +3 -0
  16. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor6_k64_multi_topkFalse_auxk_alpha0.005_full_pipeline/state.pt +3 -0
  17. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor6_k64_multi_topkFalse_auxk_alpha0.005_full_pipeline/unet.up_blocks.1.attentions.1/cfg.json +1 -0
  18. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor6_k64_multi_topkFalse_auxk_alpha0.005_full_pipeline/unet.up_blocks.1.attentions.1/sae.safetensors +3 -0
  19. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor8_k96_multi_topkFalse_auxk_alpha0.002_full_pipeline/config.json +1 -0
  20. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor8_k96_multi_topkFalse_auxk_alpha0.002_full_pipeline/lr_scheduler.pt +3 -0
  21. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor8_k96_multi_topkFalse_auxk_alpha0.002_full_pipeline/optimizer.pt +3 -0
  22. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor8_k96_multi_topkFalse_auxk_alpha0.002_full_pipeline/state.pt +3 -0
  23. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor8_k96_multi_topkFalse_auxk_alpha0.002_full_pipeline/unet.up_blocks.1.attentions.1/cfg.json +1 -0
  24. sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor8_k96_multi_topkFalse_auxk_alpha0.002_full_pipeline/unet.up_blocks.1.attentions.1/sae.safetensors +3 -0
sae_stable-diffusion-v1-4-cat/nov30_cat_patch_topk_expansion_factor8_k64_multi_topkFalse_auxk_alpha0.002_full_pipeline/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"sae": {"expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 64, "batch_topk": false, "sample_topk": false, "input_unit_norm": false, "multi_topk": false}, "dataset_path": ["/w/nobackup/385/scratch-space/expires-2025-Nov-30/pellement/SAeUron/activations/laion-coco-aesthetic/full_pipeline/output"], "effective_batch_size": 4096, "num_workers": 4, "persistent_workers": true, "prefetch_factor": 2, "grad_acc_steps": 1, "micro_acc_steps": 1, "lr": 0.0004, "lr_scheduler": "linear", "lr_warmup_steps": 0, "auxk_alpha": 0.002, "dead_feature_threshold": 10000000, "feature_sampling_window": 100, "hookpoints": ["unet.up_blocks.1.attentions.1"], "distribute_modules": false, "save_every": 5000, "log_to_wandb": true, "run_name": "patch_topk_expansion_factor8_k64_multi_topkFalse_auxk_alpha0.002_full_pipeline", "wandb_log_frequency": 500, "wandb_project": "sae_stable-diffusion-v1-4", "mixed_precision": "no", "max_examples": null, "seed": 42, "device": "cuda", "num_epochs": 15}
sae_stable-diffusion-v1-4-cat/nov30_cat_patch_topk_expansion_factor8_k64_multi_topkFalse_auxk_alpha0.002_full_pipeline/lr_scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:503c7092869333ccee6dbd538f56182caa8f525169bdfe92d45ce268bee3c65c
3
+ size 1012
sae_stable-diffusion-v1-4-cat/nov30_cat_patch_topk_expansion_factor8_k64_multi_topkFalse_auxk_alpha0.002_full_pipeline/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0227a620e8eb01b26270024b9a7c21fdf6f68d2f97f038f14e274fcad39b1526
3
+ size 209811666
sae_stable-diffusion-v1-4-cat/nov30_cat_patch_topk_expansion_factor8_k64_multi_topkFalse_auxk_alpha0.002_full_pipeline/state.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d01d9a4214b5a2cc444b2375827e0dd02dfe506656e3d82f9b257b751ab2d95
3
+ size 83090
sae_stable-diffusion-v1-4-cat/nov30_cat_patch_topk_expansion_factor8_k64_multi_topkFalse_auxk_alpha0.002_full_pipeline/unet.up_blocks.1.attentions.1/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 64, "batch_topk": false, "sample_topk": false, "input_unit_norm": false, "multi_topk": false, "d_in": 1280}
sae_stable-diffusion-v1-4-cat/nov30_cat_patch_topk_expansion_factor8_k64_multi_topkFalse_auxk_alpha0.002_full_pipeline/unet.up_blocks.1.attentions.1/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91fa0e8103b30932c5a5614779e4e2b85ee25912d69da4c479b8e1f5ba0da748
3
+ size 104904008
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor4_k64_multi_topkFalse_auxk_alpha0.015_full_pipeline/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"sae": {"expansion_factor": 4, "normalize_decoder": true, "num_latents": 0, "k": 64, "batch_topk": false, "sample_topk": false, "input_unit_norm": false, "multi_topk": false}, "dataset_path": ["/w/nobackup/385/scratch-space/expires-2025-Nov-21/pellement/SAeUron/activations/laion-coco-aesthetic/full_pipeline/output"], "effective_batch_size": 4096, "num_workers": 4, "persistent_workers": true, "prefetch_factor": 2, "grad_acc_steps": 1, "micro_acc_steps": 1, "lr": 0.0004, "lr_scheduler": "linear", "lr_warmup_steps": 0, "auxk_alpha": 0.015, "dead_feature_threshold": 10000000, "feature_sampling_window": 100, "hookpoints": ["unet.up_blocks.1.attentions.1"], "distribute_modules": false, "save_every": 5000, "log_to_wandb": true, "run_name": "patch_topk_expansion_factor4_k64_multi_topkFalse_auxk_alpha0.015_full_pipeline", "wandb_log_frequency": 500, "wandb_project": "sae_stable-diffusion-v1-4", "mixed_precision": "no", "max_examples": null, "seed": 42, "device": "cuda", "num_epochs": 15}
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor4_k64_multi_topkFalse_auxk_alpha0.015_full_pipeline/lr_scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04d6e0c3f1da6c77aa218446212ab173e331eedfeae822c363c80b139a2cd92c
3
+ size 1012
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor4_k64_multi_topkFalse_auxk_alpha0.015_full_pipeline/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eda54d39b11a2e07baa2579817564489de3387d9701622983846b98e0e9322c2
3
+ size 104913106
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor4_k64_multi_topkFalse_auxk_alpha0.015_full_pipeline/state.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b58a715ae652ce9b4f2f8a5d68bedab18a0f7b0f521279c2b34c084febb01c32
3
+ size 42130
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor4_k64_multi_topkFalse_auxk_alpha0.015_full_pipeline/unet.up_blocks.1.attentions.1/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 4, "normalize_decoder": true, "num_latents": 0, "k": 64, "batch_topk": false, "sample_topk": false, "input_unit_norm": false, "multi_topk": false, "d_in": 1280}
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor4_k64_multi_topkFalse_auxk_alpha0.015_full_pipeline/unet.up_blocks.1.attentions.1/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:899512c460ede9ef06f739eb6aa1b7b024998fc511cfd7dcb086cf51f7001834
3
+ size 52454728
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor6_k64_multi_topkFalse_auxk_alpha0.005_full_pipeline/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"sae": {"expansion_factor": 6, "normalize_decoder": true, "num_latents": 0, "k": 64, "batch_topk": false, "sample_topk": false, "input_unit_norm": false, "multi_topk": false}, "dataset_path": ["/w/nobackup/385/scratch-space/expires-2025-Nov-21/pellement/SAeUron/activations/laion-coco-aesthetic/full_pipeline/output"], "effective_batch_size": 4096, "num_workers": 4, "persistent_workers": true, "prefetch_factor": 2, "grad_acc_steps": 1, "micro_acc_steps": 1, "lr": 0.0004, "lr_scheduler": "linear", "lr_warmup_steps": 0, "auxk_alpha": 0.005, "dead_feature_threshold": 10000000, "feature_sampling_window": 100, "hookpoints": ["unet.up_blocks.1.attentions.1"], "distribute_modules": false, "save_every": 5000, "log_to_wandb": true, "run_name": "patch_topk_expansion_factor6_k64_multi_topkFalse_auxk_alpha0.005_full_pipeline", "wandb_log_frequency": 500, "wandb_project": "sae_stable-diffusion-v1-4", "mixed_precision": "no", "max_examples": null, "seed": 42, "device": "cuda", "num_epochs": 15}
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor6_k64_multi_topkFalse_auxk_alpha0.005_full_pipeline/lr_scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04d6e0c3f1da6c77aa218446212ab173e331eedfeae822c363c80b139a2cd92c
3
+ size 1012
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor6_k64_multi_topkFalse_auxk_alpha0.005_full_pipeline/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df1cade9796b2494bc95363c1a021bfc08062954dd14a339e03c4f5b5cba2f5e
3
+ size 157362386
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor6_k64_multi_topkFalse_auxk_alpha0.005_full_pipeline/state.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e28cf44134c9dff95832a3485d535974365f1bab65443d889680a68bda8ff5eb
3
+ size 62610
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor6_k64_multi_topkFalse_auxk_alpha0.005_full_pipeline/unet.up_blocks.1.attentions.1/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 6, "normalize_decoder": true, "num_latents": 0, "k": 64, "batch_topk": false, "sample_topk": false, "input_unit_norm": false, "multi_topk": false, "d_in": 1280}
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor6_k64_multi_topkFalse_auxk_alpha0.005_full_pipeline/unet.up_blocks.1.attentions.1/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a9742ce438219d96d1ceb54ef1f5fdb10a9d969d2b0712b72a694da4ea64516
3
+ size 78679368
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor8_k96_multi_topkFalse_auxk_alpha0.002_full_pipeline/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"sae": {"expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 96, "batch_topk": false, "sample_topk": false, "input_unit_norm": false, "multi_topk": false}, "dataset_path": ["/w/nobackup/385/scratch-space/expires-2025-Nov-21/pellement/SAeUron/activations/laion-coco-aesthetic/full_pipeline/output"], "effective_batch_size": 4096, "num_workers": 4, "persistent_workers": true, "prefetch_factor": 2, "grad_acc_steps": 1, "micro_acc_steps": 1, "lr": 0.0004, "lr_scheduler": "linear", "lr_warmup_steps": 0, "auxk_alpha": 0.002, "dead_feature_threshold": 10000000, "feature_sampling_window": 100, "hookpoints": ["unet.up_blocks.1.attentions.1"], "distribute_modules": false, "save_every": 5000, "log_to_wandb": true, "run_name": "patch_topk_expansion_factor8_k96_multi_topkFalse_auxk_alpha0.002_full_pipeline", "wandb_log_frequency": 500, "wandb_project": "sae_stable-diffusion-v1-4", "mixed_precision": "no", "max_examples": null, "seed": 42, "device": "cuda", "num_epochs": 15}
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor8_k96_multi_topkFalse_auxk_alpha0.002_full_pipeline/lr_scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04d6e0c3f1da6c77aa218446212ab173e331eedfeae822c363c80b139a2cd92c
3
+ size 1012
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor8_k96_multi_topkFalse_auxk_alpha0.002_full_pipeline/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbe8bf659d06df45ab84e0536a41e6c21c31cc5385f54e4156274b75d6837e4c
3
+ size 209811666
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor8_k96_multi_topkFalse_auxk_alpha0.002_full_pipeline/state.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d776f9ac9b6241dcf75df8dbf9a5ffe121bfd54734cd1468fe1fbfdca0fb25e6
3
+ size 83090
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor8_k96_multi_topkFalse_auxk_alpha0.002_full_pipeline/unet.up_blocks.1.attentions.1/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 96, "batch_topk": false, "sample_topk": false, "input_unit_norm": false, "multi_topk": false, "d_in": 1280}
sae_stable-diffusion-v1-4-vangogh/patch_topk_expansion_factor8_k96_multi_topkFalse_auxk_alpha0.002_full_pipeline/unet.up_blocks.1.attentions.1/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ec9e27e2390f850630282dae6d6a444e5aa1c2cbadadf30ecf58fecd8b7ccc7
3
+ size 104904008