Upload folder using huggingface_hub
Browse files- v2_batch_topk_expansion_factor4_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/config.json +1 -0
- v2_batch_topk_expansion_factor4_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/lr_scheduler.pt +3 -0
- v2_batch_topk_expansion_factor4_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/optimizer.pt +3 -0
- v2_batch_topk_expansion_factor4_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/state.pt +3 -0
- v2_batch_topk_expansion_factor4_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/unet.up_blocks.1.attentions.1/cfg.json +1 -0
- v2_batch_topk_expansion_factor4_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/unet.up_blocks.1.attentions.1/sae.safetensors +3 -0
v2_batch_topk_expansion_factor4_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/config.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"sae": {"expansion_factor": 4, "normalize_decoder": true, "num_latents": 0, "k": 32, "batch_topk": true, "sample_topk": false, "input_unit_norm": false, "multi_topk": false}, "dataset_path": ["/w/nobackup/385/scratch-space/expires-2025-Nov-21/pellement/11432/SAeUron/activations/laion-coco-aesthetic/full_pipeline/output"], "effective_batch_size": 4096, "num_workers": 4, "persistent_workers": true, "prefetch_factor": 2, "grad_acc_steps": 1, "micro_acc_steps": 1, "lr": 0.0004, "lr_scheduler": "linear", "lr_warmup_steps": 0, "auxk_alpha": 0.03125, "dead_feature_threshold": 10000000, "feature_sampling_window": 100, "hookpoints": ["unet.up_blocks.1.attentions.1"], "distribute_modules": false, "save_every": 5000, "log_to_wandb": true, "run_name": "batch_topk_expansion_factor4_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline", "wandb_log_frequency": 500, "wandb_project": "sae_stable-diffusion-v1-4", "mixed_precision": "no", "max_examples": null, "seed": 42, "device": "cuda", "num_epochs": 12}
|
v2_batch_topk_expansion_factor4_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/lr_scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b97be5381484503b6f3fa0c18ef0a7d7e84dae5ee3b61fd218311b0ba5febde7
|
| 3 |
+
size 1012
|
v2_batch_topk_expansion_factor4_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e9f935fe2a0f264e1ab61e3f4bf99ed624097c6fde302dca348afb37bbcbf93f
|
| 3 |
+
size 104913106
|
v2_batch_topk_expansion_factor4_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/state.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:556cd125bd92d3d2021ed56b550302a85315ea90db59af2611e499ecfc203ab7
|
| 3 |
+
size 42130
|
v2_batch_topk_expansion_factor4_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/unet.up_blocks.1.attentions.1/cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"expansion_factor": 4, "normalize_decoder": true, "num_latents": 0, "k": 32, "batch_topk": true, "sample_topk": false, "input_unit_norm": false, "multi_topk": false, "d_in": 1280}
|
v2_batch_topk_expansion_factor4_k32_multi_topkFalse_auxk_alpha0.03125_full_pipeline/unet.up_blocks.1.attentions.1/sae.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ff25286b32bf074efdba2a2c139f02130997ea474c3a155cdfc10352f090a501
|
| 3 |
+
size 52454728
|