abir-hr196 commited on
Commit
ce7c646
·
verified ·
1 Parent(s): 73623f6

Add config.json to saes_sql_interp_bm2_cs1_experiment_4.3_syn=True/k=32

Browse files
saes_sql_interp_bm2_cs1_experiment_4.3_syn=True/k=32/config.json CHANGED
@@ -1 +1 @@
1
- {"sae": {"expansion_factor": 15, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false}, "batch_size": 16, "grad_acc_steps": 1, "micro_acc_steps": 1, "lr": null, "lr_warmup_steps": 1000, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "hookpoints": ["model.layers.5.mlp", "model.layers.5.post_attention_layernorm", "model.layers.6.mlp", "model.layers.6.post_attention_layernorm", "model.layers.7.mlp", "model.layers.7.post_attention_layernorm", "model.layers.8.mlp", "model.layers.8.post_attention_layernorm", "model.layers.9.mlp", "model.layers.9.post_attention_layernorm", "model.layers.10.mlp", "model.layers.10.post_attention_layernorm", "model.layers.11.mlp", "model.layers.11.post_attention_layernorm", "model.layers.12.mlp", "model.layers.12.post_attention_layernorm", "model.layers.13.mlp", "model.layers.13.post_attention_layernorm", "model.layers.14.mlp", "model.layers.14.post_attention_layernorm", "model.layers.15.mlp", "model.layers.15.post_attention_layernorm", "model.layers.16.mlp", "model.layers.16.post_attention_layernorm", "model.layers.17.mlp", "model.layers.17.post_attention_layernorm", "model.layers.18.mlp", "model.layers.18.post_attention_layernorm", "model.layers.19.mlp", "model.layers.19.post_attention_layernorm"], "init_seeds": [0], "layers": [], "layer_stride": 1, "transcode": false, "distribute_modules": false, "save_every": 100000, "log_to_wandb": true, "run_name": "saes_sql_interp_bm2_cs1_experiment_4.3_syn=True/k=32", "wandb_log_frequency": 1}
 
1
+ {"sae": {"expansion_factor": 15, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false}, "batch_size": 16, "grad_acc_steps": 1, "micro_acc_steps": 1, "lr": null, "lr_warmup_steps": 1000, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "hookpoints": ["model.layers.0.mlp", "model.layers.0.post_attention_layernorm", "model.layers.1.mlp", "model.layers.1.post_attention_layernorm", "model.layers.2.mlp", "model.layers.2.post_attention_layernorm", "model.layers.3.mlp", "model.layers.3.post_attention_layernorm", "model.layers.4.mlp", "model.layers.4.post_attention_layernorm", "model.layers.20.mlp", "model.layers.20.post_attention_layernorm", "model.layers.21.mlp", "model.layers.21.post_attention_layernorm", "model.layers.22.mlp", "model.layers.22.post_attention_layernorm", "model.layers.23.mlp", "model.layers.23.post_attention_layernorm"], "init_seeds": [0], "layers": [], "layer_stride": 1, "transcode": false, "distribute_modules": false, "save_every": 100000, "log_to_wandb": true, "run_name": "saes_sql_interp_bm2_cs1_experiment_4.3_syn=True/k=32", "wandb_log_frequency": 1}