Upload 11 files
Browse files- cfg.json +1 -0
- gpt_neox.layers.0/cfg.json +1 -0
- gpt_neox.layers.0/sae.safetensors +3 -0
- gpt_neox.layers.1/cfg.json +1 -0
- gpt_neox.layers.1/sae.safetensors +3 -0
- gpt_neox.layers.2/cfg.json +1 -0
- gpt_neox.layers.2/sae.safetensors +3 -0
- gpt_neox.layers.4/cfg.json +1 -0
- gpt_neox.layers.4/sae.safetensors +3 -0
- gpt_neox.layers.5/cfg.json +1 -0
- gpt_neox.layers.5/sae.safetensors +3 -0
cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"sae": {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 16, "multi_topk": false}, "batch_size": 4, "grad_acc_steps": 8, "micro_acc_steps": 2, "lr": null, "lr_warmup_steps": 1000, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "hookpoints": ["gpt_neox.layers.0", "gpt_neox.layers.1", "gpt_neox.layers.2", "gpt_neox.layers.3", "gpt_neox.layers.4", "gpt_neox.layers.5"], "layers": [0, 1, 2, 3, 4, 5], "layer_stride": 1, "distribute_modules": false, "save_every": 1000, "max_tokens": 1000000000, "log_to_wandb": false, "run_name": "/home/random_sae/experiments/saved_models/pythia-70m_32_k16/redpajama-data-1t-sample_plain_text_100M_rerandomised", "wandb_log_frequency": 1}
|
gpt_neox.layers.0/cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 16, "multi_topk": false, "d_in": 512}
|
gpt_neox.layers.0/sae.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b7e5c7675fcdb3ea951b691245c48e179b5748f595153241fa540239e2e0853c
|
| 3 |
+
size 67176776
|
gpt_neox.layers.1/cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 16, "multi_topk": false, "d_in": 512}
|
gpt_neox.layers.1/sae.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c5598885e58b9388aa9b6527dcba2ca1d5f54992e212384fea9bd0704176d02d
|
| 3 |
+
size 67176776
|
gpt_neox.layers.2/cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 16, "multi_topk": false, "d_in": 512}
|
gpt_neox.layers.2/sae.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7a518b9a568e9bba0607fa9ca9e6a2b74fd07bcce8641704550b67a93901ea77
|
| 3 |
+
size 67176776
|
gpt_neox.layers.4/cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 16, "multi_topk": false, "d_in": 512}
|
gpt_neox.layers.4/sae.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d0962d0f764160901f607238f068e27ddee91bf820b27ac45524335d032dceb2
|
| 3 |
+
size 67176776
|
gpt_neox.layers.5/cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 16, "multi_topk": false, "d_in": 512}
|
gpt_neox.layers.5/sae.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a06f64e67c3da7819ba269c8fae289e0e5837c4539efeb1a1edf4a44ddfb834a
|
| 3 |
+
size 67176776
|