Upload 13 files
Browse files- cfg.json +1 -0
- gpt_neox.layers.0/cfg.json +1 -0
- gpt_neox.layers.0/sae.safetensors +3 -0
- gpt_neox.layers.1/cfg.json +1 -0
- gpt_neox.layers.1/sae.safetensors +3 -0
- gpt_neox.layers.2/cfg.json +1 -0
- gpt_neox.layers.2/sae.safetensors +3 -0
- gpt_neox.layers.3/cfg.json +1 -0
- gpt_neox.layers.3/sae.safetensors +3 -0
- gpt_neox.layers.4/cfg.json +1 -0
- gpt_neox.layers.4/sae.safetensors +3 -0
- gpt_neox.layers.5/cfg.json +1 -0
- gpt_neox.layers.5/sae.safetensors +3 -0
cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"sae": {"expansion_factor": 64, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false}, "batch_size": 4, "grad_acc_steps": 8, "micro_acc_steps": 2, "lr": null, "lr_warmup_steps": 1000, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "hookpoints": ["gpt_neox.layers.0", "gpt_neox.layers.1", "gpt_neox.layers.2", "gpt_neox.layers.3", "gpt_neox.layers.4", "gpt_neox.layers.5"], "layers": [0, 1, 2, 3, 4, 5], "layer_stride": 1, "distribute_modules": false, "save_every": 1000, "max_tokens": 1000000000, "log_to_wandb": true, "run_name": "/home/random_sae/experiments/saved_models/pythia-70m_64_k32/openwebtext_plain_text_100M_trained", "wandb_log_frequency": 1}
|
gpt_neox.layers.0/cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"expansion_factor": 64, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 512}
|
gpt_neox.layers.0/sae.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:851c8fce3bdfa509b1d722586bdf99af7ec3db2970c1c0289d7df1ebb97d901a
|
| 3 |
+
size 134351176
|
gpt_neox.layers.1/cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"expansion_factor": 64, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 512}
|
gpt_neox.layers.1/sae.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cfa5553add825136ad6c3d90fa12effcb5d178c4cd45adbb560a9ae386132ed6
|
| 3 |
+
size 134351176
|
gpt_neox.layers.2/cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"expansion_factor": 64, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 512}
|
gpt_neox.layers.2/sae.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e7fc8f1fcf55aa4a2f2d6d76172e9b786f92f8d2157daba2d4d9b2e96c42c6d4
|
| 3 |
+
size 134351176
|
gpt_neox.layers.3/cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"expansion_factor": 64, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 512}
|
gpt_neox.layers.3/sae.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dec7c72d0a3d30aa42b02ff8f6bf1cac8bc5e8580e9f8ffdd2d4cea64c1ea255
|
| 3 |
+
size 134351176
|
gpt_neox.layers.4/cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"expansion_factor": 64, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 512}
|
gpt_neox.layers.4/sae.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:271b742c2bbcefcd92d1f0164a202ad2deca0747f0982f6d5f364a73b3baedea
|
| 3 |
+
size 134351176
|
gpt_neox.layers.5/cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"expansion_factor": 64, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 512}
|
gpt_neox.layers.5/sae.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:126ca83aa52f25f8bf86865075894d9f35fc1d52d5a8970f42eeda9ed584abde
|
| 3 |
+
size 134351176
|