Upload folder using huggingface_hub
Browse files- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_0/ae.pt +3 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_0/config.json +29 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_0/eval_results.json +1 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_1/ae.pt +3 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_1/config.json +29 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_1/eval_results.json +1 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_2/ae.pt +3 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_2/config.json +29 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_2/eval_results.json +1 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_3/ae.pt +3 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_3/config.json +29 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_3/eval_results.json +1 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_4/ae.pt +3 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_4/config.json +29 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_4/eval_results.json +1 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_5/ae.pt +3 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_5/config.json +29 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_5/eval_results.json +1 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_6/ae.pt +3 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_6/config.json +32 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_6/eval_results.json +1 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_7/ae.pt +3 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_7/config.json +32 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_7/eval_results.json +1 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_8/ae.pt +3 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_8/config.json +29 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_8/eval_results.json +1 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_9/ae.pt +3 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_9/config.json +29 -0
- trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_9/eval_results.json +1 -0
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_0/ae.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:16c490d578682a47914c761e9b33177806667b97c200df037b64ddb652a7d65e
|
| 3 |
+
size 302066344
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_0/config.json
ADDED
|
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"trainer": {
|
| 3 |
+
"dict_class": "AutoEncoder",
|
| 4 |
+
"trainer_class": "StandardTrainerAprilUpdate",
|
| 5 |
+
"activation_dim": 2304,
|
| 6 |
+
"dict_size": 16384,
|
| 7 |
+
"lr": 0.0003,
|
| 8 |
+
"l1_penalty": 0.012,
|
| 9 |
+
"warmup_steps": 1000,
|
| 10 |
+
"sparsity_warmup_steps": 5000,
|
| 11 |
+
"steps": 244140,
|
| 12 |
+
"decay_start": 195312,
|
| 13 |
+
"seed": 3407,
|
| 14 |
+
"device": "cuda:1",
|
| 15 |
+
"layer": 12,
|
| 16 |
+
"lm_name": "google/gemma-2-2b",
|
| 17 |
+
"wandb_name": "StandardTrainerNew-google/gemma-2-2b-resid_post_layer_12_trainer_0",
|
| 18 |
+
"submodule_name": "resid_post_layer_12"
|
| 19 |
+
},
|
| 20 |
+
"buffer": {
|
| 21 |
+
"d_submodule": 2304,
|
| 22 |
+
"io": "out",
|
| 23 |
+
"n_ctxs": 122,
|
| 24 |
+
"ctx_len": 2048,
|
| 25 |
+
"refresh_batch_size": 4,
|
| 26 |
+
"out_batch_size": 2048,
|
| 27 |
+
"device": "cuda:1"
|
| 28 |
+
}
|
| 29 |
+
}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_0/eval_results.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"l2_loss": 53.96, "l1_loss": 1025.8, "l0": 1096.11625, "frac_variance_explained": 0.84216796875, "cossim": 0.9234375, "l2_ratio": 0.893203125, "relative_reconstruction_bias": 0.999296875, "frac_alive": 0.801025390625, "hyperparameters": {"n_inputs": 200, "context_length": 2048}}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_1/ae.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ab48e7fd50bd88ed534bec2bb0a3e377e6f1ffbcea4c56e696474b29acf71a96
|
| 3 |
+
size 302066344
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_1/config.json
ADDED
|
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"trainer": {
|
| 3 |
+
"dict_class": "AutoEncoder",
|
| 4 |
+
"trainer_class": "StandardTrainerAprilUpdate",
|
| 5 |
+
"activation_dim": 2304,
|
| 6 |
+
"dict_size": 16384,
|
| 7 |
+
"lr": 0.0003,
|
| 8 |
+
"l1_penalty": 0.015,
|
| 9 |
+
"warmup_steps": 1000,
|
| 10 |
+
"sparsity_warmup_steps": 5000,
|
| 11 |
+
"steps": 244140,
|
| 12 |
+
"decay_start": 195312,
|
| 13 |
+
"seed": 3407,
|
| 14 |
+
"device": "cuda:1",
|
| 15 |
+
"layer": 12,
|
| 16 |
+
"lm_name": "google/gemma-2-2b",
|
| 17 |
+
"wandb_name": "StandardTrainerNew-google/gemma-2-2b-resid_post_layer_12_trainer_1",
|
| 18 |
+
"submodule_name": "resid_post_layer_12"
|
| 19 |
+
},
|
| 20 |
+
"buffer": {
|
| 21 |
+
"d_submodule": 2304,
|
| 22 |
+
"io": "out",
|
| 23 |
+
"n_ctxs": 122,
|
| 24 |
+
"ctx_len": 2048,
|
| 25 |
+
"refresh_batch_size": 4,
|
| 26 |
+
"out_batch_size": 2048,
|
| 27 |
+
"device": "cuda:1"
|
| 28 |
+
}
|
| 29 |
+
}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_1/eval_results.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"l2_loss": 59.43375, "l1_loss": 816.34, "l0": 776.0542138671875, "frac_variance_explained": 0.80630859375, "cossim": 0.90646484375, "l2_ratio": 0.87634765625, "relative_reconstruction_bias": 1.0092578125, "frac_alive": 0.7996826171875, "hyperparameters": {"n_inputs": 200, "context_length": 2048}}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_2/ae.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:091e077b90ee0d0651f420ee302473fdfa5e51e40a6831f2a889bdf90a13e1f9
|
| 3 |
+
size 302066344
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_2/config.json
ADDED
|
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"trainer": {
|
| 3 |
+
"dict_class": "AutoEncoder",
|
| 4 |
+
"trainer_class": "StandardTrainerAprilUpdate",
|
| 5 |
+
"activation_dim": 2304,
|
| 6 |
+
"dict_size": 16384,
|
| 7 |
+
"lr": 0.0003,
|
| 8 |
+
"l1_penalty": 0.02,
|
| 9 |
+
"warmup_steps": 1000,
|
| 10 |
+
"sparsity_warmup_steps": 5000,
|
| 11 |
+
"steps": 244140,
|
| 12 |
+
"decay_start": 195312,
|
| 13 |
+
"seed": 3407,
|
| 14 |
+
"device": "cuda:1",
|
| 15 |
+
"layer": 12,
|
| 16 |
+
"lm_name": "google/gemma-2-2b",
|
| 17 |
+
"wandb_name": "StandardTrainerNew-google/gemma-2-2b-resid_post_layer_12_trainer_2",
|
| 18 |
+
"submodule_name": "resid_post_layer_12"
|
| 19 |
+
},
|
| 20 |
+
"buffer": {
|
| 21 |
+
"d_submodule": 2304,
|
| 22 |
+
"io": "out",
|
| 23 |
+
"n_ctxs": 122,
|
| 24 |
+
"ctx_len": 2048,
|
| 25 |
+
"refresh_batch_size": 4,
|
| 26 |
+
"out_batch_size": 2048,
|
| 27 |
+
"device": "cuda:1"
|
| 28 |
+
}
|
| 29 |
+
}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_2/eval_results.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"l2_loss": 65.825, "l1_loss": 605.46, "l0": 475.75117431640626, "frac_variance_explained": 0.7558203125, "cossim": 0.884453125, "l2_ratio": 0.8563671875, "relative_reconstruction_bias": 1.0234375, "frac_alive": 0.79644775390625, "hyperparameters": {"n_inputs": 200, "context_length": 2048}}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_3/ae.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b099cf0478630f138aa8112610cc9a26117523dfd9980452d15d9b6df8b7b0b0
|
| 3 |
+
size 302066344
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_3/config.json
ADDED
|
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"trainer": {
|
| 3 |
+
"dict_class": "AutoEncoder",
|
| 4 |
+
"trainer_class": "StandardTrainerAprilUpdate",
|
| 5 |
+
"activation_dim": 2304,
|
| 6 |
+
"dict_size": 16384,
|
| 7 |
+
"lr": 0.0003,
|
| 8 |
+
"l1_penalty": 0.03,
|
| 9 |
+
"warmup_steps": 1000,
|
| 10 |
+
"sparsity_warmup_steps": 5000,
|
| 11 |
+
"steps": 244140,
|
| 12 |
+
"decay_start": 195312,
|
| 13 |
+
"seed": 3407,
|
| 14 |
+
"device": "cuda:1",
|
| 15 |
+
"layer": 12,
|
| 16 |
+
"lm_name": "google/gemma-2-2b",
|
| 17 |
+
"wandb_name": "StandardTrainerNew-google/gemma-2-2b-resid_post_layer_12_trainer_3",
|
| 18 |
+
"submodule_name": "resid_post_layer_12"
|
| 19 |
+
},
|
| 20 |
+
"buffer": {
|
| 21 |
+
"d_submodule": 2304,
|
| 22 |
+
"io": "out",
|
| 23 |
+
"n_ctxs": 122,
|
| 24 |
+
"ctx_len": 2048,
|
| 25 |
+
"refresh_batch_size": 4,
|
| 26 |
+
"out_batch_size": 2048,
|
| 27 |
+
"device": "cuda:1"
|
| 28 |
+
}
|
| 29 |
+
}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_3/eval_results.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"l2_loss": 73.47, "l1_loss": 405.03, "l0": 230.87955810546876, "frac_variance_explained": 0.6711328125, "cossim": 0.85509765625, "l2_ratio": 0.82607421875, "relative_reconstruction_bias": 1.06302734375, "frac_alive": 0.79315185546875, "hyperparameters": {"n_inputs": 200, "context_length": 2048}}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_4/ae.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a1680f889330350bd9a3b897052e2fc67f7da58d0bdbd7b99a4a4fb6b8f08ae8
|
| 3 |
+
size 302066344
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_4/config.json
ADDED
|
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"trainer": {
|
| 3 |
+
"dict_class": "AutoEncoder",
|
| 4 |
+
"trainer_class": "StandardTrainerAprilUpdate",
|
| 5 |
+
"activation_dim": 2304,
|
| 6 |
+
"dict_size": 16384,
|
| 7 |
+
"lr": 0.0003,
|
| 8 |
+
"l1_penalty": 0.04,
|
| 9 |
+
"warmup_steps": 1000,
|
| 10 |
+
"sparsity_warmup_steps": 5000,
|
| 11 |
+
"steps": 244140,
|
| 12 |
+
"decay_start": 195312,
|
| 13 |
+
"seed": 3407,
|
| 14 |
+
"device": "cuda:1",
|
| 15 |
+
"layer": 12,
|
| 16 |
+
"lm_name": "google/gemma-2-2b",
|
| 17 |
+
"wandb_name": "StandardTrainerNew-google/gemma-2-2b-resid_post_layer_12_trainer_4",
|
| 18 |
+
"submodule_name": "resid_post_layer_12"
|
| 19 |
+
},
|
| 20 |
+
"buffer": {
|
| 21 |
+
"d_submodule": 2304,
|
| 22 |
+
"io": "out",
|
| 23 |
+
"n_ctxs": 122,
|
| 24 |
+
"ctx_len": 2048,
|
| 25 |
+
"refresh_batch_size": 4,
|
| 26 |
+
"out_batch_size": 2048,
|
| 27 |
+
"device": "cuda:1"
|
| 28 |
+
}
|
| 29 |
+
}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_4/eval_results.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"l2_loss": 78.0325, "l1_loss": 311.06, "l0": 140.126806640625, "frac_variance_explained": 0.61591796875, "cossim": 0.83541015625, "l2_ratio": 0.80708984375, "relative_reconstruction_bias": 1.08521484375, "frac_alive": 0.7899169921875, "hyperparameters": {"n_inputs": 200, "context_length": 2048}}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_5/ae.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:28dab4a789420a4cd6e0b77798a76d182ba9a28f6aaeb67da45a2d0a2144d7d1
|
| 3 |
+
size 302066344
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_5/config.json
ADDED
|
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"trainer": {
|
| 3 |
+
"dict_class": "AutoEncoder",
|
| 4 |
+
"trainer_class": "StandardTrainerAprilUpdate",
|
| 5 |
+
"activation_dim": 2304,
|
| 6 |
+
"dict_size": 16384,
|
| 7 |
+
"lr": 0.0003,
|
| 8 |
+
"l1_penalty": 0.06,
|
| 9 |
+
"warmup_steps": 1000,
|
| 10 |
+
"sparsity_warmup_steps": 5000,
|
| 11 |
+
"steps": 244140,
|
| 12 |
+
"decay_start": 195312,
|
| 13 |
+
"seed": 3407,
|
| 14 |
+
"device": "cuda:1",
|
| 15 |
+
"layer": 12,
|
| 16 |
+
"lm_name": "google/gemma-2-2b",
|
| 17 |
+
"wandb_name": "StandardTrainerNew-google/gemma-2-2b-resid_post_layer_12_trainer_5",
|
| 18 |
+
"submodule_name": "resid_post_layer_12"
|
| 19 |
+
},
|
| 20 |
+
"buffer": {
|
| 21 |
+
"d_submodule": 2304,
|
| 22 |
+
"io": "out",
|
| 23 |
+
"n_ctxs": 122,
|
| 24 |
+
"ctx_len": 2048,
|
| 25 |
+
"refresh_batch_size": 4,
|
| 26 |
+
"out_batch_size": 2048,
|
| 27 |
+
"device": "cuda:1"
|
| 28 |
+
}
|
| 29 |
+
}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_5/eval_results.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"l2_loss": 84.2, "l1_loss": 222.615, "l0": 72.55794921875, "frac_variance_explained": 0.51142578125, "cossim": 0.80783203125, "l2_ratio": 0.78095703125, "relative_reconstruction_bias": 1.15984375, "frac_alive": 0.78204345703125, "hyperparameters": {"n_inputs": 200, "context_length": 2048}}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_6/ae.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b7d89e46005c4a0f9e07cf5e3efe1f406901c6f87c26e001eab32f7e940c2a53
|
| 3 |
+
size 302066710
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_6/config.json
ADDED
|
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"trainer": {
|
| 3 |
+
"trainer_class": "BatchTopKTrainer",
|
| 4 |
+
"dict_class": "BatchTopKSAE",
|
| 5 |
+
"lr": 0.0003,
|
| 6 |
+
"steps": 244140,
|
| 7 |
+
"auxk_alpha": 0.03125,
|
| 8 |
+
"warmup_steps": 1000,
|
| 9 |
+
"decay_start": 195312,
|
| 10 |
+
"threshold_beta": 0.999,
|
| 11 |
+
"threshold_start_step": 1000,
|
| 12 |
+
"top_k_aux": 1152,
|
| 13 |
+
"seed": 3407,
|
| 14 |
+
"activation_dim": 2304,
|
| 15 |
+
"dict_size": 16384,
|
| 16 |
+
"k": 80,
|
| 17 |
+
"device": "cuda:1",
|
| 18 |
+
"layer": 12,
|
| 19 |
+
"lm_name": "google/gemma-2-2b",
|
| 20 |
+
"wandb_name": "BatchTopKTrainer-google/gemma-2-2b-resid_post_layer_12_trainer_6",
|
| 21 |
+
"submodule_name": "resid_post_layer_12"
|
| 22 |
+
},
|
| 23 |
+
"buffer": {
|
| 24 |
+
"d_submodule": 2304,
|
| 25 |
+
"io": "out",
|
| 26 |
+
"n_ctxs": 122,
|
| 27 |
+
"ctx_len": 2048,
|
| 28 |
+
"refresh_batch_size": 4,
|
| 29 |
+
"out_batch_size": 2048,
|
| 30 |
+
"device": "cuda:1"
|
| 31 |
+
}
|
| 32 |
+
}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_6/eval_results.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"l2_loss": 68.2225, "l1_loss": 653.98, "l0": 118.34955322265625, "frac_variance_explained": 0.75828125, "cossim": 0.8718359375, "l2_ratio": 0.88041015625, "relative_reconstruction_bias": 1.0090234375, "frac_alive": 0.957275390625, "hyperparameters": {"n_inputs": 200, "context_length": 2048}}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_7/ae.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dcf1b6a6f2955b893b4f31e53adbd8ff813ef7d14da048c3f44bb2f50a5a7858
|
| 3 |
+
size 302066710
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_7/config.json
ADDED
|
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"trainer": {
|
| 3 |
+
"trainer_class": "BatchTopKTrainer",
|
| 4 |
+
"dict_class": "BatchTopKSAE",
|
| 5 |
+
"lr": 0.0003,
|
| 6 |
+
"steps": 244140,
|
| 7 |
+
"auxk_alpha": 0.03125,
|
| 8 |
+
"warmup_steps": 1000,
|
| 9 |
+
"decay_start": 195312,
|
| 10 |
+
"threshold_beta": 0.999,
|
| 11 |
+
"threshold_start_step": 1000,
|
| 12 |
+
"top_k_aux": 1152,
|
| 13 |
+
"seed": 3407,
|
| 14 |
+
"activation_dim": 2304,
|
| 15 |
+
"dict_size": 16384,
|
| 16 |
+
"k": 160,
|
| 17 |
+
"device": "cuda:1",
|
| 18 |
+
"layer": 12,
|
| 19 |
+
"lm_name": "google/gemma-2-2b",
|
| 20 |
+
"wandb_name": "BatchTopKTrainer-google/gemma-2-2b-resid_post_layer_12_trainer_7",
|
| 21 |
+
"submodule_name": "resid_post_layer_12"
|
| 22 |
+
},
|
| 23 |
+
"buffer": {
|
| 24 |
+
"d_submodule": 2304,
|
| 25 |
+
"io": "out",
|
| 26 |
+
"n_ctxs": 122,
|
| 27 |
+
"ctx_len": 2048,
|
| 28 |
+
"refresh_batch_size": 4,
|
| 29 |
+
"out_batch_size": 2048,
|
| 30 |
+
"device": "cuda:1"
|
| 31 |
+
}
|
| 32 |
+
}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_7/eval_results.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"l2_loss": 60.84, "l1_loss": 1058.6, "l0": 240.37475830078125, "frac_variance_explained": 0.80884765625, "cossim": 0.898828125, "l2_ratio": 0.904375, "relative_reconstruction_bias": 1.006015625, "frac_alive": 0.93707275390625, "hyperparameters": {"n_inputs": 200, "context_length": 2048}}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_8/ae.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7ca96ae26c3395de8275063e4c47c1458ec1505e2c614ee97787c8ebd252a376
|
| 3 |
+
size 302131999
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_8/config.json
ADDED
|
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"trainer": {
|
| 3 |
+
"trainer_class": "JumpReluTrainer",
|
| 4 |
+
"dict_class": "JumpReluAutoEncoder",
|
| 5 |
+
"lr": 0.0003,
|
| 6 |
+
"steps": 244140,
|
| 7 |
+
"seed": 3407,
|
| 8 |
+
"activation_dim": 2304,
|
| 9 |
+
"dict_size": 16384,
|
| 10 |
+
"device": "cuda:1",
|
| 11 |
+
"layer": 12,
|
| 12 |
+
"lm_name": "google/gemma-2-2b",
|
| 13 |
+
"wandb_name": "JumpReluTrainer-google/gemma-2-2b-resid_post_layer_12_trainer_8",
|
| 14 |
+
"submodule_name": "resid_post_layer_12",
|
| 15 |
+
"bandwidth": 0.001,
|
| 16 |
+
"sparsity_penalty": 1.0,
|
| 17 |
+
"sparsity_warmup_steps": 5000,
|
| 18 |
+
"target_l0": 80
|
| 19 |
+
},
|
| 20 |
+
"buffer": {
|
| 21 |
+
"d_submodule": 2304,
|
| 22 |
+
"io": "out",
|
| 23 |
+
"n_ctxs": 122,
|
| 24 |
+
"ctx_len": 2048,
|
| 25 |
+
"refresh_batch_size": 4,
|
| 26 |
+
"out_batch_size": 2048,
|
| 27 |
+
"device": "cuda:1"
|
| 28 |
+
}
|
| 29 |
+
}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_8/eval_results.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"l2_loss": 68.04, "l1_loss": 670.48, "l0": 122.69531494140625, "frac_variance_explained": 0.759765625, "cossim": 0.8728515625, "l2_ratio": 0.88236328125, "relative_reconstruction_bias": 1.0092578125, "frac_alive": 0.93853759765625, "hyperparameters": {"n_inputs": 200, "context_length": 2048}}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_9/ae.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0f4199349113fc6417bb9e5569291888b7350d3a22aa0f5e3f38bb272d28f29d
|
| 3 |
+
size 302131999
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_9/config.json
ADDED
|
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"trainer": {
|
| 3 |
+
"trainer_class": "JumpReluTrainer",
|
| 4 |
+
"dict_class": "JumpReluAutoEncoder",
|
| 5 |
+
"lr": 0.0003,
|
| 6 |
+
"steps": 244140,
|
| 7 |
+
"seed": 3407,
|
| 8 |
+
"activation_dim": 2304,
|
| 9 |
+
"dict_size": 16384,
|
| 10 |
+
"device": "cuda:1",
|
| 11 |
+
"layer": 12,
|
| 12 |
+
"lm_name": "google/gemma-2-2b",
|
| 13 |
+
"wandb_name": "JumpReluTrainer-google/gemma-2-2b-resid_post_layer_12_trainer_9",
|
| 14 |
+
"submodule_name": "resid_post_layer_12",
|
| 15 |
+
"bandwidth": 0.001,
|
| 16 |
+
"sparsity_penalty": 1.0,
|
| 17 |
+
"sparsity_warmup_steps": 5000,
|
| 18 |
+
"target_l0": 160
|
| 19 |
+
},
|
| 20 |
+
"buffer": {
|
| 21 |
+
"d_submodule": 2304,
|
| 22 |
+
"io": "out",
|
| 23 |
+
"n_ctxs": 122,
|
| 24 |
+
"ctx_len": 2048,
|
| 25 |
+
"refresh_batch_size": 4,
|
| 26 |
+
"out_batch_size": 2048,
|
| 27 |
+
"device": "cuda:1"
|
| 28 |
+
}
|
| 29 |
+
}
|
trained_saes__google_gemma-2-2b_jump_relu_batch_top_k_standard_new/resid_post_layer_12/trainer_9/eval_results.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"l2_loss": 59.90375, "l1_loss": 987.2, "l0": 252.56128662109376, "frac_variance_explained": 0.81595703125, "cossim": 0.9024609375, "l2_ratio": 0.90943359375, "relative_reconstruction_bias": 1.0080859375, "frac_alive": 0.98199462890625, "hyperparameters": {"n_inputs": 200, "context_length": 2048}}
|