Add files using upload-large-folder tool
Browse files- clt/width_262k_l0_big/config.json +9 -0
- clt/width_262k_l0_big_affine/config.json +9 -0
- clt/width_262k_l0_medium/config.json +9 -0
- clt/width_262k_l0_medium/params_layer_14.safetensors +3 -0
- clt/width_262k_l0_medium/params_layer_2.safetensors +3 -0
- clt/width_262k_l0_medium/params_layer_24.safetensors +3 -0
- clt/width_262k_l0_medium_affine/config.json +9 -0
- clt/width_524k_l0_big/config.json +9 -0
- clt/width_524k_l0_big_affine/config.json +9 -0
- clt/width_524k_l0_medium/config.json +9 -0
- clt/width_524k_l0_medium_affine/config.json +9 -0
- clt/width_524k_l0_medium_affine/params_layer_10.safetensors +3 -0
- clt/width_524k_l0_medium_affine/params_layer_14.safetensors +3 -0
- clt/width_524k_l0_medium_affine/params_layer_23.safetensors +3 -0
- clt/width_524k_l0_medium_affine/params_layer_3.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_1m_l0_medium/config.json +9 -0
- crosscoder/layer_7_13_17_22_width_1m_l0_medium/params_layer_0.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_1m_l0_medium/params_layer_1.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_1m_l0_medium/params_layer_2.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_1m_l0_medium/params_layer_3.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_262k_l0_big/config.json +9 -0
- crosscoder/layer_7_13_17_22_width_262k_l0_big/params_layer_0.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_262k_l0_big/params_layer_1.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_262k_l0_big/params_layer_2.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_262k_l0_big/params_layer_3.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_262k_l0_medium/config.json +9 -0
- crosscoder/layer_7_13_17_22_width_262k_l0_medium/params_layer_0.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_262k_l0_medium/params_layer_1.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_262k_l0_medium/params_layer_2.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_262k_l0_medium/params_layer_3.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_524k_l0_big/config.json +9 -0
- crosscoder/layer_7_13_17_22_width_524k_l0_big/params_layer_0.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_524k_l0_big/params_layer_1.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_524k_l0_big/params_layer_2.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_524k_l0_big/params_layer_3.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_524k_l0_medium/config.json +9 -0
- crosscoder/layer_7_13_17_22_width_524k_l0_medium/params_layer_0.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_524k_l0_medium/params_layer_1.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_524k_l0_medium/params_layer_2.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_524k_l0_medium/params_layer_3.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_65k_l0_big/config.json +9 -0
- crosscoder/layer_7_13_17_22_width_65k_l0_big/params_layer_0.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_65k_l0_big/params_layer_1.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_65k_l0_big/params_layer_2.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_65k_l0_big/params_layer_3.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_65k_l0_medium/config.json +9 -0
- crosscoder/layer_7_13_17_22_width_65k_l0_medium/params_layer_0.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_65k_l0_medium/params_layer_1.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_65k_l0_medium/params_layer_2.safetensors +3 -0
- crosscoder/layer_7_13_17_22_width_65k_l0_medium/params_layer_3.safetensors +3 -0
clt/width_262k_l0_big/config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"hf_hook_point_in": "model.layers.{all}.pre_feedforward_layernorm.output",
|
| 3 |
+
"hf_hook_point_out": "model.layers.{all}.post_feedforward_layernorm.output",
|
| 4 |
+
"width": 262080,
|
| 5 |
+
"model_name": "gemma-v3-1b-pt",
|
| 6 |
+
"architecture": "jump_relu",
|
| 7 |
+
"l0": 150,
|
| 8 |
+
"affine_connection": false
|
| 9 |
+
}
|
clt/width_262k_l0_big_affine/config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"hf_hook_point_in": "model.layers.{all}.pre_feedforward_layernorm.output",
|
| 3 |
+
"hf_hook_point_out": "model.layers.{all}.post_feedforward_layernorm.output",
|
| 4 |
+
"width": 262080,
|
| 5 |
+
"model_name": "gemma-v3-1b-pt",
|
| 6 |
+
"architecture": "jump_relu",
|
| 7 |
+
"l0": 150,
|
| 8 |
+
"affine_connection": true
|
| 9 |
+
}
|
clt/width_262k_l0_medium/config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"hf_hook_point_in": "model.layers.{all}.pre_feedforward_layernorm.output",
|
| 3 |
+
"hf_hook_point_out": "model.layers.{all}.post_feedforward_layernorm.output",
|
| 4 |
+
"width": 262080,
|
| 5 |
+
"model_name": "gemma-v3-1b-pt",
|
| 6 |
+
"architecture": "jump_relu",
|
| 7 |
+
"l0": 50,
|
| 8 |
+
"affine_connection": false
|
| 9 |
+
}
|
clt/width_262k_l0_medium/params_layer_14.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:93fac2ba67d2c1a6a312130b4d9b3bdba3039dd2e9de8f4678d3d126bda119d1
|
| 3 |
+
size 1254198912
|
clt/width_262k_l0_medium/params_layer_2.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3c8050b99b4412b8cffcd86180b9f6ed956d485a81adecb0d32f3c538ad8940d
|
| 3 |
+
size 1254198912
|
clt/width_262k_l0_medium/params_layer_24.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c273320250e53bf94ae5d21dbba699a4b0e757caa18719bfd6946b6e6888df1f
|
| 3 |
+
size 1254198912
|
clt/width_262k_l0_medium_affine/config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"hf_hook_point_in": "model.layers.{all}.pre_feedforward_layernorm.output",
|
| 3 |
+
"hf_hook_point_out": "model.layers.{all}.post_feedforward_layernorm.output",
|
| 4 |
+
"width": 262080,
|
| 5 |
+
"model_name": "gemma-v3-1b-pt",
|
| 6 |
+
"architecture": "jump_relu",
|
| 7 |
+
"l0": 50,
|
| 8 |
+
"affine_connection": true
|
| 9 |
+
}
|
clt/width_524k_l0_big/config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"hf_hook_point_in": "model.layers.{all}.pre_feedforward_layernorm.output",
|
| 3 |
+
"hf_hook_point_out": "model.layers.{all}.post_feedforward_layernorm.output",
|
| 4 |
+
"width": 524160,
|
| 5 |
+
"model_name": "gemma-v3-1b-pt",
|
| 6 |
+
"architecture": "jump_relu",
|
| 7 |
+
"l0": 150,
|
| 8 |
+
"affine_connection": false
|
| 9 |
+
}
|
clt/width_524k_l0_big_affine/config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"hf_hook_point_in": "model.layers.{all}.pre_feedforward_layernorm.output",
|
| 3 |
+
"hf_hook_point_out": "model.layers.{all}.post_feedforward_layernorm.output",
|
| 4 |
+
"width": 524160,
|
| 5 |
+
"model_name": "gemma-v3-1b-pt",
|
| 6 |
+
"architecture": "jump_relu",
|
| 7 |
+
"l0": 150,
|
| 8 |
+
"affine_connection": true
|
| 9 |
+
}
|
clt/width_524k_l0_medium/config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"hf_hook_point_in": "model.layers.{all}.pre_feedforward_layernorm.output",
|
| 3 |
+
"hf_hook_point_out": "model.layers.{all}.post_feedforward_layernorm.output",
|
| 4 |
+
"width": 524160,
|
| 5 |
+
"model_name": "gemma-v3-1b-pt",
|
| 6 |
+
"architecture": "jump_relu",
|
| 7 |
+
"l0": 50,
|
| 8 |
+
"affine_connection": false
|
| 9 |
+
}
|
clt/width_524k_l0_medium_affine/config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"hf_hook_point_in": "model.layers.{all}.pre_feedforward_layernorm.output",
|
| 3 |
+
"hf_hook_point_out": "model.layers.{all}.post_feedforward_layernorm.output",
|
| 4 |
+
"width": 524160,
|
| 5 |
+
"model_name": "gemma-v3-1b-pt",
|
| 6 |
+
"architecture": "jump_relu",
|
| 7 |
+
"l0": 50,
|
| 8 |
+
"affine_connection": true
|
| 9 |
+
}
|
clt/width_524k_l0_medium_affine/params_layer_10.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:89096cd98215fd6e329aba96c77d84cd5975d2e68033afff9212f652ec4244ce
|
| 3 |
+
size 2513701352
|
clt/width_524k_l0_medium_affine/params_layer_14.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0ab55a70ee50fec9fcb88f6fa474809232b6c298cf9f92874af7a4b22f597353
|
| 3 |
+
size 2513701352
|
clt/width_524k_l0_medium_affine/params_layer_23.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3274855af27bd1301365ea369ab23f9f5812d4ff15448dd7a7d7effa1208fa54
|
| 3 |
+
size 2513701352
|
clt/width_524k_l0_medium_affine/params_layer_3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a61c052936dd837242969b1af4eff4d4a9a0357a51211c97fad7b3dd6c0c2c32
|
| 3 |
+
size 2513701352
|
crosscoder/layer_7_13_17_22_width_1m_l0_medium/config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"hf_hook_point_in": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
|
| 3 |
+
"hf_hook_point_out": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
|
| 4 |
+
"width": 1048576,
|
| 5 |
+
"model_name": "gemma-v3-1b-pt",
|
| 6 |
+
"architecture": "jump_relu",
|
| 7 |
+
"l0": 50,
|
| 8 |
+
"affine_connection": false
|
| 9 |
+
}
|
crosscoder/layer_7_13_17_22_width_1m_l0_medium/params_layer_0.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bdf9fec97d8cf46d6197f98f34cc3a76659feea39f43608daa59f32eec86e071
|
| 3 |
+
size 6041899912
|
crosscoder/layer_7_13_17_22_width_1m_l0_medium/params_layer_1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:74e32ae49726c632966f128b5d185e6a0f13d73bd67d5fa10b4935675269536f
|
| 3 |
+
size 6041899912
|
crosscoder/layer_7_13_17_22_width_1m_l0_medium/params_layer_2.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:51bbf0f6d57b13fd2c21ce110fe8f4da945d78375a6ae86fc78811b7cbe87ae0
|
| 3 |
+
size 6041899912
|
crosscoder/layer_7_13_17_22_width_1m_l0_medium/params_layer_3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:70821dad30b188b1ca0e089b85cc8f5ac64169a607790305e23863ba5aaf1e41
|
| 3 |
+
size 6041899912
|
crosscoder/layer_7_13_17_22_width_262k_l0_big/config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"hf_hook_point_in": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
|
| 3 |
+
"hf_hook_point_out": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
|
| 4 |
+
"width": 262144,
|
| 5 |
+
"model_name": "gemma-v3-1b-pt",
|
| 6 |
+
"architecture": "jump_relu",
|
| 7 |
+
"l0": 150,
|
| 8 |
+
"affine_connection": false
|
| 9 |
+
}
|
crosscoder/layer_7_13_17_22_width_262k_l0_big/params_layer_0.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f5a46142fbb60f62716b553c1bb30f1ca566b87fc180657c45ae4f9214937d9c
|
| 3 |
+
size 1510478720
|
crosscoder/layer_7_13_17_22_width_262k_l0_big/params_layer_1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b6d410e300090345068af567552f6cc161e9cadce350b836389fe8593590d97a
|
| 3 |
+
size 1510478720
|
crosscoder/layer_7_13_17_22_width_262k_l0_big/params_layer_2.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aa74f988831763e17aa49d2f3c995c8ff2ec02380c78a62ce8372a8a4fef32ef
|
| 3 |
+
size 1510478720
|
crosscoder/layer_7_13_17_22_width_262k_l0_big/params_layer_3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2d977fe0c290db54eecffcfd8c8d00b7e7425c063e83a61e738c58aeef4ce126
|
| 3 |
+
size 1510478720
|
crosscoder/layer_7_13_17_22_width_262k_l0_medium/config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"hf_hook_point_in": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
|
| 3 |
+
"hf_hook_point_out": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
|
| 4 |
+
"width": 262144,
|
| 5 |
+
"model_name": "gemma-v3-1b-pt",
|
| 6 |
+
"architecture": "jump_relu",
|
| 7 |
+
"l0": 50,
|
| 8 |
+
"affine_connection": false
|
| 9 |
+
}
|
crosscoder/layer_7_13_17_22_width_262k_l0_medium/params_layer_0.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0e6601e7ad1526bc36804840bb107db5d08f805301c7a588573dfc96e52a482b
|
| 3 |
+
size 1510478720
|
crosscoder/layer_7_13_17_22_width_262k_l0_medium/params_layer_1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9cf9228314d9399056fe4b1ec3aa6767d79a1538cc5e1afa977ee5d3b17779
|
| 3 |
+
size 1510478720
|
crosscoder/layer_7_13_17_22_width_262k_l0_medium/params_layer_2.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f9b9d3e7bb93e86d341a2ee2a003fd85e3dbcd42102438da8591b7b19716c5e0
|
| 3 |
+
size 1510478720
|
crosscoder/layer_7_13_17_22_width_262k_l0_medium/params_layer_3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:75db10e54428eafb0b474f4d5c86476db352bf48e5a86feb531a420f6698bd28
|
| 3 |
+
size 1510478720
|
crosscoder/layer_7_13_17_22_width_524k_l0_big/config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"hf_hook_point_in": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
|
| 3 |
+
"hf_hook_point_out": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
|
| 4 |
+
"width": 524288,
|
| 5 |
+
"model_name": "gemma-v3-1b-pt",
|
| 6 |
+
"architecture": "jump_relu",
|
| 7 |
+
"l0": 150,
|
| 8 |
+
"affine_connection": false
|
| 9 |
+
}
|
crosscoder/layer_7_13_17_22_width_524k_l0_big/params_layer_0.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:165bbce957c2b71f2685a87cff9fc9cc10679a1902eee27933b7c2a95c471a48
|
| 3 |
+
size 3020952456
|
crosscoder/layer_7_13_17_22_width_524k_l0_big/params_layer_1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c31a424e0632e00126e43d723ed4185e65065cbc3b89624363422b866bbed7d7
|
| 3 |
+
size 3020952456
|
crosscoder/layer_7_13_17_22_width_524k_l0_big/params_layer_2.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7fde4c7db3836a1dc89b9bd1f5b96f1dd23a2ba382e1c0b5c933b1f4a49de325
|
| 3 |
+
size 3020952456
|
crosscoder/layer_7_13_17_22_width_524k_l0_big/params_layer_3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e0a6418d2f8cccb10627ff2ebf3ddfec6bf476c6982ca77537edf8d71b408f2a
|
| 3 |
+
size 3020952456
|
crosscoder/layer_7_13_17_22_width_524k_l0_medium/config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"hf_hook_point_in": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
|
| 3 |
+
"hf_hook_point_out": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
|
| 4 |
+
"width": 524288,
|
| 5 |
+
"model_name": "gemma-v3-1b-pt",
|
| 6 |
+
"architecture": "jump_relu",
|
| 7 |
+
"l0": 50,
|
| 8 |
+
"affine_connection": false
|
| 9 |
+
}
|
crosscoder/layer_7_13_17_22_width_524k_l0_medium/params_layer_0.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5157dbe2bd3b436d7331c7406b260dda2751b009346b3451584d622efff21ca5
|
| 3 |
+
size 3020952456
|
crosscoder/layer_7_13_17_22_width_524k_l0_medium/params_layer_1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2572d6c53ede9a777d29207c9d46b08486ff4e64532d9c042f38bb10e54e8815
|
| 3 |
+
size 3020952456
|
crosscoder/layer_7_13_17_22_width_524k_l0_medium/params_layer_2.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:58a295a9136ea05a104be0449327fe5e8f47a8ebe9051037d426c3c22105f5fc
|
| 3 |
+
size 3020952456
|
crosscoder/layer_7_13_17_22_width_524k_l0_medium/params_layer_3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:328e90f30852353a00ae5e8988010ff9792debdccd3258b82f3c220c80f071e9
|
| 3 |
+
size 3020952456
|
crosscoder/layer_7_13_17_22_width_65k_l0_big/config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"hf_hook_point_in": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
|
| 3 |
+
"hf_hook_point_out": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
|
| 4 |
+
"width": 65536,
|
| 5 |
+
"model_name": "gemma-v3-1b-pt",
|
| 6 |
+
"architecture": "jump_relu",
|
| 7 |
+
"l0": 150,
|
| 8 |
+
"affine_connection": false
|
| 9 |
+
}
|
crosscoder/layer_7_13_17_22_width_65k_l0_big/params_layer_0.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5284453a253fbd41bef81a5794e77e3071107310e220c60a30f6dfed31f58bbc
|
| 3 |
+
size 377623424
|
crosscoder/layer_7_13_17_22_width_65k_l0_big/params_layer_1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ec492264b76ca1b7f71038a86975dec578ee2d4da6ce3ca1218bb19eebba3f7f
|
| 3 |
+
size 377623424
|
crosscoder/layer_7_13_17_22_width_65k_l0_big/params_layer_2.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6035cdb8ff1235e95bee759909119da284a383a2011f91a411e031e17984fe9b
|
| 3 |
+
size 377623424
|
crosscoder/layer_7_13_17_22_width_65k_l0_big/params_layer_3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:792b20afb0881d69c12ff323a454acb4ee93f3d37ba40bafe69a8784c100e033
|
| 3 |
+
size 377623424
|
crosscoder/layer_7_13_17_22_width_65k_l0_medium/config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"hf_hook_point_in": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
|
| 3 |
+
"hf_hook_point_out": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
|
| 4 |
+
"width": 65536,
|
| 5 |
+
"model_name": "gemma-v3-1b-pt",
|
| 6 |
+
"architecture": "jump_relu",
|
| 7 |
+
"l0": 50,
|
| 8 |
+
"affine_connection": false
|
| 9 |
+
}
|
crosscoder/layer_7_13_17_22_width_65k_l0_medium/params_layer_0.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:228dddcb01dfb89e7b3bf2ba5fd23636fed7ae5c6154b3b2f4a552bdbb4552dc
|
| 3 |
+
size 377623424
|
crosscoder/layer_7_13_17_22_width_65k_l0_medium/params_layer_1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c0a12e3e58b00bb84980118b571d40aaefa0e6b0ddbdf4c817b8c0ee67543aec
|
| 3 |
+
size 377623424
|
crosscoder/layer_7_13_17_22_width_65k_l0_medium/params_layer_2.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8af44b18e50dcd4c2228879ad7d4a0e01bd687c8c1fcd2c020cd9a877704cef3
|
| 3 |
+
size 377623424
|
crosscoder/layer_7_13_17_22_width_65k_l0_medium/params_layer_3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7115c7dfb7fa2d943da39139d7c39887d79179ad8dd5f9003a18cdc563ba31bd
|
| 3 |
+
size 377623424
|