SAELens
CallumMcDougallGDM commited on
Commit
31e7b0f
·
verified ·
1 Parent(s): b49440e

Add files using upload-large-folder tool

Browse files
Files changed (50) hide show
  1. clt/width_262k_l0_big/config.json +9 -0
  2. clt/width_262k_l0_big_affine/config.json +9 -0
  3. clt/width_262k_l0_medium/config.json +9 -0
  4. clt/width_262k_l0_medium/params_layer_14.safetensors +3 -0
  5. clt/width_262k_l0_medium/params_layer_2.safetensors +3 -0
  6. clt/width_262k_l0_medium/params_layer_24.safetensors +3 -0
  7. clt/width_262k_l0_medium_affine/config.json +9 -0
  8. clt/width_524k_l0_big/config.json +9 -0
  9. clt/width_524k_l0_big_affine/config.json +9 -0
  10. clt/width_524k_l0_medium/config.json +9 -0
  11. clt/width_524k_l0_medium_affine/config.json +9 -0
  12. clt/width_524k_l0_medium_affine/params_layer_10.safetensors +3 -0
  13. clt/width_524k_l0_medium_affine/params_layer_14.safetensors +3 -0
  14. clt/width_524k_l0_medium_affine/params_layer_23.safetensors +3 -0
  15. clt/width_524k_l0_medium_affine/params_layer_3.safetensors +3 -0
  16. crosscoder/layer_7_13_17_22_width_1m_l0_medium/config.json +9 -0
  17. crosscoder/layer_7_13_17_22_width_1m_l0_medium/params_layer_0.safetensors +3 -0
  18. crosscoder/layer_7_13_17_22_width_1m_l0_medium/params_layer_1.safetensors +3 -0
  19. crosscoder/layer_7_13_17_22_width_1m_l0_medium/params_layer_2.safetensors +3 -0
  20. crosscoder/layer_7_13_17_22_width_1m_l0_medium/params_layer_3.safetensors +3 -0
  21. crosscoder/layer_7_13_17_22_width_262k_l0_big/config.json +9 -0
  22. crosscoder/layer_7_13_17_22_width_262k_l0_big/params_layer_0.safetensors +3 -0
  23. crosscoder/layer_7_13_17_22_width_262k_l0_big/params_layer_1.safetensors +3 -0
  24. crosscoder/layer_7_13_17_22_width_262k_l0_big/params_layer_2.safetensors +3 -0
  25. crosscoder/layer_7_13_17_22_width_262k_l0_big/params_layer_3.safetensors +3 -0
  26. crosscoder/layer_7_13_17_22_width_262k_l0_medium/config.json +9 -0
  27. crosscoder/layer_7_13_17_22_width_262k_l0_medium/params_layer_0.safetensors +3 -0
  28. crosscoder/layer_7_13_17_22_width_262k_l0_medium/params_layer_1.safetensors +3 -0
  29. crosscoder/layer_7_13_17_22_width_262k_l0_medium/params_layer_2.safetensors +3 -0
  30. crosscoder/layer_7_13_17_22_width_262k_l0_medium/params_layer_3.safetensors +3 -0
  31. crosscoder/layer_7_13_17_22_width_524k_l0_big/config.json +9 -0
  32. crosscoder/layer_7_13_17_22_width_524k_l0_big/params_layer_0.safetensors +3 -0
  33. crosscoder/layer_7_13_17_22_width_524k_l0_big/params_layer_1.safetensors +3 -0
  34. crosscoder/layer_7_13_17_22_width_524k_l0_big/params_layer_2.safetensors +3 -0
  35. crosscoder/layer_7_13_17_22_width_524k_l0_big/params_layer_3.safetensors +3 -0
  36. crosscoder/layer_7_13_17_22_width_524k_l0_medium/config.json +9 -0
  37. crosscoder/layer_7_13_17_22_width_524k_l0_medium/params_layer_0.safetensors +3 -0
  38. crosscoder/layer_7_13_17_22_width_524k_l0_medium/params_layer_1.safetensors +3 -0
  39. crosscoder/layer_7_13_17_22_width_524k_l0_medium/params_layer_2.safetensors +3 -0
  40. crosscoder/layer_7_13_17_22_width_524k_l0_medium/params_layer_3.safetensors +3 -0
  41. crosscoder/layer_7_13_17_22_width_65k_l0_big/config.json +9 -0
  42. crosscoder/layer_7_13_17_22_width_65k_l0_big/params_layer_0.safetensors +3 -0
  43. crosscoder/layer_7_13_17_22_width_65k_l0_big/params_layer_1.safetensors +3 -0
  44. crosscoder/layer_7_13_17_22_width_65k_l0_big/params_layer_2.safetensors +3 -0
  45. crosscoder/layer_7_13_17_22_width_65k_l0_big/params_layer_3.safetensors +3 -0
  46. crosscoder/layer_7_13_17_22_width_65k_l0_medium/config.json +9 -0
  47. crosscoder/layer_7_13_17_22_width_65k_l0_medium/params_layer_0.safetensors +3 -0
  48. crosscoder/layer_7_13_17_22_width_65k_l0_medium/params_layer_1.safetensors +3 -0
  49. crosscoder/layer_7_13_17_22_width_65k_l0_medium/params_layer_2.safetensors +3 -0
  50. crosscoder/layer_7_13_17_22_width_65k_l0_medium/params_layer_3.safetensors +3 -0
clt/width_262k_l0_big/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hf_hook_point_in": "model.layers.{all}.pre_feedforward_layernorm.output",
3
+ "hf_hook_point_out": "model.layers.{all}.post_feedforward_layernorm.output",
4
+ "width": 262080,
5
+ "model_name": "gemma-v3-1b-pt",
6
+ "architecture": "jump_relu",
7
+ "l0": 150,
8
+ "affine_connection": false
9
+ }
clt/width_262k_l0_big_affine/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hf_hook_point_in": "model.layers.{all}.pre_feedforward_layernorm.output",
3
+ "hf_hook_point_out": "model.layers.{all}.post_feedforward_layernorm.output",
4
+ "width": 262080,
5
+ "model_name": "gemma-v3-1b-pt",
6
+ "architecture": "jump_relu",
7
+ "l0": 150,
8
+ "affine_connection": true
9
+ }
clt/width_262k_l0_medium/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hf_hook_point_in": "model.layers.{all}.pre_feedforward_layernorm.output",
3
+ "hf_hook_point_out": "model.layers.{all}.post_feedforward_layernorm.output",
4
+ "width": 262080,
5
+ "model_name": "gemma-v3-1b-pt",
6
+ "architecture": "jump_relu",
7
+ "l0": 50,
8
+ "affine_connection": false
9
+ }
clt/width_262k_l0_medium/params_layer_14.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93fac2ba67d2c1a6a312130b4d9b3bdba3039dd2e9de8f4678d3d126bda119d1
3
+ size 1254198912
clt/width_262k_l0_medium/params_layer_2.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c8050b99b4412b8cffcd86180b9f6ed956d485a81adecb0d32f3c538ad8940d
3
+ size 1254198912
clt/width_262k_l0_medium/params_layer_24.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c273320250e53bf94ae5d21dbba699a4b0e757caa18719bfd6946b6e6888df1f
3
+ size 1254198912
clt/width_262k_l0_medium_affine/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hf_hook_point_in": "model.layers.{all}.pre_feedforward_layernorm.output",
3
+ "hf_hook_point_out": "model.layers.{all}.post_feedforward_layernorm.output",
4
+ "width": 262080,
5
+ "model_name": "gemma-v3-1b-pt",
6
+ "architecture": "jump_relu",
7
+ "l0": 50,
8
+ "affine_connection": true
9
+ }
clt/width_524k_l0_big/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hf_hook_point_in": "model.layers.{all}.pre_feedforward_layernorm.output",
3
+ "hf_hook_point_out": "model.layers.{all}.post_feedforward_layernorm.output",
4
+ "width": 524160,
5
+ "model_name": "gemma-v3-1b-pt",
6
+ "architecture": "jump_relu",
7
+ "l0": 150,
8
+ "affine_connection": false
9
+ }
clt/width_524k_l0_big_affine/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hf_hook_point_in": "model.layers.{all}.pre_feedforward_layernorm.output",
3
+ "hf_hook_point_out": "model.layers.{all}.post_feedforward_layernorm.output",
4
+ "width": 524160,
5
+ "model_name": "gemma-v3-1b-pt",
6
+ "architecture": "jump_relu",
7
+ "l0": 150,
8
+ "affine_connection": true
9
+ }
clt/width_524k_l0_medium/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hf_hook_point_in": "model.layers.{all}.pre_feedforward_layernorm.output",
3
+ "hf_hook_point_out": "model.layers.{all}.post_feedforward_layernorm.output",
4
+ "width": 524160,
5
+ "model_name": "gemma-v3-1b-pt",
6
+ "architecture": "jump_relu",
7
+ "l0": 50,
8
+ "affine_connection": false
9
+ }
clt/width_524k_l0_medium_affine/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hf_hook_point_in": "model.layers.{all}.pre_feedforward_layernorm.output",
3
+ "hf_hook_point_out": "model.layers.{all}.post_feedforward_layernorm.output",
4
+ "width": 524160,
5
+ "model_name": "gemma-v3-1b-pt",
6
+ "architecture": "jump_relu",
7
+ "l0": 50,
8
+ "affine_connection": true
9
+ }
clt/width_524k_l0_medium_affine/params_layer_10.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89096cd98215fd6e329aba96c77d84cd5975d2e68033afff9212f652ec4244ce
3
+ size 2513701352
clt/width_524k_l0_medium_affine/params_layer_14.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ab55a70ee50fec9fcb88f6fa474809232b6c298cf9f92874af7a4b22f597353
3
+ size 2513701352
clt/width_524k_l0_medium_affine/params_layer_23.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3274855af27bd1301365ea369ab23f9f5812d4ff15448dd7a7d7effa1208fa54
3
+ size 2513701352
clt/width_524k_l0_medium_affine/params_layer_3.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a61c052936dd837242969b1af4eff4d4a9a0357a51211c97fad7b3dd6c0c2c32
3
+ size 2513701352
crosscoder/layer_7_13_17_22_width_1m_l0_medium/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hf_hook_point_in": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
3
+ "hf_hook_point_out": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
4
+ "width": 1048576,
5
+ "model_name": "gemma-v3-1b-pt",
6
+ "architecture": "jump_relu",
7
+ "l0": 50,
8
+ "affine_connection": false
9
+ }
crosscoder/layer_7_13_17_22_width_1m_l0_medium/params_layer_0.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdf9fec97d8cf46d6197f98f34cc3a76659feea39f43608daa59f32eec86e071
3
+ size 6041899912
crosscoder/layer_7_13_17_22_width_1m_l0_medium/params_layer_1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74e32ae49726c632966f128b5d185e6a0f13d73bd67d5fa10b4935675269536f
3
+ size 6041899912
crosscoder/layer_7_13_17_22_width_1m_l0_medium/params_layer_2.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51bbf0f6d57b13fd2c21ce110fe8f4da945d78375a6ae86fc78811b7cbe87ae0
3
+ size 6041899912
crosscoder/layer_7_13_17_22_width_1m_l0_medium/params_layer_3.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70821dad30b188b1ca0e089b85cc8f5ac64169a607790305e23863ba5aaf1e41
3
+ size 6041899912
crosscoder/layer_7_13_17_22_width_262k_l0_big/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hf_hook_point_in": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
3
+ "hf_hook_point_out": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
4
+ "width": 262144,
5
+ "model_name": "gemma-v3-1b-pt",
6
+ "architecture": "jump_relu",
7
+ "l0": 150,
8
+ "affine_connection": false
9
+ }
crosscoder/layer_7_13_17_22_width_262k_l0_big/params_layer_0.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5a46142fbb60f62716b553c1bb30f1ca566b87fc180657c45ae4f9214937d9c
3
+ size 1510478720
crosscoder/layer_7_13_17_22_width_262k_l0_big/params_layer_1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6d410e300090345068af567552f6cc161e9cadce350b836389fe8593590d97a
3
+ size 1510478720
crosscoder/layer_7_13_17_22_width_262k_l0_big/params_layer_2.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa74f988831763e17aa49d2f3c995c8ff2ec02380c78a62ce8372a8a4fef32ef
3
+ size 1510478720
crosscoder/layer_7_13_17_22_width_262k_l0_big/params_layer_3.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d977fe0c290db54eecffcfd8c8d00b7e7425c063e83a61e738c58aeef4ce126
3
+ size 1510478720
crosscoder/layer_7_13_17_22_width_262k_l0_medium/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hf_hook_point_in": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
3
+ "hf_hook_point_out": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
4
+ "width": 262144,
5
+ "model_name": "gemma-v3-1b-pt",
6
+ "architecture": "jump_relu",
7
+ "l0": 50,
8
+ "affine_connection": false
9
+ }
crosscoder/layer_7_13_17_22_width_262k_l0_medium/params_layer_0.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e6601e7ad1526bc36804840bb107db5d08f805301c7a588573dfc96e52a482b
3
+ size 1510478720
crosscoder/layer_7_13_17_22_width_262k_l0_medium/params_layer_1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b9cf9228314d9399056fe4b1ec3aa6767d79a1538cc5e1afa977ee5d3b17779
3
+ size 1510478720
crosscoder/layer_7_13_17_22_width_262k_l0_medium/params_layer_2.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9b9d3e7bb93e86d341a2ee2a003fd85e3dbcd42102438da8591b7b19716c5e0
3
+ size 1510478720
crosscoder/layer_7_13_17_22_width_262k_l0_medium/params_layer_3.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75db10e54428eafb0b474f4d5c86476db352bf48e5a86feb531a420f6698bd28
3
+ size 1510478720
crosscoder/layer_7_13_17_22_width_524k_l0_big/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hf_hook_point_in": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
3
+ "hf_hook_point_out": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
4
+ "width": 524288,
5
+ "model_name": "gemma-v3-1b-pt",
6
+ "architecture": "jump_relu",
7
+ "l0": 150,
8
+ "affine_connection": false
9
+ }
crosscoder/layer_7_13_17_22_width_524k_l0_big/params_layer_0.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:165bbce957c2b71f2685a87cff9fc9cc10679a1902eee27933b7c2a95c471a48
3
+ size 3020952456
crosscoder/layer_7_13_17_22_width_524k_l0_big/params_layer_1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c31a424e0632e00126e43d723ed4185e65065cbc3b89624363422b866bbed7d7
3
+ size 3020952456
crosscoder/layer_7_13_17_22_width_524k_l0_big/params_layer_2.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fde4c7db3836a1dc89b9bd1f5b96f1dd23a2ba382e1c0b5c933b1f4a49de325
3
+ size 3020952456
crosscoder/layer_7_13_17_22_width_524k_l0_big/params_layer_3.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0a6418d2f8cccb10627ff2ebf3ddfec6bf476c6982ca77537edf8d71b408f2a
3
+ size 3020952456
crosscoder/layer_7_13_17_22_width_524k_l0_medium/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hf_hook_point_in": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
3
+ "hf_hook_point_out": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
4
+ "width": 524288,
5
+ "model_name": "gemma-v3-1b-pt",
6
+ "architecture": "jump_relu",
7
+ "l0": 50,
8
+ "affine_connection": false
9
+ }
crosscoder/layer_7_13_17_22_width_524k_l0_medium/params_layer_0.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5157dbe2bd3b436d7331c7406b260dda2751b009346b3451584d622efff21ca5
3
+ size 3020952456
crosscoder/layer_7_13_17_22_width_524k_l0_medium/params_layer_1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2572d6c53ede9a777d29207c9d46b08486ff4e64532d9c042f38bb10e54e8815
3
+ size 3020952456
crosscoder/layer_7_13_17_22_width_524k_l0_medium/params_layer_2.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58a295a9136ea05a104be0449327fe5e8f47a8ebe9051037d426c3c22105f5fc
3
+ size 3020952456
crosscoder/layer_7_13_17_22_width_524k_l0_medium/params_layer_3.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:328e90f30852353a00ae5e8988010ff9792debdccd3258b82f3c220c80f071e9
3
+ size 3020952456
crosscoder/layer_7_13_17_22_width_65k_l0_big/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hf_hook_point_in": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
3
+ "hf_hook_point_out": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
4
+ "width": 65536,
5
+ "model_name": "gemma-v3-1b-pt",
6
+ "architecture": "jump_relu",
7
+ "l0": 150,
8
+ "affine_connection": false
9
+ }
crosscoder/layer_7_13_17_22_width_65k_l0_big/params_layer_0.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5284453a253fbd41bef81a5794e77e3071107310e220c60a30f6dfed31f58bbc
3
+ size 377623424
crosscoder/layer_7_13_17_22_width_65k_l0_big/params_layer_1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec492264b76ca1b7f71038a86975dec578ee2d4da6ce3ca1218bb19eebba3f7f
3
+ size 377623424
crosscoder/layer_7_13_17_22_width_65k_l0_big/params_layer_2.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6035cdb8ff1235e95bee759909119da284a383a2011f91a411e031e17984fe9b
3
+ size 377623424
crosscoder/layer_7_13_17_22_width_65k_l0_big/params_layer_3.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:792b20afb0881d69c12ff323a454acb4ee93f3d37ba40bafe69a8784c100e033
3
+ size 377623424
crosscoder/layer_7_13_17_22_width_65k_l0_medium/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hf_hook_point_in": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
3
+ "hf_hook_point_out": "model.layers.{7.0,13.5,17.400000000000002,22.599999999999998}.output",
4
+ "width": 65536,
5
+ "model_name": "gemma-v3-1b-pt",
6
+ "architecture": "jump_relu",
7
+ "l0": 50,
8
+ "affine_connection": false
9
+ }
crosscoder/layer_7_13_17_22_width_65k_l0_medium/params_layer_0.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:228dddcb01dfb89e7b3bf2ba5fd23636fed7ae5c6154b3b2f4a552bdbb4552dc
3
+ size 377623424
crosscoder/layer_7_13_17_22_width_65k_l0_medium/params_layer_1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0a12e3e58b00bb84980118b571d40aaefa0e6b0ddbdf4c817b8c0ee67543aec
3
+ size 377623424
crosscoder/layer_7_13_17_22_width_65k_l0_medium/params_layer_2.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8af44b18e50dcd4c2228879ad7d4a0e01bd687c8c1fcd2c020cd9a877704cef3
3
+ size 377623424
crosscoder/layer_7_13_17_22_width_65k_l0_medium/params_layer_3.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7115c7dfb7fa2d943da39139d7c39887d79179ad8dd5f9003a18cdc563ba31bd
3
+ size 377623424