nev commited on
Commit
18bd18a
·
verified ·
1 Parent(s): 46de709

Add files using upload-large-folder tool

Browse files
Files changed (50) hide show
  1. layers.0.mlp/cfg.json +1 -0
  2. layers.0.mlp/sae.safetensors +3 -0
  3. layers.1.mlp/cfg.json +1 -0
  4. layers.1.mlp/sae.safetensors +3 -0
  5. layers.10.mlp/cfg.json +1 -0
  6. layers.10.mlp/sae.safetensors +3 -0
  7. layers.11.mlp/cfg.json +1 -0
  8. layers.11.mlp/sae.safetensors +3 -0
  9. layers.12.mlp/cfg.json +1 -0
  10. layers.12.mlp/sae.safetensors +3 -0
  11. layers.13.mlp/cfg.json +1 -0
  12. layers.13.mlp/sae.safetensors +3 -0
  13. layers.14.mlp/cfg.json +1 -0
  14. layers.14.mlp/sae.safetensors +3 -0
  15. layers.15.mlp/cfg.json +1 -0
  16. layers.15.mlp/sae.safetensors +3 -0
  17. layers.16.mlp/cfg.json +1 -0
  18. layers.16.mlp/sae.safetensors +3 -0
  19. layers.17.mlp/cfg.json +1 -0
  20. layers.17.mlp/sae.safetensors +3 -0
  21. layers.18.mlp/cfg.json +1 -0
  22. layers.18.mlp/sae.safetensors +3 -0
  23. layers.19.mlp/cfg.json +1 -0
  24. layers.19.mlp/sae.safetensors +3 -0
  25. layers.2.mlp/cfg.json +1 -0
  26. layers.2.mlp/sae.safetensors +3 -0
  27. layers.20.mlp/cfg.json +1 -0
  28. layers.21.mlp/cfg.json +1 -0
  29. layers.21.mlp/sae.safetensors +3 -0
  30. layers.22.mlp/cfg.json +1 -0
  31. layers.22.mlp/sae.safetensors +3 -0
  32. layers.23.mlp/cfg.json +1 -0
  33. layers.23.mlp/sae.safetensors +3 -0
  34. layers.24.mlp/cfg.json +1 -0
  35. layers.25.mlp/cfg.json +1 -0
  36. layers.25.mlp/sae.safetensors +3 -0
  37. layers.3.mlp/cfg.json +1 -0
  38. layers.3.mlp/sae.safetensors +3 -0
  39. layers.4.mlp/cfg.json +1 -0
  40. layers.4.mlp/sae.safetensors +3 -0
  41. layers.5.mlp/cfg.json +1 -0
  42. layers.5.mlp/sae.safetensors +3 -0
  43. layers.6.mlp/cfg.json +1 -0
  44. layers.6.mlp/sae.safetensors +3 -0
  45. layers.7.mlp/cfg.json +1 -0
  46. layers.7.mlp/sae.safetensors +3 -0
  47. layers.8.mlp/cfg.json +1 -0
  48. layers.8.mlp/sae.safetensors +3 -0
  49. layers.9.mlp/cfg.json +1 -0
  50. layers.9.mlp/sae.safetensors +3 -0
layers.0.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.0.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:715cd4229b3fd1a86609f9fcdcf32691d5a5f2b0bbd859930b521293569a1448
3
+ size 1218843568
layers.1.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.1.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f60b754673c07eae9c4010e9244f92b59fdfb6d8eda669faaa5b09c0438bf294
3
+ size 1218843568
layers.10.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.10.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3cb5d7f23c6b25b22aba586d8b250bb46f174ba9aa4a1c837b4cccd1f18dba6
3
+ size 1218843568
layers.11.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.11.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6b7af929810a88f118f5be12dd64677e9a108697f2235742c609e70418e2b24
3
+ size 1218843568
layers.12.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.12.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bd3cdd8043ed9c0f562e4440cc7325192c3c6ae37f0bfb403d60a70af465bf2
3
+ size 1218843568
layers.13.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.13.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ee52b4d121b8495d813b417b059e29552113e568dbaf383dfed69c317bfed27
3
+ size 1218843568
layers.14.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.14.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f94b46fc378d11db78fc17b49f409707494dba5601ab338bbdbd8677904355f4
3
+ size 1218843568
layers.15.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.15.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8942bc7d66cab319f36597bea703333801028deadc11a5fd0c370daf13262629
3
+ size 1218843568
layers.16.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.16.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e8cae90ffae75dc13d9c8a7b559af3995f54e25f86de87671b8140a22b538f6
3
+ size 1218843568
layers.17.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.17.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd0be016a064e6fecf23af3796b9c62f9cfc16cc51160f01606122e43f360bb9
3
+ size 1218843568
layers.18.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.18.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c7c2f7a7af6e291ba3b2ccc73cb561e67598f4837d4aee62f993d3061493f49
3
+ size 1218843568
layers.19.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.19.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d7dd9e35d5d6e37c681b2e10879d8f19aaf08479f978291a3852f752c5914e7
3
+ size 1218843568
layers.2.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.2.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8b8f6214d3d7cb50e4025b2c91194d3d0589061b12ba1a77520c770f1ec1f8b
3
+ size 1218843568
layers.20.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.21.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.21.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf2b9e99dcd4f49dbc26b107190a6c546cb54476d310d459d828e7d696b6fc0a
3
+ size 1218843568
layers.22.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.22.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88467231bc410e2a35a1f2ce62d074457c3fe5f1e6e611f64f7358252bca7244
3
+ size 1218843568
layers.23.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.23.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fab5d67209fd484b26e22a0b51cc7fd81e033e731b00f2b4068e3dee5cd32c44
3
+ size 1218843568
layers.24.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.25.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.25.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c6538fe6c6fef2c0261fcae7051adce723a528d42a8613d70b7190ca7570368
3
+ size 1218843568
layers.3.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.3.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:311c523e31d9cef96ad152a77e5d6c357a5e9b57d4dc510d816f12af33e3b3ba
3
+ size 1218843568
layers.4.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.4.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1051b19edd270d27f1959f280457009d8111b03b77cbfcb6b20868bb5170138e
3
+ size 1218843568
layers.5.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.5.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d32b8edfff9e69b0cf0db7d170facc5cbc01f82d6b316963ec6ce06c01b7706b
3
+ size 1218843568
layers.6.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.6.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83978d65e7b3391f9eb7b06c4be99dbf5a336dc7ad1abe70beb629a2da48ea5d
3
+ size 1218843568
layers.7.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.7.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76f48b7db010e8090ce25b861fdd763a43f8915836ae8933ec50410039d30bbc
3
+ size 1218843568
layers.8.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.8.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:459da24e29abe98e7fe14ba382f2dc87ca9153ada6a8c1348e4477ce72c450cd
3
+ size 1218843568
layers.9.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 128, "multi_topk": false, "skip_connection": true, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.9.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e6505b91ebab5789f8339e7a97c880da32e16901277d2c67761f9fb6d834b1c
3
+ size 1218843568