nev commited on
Commit
c8422c4
·
verified ·
1 Parent(s): 57d9a53

Add files using upload-large-folder tool

Browse files
latents/layers.0.mlp/0_14744.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:598236a7535176ca16cee44366f64c2cfbb81f46b24d5e894fd4c168cb4489e8
3
+ size 420880368
latents/layers.0.mlp/14745_29490.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7f75e8e7a885ef6cf819aaa8beb4544c6ee288d75a441d62f0c25cff457caee
3
+ size 607535112
latents/layers.0.mlp/29491_44235.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0131946220d3ddafbca83d2e2c1b87dd2fdb5eecfbb414dbdfb181b363260088
3
+ size 610682336
latents/layers.0.mlp/44236_58981.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ed50a8e23993dc7cb3fdb04d091f1842eeff079958910656081906a6de2387b
3
+ size 490633344
latents/layers.0.mlp/58982_73727.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d796bd2770434fd51122c4c3145e089cad58e9b69249ff9cdfcf269eb3b94217
3
+ size 439386160
latents/layers.1.mlp/0_14744.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1aaed21f904b3d8c93e41e64d2ef2b40c89f6b0bb030b8f68ec20ddfda802a7
3
+ size 547091736
latents/layers.1.mlp/14745_29490.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0563cb177155b8174364ad735c08046145deee36e9292ac310fb09c316ff0855
3
+ size 594872344
latents/layers.1.mlp/29491_44235.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d430c29e165274a17c2ded93ed642bc4e9cd900bd571cba75b55598ab3a8d14e
3
+ size 570198728
latents/layers.1.mlp/44236_58981.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ad2558f50faa2fd941be9f5a9ff9bf2d6386144a73f5a3575cb042df406c470
3
+ size 597988448
latents/layers.1.mlp/58982_73727.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb5b78a2f862d0e7b576e2c966521f208fefda012153ab0673591aea0c76b1d2
3
+ size 648453376
latents/layers.2.mlp/0_14744.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:869b712302360e66f34c22bc6e16d29c010eac490d9fbc2a2e20bb3124456543
3
+ size 605773104
latents/layers.2.mlp/14745_29490.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b9367d555585c51bfcead8c09b9e928515185726d35a130d0a71f4f929776c9
3
+ size 612954888
latents/layers.2.mlp/29491_44235.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adb563594f8fb25fafbe496e690c5feec523ae8f4e55b051a7e60a8e68400f9c
3
+ size 593472128
latents/layers.2.mlp/44236_58981.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:722a08373f846f10ac1c0fcd90d47b82a1e7c155714a2bb39f5dac57313af3fb
3
+ size 559319504
latents/layers.2.mlp/58982_73727.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72ef9b67135ce775ffba580eb4d4aeba5aa93e42cc14f5b084712d761ca5c1ab
3
+ size 587988360
latents/layers.3.mlp/0_14744.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9d624ec821501736030683311c092bc7d75dfccf15cb21eeadb6acfec03f124
3
+ size 646849224
latents/layers.3.mlp/14745_29490.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7561e6f5b70943e2d59791e76cb8641a15e8e3acd4943b4f54f79fc54e882f5f
3
+ size 542507240
latents/layers.3.mlp/29491_44235.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebfd788597668c7538cf97b19c4e4f8e47e8edfedad3e9a3a4cd72762d677d96
3
+ size 551425832
latents/layers.3.mlp/44236_58981.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:daa97b583b9850b6776c6cf07cda17bdf4ad4889f7017767bd03703717cad165
3
+ size 577676304
latents/layers.3.mlp/58982_73727.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:679b664ad1a7ec038f121e06cbb44784feadcbd99bedc08d6956ad1a5624261d
3
+ size 641050024
run_config.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cache_cfg": {
3
+ "dataset_repo": "EleutherAI/fineweb-edu-dedup-10b",
4
+ "dataset_split": "train",
5
+ "dataset_name": "",
6
+ "dataset_column": "text",
7
+ "batch_size": 16,
8
+ "cache_ctx_len": 256,
9
+ "n_tokens": 10000000,
10
+ "n_splits": 5
11
+ },
12
+ "constructor_cfg": {
13
+ "faiss_embedding_model": "sentence-transformers/all-MiniLM-L6-v2",
14
+ "faiss_embedding_cache_dir": ".embedding_cache",
15
+ "faiss_embedding_cache_enabled": true,
16
+ "example_ctx_len": 32,
17
+ "min_examples": 200,
18
+ "n_non_activating": 50,
19
+ "center_examples": true,
20
+ "non_activating_source": "random",
21
+ "neighbours_type": "co-occurrence"
22
+ },
23
+ "sampler_cfg": {
24
+ "n_examples_train": 40,
25
+ "n_examples_test": 50,
26
+ "n_quantiles": 10,
27
+ "train_type": "quantiles",
28
+ "test_type": "quantiles",
29
+ "ratio_top": 0.2
30
+ },
31
+ "model": "HuggingFaceTB/SmolLM2-135M",
32
+ "sparse_model": "/mnt/ssd-1/nev/e2e/checkpoints/smollm/k32",
33
+ "hookpoints": [
34
+ "layers.0.mlp",
35
+ "layers.1.mlp",
36
+ "layers.2.mlp",
37
+ "layers.3.mlp",
38
+ "layers.4.mlp",
39
+ "layers.5.mlp",
40
+ "layers.6.mlp",
41
+ "layers.7.mlp",
42
+ "layers.8.mlp",
43
+ "layers.9.mlp",
44
+ "layers.10.mlp",
45
+ "layers.11.mlp",
46
+ "layers.12.mlp",
47
+ "layers.13.mlp",
48
+ "layers.14.mlp",
49
+ "layers.15.mlp",
50
+ "layers.16.mlp",
51
+ "layers.17.mlp",
52
+ "layers.18.mlp",
53
+ "layers.19.mlp",
54
+ "layers.20.mlp",
55
+ "layers.21.mlp",
56
+ "layers.22.mlp",
57
+ "layers.23.mlp",
58
+ "layers.24.mlp",
59
+ "layers.25.mlp",
60
+ "layers.26.mlp",
61
+ "layers.27.mlp",
62
+ "layers.28.mlp",
63
+ "layers.29.mlp"
64
+ ],
65
+ "explainer_model": "hugging-quants/Meta-Llama-3.1-70B-Instruct-AWQ-INT4",
66
+ "explainer_model_max_len": 5120,
67
+ "explainer_provider": "offline",
68
+ "explainer": "default",
69
+ "scorers": [
70
+ "fuzz",
71
+ "detection"
72
+ ],
73
+ "name": "smollm-v1",
74
+ "max_latents": null,
75
+ "filter_bos": false,
76
+ "log_probs": false,
77
+ "load_in_8bit": false,
78
+ "hf_token": null,
79
+ "pipeline_num_proc": 48,
80
+ "num_gpus": 1,
81
+ "seed": 22,
82
+ "verbose": true,
83
+ "num_examples_per_scorer_prompt": 5,
84
+ "overwrite": []
85
+ }