canrager commited on
Commit
0a98929
·
verified ·
1 Parent(s): db07a7f

Upload folder using huggingface_hub

Browse files
gemma_scope/scr_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_107_eval_results.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_type_id": "scr",
3
+ "eval_config": {
4
+ "random_seed": 42,
5
+ "dataset_names": [
6
+ "LabHC/bias_in_bios_class_set1",
7
+ "canrager/amazon_reviews_mcauley_1and5"
8
+ ],
9
+ "perform_scr": true,
10
+ "early_stopping_patience": 20,
11
+ "train_set_size": 4000,
12
+ "test_set_size": 1000,
13
+ "context_length": 128,
14
+ "probe_train_batch_size": 16,
15
+ "probe_test_batch_size": 500,
16
+ "probe_epochs": 20,
17
+ "probe_lr": 0.001,
18
+ "probe_l1_penalty": 0.001,
19
+ "sae_batch_size": 5,
20
+ "llm_batch_size": 32,
21
+ "llm_dtype": "bfloat16",
22
+ "lower_vram_usage": true,
23
+ "model_name": "gemma-2-2b",
24
+ "n_values": [
25
+ 2,
26
+ 5,
27
+ 10,
28
+ 20,
29
+ 50,
30
+ 100,
31
+ 500
32
+ ],
33
+ "column1_vals_lookup": {
34
+ "LabHC/bias_in_bios_class_set1": [
35
+ [
36
+ "professor",
37
+ "nurse"
38
+ ],
39
+ [
40
+ "architect",
41
+ "journalist"
42
+ ],
43
+ [
44
+ "surgeon",
45
+ "psychologist"
46
+ ],
47
+ [
48
+ "attorney",
49
+ "teacher"
50
+ ]
51
+ ],
52
+ "canrager/amazon_reviews_mcauley_1and5": [
53
+ [
54
+ "Books",
55
+ "CDs_and_Vinyl"
56
+ ],
57
+ [
58
+ "Software",
59
+ "Electronics"
60
+ ],
61
+ [
62
+ "Pet_Supplies",
63
+ "Office_Products"
64
+ ],
65
+ [
66
+ "Industrial_and_Scientific",
67
+ "Toys_and_Games"
68
+ ]
69
+ ]
70
+ }
71
+ },
72
+ "eval_id": "62aa6a87-f19c-4661-af6b-3cf92ac303eb",
73
+ "datetime_epoch_millis": 1732735217643,
74
+ "eval_result_metrics": {
75
+ "scr_metrics": {
76
+ "scr_dir1_threshold_2": 0.1952003481008311,
77
+ "scr_metric_threshold_2": 0.02917879276761306,
78
+ "scr_dir2_threshold_2": 0.026848935672591483,
79
+ "scr_dir1_threshold_5": 0.20431255623082573,
80
+ "scr_metric_threshold_5": 0.05379061133677988,
81
+ "scr_dir2_threshold_5": 0.05336396198727458,
82
+ "scr_dir1_threshold_10": 0.22934789158586028,
83
+ "scr_metric_threshold_10": 0.07765409422399916,
84
+ "scr_dir2_threshold_10": 0.07811649736013061,
85
+ "scr_dir1_threshold_20": 0.2473006631023363,
86
+ "scr_metric_threshold_20": 0.11603065634829988,
87
+ "scr_dir2_threshold_20": 0.11589013672422316,
88
+ "scr_dir1_threshold_50": 0.31414080783599213,
89
+ "scr_metric_threshold_50": 0.1950069114030944,
90
+ "scr_dir2_threshold_50": 0.20039980906054766,
91
+ "scr_dir1_threshold_100": 0.34433168446809054,
92
+ "scr_metric_threshold_100": 0.2563312840537805,
93
+ "scr_dir2_threshold_100": 0.25838013737959736,
94
+ "scr_dir1_threshold_500": 0.2132407243335966,
95
+ "scr_metric_threshold_500": 0.3111206132737665,
96
+ "scr_dir2_threshold_500": 0.3042843125498677
97
+ }
98
+ },
99
+ "eval_result_details": [
100
+ {
101
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
102
+ "scr_dir1_threshold_2": 0.42187508731147705,
103
+ "scr_metric_threshold_2": 0.0,
104
+ "scr_dir2_threshold_2": 0.0,
105
+ "scr_dir1_threshold_5": 0.43749988358469727,
106
+ "scr_metric_threshold_5": 0.0024630223664217393,
107
+ "scr_dir2_threshold_5": 0.0024630223664217393,
108
+ "scr_dir1_threshold_10": 0.4531256111803394,
109
+ "scr_metric_threshold_10": 0.012315258641572036,
110
+ "scr_dir2_threshold_10": 0.012315258641572036,
111
+ "scr_dir1_threshold_20": 0.42187508731147705,
112
+ "scr_metric_threshold_20": 0.03694577592471611,
113
+ "scr_dir2_threshold_20": 0.03694577592471611,
114
+ "scr_dir1_threshold_50": 0.4687504074535596,
115
+ "scr_metric_threshold_50": 0.05911327084143844,
116
+ "scr_dir2_threshold_50": 0.05911327084143844,
117
+ "scr_dir1_threshold_100": 0.4843752037267798,
118
+ "scr_metric_threshold_100": 0.07389155184943222,
119
+ "scr_dir2_threshold_100": 0.07389155184943222,
120
+ "scr_dir1_threshold_500": 0.3749997671693945,
121
+ "scr_metric_threshold_500": 0.12807877795802716,
122
+ "scr_dir2_threshold_500": 0.12807877795802716
123
+ },
124
+ {
125
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
126
+ "scr_dir1_threshold_2": 0.21782190488161987,
127
+ "scr_metric_threshold_2": 0.03988603456423997,
128
+ "scr_dir2_threshold_2": 0.03988603456423997,
129
+ "scr_dir1_threshold_5": 0.18811873007892008,
130
+ "scr_metric_threshold_5": 0.08831913718288759,
131
+ "scr_dir2_threshold_5": 0.08831913718288759,
132
+ "scr_dir1_threshold_10": 0.19801959163150878,
133
+ "scr_metric_threshold_10": 0.10826223937190631,
134
+ "scr_dir2_threshold_10": 0.10826223937190631,
135
+ "scr_dir1_threshold_20": 0.22772276643420858,
136
+ "scr_metric_threshold_20": 0.15099724001634968,
137
+ "scr_dir2_threshold_20": 0.15099724001634968,
138
+ "scr_dir1_threshold_50": 0.18811873007892008,
139
+ "scr_metric_threshold_50": 0.2564103434942551,
140
+ "scr_dir2_threshold_50": 0.2564103434942551,
141
+ "scr_dir1_threshold_100": 0.15841614542115398,
142
+ "scr_metric_threshold_100": 0.3304843106485308,
143
+ "scr_dir2_threshold_100": 0.3304843106485308,
144
+ "scr_dir1_threshold_500": 0.09900979581575439,
145
+ "scr_metric_threshold_500": 0.15384620609655306,
146
+ "scr_dir2_threshold_500": 0.15384620609655306
147
+ },
148
+ {
149
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
150
+ "scr_dir1_threshold_2": 0.4603179859314168,
151
+ "scr_metric_threshold_2": 0.01015230346203094,
152
+ "scr_dir2_threshold_2": 0.01015230346203094,
153
+ "scr_dir1_threshold_5": 0.5396829601737049,
154
+ "scr_metric_threshold_5": 0.015228530833448307,
155
+ "scr_dir2_threshold_5": 0.015228530833448307,
156
+ "scr_dir1_threshold_10": 0.5555557658011382,
157
+ "scr_metric_threshold_10": 0.022842644969368664,
158
+ "scr_dir2_threshold_10": 0.022842644969368664,
159
+ "scr_dir1_threshold_20": 0.5714285714285714,
160
+ "scr_metric_threshold_20": 0.05076151731015469,
161
+ "scr_dir2_threshold_20": 0.05076151731015469,
162
+ "scr_dir1_threshold_50": 0.5714285714285714,
163
+ "scr_metric_threshold_50": 0.12690356635418104,
164
+ "scr_dir2_threshold_50": 0.12690356635418104,
165
+ "scr_dir1_threshold_100": 0.5555557658011382,
166
+ "scr_metric_threshold_100": 0.15989851478558065,
167
+ "scr_dir2_threshold_100": 0.15989851478558065,
168
+ "scr_dir1_threshold_500": -0.07936497424228806,
169
+ "scr_metric_threshold_500": 0.06852793490810599,
170
+ "scr_dir2_threshold_500": 0.06852793490810599
171
+ },
172
+ {
173
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
174
+ "scr_dir1_threshold_2": 0.31007766623341637,
175
+ "scr_metric_threshold_2": 0.03273803188281158,
176
+ "scr_dir2_threshold_2": 0.03273803188281158,
177
+ "scr_dir1_threshold_5": 0.3023258285391642,
178
+ "scr_metric_threshold_5": 0.06249991130260288,
179
+ "scr_dir2_threshold_5": 0.06249991130260288,
180
+ "scr_dir1_threshold_10": 0.31007766623341637,
181
+ "scr_metric_threshold_10": 0.07440469854947825,
182
+ "scr_dir2_threshold_10": 0.07440469854947825,
183
+ "scr_dir1_threshold_20": 0.3410854790619061,
184
+ "scr_metric_threshold_20": 0.11011906029010435,
185
+ "scr_dir2_threshold_20": 0.11011906029010435,
186
+ "scr_dir1_threshold_50": 0.5116277565413784,
187
+ "scr_metric_threshold_50": 0.1488093970989565,
188
+ "scr_dir2_threshold_50": 0.1488093970989565,
189
+ "scr_dir1_threshold_100": 0.6821704960723314,
190
+ "scr_metric_threshold_100": 0.22916657796926954,
191
+ "scr_dir2_threshold_100": 0.22916657796926954,
192
+ "scr_dir1_threshold_500": 0.3255813416219208,
193
+ "scr_metric_threshold_500": 0.12797615246302027,
194
+ "scr_dir2_threshold_500": 0.12797615246302027
195
+ },
196
+ {
197
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
198
+ "scr_dir1_threshold_2": 0.0,
199
+ "scr_metric_threshold_2": 0.06225692845935926,
200
+ "scr_dir2_threshold_2": 0.06225692845935926,
201
+ "scr_dir1_threshold_5": -0.016393554752069144,
202
+ "scr_metric_threshold_5": 0.12062262494651238,
203
+ "scr_dir2_threshold_5": 0.12062262494651238,
204
+ "scr_dir1_threshold_10": 0.01092881936245188,
205
+ "scr_metric_threshold_10": 0.16342408931882568,
206
+ "scr_dir2_threshold_10": 0.16342408931882568,
207
+ "scr_dir1_threshold_20": 0.021857964433295084,
208
+ "scr_metric_threshold_20": 0.20622578561580057,
209
+ "scr_dir2_threshold_20": 0.20622578561580057,
210
+ "scr_dir1_threshold_50": 0.08196712234356307,
211
+ "scr_metric_threshold_50": 0.31906617854256225,
212
+ "scr_dir2_threshold_50": 0.31906617854256225,
213
+ "scr_dir1_threshold_100": 0.05464474822904205,
214
+ "scr_metric_threshold_100": 0.3891051070970106,
215
+ "scr_dir2_threshold_100": 0.3891051070970106,
216
+ "scr_dir1_threshold_500": -0.021857964433295084,
217
+ "scr_metric_threshold_500": 0.6381323570851245,
218
+ "scr_dir2_threshold_500": 0.6381323570851245
219
+ },
220
+ {
221
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
222
+ "scr_dir1_threshold_2": 0.04123688851473832,
223
+ "scr_metric_threshold_2": 0.032258126539686356,
224
+ "scr_dir2_threshold_2": 0.032258126539686356,
225
+ "scr_dir1_threshold_5": 0.04639146117402712,
226
+ "scr_metric_threshold_5": 0.036290332271825994,
227
+ "scr_dir2_threshold_5": 0.036290332271825994,
228
+ "scr_dir1_threshold_10": 0.07216493895132697,
229
+ "scr_metric_threshold_10": 0.06451625307937271,
230
+ "scr_dir2_threshold_10": 0.06451625307937271,
231
+ "scr_dir1_threshold_20": 0.1030926821474877,
232
+ "scr_metric_threshold_20": 0.09677413927777444,
233
+ "scr_dir2_threshold_20": 0.09677413927777444,
234
+ "scr_dir1_threshold_50": 0.2113399369542642,
235
+ "scr_metric_threshold_50": 0.17741945562699032,
236
+ "scr_dir2_threshold_50": 0.17741945562699032,
237
+ "scr_dir1_threshold_100": 0.18041219375810347,
238
+ "scr_metric_threshold_100": 0.2419354683650784,
239
+ "scr_dir2_threshold_100": 0.2419354683650784,
240
+ "scr_dir1_threshold_500": 0.2783503032463024,
241
+ "scr_metric_threshold_500": 0.4596774619960344,
242
+ "scr_dir2_threshold_500": 0.4596774619960344
243
+ },
244
+ {
245
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
246
+ "scr_dir1_threshold_2": 0.06306305580659313,
247
+ "scr_metric_threshold_2": 0.008928721105389243,
248
+ "scr_dir2_threshold_2": 0.008928721105389243,
249
+ "scr_dir1_threshold_5": 0.0810811101069608,
250
+ "scr_metric_threshold_5": 0.04910730084933944,
251
+ "scr_dir2_threshold_5": 0.04910730084933944,
252
+ "scr_dir1_threshold_10": 0.15765750527178912,
253
+ "scr_metric_threshold_10": 0.09821433560655832,
254
+ "scr_dir2_threshold_10": 0.09821433560655832,
255
+ "scr_dir1_threshold_20": 0.17117111311941172,
256
+ "scr_metric_threshold_20": 0.15625009146916644,
257
+ "scr_dir2_threshold_20": 0.15625009146916644,
258
+ "scr_dir1_threshold_50": 0.34684667269156855,
259
+ "scr_metric_threshold_50": 0.3392858140702595,
260
+ "scr_dir2_threshold_50": 0.3392858140702595,
261
+ "scr_dir1_threshold_100": 0.445945837098897,
262
+ "scr_metric_threshold_100": 0.43303565607806294,
263
+ "scr_dir2_threshold_100": 0.43303565607806294,
264
+ "scr_dir1_threshold_500": 0.445945837098897,
265
+ "scr_metric_threshold_500": 0.6294643272911795,
266
+ "scr_dir2_threshold_500": 0.6294643272911795
267
+ },
268
+ {
269
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
270
+ "scr_dir1_threshold_2": 0.047210196127387125,
271
+ "scr_metric_threshold_2": 0.047210196127387125,
272
+ "scr_dir2_threshold_2": 0.0285713393672145,
273
+ "scr_dir1_threshold_5": 0.05579403094120067,
274
+ "scr_metric_threshold_5": 0.05579403094120067,
275
+ "scr_dir2_threshold_5": 0.05238083614515829,
276
+ "scr_dir1_threshold_10": 0.07725323425491137,
277
+ "scr_metric_threshold_10": 0.07725323425491137,
278
+ "scr_dir2_threshold_10": 0.08095245934396302,
279
+ "scr_dir1_threshold_20": 0.12017164088233277,
280
+ "scr_metric_threshold_20": 0.12017164088233277,
281
+ "scr_dir2_threshold_20": 0.11904748388971893,
282
+ "scr_dir1_threshold_50": 0.13304726519611204,
283
+ "scr_metric_threshold_50": 0.13304726519611204,
284
+ "scr_dir2_threshold_50": 0.17619044645573817,
285
+ "scr_dir1_threshold_100": 0.19313308563727843,
286
+ "scr_metric_threshold_100": 0.19313308563727843,
287
+ "scr_dir2_threshold_100": 0.20952391224381361,
288
+ "scr_dir1_threshold_500": 0.283261688392087,
289
+ "scr_metric_threshold_500": 0.283261688392087,
290
+ "scr_dir2_threshold_500": 0.22857128260089646
291
+ }
292
+ ],
293
+ "sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
294
+ "sae_lens_id": "layer_12/width_1m/average_l0_107",
295
+ "sae_lens_release_id": "gemma-scope-2b-pt-res",
296
+ "sae_lens_version": "4.4.5",
297
+ "eval_result_unstructured": null
298
+ }
gemma_scope/scr_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_19_eval_results.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_type_id": "scr",
3
+ "eval_config": {
4
+ "random_seed": 42,
5
+ "dataset_names": [
6
+ "LabHC/bias_in_bios_class_set1",
7
+ "canrager/amazon_reviews_mcauley_1and5"
8
+ ],
9
+ "perform_scr": true,
10
+ "early_stopping_patience": 20,
11
+ "train_set_size": 4000,
12
+ "test_set_size": 1000,
13
+ "context_length": 128,
14
+ "probe_train_batch_size": 16,
15
+ "probe_test_batch_size": 500,
16
+ "probe_epochs": 20,
17
+ "probe_lr": 0.001,
18
+ "probe_l1_penalty": 0.001,
19
+ "sae_batch_size": 5,
20
+ "llm_batch_size": 32,
21
+ "llm_dtype": "bfloat16",
22
+ "lower_vram_usage": true,
23
+ "model_name": "gemma-2-2b",
24
+ "n_values": [
25
+ 2,
26
+ 5,
27
+ 10,
28
+ 20,
29
+ 50,
30
+ 100,
31
+ 500
32
+ ],
33
+ "column1_vals_lookup": {
34
+ "LabHC/bias_in_bios_class_set1": [
35
+ [
36
+ "professor",
37
+ "nurse"
38
+ ],
39
+ [
40
+ "architect",
41
+ "journalist"
42
+ ],
43
+ [
44
+ "surgeon",
45
+ "psychologist"
46
+ ],
47
+ [
48
+ "attorney",
49
+ "teacher"
50
+ ]
51
+ ],
52
+ "canrager/amazon_reviews_mcauley_1and5": [
53
+ [
54
+ "Books",
55
+ "CDs_and_Vinyl"
56
+ ],
57
+ [
58
+ "Software",
59
+ "Electronics"
60
+ ],
61
+ [
62
+ "Pet_Supplies",
63
+ "Office_Products"
64
+ ],
65
+ [
66
+ "Industrial_and_Scientific",
67
+ "Toys_and_Games"
68
+ ]
69
+ ]
70
+ }
71
+ },
72
+ "eval_id": "ccb94d0b-5b7c-4bae-83f4-12bebdffd7fa",
73
+ "datetime_epoch_millis": 1732747908562,
74
+ "eval_result_metrics": {
75
+ "scr_metrics": {
76
+ "scr_dir1_threshold_2": 0.061243707307435294,
77
+ "scr_metric_threshold_2": 0.013333291563881291,
78
+ "scr_dir2_threshold_2": 0.015831761609791675,
79
+ "scr_dir1_threshold_5": 0.09168076259777322,
80
+ "scr_metric_threshold_5": 0.038698233129179124,
81
+ "scr_dir2_threshold_5": 0.03815410248408284,
82
+ "scr_dir1_threshold_10": 0.13691496896076732,
83
+ "scr_metric_threshold_10": 0.06056341775485053,
84
+ "scr_dir2_threshold_10": 0.05858610748123473,
85
+ "scr_dir1_threshold_20": 0.030612087200795557,
86
+ "scr_metric_threshold_20": 0.0768146610397201,
87
+ "scr_dir2_threshold_20": 0.08210539131678352,
88
+ "scr_dir1_threshold_50": -0.226625297148567,
89
+ "scr_metric_threshold_50": 0.14296252707697382,
90
+ "scr_dir2_threshold_50": 0.13758238035660975,
91
+ "scr_dir1_threshold_100": -0.20399739480516063,
92
+ "scr_metric_threshold_100": 0.19868579375350148,
93
+ "scr_dir2_threshold_100": 0.1996105645468156,
94
+ "scr_dir1_threshold_500": -0.23518319131373794,
95
+ "scr_metric_threshold_500": 0.2517437783396029,
96
+ "scr_dir2_threshold_500": 0.2693275541114566
97
+ }
98
+ },
99
+ "eval_result_details": [
100
+ {
101
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
102
+ "scr_dir1_threshold_2": 0.14062502910382568,
103
+ "scr_metric_threshold_2": 0.012315258641572036,
104
+ "scr_dir2_threshold_2": 0.012315258641572036,
105
+ "scr_dir1_threshold_5": 0.20312514551912844,
106
+ "scr_metric_threshold_5": 0.02955656201598755,
107
+ "scr_dir2_threshold_5": 0.02955656201598755,
108
+ "scr_dir1_threshold_10": 0.2656252619344312,
109
+ "scr_metric_threshold_10": 0.06650233794070366,
110
+ "scr_dir2_threshold_10": 0.06650233794070366,
111
+ "scr_dir1_threshold_20": 0.2656252619344312,
112
+ "scr_metric_threshold_20": 0.11576351931645514,
113
+ "scr_dir2_threshold_20": 0.11576351931645514,
114
+ "scr_dir1_threshold_50": 0.2343756693879908,
115
+ "scr_metric_threshold_50": 0.16502455388274329,
116
+ "scr_dir2_threshold_50": 0.16502455388274329,
117
+ "scr_dir1_threshold_100": 0.17187555297268803,
118
+ "scr_metric_threshold_100": 0.18965507116588734,
119
+ "scr_dir2_threshold_100": 0.18965507116588734,
120
+ "scr_dir1_threshold_500": 0.1250002328306055,
121
+ "scr_metric_threshold_500": 0.21921177999133823,
122
+ "scr_dir2_threshold_500": 0.21921177999133823
123
+ },
124
+ {
125
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
126
+ "scr_dir1_threshold_2": 0.15841614542115398,
127
+ "scr_metric_threshold_2": 0.03133913632362978,
128
+ "scr_dir2_threshold_2": 0.03133913632362978,
129
+ "scr_dir1_threshold_5": 0.17821786852633137,
130
+ "scr_metric_threshold_5": 0.048433102618647625,
131
+ "scr_dir2_threshold_5": 0.048433102618647625,
132
+ "scr_dir1_threshold_10": 0.21782190488161987,
133
+ "scr_metric_threshold_10": 0.06267810283346208,
134
+ "scr_dir2_threshold_10": 0.06267810283346208,
135
+ "scr_dir1_threshold_20": 0.2376236279867973,
136
+ "scr_metric_threshold_20": 0.0826212050224808,
137
+ "scr_dir2_threshold_20": 0.0826212050224808,
138
+ "scr_dir1_threshold_50": -0.514851292328883,
139
+ "scr_metric_threshold_50": 0.2364672413052364,
140
+ "scr_dir2_threshold_50": 0.2364672413052364,
141
+ "scr_dir1_threshold_100": -0.7128708839603919,
142
+ "scr_metric_threshold_100": 0.273504309789273,
143
+ "scr_dir2_threshold_100": 0.273504309789273,
144
+ "scr_dir1_threshold_500": -1.0,
145
+ "scr_metric_threshold_500": 0.2820513778436806,
146
+ "scr_dir2_threshold_500": 0.2820513778436806
147
+ },
148
+ {
149
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
150
+ "scr_dir1_threshold_2": 0.1587299484845761,
151
+ "scr_metric_threshold_2": 0.007614265416724154,
152
+ "scr_dir2_threshold_2": 0.007614265416724154,
153
+ "scr_dir1_threshold_5": 0.20634931147199764,
154
+ "scr_metric_threshold_5": 0.012690492788141522,
155
+ "scr_dir2_threshold_5": 0.012690492788141522,
156
+ "scr_dir1_threshold_10": 0.2222221170994309,
157
+ "scr_metric_threshold_10": 0.015228530833448307,
158
+ "scr_dir2_threshold_10": 0.015228530833448307,
159
+ "scr_dir1_threshold_20": -0.4761898454537284,
160
+ "scr_metric_threshold_20": 0.05837563144607505,
161
+ "scr_dir2_threshold_20": 0.05837563144607505,
162
+ "scr_dir1_threshold_50": -1.285713339609164,
163
+ "scr_metric_threshold_50": 0.08121827641544371,
164
+ "scr_dir2_threshold_50": 0.08121827641544371,
165
+ "scr_dir1_threshold_100": -1.126983391124588,
166
+ "scr_metric_threshold_100": 0.2157361081863489,
167
+ "scr_dir2_threshold_100": 0.2157361081863489,
168
+ "scr_dir1_threshold_500": -0.2380949227268642,
169
+ "scr_metric_threshold_500": 0.30964472610913035,
170
+ "scr_dir2_threshold_500": 0.30964472610913035
171
+ },
172
+ {
173
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
174
+ "scr_dir1_threshold_2": 0.09302343848546918,
175
+ "scr_metric_threshold_2": 0.008928457389060843,
176
+ "scr_dir2_threshold_2": 0.008928457389060843,
177
+ "scr_dir1_threshold_5": 0.10852711387397361,
178
+ "scr_metric_threshold_5": 0.026785726956771016,
179
+ "scr_dir2_threshold_5": 0.026785726956771016,
180
+ "scr_dir1_threshold_10": 0.13953492670246334,
181
+ "scr_metric_threshold_10": 0.056547606376562316,
182
+ "scr_dir2_threshold_10": 0.056547606376562316,
183
+ "scr_dir1_threshold_20": 0.17054273953095306,
184
+ "scr_metric_threshold_20": 0.125,
185
+ "scr_dir2_threshold_20": 0.125,
186
+ "scr_dir1_threshold_50": -0.6046511950268475,
187
+ "scr_metric_threshold_50": 0.13392845738906084,
188
+ "scr_dir2_threshold_50": 0.13392845738906084,
189
+ "scr_dir1_threshold_100": -0.372092829838915,
190
+ "scr_metric_threshold_100": 0.13095230492604057,
191
+ "scr_dir2_threshold_100": 0.13095230492604057,
192
+ "scr_dir1_threshold_500": -1.3643409921446628,
193
+ "scr_metric_threshold_500": 0.10416657796926955,
194
+ "scr_dir2_threshold_500": 0.10416657796926955
195
+ },
196
+ {
197
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
198
+ "scr_dir1_threshold_2": -0.06557389329988525,
199
+ "scr_metric_threshold_2": 0.015564232114839816,
200
+ "scr_dir2_threshold_2": 0.015564232114839816,
201
+ "scr_dir1_threshold_5": 0.016393554752069144,
202
+ "scr_metric_threshold_5": 0.050583696392064004,
203
+ "scr_dir2_threshold_5": 0.050583696392064004,
204
+ "scr_dir1_threshold_10": 0.16939898007674342,
205
+ "scr_metric_threshold_10": 0.06614792850690382,
206
+ "scr_dir2_threshold_10": 0.06614792850690382,
207
+ "scr_dir1_threshold_20": 0.00546440968122594,
208
+ "scr_metric_threshold_20": 0.007782232019750705,
209
+ "scr_dir2_threshold_20": 0.007782232019750705,
210
+ "scr_dir1_threshold_50": 0.06557389329988525,
211
+ "scr_metric_threshold_50": 0.16731532129103183,
212
+ "scr_dir2_threshold_50": 0.16731532129103183,
213
+ "scr_dir1_threshold_100": 0.05464474822904205,
214
+ "scr_metric_threshold_100": 0.18287955340587164,
215
+ "scr_dir2_threshold_100": 0.18287955340587164,
216
+ "scr_dir1_threshold_500": -0.05464474822904205,
217
+ "scr_metric_threshold_500": 0.33463041065740207,
218
+ "scr_dir2_threshold_500": 0.33463041065740207
219
+ },
220
+ {
221
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
222
+ "scr_dir1_threshold_2": 0.0051545726592888,
223
+ "scr_metric_threshold_2": 0.0,
224
+ "scr_dir2_threshold_2": 0.0,
225
+ "scr_dir1_threshold_5": 0.0051545726592888,
226
+ "scr_metric_threshold_5": 0.036290332271825994,
227
+ "scr_dir2_threshold_5": 0.036290332271825994,
228
+ "scr_dir1_threshold_10": 0.025773170536871923,
229
+ "scr_metric_threshold_10": 0.06854845881151235,
230
+ "scr_dir2_threshold_10": 0.06854845881151235,
231
+ "scr_dir1_threshold_20": -0.010309452559005524,
232
+ "scr_metric_threshold_20": 0.05241939554166917,
233
+ "scr_dir2_threshold_20": 0.05241939554166917,
234
+ "scr_dir1_threshold_50": 0.04639146117402712,
235
+ "scr_metric_threshold_50": 0.05241939554166917,
236
+ "scr_dir2_threshold_50": 0.05241939554166917,
237
+ "scr_dir1_threshold_100": 0.07731951161061577,
238
+ "scr_metric_threshold_100": 0.08064531634921589,
239
+ "scr_dir2_threshold_100": 0.08064531634921589,
240
+ "scr_dir1_threshold_500": 0.12886585268435963,
241
+ "scr_metric_threshold_500": 0.08467752208135552,
242
+ "scr_dir2_threshold_500": 0.08467752208135552
243
+ },
244
+ {
245
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
246
+ "scr_dir1_threshold_2": -0.009009161394877545,
247
+ "scr_metric_threshold_2": 0.02232140362529227,
248
+ "scr_dir2_threshold_2": 0.02232140362529227,
249
+ "scr_dir1_threshold_5": -0.031531662147990275,
250
+ "scr_metric_threshold_5": 0.058035755862608125,
251
+ "scr_dir2_threshold_5": 0.058035755862608125,
252
+ "scr_dir1_threshold_10": -0.01801805430036767,
253
+ "scr_metric_threshold_10": 0.07589293198126605,
254
+ "scr_dir2_threshold_10": 0.07589293198126605,
255
+ "scr_dir1_threshold_20": 0.013513339358235187,
256
+ "scr_metric_threshold_20": 0.13392868784387418,
257
+ "scr_dir2_threshold_20": 0.13392868784387418,
258
+ "scr_dir1_threshold_50": 0.11711695021830872,
259
+ "scr_metric_threshold_50": 0.17857149509445872,
260
+ "scr_dir2_threshold_50": 0.17857149509445872,
261
+ "scr_dir1_threshold_100": 0.1216216651604412,
262
+ "scr_metric_threshold_100": 0.36160721769555176,
263
+ "scr_dir2_threshold_100": 0.36160721769555176,
264
+ "scr_dir1_threshold_500": 0.3243241719384558,
265
+ "scr_metric_threshold_500": 0.48214295692740233,
266
+ "scr_dir2_threshold_500": 0.48214295692740233
267
+ },
268
+ {
269
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
270
+ "scr_dir1_threshold_2": 0.008583578999931441,
271
+ "scr_metric_threshold_2": 0.008583578999931441,
272
+ "scr_dir2_threshold_2": 0.0285713393672145,
273
+ "scr_dir1_threshold_5": 0.047210196127387125,
274
+ "scr_metric_threshold_5": 0.047210196127387125,
275
+ "scr_dir2_threshold_5": 0.042857150966616867,
276
+ "scr_dir1_threshold_10": 0.07296144475494565,
277
+ "scr_metric_threshold_10": 0.07296144475494565,
278
+ "scr_dir2_threshold_10": 0.05714296256601923,
279
+ "scr_dir1_threshold_20": 0.03862661712745569,
280
+ "scr_metric_threshold_20": 0.03862661712745569,
281
+ "scr_dir2_threshold_20": 0.08095245934396302,
282
+ "scr_dir1_threshold_50": 0.12875547569614632,
283
+ "scr_metric_threshold_50": 0.12875547569614632,
284
+ "scr_dir2_threshold_50": 0.08571430193323373,
285
+ "scr_dir1_threshold_100": 0.15450646850982275,
286
+ "scr_metric_threshold_100": 0.15450646850982275,
287
+ "scr_dir2_threshold_100": 0.1619046348563358,
288
+ "scr_dir1_threshold_500": 0.19742487513724416,
289
+ "scr_metric_threshold_500": 0.19742487513724416,
290
+ "scr_dir2_threshold_500": 0.33809508131207394
291
+ }
292
+ ],
293
+ "sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
294
+ "sae_lens_id": "layer_12/width_1m/average_l0_19",
295
+ "sae_lens_release_id": "gemma-scope-2b-pt-res",
296
+ "sae_lens_version": "4.4.5",
297
+ "eval_result_unstructured": null
298
+ }
gemma_scope/scr_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_207_eval_results.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_type_id": "scr",
3
+ "eval_config": {
4
+ "random_seed": 42,
5
+ "dataset_names": [
6
+ "LabHC/bias_in_bios_class_set1",
7
+ "canrager/amazon_reviews_mcauley_1and5"
8
+ ],
9
+ "perform_scr": true,
10
+ "early_stopping_patience": 20,
11
+ "train_set_size": 4000,
12
+ "test_set_size": 1000,
13
+ "context_length": 128,
14
+ "probe_train_batch_size": 16,
15
+ "probe_test_batch_size": 500,
16
+ "probe_epochs": 20,
17
+ "probe_lr": 0.001,
18
+ "probe_l1_penalty": 0.001,
19
+ "sae_batch_size": 5,
20
+ "llm_batch_size": 32,
21
+ "llm_dtype": "bfloat16",
22
+ "lower_vram_usage": true,
23
+ "model_name": "gemma-2-2b",
24
+ "n_values": [
25
+ 2,
26
+ 5,
27
+ 10,
28
+ 20,
29
+ 50,
30
+ 100,
31
+ 500
32
+ ],
33
+ "column1_vals_lookup": {
34
+ "LabHC/bias_in_bios_class_set1": [
35
+ [
36
+ "professor",
37
+ "nurse"
38
+ ],
39
+ [
40
+ "architect",
41
+ "journalist"
42
+ ],
43
+ [
44
+ "surgeon",
45
+ "psychologist"
46
+ ],
47
+ [
48
+ "attorney",
49
+ "teacher"
50
+ ]
51
+ ],
52
+ "canrager/amazon_reviews_mcauley_1and5": [
53
+ [
54
+ "Books",
55
+ "CDs_and_Vinyl"
56
+ ],
57
+ [
58
+ "Software",
59
+ "Electronics"
60
+ ],
61
+ [
62
+ "Pet_Supplies",
63
+ "Office_Products"
64
+ ],
65
+ [
66
+ "Industrial_and_Scientific",
67
+ "Toys_and_Games"
68
+ ]
69
+ ]
70
+ }
71
+ },
72
+ "eval_id": "53a69d26-160f-4532-ad41-19e238470fee",
73
+ "datetime_epoch_millis": 1732760566948,
74
+ "eval_result_metrics": {
75
+ "scr_metrics": {
76
+ "scr_dir1_threshold_2": 0.21400088767581077,
77
+ "scr_metric_threshold_2": 0.08818504177842121,
78
+ "scr_dir2_threshold_2": 0.09068351182433158,
79
+ "scr_dir1_threshold_5": 0.21606008564343998,
80
+ "scr_metric_threshold_5": 0.11629804977933939,
81
+ "scr_dir2_threshold_5": 0.11861259295915984,
82
+ "scr_dir1_threshold_10": 0.26548402770414853,
83
+ "scr_metric_threshold_10": 0.1362917242579591,
84
+ "scr_dir2_threshold_10": 0.14116346214311778,
85
+ "scr_dir1_threshold_20": 0.3097537166411326,
86
+ "scr_metric_threshold_20": 0.17659663236553996,
87
+ "scr_dir2_threshold_20": 0.1852160930590386,
88
+ "scr_dir1_threshold_50": 0.3351538356911903,
89
+ "scr_metric_threshold_50": 0.2415669398330668,
90
+ "scr_dir2_threshold_50": 0.24695983749052006,
91
+ "scr_dir1_threshold_100": 0.3424757178170776,
92
+ "scr_metric_threshold_100": 0.2899064945198467,
93
+ "scr_dir2_threshold_100": 0.28641427360653304,
94
+ "scr_dir1_threshold_500": 0.0691028639978861,
95
+ "scr_metric_threshold_500": 0.3050465739364306,
96
+ "scr_dir2_threshold_500": 0.3012375634905508
97
+ }
98
+ },
99
+ "eval_result_details": [
100
+ {
101
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
102
+ "scr_dir1_threshold_2": 0.5,
103
+ "scr_metric_threshold_2": 0.0073890670992652185,
104
+ "scr_dir2_threshold_2": 0.0073890670992652185,
105
+ "scr_dir1_threshold_5": 0.5468753201420825,
106
+ "scr_metric_threshold_5": 0.03448275355829437,
107
+ "scr_dir2_threshold_5": 0.03448275355829437,
108
+ "scr_dir1_threshold_10": 0.5625001164153027,
109
+ "scr_metric_threshold_10": 0.04187182065755959,
110
+ "scr_dir2_threshold_10": 0.04187182065755959,
111
+ "scr_dir1_threshold_20": 0.578124912688523,
112
+ "scr_metric_threshold_20": 0.05418707929913162,
113
+ "scr_dir2_threshold_20": 0.05418707929913162,
114
+ "scr_dir1_threshold_50": 0.5156247962732202,
115
+ "scr_metric_threshold_50": 0.07389155184943222,
116
+ "scr_dir2_threshold_50": 0.07389155184943222,
117
+ "scr_dir1_threshold_100": 0.39062549476503666,
118
+ "scr_metric_threshold_100": 0.13054180032444893,
119
+ "scr_dir2_threshold_100": 0.13054180032444893,
120
+ "scr_dir1_threshold_500": -0.21874994179234863,
121
+ "scr_metric_threshold_500": 0.02709353964956581,
122
+ "scr_dir2_threshold_500": 0.02709353964956581
123
+ },
124
+ {
125
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
126
+ "scr_dir1_threshold_2": 0.19801959163150878,
127
+ "scr_metric_threshold_2": 0.091168103263091,
128
+ "scr_dir2_threshold_2": 0.091168103263091,
129
+ "scr_dir1_threshold_5": 0.10891124751327678,
130
+ "scr_metric_threshold_5": 0.1168091376125165,
131
+ "scr_dir2_threshold_5": 0.1168091376125165,
132
+ "scr_dir1_threshold_10": 0.20792104332903116,
133
+ "scr_metric_threshold_10": 0.142450171961942,
134
+ "scr_dir2_threshold_10": 0.142450171961942,
135
+ "scr_dir1_threshold_20": 0.35643573705266274,
136
+ "scr_metric_threshold_20": 0.19658120674099644,
137
+ "scr_dir2_threshold_20": 0.19658120674099644,
138
+ "scr_dir1_threshold_50": 0.32673256224996294,
139
+ "scr_metric_threshold_50": 0.3076924121931061,
140
+ "scr_dir2_threshold_50": 0.3076924121931061,
141
+ "scr_dir1_threshold_100": 0.3069308391447856,
142
+ "scr_metric_threshold_100": 0.37037034521277074,
143
+ "scr_dir2_threshold_100": 0.37037034521277074,
144
+ "scr_dir1_threshold_500": -0.32673256224996294,
145
+ "scr_metric_threshold_500": 0.14529913804214542,
146
+ "scr_dir2_threshold_500": 0.14529913804214542
147
+ },
148
+ {
149
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
150
+ "scr_dir1_threshold_2": 0.5714285714285714,
151
+ "scr_metric_threshold_2": 0.027918872340786032,
152
+ "scr_dir2_threshold_2": 0.027918872340786032,
153
+ "scr_dir1_threshold_5": 0.5238092084411499,
154
+ "scr_metric_threshold_5": 0.04060921384812376,
155
+ "scr_dir2_threshold_5": 0.04060921384812376,
156
+ "scr_dir1_threshold_10": 0.5079364028137167,
157
+ "scr_metric_threshold_10": 0.07614220032483014,
158
+ "scr_dir2_threshold_10": 0.07614220032483014,
159
+ "scr_dir1_threshold_20": 0.4761907915588501,
160
+ "scr_metric_threshold_20": 0.09644680724889201,
161
+ "scr_dir2_threshold_20": 0.09644680724889201,
162
+ "scr_dir1_threshold_50": 0.5238092084411499,
163
+ "scr_metric_threshold_50": 0.13705586981621198,
164
+ "scr_dir2_threshold_50": 0.13705586981621198,
165
+ "scr_dir1_threshold_100": 0.31746084307427397,
166
+ "scr_metric_threshold_100": 0.017766568878755093,
167
+ "scr_dir2_threshold_100": 0.017766568878755093,
168
+ "scr_dir1_threshold_500": -0.380952065584007,
169
+ "scr_metric_threshold_500": 0.027918872340786032,
170
+ "scr_dir2_threshold_500": 0.027918872340786032
171
+ },
172
+ {
173
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
174
+ "scr_dir1_threshold_2": 0.31007766623341637,
175
+ "scr_metric_threshold_2": 0.029761879419791297,
176
+ "scr_dir2_threshold_2": 0.029761879419791297,
177
+ "scr_dir1_threshold_5": 0.17054273953095306,
178
+ "scr_metric_threshold_5": 0.06845239362343769,
179
+ "scr_dir2_threshold_5": 0.06845239362343769,
180
+ "scr_dir1_threshold_10": 0.3178295039276686,
181
+ "scr_metric_threshold_10": 0.11904751767916519,
182
+ "scr_dir2_threshold_10": 0.11904751767916519,
183
+ "scr_dir1_threshold_20": 0.4108529424131378,
184
+ "scr_metric_threshold_20": 0.19940469854947826,
185
+ "scr_dir2_threshold_20": 0.19940469854947826,
186
+ "scr_dir1_threshold_50": 0.36434099214466276,
187
+ "scr_metric_threshold_50": 0.1755951240557275,
188
+ "scr_dir2_threshold_50": 0.1755951240557275,
189
+ "scr_dir1_threshold_100": 0.5271318939813636,
190
+ "scr_metric_threshold_100": 0.2202381205802087,
191
+ "scr_dir2_threshold_100": 0.2202381205802087,
192
+ "scr_dir1_threshold_500": 0.33333317931617307,
193
+ "scr_metric_threshold_500": -0.020833422030730454,
194
+ "scr_dir2_threshold_500": -0.020833422030730454
195
+ },
196
+ {
197
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
198
+ "scr_dir1_threshold_2": 0.016393554752069144,
199
+ "scr_metric_threshold_2": 0.4669260357465481,
200
+ "scr_dir2_threshold_2": 0.4669260357465481,
201
+ "scr_dir1_threshold_5": 0.1092894964580841,
202
+ "scr_metric_threshold_5": 0.4941634999286832,
203
+ "scr_dir2_threshold_5": 0.4941634999286832,
204
+ "scr_dir1_threshold_10": 0.1311474608913792,
205
+ "scr_metric_threshold_10": 0.46303503569900356,
206
+ "scr_dir2_threshold_10": 0.46303503569900356,
207
+ "scr_dir1_threshold_20": 0.16393424468712614,
208
+ "scr_metric_threshold_20": 0.5058367319959785,
209
+ "scr_dir2_threshold_20": 0.5058367319959785,
210
+ "scr_dir1_threshold_50": 0.16939898007674342,
211
+ "scr_metric_threshold_50": 0.5758754286257652,
212
+ "scr_dir2_threshold_50": 0.5758754286257652,
213
+ "scr_dir1_threshold_100": 0.14207660596222238,
214
+ "scr_metric_threshold_100": 0.6225681249702847,
215
+ "scr_dir2_threshold_100": 0.6225681249702847,
216
+ "scr_dir1_threshold_500": -0.04918033854781611,
217
+ "scr_metric_threshold_500": 0.7704279821742706,
218
+ "scr_dir2_threshold_500": 0.7704279821742706
219
+ },
220
+ {
221
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
222
+ "scr_dir1_threshold_2": 0.030927743196160724,
223
+ "scr_metric_threshold_2": 0.020161269001982816,
224
+ "scr_dir2_threshold_2": 0.020161269001982816,
225
+ "scr_dir1_threshold_5": 0.04123688851473832,
226
+ "scr_metric_threshold_5": 0.05241939554166917,
227
+ "scr_dir2_threshold_5": 0.05241939554166917,
228
+ "scr_dir1_threshold_10": 0.09278322958848217,
229
+ "scr_metric_threshold_10": 0.08467752208135552,
230
+ "scr_dir2_threshold_10": 0.08467752208135552,
231
+ "scr_dir1_threshold_20": 0.14432987790265395,
232
+ "scr_metric_threshold_20": 0.11290320254761761,
233
+ "scr_dir2_threshold_20": 0.11290320254761761,
234
+ "scr_dir1_threshold_50": 0.2113399369542642,
235
+ "scr_metric_threshold_50": 0.22580640509523522,
236
+ "scr_dir2_threshold_50": 0.22580640509523522,
237
+ "scr_dir1_threshold_100": 0.2783503032463024,
238
+ "scr_metric_threshold_100": 0.31048392717659073,
239
+ "scr_dir2_threshold_100": 0.31048392717659073,
240
+ "scr_dir1_threshold_500": 0.3195874990014686,
241
+ "scr_metric_threshold_500": 0.5483871898095295,
242
+ "scr_dir2_threshold_500": 0.5483871898095295
243
+ },
244
+ {
245
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
246
+ "scr_dir1_threshold_2": 0.07657639516482832,
247
+ "scr_metric_threshold_2": 0.05357152835597378,
248
+ "scr_dir2_threshold_2": 0.05357152835597378,
249
+ "scr_dir1_threshold_5": 0.1891891674197794,
250
+ "scr_metric_threshold_5": 0.08482138699453473,
251
+ "scr_dir2_threshold_5": 0.08482138699453473,
252
+ "scr_dir1_threshold_10": 0.25225222322637253,
253
+ "scr_metric_threshold_10": 0.1116072842185819,
254
+ "scr_dir2_threshold_10": 0.1116072842185819,
255
+ "scr_dir1_threshold_20": 0.2837836168849754,
256
+ "scr_metric_threshold_20": 0.18303572260109305,
257
+ "scr_dir2_threshold_20": 0.18303572260109305,
258
+ "scr_dir1_threshold_50": 0.43693694419340684,
259
+ "scr_metric_threshold_50": 0.30357146183294365,
260
+ "scr_dir2_threshold_50": 0.30357146183294365,
261
+ "scr_dir1_threshold_100": 0.5540538944117156,
262
+ "scr_metric_threshold_100": 0.4241072010647942,
263
+ "scr_dir2_threshold_100": 0.4241072010647942,
264
+ "scr_dir1_threshold_500": 0.5450450015062255,
265
+ "scr_metric_threshold_500": 0.6116071511725216,
266
+ "scr_dir2_threshold_500": 0.6116071511725216
267
+ },
268
+ {
269
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
270
+ "scr_dir1_threshold_2": 0.008583578999931441,
271
+ "scr_metric_threshold_2": 0.008583578999931441,
272
+ "scr_dir2_threshold_2": 0.0285713393672145,
273
+ "scr_dir1_threshold_5": 0.03862661712745569,
274
+ "scr_metric_threshold_5": 0.03862661712745569,
275
+ "scr_dir2_threshold_5": 0.05714296256601923,
276
+ "scr_dir1_threshold_10": 0.05150224144123495,
277
+ "scr_metric_threshold_10": 0.05150224144123495,
278
+ "scr_dir2_threshold_10": 0.09047614452250444,
279
+ "scr_dir1_threshold_20": 0.0643776099411321,
280
+ "scr_metric_threshold_20": 0.0643776099411321,
281
+ "scr_dir2_threshold_20": 0.1333332954891213,
282
+ "scr_dir1_threshold_50": 0.13304726519611204,
283
+ "scr_metric_threshold_50": 0.13304726519611204,
284
+ "scr_dir2_threshold_50": 0.17619044645573817,
285
+ "scr_dir1_threshold_100": 0.22317586795092056,
286
+ "scr_metric_threshold_100": 0.22317586795092056,
287
+ "scr_dir2_threshold_100": 0.19523810064441124,
288
+ "scr_dir1_threshold_500": 0.3304721403333562,
289
+ "scr_metric_threshold_500": 0.3304721403333562,
290
+ "scr_dir2_threshold_500": 0.300000056766318
291
+ }
292
+ ],
293
+ "sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
294
+ "sae_lens_id": "layer_12/width_1m/average_l0_207",
295
+ "sae_lens_release_id": "gemma-scope-2b-pt-res",
296
+ "sae_lens_version": "4.4.5",
297
+ "eval_result_unstructured": null
298
+ }
gemma_scope/scr_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_26_eval_results.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_type_id": "scr",
3
+ "eval_config": {
4
+ "random_seed": 42,
5
+ "dataset_names": [
6
+ "LabHC/bias_in_bios_class_set1",
7
+ "canrager/amazon_reviews_mcauley_1and5"
8
+ ],
9
+ "perform_scr": true,
10
+ "early_stopping_patience": 20,
11
+ "train_set_size": 4000,
12
+ "test_set_size": 1000,
13
+ "context_length": 128,
14
+ "probe_train_batch_size": 16,
15
+ "probe_test_batch_size": 500,
16
+ "probe_epochs": 20,
17
+ "probe_lr": 0.001,
18
+ "probe_l1_penalty": 0.001,
19
+ "sae_batch_size": 5,
20
+ "llm_batch_size": 32,
21
+ "llm_dtype": "bfloat16",
22
+ "lower_vram_usage": true,
23
+ "model_name": "gemma-2-2b",
24
+ "n_values": [
25
+ 2,
26
+ 5,
27
+ 10,
28
+ 20,
29
+ 50,
30
+ 100,
31
+ 500
32
+ ],
33
+ "column1_vals_lookup": {
34
+ "LabHC/bias_in_bios_class_set1": [
35
+ [
36
+ "professor",
37
+ "nurse"
38
+ ],
39
+ [
40
+ "architect",
41
+ "journalist"
42
+ ],
43
+ [
44
+ "surgeon",
45
+ "psychologist"
46
+ ],
47
+ [
48
+ "attorney",
49
+ "teacher"
50
+ ]
51
+ ],
52
+ "canrager/amazon_reviews_mcauley_1and5": [
53
+ [
54
+ "Books",
55
+ "CDs_and_Vinyl"
56
+ ],
57
+ [
58
+ "Software",
59
+ "Electronics"
60
+ ],
61
+ [
62
+ "Pet_Supplies",
63
+ "Office_Products"
64
+ ],
65
+ [
66
+ "Industrial_and_Scientific",
67
+ "Toys_and_Games"
68
+ ]
69
+ ]
70
+ }
71
+ },
72
+ "eval_id": "0f3e4ae3-27f1-4250-90e1-3b28d213b483",
73
+ "datetime_epoch_millis": 1732773253161,
74
+ "eval_result_metrics": {
75
+ "scr_metrics": {
76
+ "scr_dir1_threshold_2": 0.06148361920570596,
77
+ "scr_metric_threshold_2": 0.011229464370979547,
78
+ "scr_dir2_threshold_2": 0.010626577089720141,
79
+ "scr_dir1_threshold_5": 0.089867895804999,
80
+ "scr_metric_threshold_5": 0.025225298896936404,
81
+ "scr_dir2_threshold_5": 0.0307587161784664,
82
+ "scr_dir1_threshold_10": 0.14557085343617018,
83
+ "scr_metric_threshold_10": 0.05736782053918504,
84
+ "scr_dir2_threshold_10": 0.06087282436632316,
85
+ "scr_dir1_threshold_20": 0.17403282007686996,
86
+ "scr_metric_threshold_20": 0.08763637119243324,
87
+ "scr_dir2_threshold_20": 0.09024470105528282,
88
+ "scr_dir1_threshold_50": 0.15470056611421365,
89
+ "scr_metric_threshold_50": 0.1489507702555035,
90
+ "scr_dir2_threshold_50": 0.15303569049109933,
91
+ "scr_dir1_threshold_100": 0.13664726925200954,
92
+ "scr_metric_threshold_100": 0.20090026795675153,
93
+ "scr_dir2_threshold_100": 0.20152359510943596,
94
+ "scr_dir1_threshold_500": 0.010855836901899757,
95
+ "scr_metric_threshold_500": 0.29273813146866784,
96
+ "scr_dir2_threshold_500": 0.31925050401119925
97
+ }
98
+ },
99
+ "eval_result_details": [
100
+ {
101
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
102
+ "scr_dir1_threshold_2": 0.0937506402841651,
103
+ "scr_metric_threshold_2": -0.01724145018387885,
104
+ "scr_dir2_threshold_2": -0.01724145018387885,
105
+ "scr_dir1_threshold_5": 0.1250002328306055,
106
+ "scr_metric_threshold_5": 0.012315258641572036,
107
+ "scr_dir2_threshold_5": 0.012315258641572036,
108
+ "scr_dir1_threshold_10": 0.28125005820765137,
109
+ "scr_metric_threshold_10": 0.039408798291137845,
110
+ "scr_dir2_threshold_10": 0.039408798291137845,
111
+ "scr_dir1_threshold_20": 0.32812537834973393,
112
+ "scr_metric_threshold_20": 0.06403931557428191,
113
+ "scr_dir2_threshold_20": 0.06403931557428191,
114
+ "scr_dir1_threshold_50": 0.1250002328306055,
115
+ "scr_metric_threshold_50": 0.09852206913257629,
116
+ "scr_dir2_threshold_50": 0.09852206913257629,
117
+ "scr_dir1_threshold_100": 0.3125005820765137,
118
+ "scr_metric_threshold_100": 0.14285705896602094,
119
+ "scr_dir2_threshold_100": 0.14285705896602094,
120
+ "scr_dir1_threshold_500": 0.17187555297268803,
121
+ "scr_metric_threshold_500": 0.15763533997401472,
122
+ "scr_dir2_threshold_500": 0.15763533997401472
123
+ },
124
+ {
125
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
126
+ "scr_dir1_threshold_2": 0.11881210906586549,
127
+ "scr_metric_threshold_2": 0.028490000429628907,
128
+ "scr_dir2_threshold_2": 0.028490000429628907,
129
+ "scr_dir1_threshold_5": 0.19801959163150878,
130
+ "scr_metric_threshold_5": 0.05698017067305529,
131
+ "scr_dir2_threshold_5": 0.05698017067305529,
132
+ "scr_dir1_threshold_10": 0.19801959163150878,
133
+ "scr_metric_threshold_10": 0.07977206912847994,
134
+ "scr_dir2_threshold_10": 0.07977206912847994,
135
+ "scr_dir1_threshold_20": 0.20792104332903116,
136
+ "scr_metric_threshold_20": 0.142450171961942,
137
+ "scr_dir2_threshold_20": 0.142450171961942,
138
+ "scr_dir1_threshold_50": -0.0594057594604659,
139
+ "scr_metric_threshold_50": 0.15384620609655306,
140
+ "scr_dir2_threshold_50": 0.15384620609655306,
141
+ "scr_dir1_threshold_100": 0.11881210906586549,
142
+ "scr_metric_threshold_100": 0.23931637719923726,
143
+ "scr_dir2_threshold_100": 0.23931637719923726,
144
+ "scr_dir1_threshold_500": -0.514851292328883,
145
+ "scr_metric_threshold_500": 0.38461551524138266,
146
+ "scr_dir2_threshold_500": 0.38461551524138266
147
+ },
148
+ {
149
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
150
+ "scr_dir1_threshold_2": 0.1269843372297096,
151
+ "scr_metric_threshold_2": 0.0025381893261105827,
152
+ "scr_dir2_threshold_2": 0.0025381893261105827,
153
+ "scr_dir1_threshold_5": 0.19047650584456438,
154
+ "scr_metric_threshold_5": 0.005076227371417368,
155
+ "scr_dir2_threshold_5": 0.005076227371417368,
156
+ "scr_dir1_threshold_10": 0.2857142857142857,
157
+ "scr_metric_threshold_10": 0.01015230346203094,
158
+ "scr_dir2_threshold_10": 0.01015230346203094,
159
+ "scr_dir1_threshold_20": 0.33333364870170723,
160
+ "scr_metric_threshold_20": 0.04314725189343054,
161
+ "scr_dir2_threshold_20": 0.04314725189343054,
162
+ "scr_dir1_threshold_50": 0.4126986229439953,
163
+ "scr_metric_threshold_50": 0.08121827641544371,
164
+ "scr_dir2_threshold_50": 0.08121827641544371,
165
+ "scr_dir1_threshold_100": 0.30158709134171896,
166
+ "scr_metric_threshold_100": 0.1319797937255984,
167
+ "scr_dir2_threshold_100": 0.1319797937255984,
168
+ "scr_dir1_threshold_500": 0.3492064543291405,
169
+ "scr_metric_threshold_500": 0.2791878157230375,
170
+ "scr_dir2_threshold_500": 0.2791878157230375
171
+ },
172
+ {
173
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
174
+ "scr_dir1_threshold_2": 0.0852711387397361,
175
+ "scr_metric_threshold_2": 0.008928457389060843,
176
+ "scr_dir2_threshold_2": 0.008928457389060843,
177
+ "scr_dir1_threshold_5": 0.13178308900821112,
178
+ "scr_metric_threshold_5": 0.014880939709895648,
179
+ "scr_dir2_threshold_5": 0.014880939709895648,
180
+ "scr_dir1_threshold_10": 0.17829457722520528,
181
+ "scr_metric_threshold_10": 0.05357145391354203,
182
+ "scr_dir2_threshold_10": 0.05357145391354203,
183
+ "scr_dir1_threshold_20": 0.209302390053695,
184
+ "scr_metric_threshold_20": 0.10714273043228982,
185
+ "scr_dir2_threshold_20": 0.10714273043228982,
186
+ "scr_dir1_threshold_50": 0.17829457722520528,
187
+ "scr_metric_threshold_50": 0.19940469854947826,
188
+ "scr_dir2_threshold_50": 0.19940469854947826,
189
+ "scr_dir1_threshold_100": -0.2945735287934311,
190
+ "scr_metric_threshold_100": 0.303571453913542,
191
+ "scr_dir2_threshold_100": 0.303571453913542,
192
+ "scr_dir1_threshold_500": -0.5271318939813636,
193
+ "scr_metric_threshold_500": 0.38095230492604054,
194
+ "scr_dir2_threshold_500": 0.38095230492604054
195
+ },
196
+ {
197
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
198
+ "scr_dir1_threshold_2": 0.00546440968122594,
199
+ "scr_metric_threshold_2": 0.01167323206729526,
200
+ "scr_dir2_threshold_2": 0.01167323206729526,
201
+ "scr_dir1_threshold_5": 0.016393554752069144,
202
+ "scr_metric_threshold_5": 0.027237464182135074,
203
+ "scr_dir2_threshold_5": 0.027237464182135074,
204
+ "scr_dir1_threshold_10": 0.1092894964580841,
205
+ "scr_metric_threshold_10": 0.11673162489896782,
206
+ "scr_dir2_threshold_10": 0.11673162489896782,
207
+ "scr_dir1_threshold_20": 0.08743153202478901,
208
+ "scr_metric_threshold_20": 0.13229585701380764,
209
+ "scr_dir2_threshold_20": 0.13229585701380764,
210
+ "scr_dir1_threshold_50": 0.09289626741440628,
211
+ "scr_metric_threshold_50": 0.2295720178257295,
212
+ "scr_dir2_threshold_50": 0.2295720178257295,
213
+ "scr_dir1_threshold_100": 0.01092881936245188,
214
+ "scr_metric_threshold_100": 0.24513624994056932,
215
+ "scr_dir2_threshold_100": 0.24513624994056932,
216
+ "scr_dir1_threshold_500": -0.09836067709563222,
217
+ "scr_metric_threshold_500": 0.3735408749821708,
218
+ "scr_dir2_threshold_500": 0.3735408749821708
219
+ },
220
+ {
221
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
222
+ "scr_dir1_threshold_2": 0.0051545726592888,
223
+ "scr_metric_threshold_2": 0.008064651805563901,
224
+ "scr_dir2_threshold_2": 0.008064651805563901,
225
+ "scr_dir1_threshold_5": -0.005154879899716724,
226
+ "scr_metric_threshold_5": 0.032258126539686356,
227
+ "scr_dir2_threshold_5": 0.032258126539686356,
228
+ "scr_dir1_threshold_10": 0.0103091453185776,
229
+ "scr_metric_threshold_10": 0.04032253800396563,
230
+ "scr_dir2_threshold_10": 0.04032253800396563,
231
+ "scr_dir1_threshold_20": 0.030927743196160724,
232
+ "scr_metric_threshold_20": 0.04032253800396563,
233
+ "scr_dir2_threshold_20": 0.04032253800396563,
234
+ "scr_dir1_threshold_50": 0.06185548639232145,
235
+ "scr_metric_threshold_50": 0.032258126539686356,
236
+ "scr_dir2_threshold_50": 0.032258126539686356,
237
+ "scr_dir1_threshold_100": 0.1030926821474877,
238
+ "scr_metric_threshold_100": 0.020161269001982816,
239
+ "scr_dir2_threshold_100": 0.020161269001982816,
240
+ "scr_dir1_threshold_500": 0.24742256005014163,
241
+ "scr_metric_threshold_500": 0.06854845881151235,
242
+ "scr_dir2_threshold_500": 0.06854845881151235
243
+ },
244
+ {
245
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
246
+ "scr_dir1_threshold_2": 0.013513339358235187,
247
+ "scr_metric_threshold_2": 0.0044642275066343415,
248
+ "scr_dir2_threshold_2": 0.0044642275066343415,
249
+ "scr_dir1_threshold_5": 0.049549447958970524,
250
+ "scr_metric_threshold_5": 0.0401785797439502,
251
+ "scr_dir2_threshold_5": 0.0401785797439502,
252
+ "scr_dir1_threshold_10": 0.06306305580659313,
253
+ "scr_metric_threshold_10": 0.0803571594879004,
254
+ "scr_dir2_threshold_10": 0.0803571594879004,
255
+ "scr_dir1_threshold_20": 0.13513500451867638,
256
+ "scr_metric_threshold_20": 0.1116072842185819,
257
+ "scr_dir2_threshold_20": 0.1116072842185819,
258
+ "scr_dir1_threshold_50": 0.30180167118534307,
259
+ "scr_metric_threshold_50": 0.27232133710226214,
260
+ "scr_dir2_threshold_50": 0.27232133710226214,
261
+ "scr_dir1_threshold_100": 0.3648647269919362,
262
+ "scr_metric_threshold_100": 0.3482142690835282,
263
+ "scr_dir2_threshold_100": 0.3482142690835282,
264
+ "scr_dir1_threshold_500": 0.26126111613186265,
265
+ "scr_metric_threshold_500": 0.4999998669539397,
266
+ "scr_dir2_threshold_500": 0.4999998669539397
267
+ },
268
+ {
269
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
270
+ "scr_dir1_threshold_2": 0.042918406627421406,
271
+ "scr_metric_threshold_2": 0.042918406627421406,
272
+ "scr_dir2_threshold_2": 0.03809530837734615,
273
+ "scr_dir1_threshold_5": 0.012875624313779262,
274
+ "scr_metric_threshold_5": 0.012875624313779262,
275
+ "scr_dir2_threshold_5": 0.05714296256601923,
276
+ "scr_dir1_threshold_10": 0.03862661712745569,
277
+ "scr_metric_threshold_10": 0.03862661712745569,
278
+ "scr_dir2_threshold_10": 0.06666664774456064,
279
+ "scr_dir1_threshold_20": 0.060085820441166386,
280
+ "scr_metric_threshold_20": 0.060085820441166386,
281
+ "scr_dir2_threshold_20": 0.08095245934396302,
282
+ "scr_dir1_threshold_50": 0.1244634303822985,
283
+ "scr_metric_threshold_50": 0.1244634303822985,
284
+ "scr_dir2_threshold_50": 0.1571427922670651,
285
+ "scr_dir1_threshold_100": 0.17596567182353345,
286
+ "scr_metric_threshold_100": 0.17596567182353345,
287
+ "scr_dir2_threshold_100": 0.18095228904500887,
288
+ "scr_dir1_threshold_500": 0.19742487513724416,
289
+ "scr_metric_threshold_500": 0.19742487513724416,
290
+ "scr_dir2_threshold_500": 0.40952385547749554
291
+ }
292
+ ],
293
+ "sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
294
+ "sae_lens_id": "layer_12/width_1m/average_l0_26",
295
+ "sae_lens_release_id": "gemma-scope-2b-pt-res",
296
+ "sae_lens_version": "4.4.5",
297
+ "eval_result_unstructured": null
298
+ }
gemma_scope/scr_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_58_eval_results.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_type_id": "scr",
3
+ "eval_config": {
4
+ "random_seed": 42,
5
+ "dataset_names": [
6
+ "LabHC/bias_in_bios_class_set1",
7
+ "canrager/amazon_reviews_mcauley_1and5"
8
+ ],
9
+ "perform_scr": true,
10
+ "early_stopping_patience": 20,
11
+ "train_set_size": 4000,
12
+ "test_set_size": 1000,
13
+ "context_length": 128,
14
+ "probe_train_batch_size": 16,
15
+ "probe_test_batch_size": 500,
16
+ "probe_epochs": 20,
17
+ "probe_lr": 0.001,
18
+ "probe_l1_penalty": 0.001,
19
+ "sae_batch_size": 5,
20
+ "llm_batch_size": 32,
21
+ "llm_dtype": "bfloat16",
22
+ "lower_vram_usage": true,
23
+ "model_name": "gemma-2-2b",
24
+ "n_values": [
25
+ 2,
26
+ 5,
27
+ 10,
28
+ 20,
29
+ 50,
30
+ 100,
31
+ 500
32
+ ],
33
+ "column1_vals_lookup": {
34
+ "LabHC/bias_in_bios_class_set1": [
35
+ [
36
+ "professor",
37
+ "nurse"
38
+ ],
39
+ [
40
+ "architect",
41
+ "journalist"
42
+ ],
43
+ [
44
+ "surgeon",
45
+ "psychologist"
46
+ ],
47
+ [
48
+ "attorney",
49
+ "teacher"
50
+ ]
51
+ ],
52
+ "canrager/amazon_reviews_mcauley_1and5": [
53
+ [
54
+ "Books",
55
+ "CDs_and_Vinyl"
56
+ ],
57
+ [
58
+ "Software",
59
+ "Electronics"
60
+ ],
61
+ [
62
+ "Pet_Supplies",
63
+ "Office_Products"
64
+ ],
65
+ [
66
+ "Industrial_and_Scientific",
67
+ "Toys_and_Games"
68
+ ]
69
+ ]
70
+ }
71
+ },
72
+ "eval_id": "511c63f4-f68b-4f42-a03d-ed48cfc19974",
73
+ "datetime_epoch_millis": 1732785765426,
74
+ "eval_result_metrics": {
75
+ "scr_metrics": {
76
+ "scr_dir1_threshold_2": 0.12996131563746238,
77
+ "scr_metric_threshold_2": 0.018736419911511672,
78
+ "scr_dir2_threshold_2": 0.018728762953911106,
79
+ "scr_dir1_threshold_5": 0.16304298752291166,
80
+ "scr_metric_threshold_5": 0.037932488480724294,
81
+ "scr_dir2_threshold_5": 0.03815986156998974,
82
+ "scr_dir1_threshold_10": 0.15040373835049653,
83
+ "scr_metric_threshold_10": 0.059289042368358545,
84
+ "scr_dir2_threshold_10": 0.05915617970188238,
85
+ "scr_dir1_threshold_20": 0.07409133214850107,
86
+ "scr_metric_threshold_20": 0.10703729874713164,
87
+ "scr_dir2_threshold_20": 0.10725701487879652,
88
+ "scr_dir1_threshold_50": -0.42047175448263635,
89
+ "scr_metric_threshold_50": 0.17415281425475848,
90
+ "scr_dir2_threshold_50": 0.18235328255635241,
91
+ "scr_dir1_threshold_100": -0.3344911869660311,
92
+ "scr_metric_threshold_100": 0.22845328392819195,
93
+ "scr_dir2_threshold_100": 0.2524671509458642,
94
+ "scr_dir1_threshold_500": -0.22203840408346326,
95
+ "scr_metric_threshold_500": 0.3028448292297312,
96
+ "scr_dir2_threshold_500": 0.3398364594670509
97
+ }
98
+ },
99
+ "eval_result_details": [
100
+ {
101
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
102
+ "scr_dir1_threshold_2": 0.2968748544808716,
103
+ "scr_metric_threshold_2": -0.0024631691758850776,
104
+ "scr_dir2_threshold_2": -0.0024631691758850776,
105
+ "scr_dir1_threshold_5": 0.3437501746229541,
106
+ "scr_metric_threshold_5": 0.0073890670992652185,
107
+ "scr_dir2_threshold_5": 0.0073890670992652185,
108
+ "scr_dir1_threshold_10": 0.3749997671693945,
109
+ "scr_metric_threshold_10": 0.012315258641572036,
110
+ "scr_dir2_threshold_10": 0.012315258641572036,
111
+ "scr_dir1_threshold_20": 0.250000465661211,
112
+ "scr_metric_threshold_20": 0.04187182065755959,
113
+ "scr_dir2_threshold_20": 0.04187182065755959,
114
+ "scr_dir1_threshold_50": -0.4062493597158349,
115
+ "scr_metric_threshold_50": 0.08620681049100425,
116
+ "scr_dir2_threshold_50": 0.08620681049100425,
117
+ "scr_dir1_threshold_100": 0.2343756693879908,
118
+ "scr_metric_threshold_100": 0.08128076575816078,
119
+ "scr_dir2_threshold_100": 0.08128076575816078,
120
+ "scr_dir1_threshold_500": -0.17187462165026607,
121
+ "scr_metric_threshold_500": 0.28817728710792695,
122
+ "scr_dir2_threshold_500": 0.28817728710792695
123
+ },
124
+ {
125
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
126
+ "scr_dir1_threshold_2": 0.07920807271057699,
127
+ "scr_metric_threshold_2": 0.025641034349425513,
128
+ "scr_dir2_threshold_2": 0.025641034349425513,
129
+ "scr_dir1_threshold_5": 0.11881210906586549,
130
+ "scr_metric_threshold_5": 0.048433102618647625,
131
+ "scr_dir2_threshold_5": 0.048433102618647625,
132
+ "scr_dir1_threshold_10": 0.20792104332903116,
133
+ "scr_metric_threshold_10": 0.09401706934329439,
134
+ "scr_dir2_threshold_10": 0.09401706934329439,
135
+ "scr_dir1_threshold_20": 0.5544553286841716,
136
+ "scr_metric_threshold_20": 0.13390327372133182,
137
+ "scr_dir2_threshold_20": 0.13390327372133182,
138
+ "scr_dir1_threshold_50": 0.049504897907877196,
139
+ "scr_metric_threshold_50": 0.21652430893001515,
140
+ "scr_dir2_threshold_50": 0.21652430893001515,
141
+ "scr_dir1_threshold_100": -0.6534651244999259,
142
+ "scr_metric_threshold_100": 0.2535613774140517,
143
+ "scr_dir2_threshold_100": 0.2535613774140517,
144
+ "scr_dir1_threshold_500": -0.9108910657368343,
145
+ "scr_metric_threshold_500": 0.25925930957445853,
146
+ "scr_dir2_threshold_500": 0.25925930957445853
147
+ },
148
+ {
149
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
150
+ "scr_dir1_threshold_2": 0.380952065584007,
151
+ "scr_metric_threshold_2": 0.01015230346203094,
152
+ "scr_dir2_threshold_2": 0.01015230346203094,
153
+ "scr_dir1_threshold_5": 0.42857142857142855,
154
+ "scr_metric_threshold_5": 0.01015230346203094,
155
+ "scr_dir2_threshold_5": 0.01015230346203094,
156
+ "scr_dir1_threshold_10": 0.0,
157
+ "scr_metric_threshold_10": 0.012690492788141522,
158
+ "scr_dir2_threshold_10": 0.012690492788141522,
159
+ "scr_dir1_threshold_20": -0.7460313255405808,
160
+ "scr_metric_threshold_20": 0.07868023837013692,
161
+ "scr_dir2_threshold_20": 0.07868023837013692,
162
+ "scr_dir1_threshold_50": -2.222221170994309,
163
+ "scr_metric_threshold_50": 0.1827411597549493,
164
+ "scr_dir2_threshold_50": 0.1827411597549493,
165
+ "scr_dir1_threshold_100": -1.3333327025965855,
166
+ "scr_metric_threshold_100": 0.17766508366433573,
167
+ "scr_dir2_threshold_100": 0.17766508366433573,
168
+ "scr_dir1_threshold_500": -1.1904755597394427,
169
+ "scr_metric_threshold_500": 0.14467013523293612,
170
+ "scr_dir2_threshold_500": 0.14467013523293612
171
+ },
172
+ {
173
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
174
+ "scr_dir1_threshold_2": 0.10077527617972139,
175
+ "scr_metric_threshold_2": 0.014880939709895648,
176
+ "scr_dir2_threshold_2": 0.014880939709895648,
177
+ "scr_dir1_threshold_5": 0.13953492670246334,
178
+ "scr_metric_threshold_5": 0.029761879419791297,
179
+ "scr_dir2_threshold_5": 0.029761879419791297,
180
+ "scr_dir1_threshold_10": 0.3255813416219208,
181
+ "scr_metric_threshold_10": 0.059523758839582594,
182
+ "scr_dir2_threshold_10": 0.059523758839582594,
183
+ "scr_dir1_threshold_20": 0.1240312513139589,
184
+ "scr_metric_threshold_20": 0.08630948579635361,
185
+ "scr_dir2_threshold_20": 0.08630948579635361,
186
+ "scr_dir1_threshold_50": -1.286821691099179,
187
+ "scr_metric_threshold_50": 0.13392845738906084,
188
+ "scr_dir2_threshold_50": 0.13392845738906084,
189
+ "scr_dir1_threshold_100": -1.6666668206838269,
190
+ "scr_metric_threshold_100": 0.2708332446359362,
191
+ "scr_dir2_threshold_100": 0.2708332446359362,
192
+ "scr_dir1_threshold_500": -0.2713175536591936,
193
+ "scr_metric_threshold_500": 0.20238085101249853,
194
+ "scr_dir2_threshold_500": 0.20238085101249853
195
+ },
196
+ {
197
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
198
+ "scr_dir1_threshold_2": 0.07650271266233713,
199
+ "scr_metric_threshold_2": 0.007782232019750705,
200
+ "scr_dir2_threshold_2": 0.007782232019750705,
201
+ "scr_dir1_threshold_5": 0.1092894964580841,
202
+ "scr_metric_threshold_5": 0.03891069624943033,
203
+ "scr_dir2_threshold_5": 0.03891069624943033,
204
+ "scr_dir1_threshold_10": 0.09836067709563222,
205
+ "scr_metric_threshold_10": 0.07003892855444838,
206
+ "scr_dir2_threshold_10": 0.07003892855444838,
207
+ "scr_dir1_threshold_20": 0.10382508677685816,
208
+ "scr_metric_threshold_20": 0.12062262494651238,
209
+ "scr_dir2_threshold_20": 0.12062262494651238,
210
+ "scr_dir1_threshold_50": 0.1202186415289273,
211
+ "scr_metric_threshold_50": 0.1945525535485053,
212
+ "scr_dir2_threshold_50": 0.1945525535485053,
213
+ "scr_dir1_threshold_100": 0.021857964433295084,
214
+ "scr_metric_threshold_100": 0.24124524989302476,
215
+ "scr_dir2_threshold_100": 0.24124524989302476,
216
+ "scr_dir1_threshold_500": -0.1311474608913792,
217
+ "scr_metric_threshold_500": 0.5214007321861567,
218
+ "scr_dir2_threshold_500": 0.5214007321861567
219
+ },
220
+ {
221
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
222
+ "scr_dir1_threshold_2": 0.030927743196160724,
223
+ "scr_metric_threshold_2": 0.024193715075407077,
224
+ "scr_dir2_threshold_2": 0.024193715075407077,
225
+ "scr_dir1_threshold_5": 0.04123688851473832,
226
+ "scr_metric_threshold_5": 0.06854845881151235,
227
+ "scr_dir2_threshold_5": 0.06854845881151235,
228
+ "scr_dir1_threshold_10": 0.05154634107374385,
229
+ "scr_metric_threshold_10": 0.07258066454365199,
230
+ "scr_dir2_threshold_10": 0.07258066454365199,
231
+ "scr_dir1_threshold_20": 0.07731951161061577,
232
+ "scr_metric_threshold_20": 0.10887099681547797,
233
+ "scr_dir2_threshold_20": 0.10887099681547797,
234
+ "scr_dir1_threshold_50": 0.12886585268435963,
235
+ "scr_metric_threshold_50": 0.18951607282340924,
236
+ "scr_dir2_threshold_50": 0.18951607282340924,
237
+ "scr_dir1_threshold_100": 0.2216493895132697,
238
+ "scr_metric_threshold_100": 0.22580640509523522,
239
+ "scr_dir2_threshold_100": 0.22580640509523522,
240
+ "scr_dir1_threshold_500": 0.2731957305870136,
241
+ "scr_metric_threshold_500": 0.32661299044643394,
242
+ "scr_dir2_threshold_500": 0.32661299044643394
243
+ },
244
+ {
245
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
246
+ "scr_dir1_threshold_2": 0.03153139365860286,
247
+ "scr_metric_threshold_2": 0.02678589722404717,
248
+ "scr_dir2_threshold_2": 0.02678589722404717,
249
+ "scr_dir1_threshold_5": 0.06306305580659313,
250
+ "scr_metric_threshold_5": 0.0401785797439502,
251
+ "scr_dir2_threshold_5": 0.0401785797439502,
252
+ "scr_dir1_threshold_10": 0.06756750225933819,
253
+ "scr_metric_threshold_10": 0.07589293198126605,
254
+ "scr_dir2_threshold_10": 0.07589293198126605,
255
+ "scr_dir1_threshold_20": 0.12612611161318626,
256
+ "scr_metric_threshold_20": 0.18303572260109305,
257
+ "scr_dir2_threshold_20": 0.18303572260109305,
258
+ "scr_dir1_threshold_50": 0.1756755595721568,
259
+ "scr_metric_threshold_50": 0.3124999168462123,
260
+ "scr_dir2_threshold_50": 0.3124999168462123,
261
+ "scr_dir1_threshold_100": 0.3108108325802206,
262
+ "scr_metric_threshold_100": 0.3883928488274784,
263
+ "scr_dir2_threshold_100": 0.3883928488274784,
264
+ "scr_dir1_threshold_500": 0.445945837098897,
265
+ "scr_metric_threshold_500": 0.4999998669539397,
266
+ "scr_dir2_threshold_500": 0.4999998669539397
267
+ },
268
+ {
269
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
270
+ "scr_dir1_threshold_2": 0.042918406627421406,
271
+ "scr_metric_threshold_2": 0.042918406627421406,
272
+ "scr_dir2_threshold_2": 0.042857150966616867,
273
+ "scr_dir1_threshold_5": 0.060085820441166386,
274
+ "scr_metric_threshold_5": 0.060085820441166386,
275
+ "scr_dir2_threshold_5": 0.06190480515528994,
276
+ "scr_dir1_threshold_10": 0.07725323425491137,
277
+ "scr_metric_threshold_10": 0.07725323425491137,
278
+ "scr_dir2_threshold_10": 0.07619033292310208,
279
+ "scr_dir1_threshold_20": 0.10300422706858779,
280
+ "scr_metric_threshold_20": 0.10300422706858779,
281
+ "scr_dir2_threshold_20": 0.10476195612190681,
282
+ "scr_dir1_threshold_50": 0.07725323425491137,
283
+ "scr_metric_threshold_50": 0.07725323425491137,
284
+ "scr_dir2_threshold_50": 0.14285698066766273,
285
+ "scr_dir1_threshold_100": 0.1888412961373127,
286
+ "scr_metric_threshold_100": 0.1888412961373127,
287
+ "scr_dir2_threshold_100": 0.38095223227869085,
288
+ "scr_dir1_threshold_500": 0.18025746132349915,
289
+ "scr_metric_threshold_500": 0.18025746132349915,
290
+ "scr_dir2_threshold_500": 0.4761905032220562
291
+ }
292
+ ],
293
+ "sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
294
+ "sae_lens_id": "layer_12/width_1m/average_l0_58",
295
+ "sae_lens_release_id": "gemma-scope-2b-pt-res",
296
+ "sae_lens_version": "4.4.5",
297
+ "eval_result_unstructured": null
298
+ }
gemma_scope/scr_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_73_eval_results.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_type_id": "scr",
3
+ "eval_config": {
4
+ "random_seed": 42,
5
+ "dataset_names": [
6
+ "LabHC/bias_in_bios_class_set1",
7
+ "canrager/amazon_reviews_mcauley_1and5"
8
+ ],
9
+ "perform_scr": true,
10
+ "early_stopping_patience": 20,
11
+ "train_set_size": 4000,
12
+ "test_set_size": 1000,
13
+ "context_length": 128,
14
+ "probe_train_batch_size": 16,
15
+ "probe_test_batch_size": 500,
16
+ "probe_epochs": 20,
17
+ "probe_lr": 0.001,
18
+ "probe_l1_penalty": 0.001,
19
+ "sae_batch_size": 5,
20
+ "llm_batch_size": 32,
21
+ "llm_dtype": "bfloat16",
22
+ "lower_vram_usage": true,
23
+ "model_name": "gemma-2-2b",
24
+ "n_values": [
25
+ 2,
26
+ 5,
27
+ 10,
28
+ 20,
29
+ 50,
30
+ 100,
31
+ 500
32
+ ],
33
+ "column1_vals_lookup": {
34
+ "LabHC/bias_in_bios_class_set1": [
35
+ [
36
+ "professor",
37
+ "nurse"
38
+ ],
39
+ [
40
+ "architect",
41
+ "journalist"
42
+ ],
43
+ [
44
+ "surgeon",
45
+ "psychologist"
46
+ ],
47
+ [
48
+ "attorney",
49
+ "teacher"
50
+ ]
51
+ ],
52
+ "canrager/amazon_reviews_mcauley_1and5": [
53
+ [
54
+ "Books",
55
+ "CDs_and_Vinyl"
56
+ ],
57
+ [
58
+ "Software",
59
+ "Electronics"
60
+ ],
61
+ [
62
+ "Pet_Supplies",
63
+ "Office_Products"
64
+ ],
65
+ [
66
+ "Industrial_and_Scientific",
67
+ "Toys_and_Games"
68
+ ]
69
+ ]
70
+ }
71
+ },
72
+ "eval_id": "533dc60a-e587-4dbd-b20a-b7b40c569c07",
73
+ "datetime_epoch_millis": 1732798228842,
74
+ "eval_result_metrics": {
75
+ "scr_metrics": {
76
+ "scr_dir1_threshold_2": 0.14804548941247625,
77
+ "scr_metric_threshold_2": 0.024367656824407287,
78
+ "scr_dir2_threshold_2": 0.022633030053044546,
79
+ "scr_dir1_threshold_5": 0.1627032747183734,
80
+ "scr_metric_threshold_5": 0.04136765752915408,
81
+ "scr_dir2_threshold_5": 0.04034577785598994,
82
+ "scr_dir1_threshold_10": 0.16080778512027716,
83
+ "scr_metric_threshold_10": 0.0654848621083852,
84
+ "scr_dir2_threshold_10": 0.06594726524451666,
85
+ "scr_dir1_threshold_20": 0.10296406245633792,
86
+ "scr_metric_threshold_20": 0.09620359678647643,
87
+ "scr_dir2_threshold_20": 0.09928195126410926,
88
+ "scr_dir1_threshold_50": 0.18945984388401493,
89
+ "scr_metric_threshold_50": 0.16215599349939547,
90
+ "scr_dir2_threshold_50": 0.16457270405960575,
91
+ "scr_dir1_threshold_100": 0.3281586203173022,
92
+ "scr_metric_threshold_100": 0.2354526072505888,
93
+ "scr_dir2_threshold_100": 0.23642851320141428,
94
+ "scr_dir1_threshold_500": 0.25805203171677615,
95
+ "scr_metric_threshold_500": 0.27879606577536864,
96
+ "scr_dir2_threshold_500": 0.27118057238277227
97
+ }
98
+ },
99
+ "eval_result_details": [
100
+ {
101
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
102
+ "scr_dir1_threshold_2": 0.3437501746229541,
103
+ "scr_metric_threshold_2": 0.0073890670992652185,
104
+ "scr_dir2_threshold_2": 0.0073890670992652185,
105
+ "scr_dir1_threshold_5": 0.4062502910382569,
106
+ "scr_metric_threshold_5": 0.009852089465686957,
107
+ "scr_dir2_threshold_5": 0.009852089465686957,
108
+ "scr_dir1_threshold_10": 0.42187508731147705,
109
+ "scr_metric_threshold_10": 0.012315258641572036,
110
+ "scr_dir2_threshold_10": 0.012315258641572036,
111
+ "scr_dir1_threshold_20": 0.3437501746229541,
112
+ "scr_metric_threshold_20": 0.017241303374415515,
113
+ "scr_dir2_threshold_20": 0.017241303374415515,
114
+ "scr_dir1_threshold_50": 0.4062502910382569,
115
+ "scr_metric_threshold_50": 0.04187182065755959,
116
+ "scr_dir2_threshold_50": 0.04187182065755959,
117
+ "scr_dir1_threshold_100": 0.42187508731147705,
118
+ "scr_metric_threshold_100": 0.07389155184943222,
119
+ "scr_dir2_threshold_100": 0.07389155184943222,
120
+ "scr_dir1_threshold_500": 0.4531256111803394,
121
+ "scr_metric_threshold_500": 0.13300482269087066,
122
+ "scr_dir2_threshold_500": 0.13300482269087066
123
+ },
124
+ {
125
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
126
+ "scr_dir1_threshold_2": 0.08910893426316568,
127
+ "scr_metric_threshold_2": 0.019943102189018718,
128
+ "scr_dir2_threshold_2": 0.019943102189018718,
129
+ "scr_dir1_threshold_5": 0.1386138321710429,
130
+ "scr_metric_threshold_5": 0.04273517045824084,
131
+ "scr_dir2_threshold_5": 0.04273517045824084,
132
+ "scr_dir1_threshold_10": 0.1386138321710429,
133
+ "scr_metric_threshold_10": 0.07122517088786974,
134
+ "scr_dir2_threshold_10": 0.07122517088786974,
135
+ "scr_dir1_threshold_20": -0.08910893426316568,
136
+ "scr_metric_threshold_20": 0.09686620523729525,
137
+ "scr_dir2_threshold_20": 0.09686620523729525,
138
+ "scr_dir1_threshold_50": -0.3069308391447856,
139
+ "scr_metric_threshold_50": 0.15669517217675646,
140
+ "scr_dir2_threshold_50": 0.15669517217675646,
141
+ "scr_dir1_threshold_100": 0.24752507968431967,
142
+ "scr_metric_threshold_100": 0.23931637719923726,
143
+ "scr_dir2_threshold_100": 0.23931637719923726,
144
+ "scr_dir1_threshold_500": 0.45544553286841716,
145
+ "scr_metric_threshold_500": 0.1396012058817386,
146
+ "scr_dir2_threshold_500": 0.1396012058817386
147
+ },
148
+ {
149
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
150
+ "scr_dir1_threshold_2": 0.5079364028137167,
151
+ "scr_metric_threshold_2": 0.007614265416724154,
152
+ "scr_dir2_threshold_2": 0.007614265416724154,
153
+ "scr_dir1_threshold_5": 0.5079364028137167,
154
+ "scr_metric_threshold_5": 0.01015230346203094,
155
+ "scr_dir2_threshold_5": 0.01015230346203094,
156
+ "scr_dir1_threshold_10": 0.4761907915588501,
157
+ "scr_metric_threshold_10": 0.022842644969368664,
158
+ "scr_dir2_threshold_10": 0.022842644969368664,
159
+ "scr_dir1_threshold_20": 0.31746084307427397,
160
+ "scr_metric_threshold_20": 0.03807117580281697,
161
+ "scr_dir2_threshold_20": 0.03807117580281697,
162
+ "scr_dir1_threshold_50": 0.5873013770560047,
163
+ "scr_metric_threshold_50": 0.12944175568029162,
164
+ "scr_dir2_threshold_50": 0.12944175568029162,
165
+ "scr_dir1_threshold_100": 0.5714285714285714,
166
+ "scr_metric_threshold_100": 0.15736047674027384,
167
+ "scr_dir2_threshold_100": 0.15736047674027384,
168
+ "scr_dir1_threshold_500": -0.2380949227268642,
169
+ "scr_metric_threshold_500": 0.08121827641544371,
170
+ "scr_dir2_threshold_500": 0.08121827641544371
171
+ },
172
+ {
173
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
174
+ "scr_dir1_threshold_2": 0.10852711387397361,
175
+ "scr_metric_threshold_2": 0.01785709217291593,
176
+ "scr_dir2_threshold_2": 0.01785709217291593,
177
+ "scr_dir1_threshold_5": 0.13178308900821112,
178
+ "scr_metric_threshold_5": 0.038690336808852144,
179
+ "scr_dir2_threshold_5": 0.038690336808852144,
180
+ "scr_dir1_threshold_10": 0.07751930104548388,
181
+ "scr_metric_threshold_10": 0.07142854608645796,
182
+ "scr_dir2_threshold_10": 0.07142854608645796,
183
+ "scr_dir1_threshold_20": 0.0,
184
+ "scr_metric_threshold_20": 0.09821427304322898,
185
+ "scr_dir2_threshold_20": 0.09821427304322898,
186
+ "scr_dir1_threshold_50": 0.37984512958464806,
187
+ "scr_metric_threshold_50": 0.11607136521614492,
188
+ "scr_dir2_threshold_50": 0.11607136521614492,
189
+ "scr_dir1_threshold_100": 0.6511626832438416,
190
+ "scr_metric_threshold_100": 0.1904762411604174,
191
+ "scr_dir2_threshold_100": 0.1904762411604174,
192
+ "scr_dir1_threshold_500": 0.35658915445041056,
193
+ "scr_metric_threshold_500": 0.059523758839582594,
194
+ "scr_dir2_threshold_500": 0.059523758839582594
195
+ },
196
+ {
197
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
198
+ "scr_dir1_threshold_2": 0.027322374114521025,
199
+ "scr_metric_threshold_2": 0.03501946427722419,
200
+ "scr_dir2_threshold_2": 0.03501946427722419,
201
+ "scr_dir1_threshold_5": -0.06557389329988525,
202
+ "scr_metric_threshold_5": 0.058365928411814705,
203
+ "scr_dir2_threshold_5": 0.058365928411814705,
204
+ "scr_dir1_threshold_10": -0.04918033854781611,
205
+ "scr_metric_threshold_10": 0.12840485696626308,
206
+ "scr_dir2_threshold_10": 0.12840485696626308,
207
+ "scr_dir1_threshold_20": -0.06557389329988525,
208
+ "scr_metric_threshold_20": 0.18287955340587164,
209
+ "scr_dir2_threshold_20": 0.18287955340587164,
210
+ "scr_dir1_threshold_50": -0.021857964433295084,
211
+ "scr_metric_threshold_50": 0.30350194642772244,
212
+ "scr_dir2_threshold_50": 0.30350194642772244,
213
+ "scr_dir1_threshold_100": -0.027322374114521025,
214
+ "scr_metric_threshold_100": 0.36964987493462625,
215
+ "scr_dir2_threshold_100": 0.36964987493462625,
216
+ "scr_dir1_threshold_500": -0.06010915791026799,
217
+ "scr_metric_threshold_500": 0.560311428435587,
218
+ "scr_dir2_threshold_500": 0.560311428435587
219
+ },
220
+ {
221
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
222
+ "scr_dir1_threshold_2": 0.015463717977866399,
223
+ "scr_metric_threshold_2": 0.024193715075407077,
224
+ "scr_dir2_threshold_2": 0.024193715075407077,
225
+ "scr_dir1_threshold_5": 0.04123688851473832,
226
+ "scr_metric_threshold_5": 0.04838718980952953,
227
+ "scr_dir2_threshold_5": 0.04838718980952953,
228
+ "scr_dir1_threshold_10": 0.03608231585544952,
229
+ "scr_metric_threshold_10": 0.06451625307937271,
230
+ "scr_dir2_threshold_10": 0.06451625307937271,
231
+ "scr_dir1_threshold_20": 0.08762865692919337,
232
+ "scr_metric_threshold_20": 0.11693564862104187,
233
+ "scr_dir2_threshold_20": 0.11693564862104187,
234
+ "scr_dir1_threshold_50": 0.14432987790265395,
235
+ "scr_metric_threshold_50": 0.15322574055158325,
236
+ "scr_dir2_threshold_50": 0.15322574055158325,
237
+ "scr_dir1_threshold_100": 0.25257713270943044,
238
+ "scr_metric_threshold_100": 0.2137097878988163,
239
+ "scr_dir2_threshold_100": 0.2137097878988163,
240
+ "scr_dir1_threshold_500": 0.2886597558053079,
241
+ "scr_metric_threshold_500": 0.40322586072222555,
242
+ "scr_dir2_threshold_500": 0.40322586072222555
243
+ },
244
+ {
245
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
246
+ "scr_dir1_threshold_2": 0.045045001506225466,
247
+ "scr_metric_threshold_2": 0.03571435223731585,
248
+ "scr_dir2_threshold_2": 0.03571435223731585,
249
+ "scr_dir1_threshold_5": 0.08558555655970586,
250
+ "scr_metric_threshold_5": 0.06696421087587681,
251
+ "scr_dir2_threshold_5": 0.06696421087587681,
252
+ "scr_dir1_threshold_10": 0.1081080573128186,
253
+ "scr_metric_threshold_10": 0.07589293198126605,
254
+ "scr_dir2_threshold_10": 0.07589293198126605,
255
+ "scr_dir1_threshold_20": 0.13513500451867638,
256
+ "scr_metric_threshold_20": 0.12499996673848493,
257
+ "scr_dir2_threshold_20": 0.12499996673848493,
258
+ "scr_dir1_threshold_50": 0.19369361387252446,
259
+ "scr_metric_threshold_50": 0.2633928820889934,
260
+ "scr_dir2_threshold_50": 0.2633928820889934,
261
+ "scr_dir1_threshold_100": 0.3063061176380881,
262
+ "scr_metric_threshold_100": 0.43749988358469727,
263
+ "scr_dir2_threshold_100": 0.43749988358469727,
264
+ "scr_dir1_threshold_500": 0.49549555354725494,
265
+ "scr_metric_threshold_500": 0.54017844669789,
266
+ "scr_dir2_threshold_500": 0.54017844669789
267
+ },
268
+ {
269
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
270
+ "scr_dir1_threshold_2": 0.047210196127387125,
271
+ "scr_metric_threshold_2": 0.047210196127387125,
272
+ "scr_dir2_threshold_2": 0.033333181956485214,
273
+ "scr_dir1_threshold_5": 0.05579403094120067,
274
+ "scr_metric_threshold_5": 0.05579403094120067,
275
+ "scr_dir2_threshold_5": 0.04761899355588758,
276
+ "scr_dir1_threshold_10": 0.07725323425491137,
277
+ "scr_metric_threshold_10": 0.07725323425491137,
278
+ "scr_dir2_threshold_10": 0.08095245934396302,
279
+ "scr_dir1_threshold_20": 0.09442064806865635,
280
+ "scr_metric_threshold_20": 0.09442064806865635,
281
+ "scr_dir2_threshold_20": 0.11904748388971893,
282
+ "scr_dir1_threshold_50": 0.13304726519611204,
283
+ "scr_metric_threshold_50": 0.13304726519611204,
284
+ "scr_dir2_threshold_50": 0.1523809496777944,
285
+ "scr_dir1_threshold_100": 0.20171666463720986,
286
+ "scr_metric_threshold_100": 0.20171666463720986,
287
+ "scr_dir2_threshold_100": 0.20952391224381361,
288
+ "scr_dir1_threshold_500": 0.3133047265196112,
289
+ "scr_metric_threshold_500": 0.3133047265196112,
290
+ "scr_dir2_threshold_500": 0.25238077937884024
291
+ }
292
+ ],
293
+ "sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
294
+ "sae_lens_id": "layer_12/width_1m/average_l0_73",
295
+ "sae_lens_release_id": "gemma-scope-2b-pt-res",
296
+ "sae_lens_version": "4.4.5",
297
+ "eval_result_unstructured": null
298
+ }
gemma_scope/scr_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_157_eval_results.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_type_id": "scr",
3
+ "eval_config": {
4
+ "random_seed": 42,
5
+ "dataset_names": [
6
+ "LabHC/bias_in_bios_class_set1",
7
+ "canrager/amazon_reviews_mcauley_1and5"
8
+ ],
9
+ "perform_scr": true,
10
+ "early_stopping_patience": 20,
11
+ "train_set_size": 4000,
12
+ "test_set_size": 1000,
13
+ "context_length": 128,
14
+ "probe_train_batch_size": 16,
15
+ "probe_test_batch_size": 500,
16
+ "probe_epochs": 20,
17
+ "probe_lr": 0.001,
18
+ "probe_l1_penalty": 0.001,
19
+ "sae_batch_size": 5,
20
+ "llm_batch_size": 32,
21
+ "llm_dtype": "bfloat16",
22
+ "lower_vram_usage": true,
23
+ "model_name": "gemma-2-2b",
24
+ "n_values": [
25
+ 2,
26
+ 5,
27
+ 10,
28
+ 20,
29
+ 50,
30
+ 100,
31
+ 500
32
+ ],
33
+ "column1_vals_lookup": {
34
+ "LabHC/bias_in_bios_class_set1": [
35
+ [
36
+ "professor",
37
+ "nurse"
38
+ ],
39
+ [
40
+ "architect",
41
+ "journalist"
42
+ ],
43
+ [
44
+ "surgeon",
45
+ "psychologist"
46
+ ],
47
+ [
48
+ "attorney",
49
+ "teacher"
50
+ ]
51
+ ],
52
+ "canrager/amazon_reviews_mcauley_1and5": [
53
+ [
54
+ "Books",
55
+ "CDs_and_Vinyl"
56
+ ],
57
+ [
58
+ "Software",
59
+ "Electronics"
60
+ ],
61
+ [
62
+ "Pet_Supplies",
63
+ "Office_Products"
64
+ ],
65
+ [
66
+ "Industrial_and_Scientific",
67
+ "Toys_and_Games"
68
+ ]
69
+ ]
70
+ }
71
+ },
72
+ "eval_id": "7f5ed994-f91b-497f-bf58-044e9cdcc0ae",
73
+ "datetime_epoch_millis": 1732735620345,
74
+ "eval_result_metrics": {
75
+ "scr_metrics": {
76
+ "scr_dir1_threshold_2": 0.20087590911464995,
77
+ "scr_metric_threshold_2": 0.21201733061264716,
78
+ "scr_dir2_threshold_2": 0.21201733061264716,
79
+ "scr_dir1_threshold_5": 0.3031110220011594,
80
+ "scr_metric_threshold_5": 0.28525899022838364,
81
+ "scr_dir2_threshold_5": 0.28525899022838364,
82
+ "scr_dir1_threshold_10": 0.31985864321929136,
83
+ "scr_metric_threshold_10": 0.3326241129375694,
84
+ "scr_dir2_threshold_10": 0.3326241129375694,
85
+ "scr_dir1_threshold_20": 0.2790153396234907,
86
+ "scr_metric_threshold_20": 0.38807959834035755,
87
+ "scr_dir2_threshold_20": 0.38807959834035755,
88
+ "scr_dir1_threshold_50": 0.22696246720827495,
89
+ "scr_metric_threshold_50": 0.4644569374527007,
90
+ "scr_dir2_threshold_50": 0.4644569374527007,
91
+ "scr_dir1_threshold_100": 0.09578760809133437,
92
+ "scr_metric_threshold_100": 0.48887880260649363,
93
+ "scr_dir2_threshold_100": 0.48887880260649363,
94
+ "scr_dir1_threshold_500": -0.23202806204917542,
95
+ "scr_metric_threshold_500": 0.48898892740962463,
96
+ "scr_dir2_threshold_500": 0.48898892740962463
97
+ }
98
+ },
99
+ "eval_result_details": [
100
+ {
101
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
102
+ "scr_dir1_threshold_2": 0.2941181626697965,
103
+ "scr_metric_threshold_2": 0.055555589003726494,
104
+ "scr_dir2_threshold_2": 0.055555589003726494,
105
+ "scr_dir1_threshold_5": 0.4117654277377151,
106
+ "scr_metric_threshold_5": 0.09848492830434731,
107
+ "scr_dir2_threshold_5": 0.09848492830434731,
108
+ "scr_dir1_threshold_10": 0.500000438269327,
109
+ "scr_metric_threshold_10": 0.1060607383329184,
110
+ "scr_dir2_threshold_10": 0.1060607383329184,
111
+ "scr_dir1_threshold_20": 0.47058818373302036,
112
+ "scr_metric_threshold_20": 0.14898992711677,
113
+ "scr_dir2_threshold_20": 0.14898992711677,
114
+ "scr_dir1_threshold_50": 0.500000438269327,
115
+ "scr_metric_threshold_50": 0.1641415471739122,
116
+ "scr_dir2_threshold_50": 0.1641415471739122,
117
+ "scr_dir1_threshold_100": 0.14705951960422528,
118
+ "scr_metric_threshold_100": 0.1666667670111795,
119
+ "scr_dir2_threshold_100": 0.1666667670111795,
120
+ "scr_dir1_threshold_500": -2.1764700210632237,
121
+ "scr_metric_threshold_500": 0.24242426522981353,
122
+ "scr_dir2_threshold_500": 0.24242426522981353
123
+ },
124
+ {
125
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
126
+ "scr_dir1_threshold_2": 0.29464281912968693,
127
+ "scr_metric_threshold_2": 0.15451902187229455,
128
+ "scr_dir2_threshold_2": 0.15451902187229455,
129
+ "scr_dir1_threshold_5": 0.5357143617406261,
130
+ "scr_metric_threshold_5": 0.27113696797649733,
131
+ "scr_dir2_threshold_5": 0.27113696797649733,
132
+ "scr_dir1_threshold_10": 0.5267859043515652,
133
+ "scr_metric_threshold_10": 0.39358608942502554,
134
+ "scr_dir2_threshold_10": 0.39358608942502554,
135
+ "scr_dir1_threshold_20": 0.5535718087031305,
136
+ "scr_metric_threshold_20": 0.521865864894538,
137
+ "scr_dir2_threshold_20": 0.521865864894538,
138
+ "scr_dir1_threshold_50": 0.6339289895734436,
139
+ "scr_metric_threshold_50": 0.6355685708754721,
140
+ "scr_dir2_threshold_50": 0.6355685708754721,
141
+ "scr_dir1_threshold_100": 0.6428574469625045,
142
+ "scr_metric_threshold_100": 0.7055393732928833,
143
+ "scr_dir2_threshold_100": 0.7055393732928833,
144
+ "scr_dir1_threshold_500": 0.4732146278328175,
145
+ "scr_metric_threshold_500": 0.3206996993354518,
146
+ "scr_dir2_threshold_500": 0.3206996993354518
147
+ },
148
+ {
149
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
150
+ "scr_dir1_threshold_2": 0.33999911785146764,
151
+ "scr_metric_threshold_2": 0.0679611580256569,
152
+ "scr_dir2_threshold_2": 0.0679611580256569,
153
+ "scr_dir1_threshold_5": 0.47999966621406887,
154
+ "scr_metric_threshold_5": 0.13349516283714058,
155
+ "scr_dir2_threshold_5": 0.13349516283714058,
156
+ "scr_dir1_threshold_10": 0.5199991416933198,
157
+ "scr_metric_threshold_10": 0.19174756333463683,
158
+ "scr_dir2_threshold_10": 0.19174756333463683,
159
+ "scr_dir1_threshold_20": 0.47999966621406887,
160
+ "scr_metric_threshold_20": 0.24999996383213308,
161
+ "scr_dir2_threshold_20": 0.24999996383213308,
162
+ "scr_dir1_threshold_50": 0.33999911785146764,
163
+ "scr_metric_threshold_50": 0.3543688540847917,
164
+ "scr_dir2_threshold_50": 0.3543688540847917,
165
+ "scr_dir1_threshold_100": 0.19999976158147775,
166
+ "scr_metric_threshold_100": 0.4247573099960895,
167
+ "scr_dir2_threshold_100": 0.4247573099960895,
168
+ "scr_dir1_threshold_500": -0.16000028610222672,
169
+ "scr_metric_threshold_500": 0.22330098913329205,
170
+ "scr_dir2_threshold_500": 0.22330098913329205
171
+ },
172
+ {
173
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
174
+ "scr_dir1_threshold_2": 0.20454557769557552,
175
+ "scr_metric_threshold_2": 0.1975683483084694,
176
+ "scr_dir2_threshold_2": 0.1975683483084694,
177
+ "scr_dir1_threshold_5": 0.4545455776955755,
178
+ "scr_metric_threshold_5": 0.3738601972388017,
179
+ "scr_dir2_threshold_5": 0.3738601972388017,
180
+ "scr_dir1_threshold_10": 0.49242433820766984,
181
+ "scr_metric_threshold_10": 0.44072953174127966,
182
+ "scr_dir2_threshold_10": 0.44072953174127966,
183
+ "scr_dir1_threshold_20": 0.13636371846371703,
184
+ "scr_metric_threshold_20": 0.480243237636794,
185
+ "scr_dir2_threshold_20": 0.480243237636794,
186
+ "scr_dir1_threshold_50": 0.25,
187
+ "scr_metric_threshold_50": 0.5623099422885824,
188
+ "scr_dir2_threshold_50": 0.5623099422885824,
189
+ "scr_dir1_threshold_100": 0.40909115539115104,
190
+ "scr_metric_threshold_100": 0.5288753656218946,
191
+ "scr_dir2_threshold_100": 0.5288753656218946,
192
+ "scr_dir1_threshold_500": 0.3560606197439528,
193
+ "scr_metric_threshold_500": 0.5744680195771332,
194
+ "scr_dir2_threshold_500": 0.5744680195771332
195
+ },
196
+ {
197
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
198
+ "scr_dir1_threshold_2": 0.017241497432889975,
199
+ "scr_metric_threshold_2": 0.6328125454747382,
200
+ "scr_dir2_threshold_2": 0.6328125454747382,
201
+ "scr_dir1_threshold_5": 0.045977098117452646,
202
+ "scr_metric_threshold_5": 0.6640623781277015,
203
+ "scr_dir2_threshold_5": 0.6640623781277015,
204
+ "scr_dir1_threshold_10": 0.14367800342281337,
205
+ "scr_metric_threshold_10": 0.6679688109361492,
206
+ "scr_dir2_threshold_10": 0.6679688109361492,
207
+ "scr_dir1_threshold_20": 0.137930951796977,
208
+ "scr_metric_threshold_20": 0.7187501091393718,
209
+ "scr_dir2_threshold_20": 0.7187501091393718,
210
+ "scr_dir1_threshold_50": -0.5,
211
+ "scr_metric_threshold_50": 0.7656249745341466,
212
+ "scr_dir2_threshold_50": 0.7656249745341466,
213
+ "scr_dir1_threshold_100": -0.551724149743289,
214
+ "scr_metric_threshold_100": 0.6953124436113246,
215
+ "scr_dir2_threshold_100": 0.6953124436113246,
216
+ "scr_dir1_threshold_500": -0.7183907022248287,
217
+ "scr_metric_threshold_500": 0.6953124436113246,
218
+ "scr_dir2_threshold_500": 0.6953124436113246
219
+ },
220
+ {
221
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
222
+ "scr_dir1_threshold_2": 0.09677420388564499,
223
+ "scr_metric_threshold_2": 0.13888897430117703,
224
+ "scr_dir2_threshold_2": 0.13888897430117703,
225
+ "scr_dir1_threshold_5": 0.09677420388564499,
226
+ "scr_metric_threshold_5": 0.14682537899206194,
227
+ "scr_dir2_threshold_5": 0.14682537899206194,
228
+ "scr_dir1_threshold_10": 0.16666672007583247,
229
+ "scr_metric_threshold_10": 0.17857147080827435,
230
+ "scr_dir2_threshold_10": 0.17857147080827435,
231
+ "scr_dir1_threshold_20": 0.1344087589322824,
232
+ "scr_metric_threshold_20": 0.2500000591315841,
233
+ "scr_dir2_threshold_20": 0.2500000591315841,
234
+ "scr_dir1_threshold_50": 0.02688168769545752,
235
+ "scr_metric_threshold_50": 0.35317473927110626,
236
+ "scr_dir2_threshold_50": 0.35317473927110626,
237
+ "scr_dir1_threshold_100": 0.00537659390308736,
238
+ "scr_metric_threshold_100": 0.396825438123646,
239
+ "scr_dir2_threshold_100": 0.396825438123646,
240
+ "scr_dir1_threshold_500": 0.14516130582846748,
241
+ "scr_metric_threshold_500": 0.5793651112773629,
242
+ "scr_dir2_threshold_500": 0.5793651112773629
243
+ },
244
+ {
245
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
246
+ "scr_dir1_threshold_2": 0.23134341304194075,
247
+ "scr_metric_threshold_2": 0.3344262154119629,
248
+ "scr_dir2_threshold_2": 0.3344262154119629,
249
+ "scr_dir1_threshold_5": 0.2985076153828019,
250
+ "scr_metric_threshold_5": 0.4459017508326573,
251
+ "scr_dir2_threshold_5": 0.4459017508326573,
252
+ "scr_dir1_threshold_10": 0.09701500836021844,
253
+ "scr_metric_threshold_10": 0.5213115913833964,
254
+ "scr_dir2_threshold_10": 0.5213115913833964,
255
+ "scr_dir1_threshold_20": 0.1641792107010796,
256
+ "scr_metric_threshold_20": 0.5737705174327948,
257
+ "scr_dir2_threshold_20": 0.5737705174327948,
258
+ "scr_dir1_threshold_50": 0.27611939966559124,
259
+ "scr_metric_threshold_50": 0.6262296389072534,
260
+ "scr_dir2_threshold_50": 0.6262296389072534,
261
+ "scr_dir1_threshold_100": -0.4179103946494602,
262
+ "scr_metric_threshold_100": 0.7049181256938811,
263
+ "scr_dir2_threshold_100": 0.7049181256938811,
264
+ "scr_dir1_threshold_500": -0.2089551973247301,
265
+ "scr_metric_threshold_500": 0.8229509535863527,
266
+ "scr_dir2_threshold_500": 0.8229509535863527
267
+ },
268
+ {
269
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
270
+ "scr_dir1_threshold_2": 0.12834248121019717,
271
+ "scr_metric_threshold_2": 0.11440679250315197,
272
+ "scr_dir2_threshold_2": 0.11440679250315197,
273
+ "scr_dir1_threshold_5": 0.10160422523539049,
274
+ "scr_metric_threshold_5": 0.1483051575178612,
275
+ "scr_dir2_threshold_5": 0.1483051575178612,
276
+ "scr_dir1_threshold_10": 0.11229959137358399,
277
+ "scr_metric_threshold_10": 0.16101710753887435,
278
+ "scr_dir2_threshold_10": 0.16101710753887435,
279
+ "scr_dir1_threshold_20": 0.15508041844364973,
280
+ "scr_metric_threshold_20": 0.16101710753887435,
281
+ "scr_dir2_threshold_20": 0.16101710753887435,
282
+ "scr_dir1_threshold_50": 0.2887701046109124,
283
+ "scr_metric_threshold_50": 0.25423723248634145,
284
+ "scr_dir2_threshold_50": 0.25423723248634145,
285
+ "scr_dir1_threshold_100": 0.33155093168097816,
286
+ "scr_metric_threshold_100": 0.28813559750105067,
287
+ "scr_dir2_threshold_100": 0.28813559750105067,
288
+ "scr_dir1_threshold_500": 0.43315515691636863,
289
+ "scr_metric_threshold_500": 0.45338993752626644,
290
+ "scr_dir2_threshold_500": 0.45338993752626644
291
+ }
292
+ ],
293
+ "sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
294
+ "sae_lens_id": "layer_19/width_1m/average_l0_157",
295
+ "sae_lens_release_id": "gemma-scope-2b-pt-res",
296
+ "sae_lens_version": "4.4.5",
297
+ "eval_result_unstructured": null
298
+ }
gemma_scope/scr_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_16_eval_results.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_type_id": "scr",
3
+ "eval_config": {
4
+ "random_seed": 42,
5
+ "dataset_names": [
6
+ "LabHC/bias_in_bios_class_set1",
7
+ "canrager/amazon_reviews_mcauley_1and5"
8
+ ],
9
+ "perform_scr": true,
10
+ "early_stopping_patience": 20,
11
+ "train_set_size": 4000,
12
+ "test_set_size": 1000,
13
+ "context_length": 128,
14
+ "probe_train_batch_size": 16,
15
+ "probe_test_batch_size": 500,
16
+ "probe_epochs": 20,
17
+ "probe_lr": 0.001,
18
+ "probe_l1_penalty": 0.001,
19
+ "sae_batch_size": 5,
20
+ "llm_batch_size": 32,
21
+ "llm_dtype": "bfloat16",
22
+ "lower_vram_usage": true,
23
+ "model_name": "gemma-2-2b",
24
+ "n_values": [
25
+ 2,
26
+ 5,
27
+ 10,
28
+ 20,
29
+ 50,
30
+ 100,
31
+ 500
32
+ ],
33
+ "column1_vals_lookup": {
34
+ "LabHC/bias_in_bios_class_set1": [
35
+ [
36
+ "professor",
37
+ "nurse"
38
+ ],
39
+ [
40
+ "architect",
41
+ "journalist"
42
+ ],
43
+ [
44
+ "surgeon",
45
+ "psychologist"
46
+ ],
47
+ [
48
+ "attorney",
49
+ "teacher"
50
+ ]
51
+ ],
52
+ "canrager/amazon_reviews_mcauley_1and5": [
53
+ [
54
+ "Books",
55
+ "CDs_and_Vinyl"
56
+ ],
57
+ [
58
+ "Software",
59
+ "Electronics"
60
+ ],
61
+ [
62
+ "Pet_Supplies",
63
+ "Office_Products"
64
+ ],
65
+ [
66
+ "Industrial_and_Scientific",
67
+ "Toys_and_Games"
68
+ ]
69
+ ]
70
+ }
71
+ },
72
+ "eval_id": "cb05704d-7ca9-4b45-b7fb-b57d5813d1d3",
73
+ "datetime_epoch_millis": 1732748745968,
74
+ "eval_result_metrics": {
75
+ "scr_metrics": {
76
+ "scr_dir1_threshold_2": 0.04439077933157716,
77
+ "scr_metric_threshold_2": 0.033677728834152036,
78
+ "scr_dir2_threshold_2": 0.033677728834152036,
79
+ "scr_dir1_threshold_5": 0.0868800502249312,
80
+ "scr_metric_threshold_5": 0.06234706386294911,
81
+ "scr_dir2_threshold_5": 0.06234706386294911,
82
+ "scr_dir1_threshold_10": 0.14059626069536535,
83
+ "scr_metric_threshold_10": 0.09179859296095819,
84
+ "scr_dir2_threshold_10": 0.09179859296095819,
85
+ "scr_dir1_threshold_20": 0.17316656822808946,
86
+ "scr_metric_threshold_20": 0.1338064501359954,
87
+ "scr_dir2_threshold_20": 0.1338064501359954,
88
+ "scr_dir1_threshold_50": 0.19808046116344458,
89
+ "scr_metric_threshold_50": 0.20275272624453913,
90
+ "scr_dir2_threshold_50": 0.20275272624453913,
91
+ "scr_dir1_threshold_100": 0.17633950151105227,
92
+ "scr_metric_threshold_100": 0.24692504270711876,
93
+ "scr_dir2_threshold_100": 0.24692504270711876,
94
+ "scr_dir1_threshold_500": -0.18619147782445633,
95
+ "scr_metric_threshold_500": 0.3369847816425646,
96
+ "scr_dir2_threshold_500": 0.3369847816425646
97
+ }
98
+ },
99
+ "eval_result_details": [
100
+ {
101
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
102
+ "scr_dir1_threshold_2": 0.1176472650679186,
103
+ "scr_metric_threshold_2": 0.04545455913788811,
104
+ "scr_dir2_threshold_2": 0.04545455913788811,
105
+ "scr_dir1_threshold_5": 0.1764708976018779,
106
+ "scr_metric_threshold_5": 0.07323242889813597,
107
+ "scr_dir2_threshold_5": 0.07323242889813597,
108
+ "scr_dir1_threshold_10": 0.20588315213818456,
109
+ "scr_metric_threshold_10": 0.08838389843850893,
110
+ "scr_dir2_threshold_10": 0.08838389843850893,
111
+ "scr_dir1_threshold_20": 0.27941247367097016,
112
+ "scr_metric_threshold_20": 0.1085859581701857,
113
+ "scr_dir2_threshold_20": 0.1085859581701857,
114
+ "scr_dir1_threshold_50": 0.029412254536306668,
115
+ "scr_metric_threshold_50": 0.13131323773912976,
116
+ "scr_dir2_threshold_50": 0.13131323773912976,
117
+ "scr_dir1_threshold_100": -0.13235207752809086,
118
+ "scr_metric_threshold_100": 0.25000007525838464,
119
+ "scr_dir2_threshold_100": 0.25000007525838464,
120
+ "scr_dir1_threshold_500": -1.1911757100620501,
121
+ "scr_metric_threshold_500": 0.29545463439627273,
122
+ "scr_dir2_threshold_500": 0.29545463439627273
123
+ },
124
+ {
125
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
126
+ "scr_dir1_threshold_2": 0.0892856382593739,
127
+ "scr_metric_threshold_2": 0.06705538851969542,
128
+ "scr_dir2_threshold_2": 0.06705538851969542,
129
+ "scr_dir1_threshold_5": 0.16071436174062612,
130
+ "scr_metric_threshold_5": 0.10787170441105581,
131
+ "scr_dir2_threshold_5": 0.10787170441105581,
132
+ "scr_dir1_threshold_10": 0.23214308522187832,
133
+ "scr_metric_threshold_10": 0.13119536314167518,
134
+ "scr_dir2_threshold_10": 0.13119536314167518,
135
+ "scr_dir1_threshold_20": 0.22321462783281748,
136
+ "scr_metric_threshold_20": 0.19241992386593926,
137
+ "scr_dir2_threshold_20": 0.19241992386593926,
138
+ "scr_dir1_threshold_50": 0.3482146278328175,
139
+ "scr_metric_threshold_50": 0.26239072628335036,
140
+ "scr_dir2_threshold_50": 0.26239072628335036,
141
+ "scr_dir1_threshold_100": 0.3482146278328175,
142
+ "scr_metric_threshold_100": 0.16618085123760423,
143
+ "scr_dir2_threshold_100": 0.16618085123760423,
144
+ "scr_dir1_threshold_500": -1.2232146278328175,
145
+ "scr_metric_threshold_500": 0.09620987504574613,
146
+ "scr_dir2_threshold_500": 0.09620987504574613
147
+ },
148
+ {
149
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
150
+ "scr_dir1_threshold_2": 0.039999475479251034,
151
+ "scr_metric_threshold_2": 0.007281459642519697,
152
+ "scr_dir2_threshold_2": 0.007281459642519697,
153
+ "scr_dir1_threshold_5": 0.08000014305111336,
154
+ "scr_metric_threshold_5": 0.014563063956507128,
155
+ "scr_dir2_threshold_5": 0.014563063956507128,
156
+ "scr_dir1_threshold_10": 0.19999976158147775,
157
+ "scr_metric_threshold_10": 0.01699021717068036,
158
+ "scr_dir2_threshold_10": 0.01699021717068036,
159
+ "scr_dir1_threshold_20": 0.2599989748003543,
160
+ "scr_metric_threshold_20": 0.031553281127187485,
161
+ "scr_dir2_threshold_20": 0.031553281127187485,
162
+ "scr_dir1_threshold_50": 0.15999909400961543,
163
+ "scr_metric_threshold_50": 0.10679604346683183,
164
+ "scr_dir2_threshold_50": 0.10679604346683183,
165
+ "scr_dir1_threshold_100": 0.15999909400961543,
166
+ "scr_metric_threshold_100": 0.19902902297715652,
167
+ "scr_dir2_threshold_100": 0.19902902297715652,
168
+ "scr_dir1_threshold_500": -0.06000040531148784,
169
+ "scr_metric_threshold_500": 0.3082523643296293,
170
+ "scr_dir2_threshold_500": 0.3082523643296293
171
+ },
172
+ {
173
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
174
+ "scr_dir1_threshold_2": 0.053030084096754616,
175
+ "scr_metric_threshold_2": 0.009118603258688643,
176
+ "scr_dir2_threshold_2": 0.009118603258688643,
177
+ "scr_dir1_threshold_5": 0.12121194332861313,
178
+ "scr_metric_threshold_5": 0.05167178318406509,
179
+ "scr_dir2_threshold_5": 0.05167178318406509,
180
+ "scr_dir1_threshold_10": 0.2348486764153397,
181
+ "scr_metric_threshold_10": 0.11550164365668093,
182
+ "scr_dir2_threshold_10": 0.11550164365668093,
183
+ "scr_dir1_threshold_20": 0.2348486764153397,
184
+ "scr_metric_threshold_20": 0.18844992621888307,
185
+ "scr_dir2_threshold_20": 0.18844992621888307,
186
+ "scr_dir1_threshold_50": 0.47727256307256594,
187
+ "scr_metric_threshold_50": 0.2583587347512231,
188
+ "scr_dir2_threshold_50": 0.2583587347512231,
189
+ "scr_dir1_threshold_100": 0.4545455776955755,
190
+ "scr_metric_threshold_100": 0.23404258017412163,
191
+ "scr_dir2_threshold_100": 0.23404258017412163,
192
+ "scr_dir1_threshold_500": 0.3181818592318585,
193
+ "scr_metric_threshold_500": 0.3586626459203889,
194
+ "scr_dir2_threshold_500": 0.3586626459203889
195
+ },
196
+ {
197
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
198
+ "scr_dir1_threshold_2": -0.08045975042785167,
199
+ "scr_metric_threshold_2": 0.019531232719599476,
200
+ "scr_dir2_threshold_2": 0.019531232719599476,
201
+ "scr_dir1_threshold_5": -0.0632182529949617,
202
+ "scr_metric_threshold_5": 0.070312530922822,
203
+ "scr_dir2_threshold_5": 0.070312530922822,
204
+ "scr_dir1_threshold_10": -0.03448265231039902,
205
+ "scr_metric_threshold_10": 0.09374996362020942,
206
+ "scr_dir2_threshold_10": 0.09374996362020942,
207
+ "scr_dir1_threshold_20": 0.0,
208
+ "scr_metric_threshold_20": 0.1132811963398089,
209
+ "scr_dir2_threshold_20": 0.1132811963398089,
210
+ "scr_dir1_threshold_50": -0.011494445807053626,
211
+ "scr_metric_threshold_50": 0.2265623926796178,
212
+ "scr_dir2_threshold_50": 0.2265623926796178,
213
+ "scr_dir1_threshold_100": -0.017241497432889975,
214
+ "scr_metric_threshold_100": 0.32031235629982724,
215
+ "scr_dir2_threshold_100": 0.32031235629982724,
216
+ "scr_dir1_threshold_500": -0.011494445807053626,
217
+ "scr_metric_threshold_500": 0.5156249163264817,
218
+ "scr_dir2_threshold_500": 0.5156249163264817
219
+ },
220
+ {
221
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
222
+ "scr_dir1_threshold_2": 0.010752867351179899,
223
+ "scr_metric_threshold_2": 0.019841248253548664,
224
+ "scr_dir2_threshold_2": 0.019841248253548664,
225
+ "scr_dir1_threshold_5": -0.00537627344809254,
226
+ "scr_metric_threshold_5": 0.039682733033433705,
227
+ "scr_dir2_threshold_5": 0.039682733033433705,
228
+ "scr_dir1_threshold_10": 0.037634555046637416,
229
+ "scr_metric_threshold_10": 0.07539702719508858,
230
+ "scr_dir2_threshold_10": 0.07539702719508858,
231
+ "scr_dir1_threshold_20": 0.0591399692940024,
232
+ "scr_metric_threshold_20": 0.09920647779407969,
233
+ "scr_dir2_threshold_20": 0.09920647779407969,
234
+ "scr_dir1_threshold_50": 0.10215047733373753,
235
+ "scr_metric_threshold_50": 0.10317468013952215,
236
+ "scr_dir2_threshold_50": 0.10317468013952215,
237
+ "scr_dir1_threshold_100": 0.1075270712368249,
238
+ "scr_metric_threshold_100": 0.16269842490016814,
239
+ "scr_dir2_threshold_100": 0.16269842490016814,
240
+ "scr_dir1_threshold_500": 0.16129044662773992,
241
+ "scr_metric_threshold_500": 0.3253968498003363,
242
+ "scr_dir2_threshold_500": 0.3253968498003363
243
+ },
244
+ {
245
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
246
+ "scr_dir1_threshold_2": 0.08208982775592522,
247
+ "scr_metric_threshold_2": 0.01639342660450326,
248
+ "scr_dir2_threshold_2": 0.01639342660450326,
249
+ "scr_dir1_threshold_5": 0.13432840468172233,
250
+ "scr_metric_threshold_5": 0.05245912147445863,
251
+ "scr_dir2_threshold_5": 0.05245912147445863,
252
+ "scr_dir1_threshold_10": 0.14179143979463976,
253
+ "scr_metric_threshold_10": 0.11147553542069445,
254
+ "scr_dir2_threshold_10": 0.11147553542069445,
255
+ "scr_dir1_threshold_20": 0.21641823243764755,
256
+ "scr_metric_threshold_20": 0.209836095047714,
257
+ "scr_dir2_threshold_20": 0.209836095047714,
258
+ "scr_dir1_threshold_50": 0.3507466371193699,
259
+ "scr_metric_threshold_50": 0.3639344223850808,
260
+ "scr_dir2_threshold_50": 0.3639344223850808,
261
+ "scr_dir1_threshold_100": 0.27611939966559124,
262
+ "scr_metric_threshold_100": 0.45245904330443454,
263
+ "scr_dir2_threshold_100": 0.45245904330443454,
264
+ "scr_dir1_threshold_500": 0.2388060033440874,
265
+ "scr_metric_threshold_500": 0.6098362123027502,
266
+ "scr_dir2_threshold_500": 0.6098362123027502
267
+ },
268
+ {
269
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
270
+ "scr_dir1_threshold_2": 0.042780827070065724,
271
+ "scr_metric_threshold_2": 0.08474591253677302,
272
+ "scr_dir2_threshold_2": 0.08474591253677302,
273
+ "scr_dir1_threshold_5": 0.09090917783855113,
274
+ "scr_metric_threshold_5": 0.08898314502311447,
275
+ "scr_dir2_threshold_5": 0.08898314502311447,
276
+ "scr_dir1_threshold_10": 0.10695206767516431,
277
+ "scr_metric_threshold_10": 0.10169509504412763,
278
+ "scr_dir2_threshold_10": 0.10169509504412763,
279
+ "scr_dir1_threshold_20": 0.11229959137358399,
280
+ "scr_metric_threshold_20": 0.12711874252416513,
281
+ "scr_dir2_threshold_20": 0.12711874252416513,
282
+ "scr_dir1_threshold_50": 0.12834248121019717,
283
+ "scr_metric_threshold_50": 0.16949157251155725,
284
+ "scr_dir2_threshold_50": 0.16949157251155725,
285
+ "scr_dir1_threshold_100": 0.21390381660897448,
286
+ "scr_metric_threshold_100": 0.19067798750525328,
287
+ "scr_dir2_threshold_100": 0.19067798750525328,
288
+ "scr_dir1_threshold_500": 0.27807505721407305,
289
+ "scr_metric_threshold_500": 0.18644075501891183,
290
+ "scr_dir2_threshold_500": 0.18644075501891183
291
+ }
292
+ ],
293
+ "sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
294
+ "sae_lens_id": "layer_19/width_1m/average_l0_16",
295
+ "sae_lens_release_id": "gemma-scope-2b-pt-res",
296
+ "sae_lens_version": "4.4.5",
297
+ "eval_result_unstructured": null
298
+ }
gemma_scope/scr_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_18_eval_results.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_type_id": "scr",
3
+ "eval_config": {
4
+ "random_seed": 42,
5
+ "dataset_names": [
6
+ "LabHC/bias_in_bios_class_set1",
7
+ "canrager/amazon_reviews_mcauley_1and5"
8
+ ],
9
+ "perform_scr": true,
10
+ "early_stopping_patience": 20,
11
+ "train_set_size": 4000,
12
+ "test_set_size": 1000,
13
+ "context_length": 128,
14
+ "probe_train_batch_size": 16,
15
+ "probe_test_batch_size": 500,
16
+ "probe_epochs": 20,
17
+ "probe_lr": 0.001,
18
+ "probe_l1_penalty": 0.001,
19
+ "sae_batch_size": 5,
20
+ "llm_batch_size": 32,
21
+ "llm_dtype": "bfloat16",
22
+ "lower_vram_usage": true,
23
+ "model_name": "gemma-2-2b",
24
+ "n_values": [
25
+ 2,
26
+ 5,
27
+ 10,
28
+ 20,
29
+ 50,
30
+ 100,
31
+ 500
32
+ ],
33
+ "column1_vals_lookup": {
34
+ "LabHC/bias_in_bios_class_set1": [
35
+ [
36
+ "professor",
37
+ "nurse"
38
+ ],
39
+ [
40
+ "architect",
41
+ "journalist"
42
+ ],
43
+ [
44
+ "surgeon",
45
+ "psychologist"
46
+ ],
47
+ [
48
+ "attorney",
49
+ "teacher"
50
+ ]
51
+ ],
52
+ "canrager/amazon_reviews_mcauley_1and5": [
53
+ [
54
+ "Books",
55
+ "CDs_and_Vinyl"
56
+ ],
57
+ [
58
+ "Software",
59
+ "Electronics"
60
+ ],
61
+ [
62
+ "Pet_Supplies",
63
+ "Office_Products"
64
+ ],
65
+ [
66
+ "Industrial_and_Scientific",
67
+ "Toys_and_Games"
68
+ ]
69
+ ]
70
+ }
71
+ },
72
+ "eval_id": "2a59aca2-c2bb-4c84-8c8b-4ab2de8b7019",
73
+ "datetime_epoch_millis": 1732761952025,
74
+ "eval_result_metrics": {
75
+ "scr_metrics": {
76
+ "scr_dir1_threshold_2": 0.07190180636775037,
77
+ "scr_metric_threshold_2": 0.047403187304846,
78
+ "scr_dir2_threshold_2": 0.047403187304846,
79
+ "scr_dir1_threshold_5": 0.10590597762225093,
80
+ "scr_metric_threshold_5": 0.07198432795760676,
81
+ "scr_dir2_threshold_5": 0.07198432795760676,
82
+ "scr_dir1_threshold_10": 0.1541878575146368,
83
+ "scr_metric_threshold_10": 0.11080595227866616,
84
+ "scr_dir2_threshold_10": 0.11080595227866616,
85
+ "scr_dir1_threshold_20": 0.20802151886686052,
86
+ "scr_metric_threshold_20": 0.16538215325135666,
87
+ "scr_dir2_threshold_20": 0.16538215325135666,
88
+ "scr_dir1_threshold_50": 0.19118555610992977,
89
+ "scr_metric_threshold_50": 0.23339004770166508,
90
+ "scr_dir2_threshold_50": 0.23339004770166508,
91
+ "scr_dir1_threshold_100": 0.036211319372284625,
92
+ "scr_metric_threshold_100": 0.3122091683712367,
93
+ "scr_dir2_threshold_100": 0.3122091683712367,
94
+ "scr_dir1_threshold_500": 0.023250959650155514,
95
+ "scr_metric_threshold_500": 0.43515112181516163,
96
+ "scr_dir2_threshold_500": 0.43515112181516163
97
+ }
98
+ },
99
+ "eval_result_details": [
100
+ {
101
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
102
+ "scr_dir1_threshold_2": 0.1323529540667449,
103
+ "scr_metric_threshold_2": 0.055555589003726494,
104
+ "scr_dir2_threshold_2": 0.055555589003726494,
105
+ "scr_dir1_threshold_5": 0.1764708976018779,
106
+ "scr_metric_threshold_5": 0.08080808840993783,
107
+ "scr_dir2_threshold_5": 0.08080808840993783,
108
+ "scr_dir1_threshold_10": 0.2352945301358372,
109
+ "scr_metric_threshold_10": 0.08585867860124165,
110
+ "scr_dir2_threshold_10": 0.08585867860124165,
111
+ "scr_dir1_threshold_20": 0.27941247367097016,
112
+ "scr_metric_threshold_20": 0.08838389843850893,
113
+ "scr_dir2_threshold_20": 0.08838389843850893,
114
+ "scr_dir1_threshold_50": 0.029412254536306668,
115
+ "scr_metric_threshold_50": 0.12373742771055866,
116
+ "scr_dir2_threshold_50": 0.12373742771055866,
117
+ "scr_dir1_threshold_100": -0.8970584239309077,
118
+ "scr_metric_threshold_100": 0.13131323773912976,
119
+ "scr_dir2_threshold_100": 0.13131323773912976,
120
+ "scr_dir1_threshold_500": -0.867646169394601,
121
+ "scr_metric_threshold_500": 0.23484860571801164,
122
+ "scr_dir2_threshold_500": 0.23484860571801164
123
+ },
124
+ {
125
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
126
+ "scr_dir1_threshold_2": 0.11607154261093916,
127
+ "scr_metric_threshold_2": 0.11661811987864983,
128
+ "scr_dir2_threshold_2": 0.11661811987864983,
129
+ "scr_dir1_threshold_5": 0.16964281912968696,
130
+ "scr_metric_threshold_5": 0.1603498496677259,
131
+ "scr_dir2_threshold_5": 0.1603498496677259,
132
+ "scr_dir1_threshold_10": 0.21428563825937388,
133
+ "scr_metric_threshold_10": 0.19533533776365494,
134
+ "scr_dir2_threshold_10": 0.19533533776365494,
135
+ "scr_dir1_threshold_20": 0.27678590435156525,
136
+ "scr_metric_threshold_20": 0.2448978953481623,
137
+ "scr_dir2_threshold_20": 0.2448978953481623,
138
+ "scr_dir1_threshold_50": 0.32142872348125223,
139
+ "scr_metric_threshold_50": 0.30320704217471073,
140
+ "scr_dir2_threshold_50": 0.30320704217471073,
141
+ "scr_dir1_threshold_100": 0.23214308522187832,
142
+ "scr_metric_threshold_100": 0.37900867238755315,
143
+ "scr_dir2_threshold_100": 0.37900867238755315,
144
+ "scr_dir1_threshold_500": 0.18750026609219136,
145
+ "scr_metric_threshold_500": 0.43148681764422325,
146
+ "scr_dir2_threshold_500": 0.43148681764422325
147
+ },
148
+ {
149
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
150
+ "scr_dir1_threshold_2": 0.09999988079073888,
151
+ "scr_metric_threshold_2": 0.02427182148466779,
152
+ "scr_dir2_threshold_2": 0.02427182148466779,
153
+ "scr_dir1_threshold_5": 0.08000014305111336,
154
+ "scr_metric_threshold_5": 0.029126127913014256,
155
+ "scr_dir2_threshold_5": 0.029126127913014256,
156
+ "scr_dir1_threshold_10": 0.18000002384185224,
157
+ "scr_metric_threshold_10": 0.04368933654098912,
158
+ "scr_dir2_threshold_10": 0.04368933654098912,
159
+ "scr_dir1_threshold_20": 0.2799999046325911,
160
+ "scr_metric_threshold_20": 0.07281546445400337,
161
+ "scr_dir2_threshold_20": 0.07281546445400337,
162
+ "scr_dir1_threshold_50": 0.31999938011184215,
163
+ "scr_metric_threshold_50": 0.16747574184996902,
164
+ "scr_dir2_threshold_50": 0.16747574184996902,
165
+ "scr_dir1_threshold_100": 0.1199996185303644,
166
+ "scr_metric_threshold_100": 0.24271835951814563,
167
+ "scr_dir2_threshold_100": 0.24271835951814563,
168
+ "scr_dir1_threshold_500": -0.040000667571862325,
169
+ "scr_metric_threshold_500": 0.3155339686436167,
170
+ "scr_dir2_threshold_500": 0.3155339686436167
171
+ },
172
+ {
173
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
174
+ "scr_dir1_threshold_2": 0.0833331828165188,
175
+ "scr_metric_threshold_2": 0.01519755131841285,
176
+ "scr_dir2_threshold_2": 0.01519755131841285,
177
+ "scr_dir1_threshold_5": 0.14393938025604716,
178
+ "scr_metric_threshold_5": 0.021276680547239392,
179
+ "scr_dir2_threshold_5": 0.021276680547239392,
180
+ "scr_dir1_threshold_10": 0.19696946435280177,
181
+ "scr_metric_threshold_10": 0.08814601504971734,
182
+ "scr_dir2_threshold_10": 0.08814601504971734,
183
+ "scr_dir1_threshold_20": 0.2651513235846603,
184
+ "scr_metric_threshold_20": 0.16717324567164368,
185
+ "scr_dir2_threshold_20": 0.16717324567164368,
186
+ "scr_dir1_threshold_50": 0.21212123948790568,
187
+ "scr_metric_threshold_50": 0.27051681203977385,
188
+ "scr_dir2_threshold_50": 0.27051681203977385,
189
+ "scr_dir1_threshold_100": 0.19696946435280177,
190
+ "scr_metric_threshold_100": 0.38601827452735243,
191
+ "scr_dir2_threshold_100": 0.38601827452735243,
192
+ "scr_dir1_threshold_500": 0.12878760512094328,
193
+ "scr_metric_threshold_500": 0.6079027774129233,
194
+ "scr_dir2_threshold_500": 0.6079027774129233
195
+ },
196
+ {
197
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
198
+ "scr_dir1_threshold_2": 0.005747051625836349,
199
+ "scr_metric_threshold_2": 0.019531232719599476,
200
+ "scr_dir2_threshold_2": 0.019531232719599476,
201
+ "scr_dir1_threshold_5": 0.022988549058726323,
202
+ "scr_metric_threshold_5": 0.07421873090060994,
203
+ "scr_dir2_threshold_5": 0.07421873090060994,
204
+ "scr_dir1_threshold_10": 0.045977098117452646,
205
+ "scr_metric_threshold_10": 0.13281242905940838,
206
+ "scr_dir2_threshold_10": 0.13281242905940838,
207
+ "scr_dir1_threshold_20": 0.05747120136912535,
208
+ "scr_metric_threshold_20": 0.20312495998223037,
209
+ "scr_dir2_threshold_20": 0.20312495998223037,
210
+ "scr_dir1_threshold_50": 0.07471269880201532,
211
+ "scr_metric_threshold_50": 0.28906252364686386,
212
+ "scr_dir2_threshold_50": 0.28906252364686386,
213
+ "scr_dir1_threshold_100": 0.08045975042785167,
214
+ "scr_metric_threshold_100": 0.4140625527506963,
215
+ "scr_dir2_threshold_100": 0.4140625527506963,
216
+ "scr_dir1_threshold_500": -0.005747051625836349,
217
+ "scr_metric_threshold_500": 0.6054686799689032,
218
+ "scr_dir2_threshold_500": 0.6054686799689032
219
+ },
220
+ {
221
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
222
+ "scr_dir1_threshold_2": 0.00537659390308736,
223
+ "scr_metric_threshold_2": 0.03571429416165487,
224
+ "scr_dir2_threshold_2": 0.03571429416165487,
225
+ "scr_dir1_threshold_5": 0.02150541424736498,
226
+ "scr_metric_threshold_5": 0.04761913772431862,
227
+ "scr_dir2_threshold_5": 0.04761913772431862,
228
+ "scr_dir1_threshold_10": 0.03225828159854488,
229
+ "scr_metric_threshold_10": 0.06746038597786728,
230
+ "scr_dir2_threshold_10": 0.06746038597786728,
231
+ "scr_dir1_threshold_20": 0.07526878963828001,
232
+ "scr_metric_threshold_20": 0.07936522954053103,
233
+ "scr_dir2_threshold_20": 0.07936522954053103,
234
+ "scr_dir1_threshold_50": 0.1075270712368249,
235
+ "scr_metric_threshold_50": 0.11507952370218591,
236
+ "scr_dir2_threshold_50": 0.11507952370218591,
237
+ "scr_dir1_threshold_100": 0.13978503238037496,
238
+ "scr_metric_threshold_100": 0.18650811202549564,
239
+ "scr_dir2_threshold_100": 0.18650811202549564,
240
+ "scr_dir1_threshold_500": 0.22580636891484004,
241
+ "scr_metric_threshold_500": 0.33730169336300003,
242
+ "scr_dir2_threshold_500": 0.33730169336300003
243
+ },
244
+ {
245
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
246
+ "scr_dir1_threshold_2": 0.08955241805807182,
247
+ "scr_metric_threshold_2": 0.036065694869955374,
248
+ "scr_dir2_threshold_2": 0.036065694869955374,
249
+ "scr_dir1_threshold_5": 0.14179143979463976,
250
+ "scr_metric_threshold_5": 0.08196732844757654,
251
+ "scr_dir2_threshold_5": 0.08196732844757654,
252
+ "scr_dir1_threshold_10": 0.21641823243764755,
253
+ "scr_metric_threshold_10": 0.16721310770598144,
254
+ "scr_dir2_threshold_10": 0.16721310770598144,
255
+ "scr_dir1_threshold_20": 0.29104502508065533,
256
+ "scr_metric_threshold_20": 0.314754142571571,
257
+ "scr_dir2_threshold_20": 0.314754142571571,
258
+ "scr_dir1_threshold_50": 0.2985076153828019,
259
+ "scr_metric_threshold_50": 0.4327869704640427,
260
+ "scr_dir2_threshold_50": 0.4327869704640427,
261
+ "scr_dir1_threshold_100": 0.24626903845700482,
262
+ "scr_metric_threshold_100": 0.5377050179878997,
263
+ "scr_dir2_threshold_100": 0.5377050179878997,
264
+ "scr_dir1_threshold_500": 0.30597020568494854,
265
+ "scr_metric_threshold_500": 0.639344419275868,
266
+ "scr_dir2_threshold_500": 0.639344419275868
267
+ },
268
+ {
269
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
270
+ "scr_dir1_threshold_2": 0.042780827070065724,
271
+ "scr_metric_threshold_2": 0.07627119500210132,
272
+ "scr_dir2_threshold_2": 0.07627119500210132,
273
+ "scr_dir1_threshold_5": 0.09090917783855113,
274
+ "scr_metric_threshold_5": 0.08050868005043157,
275
+ "scr_dir2_threshold_5": 0.08050868005043157,
276
+ "scr_dir1_threshold_10": 0.11229959137358399,
277
+ "scr_metric_threshold_10": 0.10593232753046908,
278
+ "scr_dir2_threshold_10": 0.10593232753046908,
279
+ "scr_dir1_threshold_20": 0.13903752860703653,
280
+ "scr_metric_threshold_20": 0.15254239000420264,
281
+ "scr_dir2_threshold_20": 0.15254239000420264,
282
+ "scr_dir1_threshold_50": 0.16577546584048908,
283
+ "scr_metric_threshold_50": 0.1652543400252158,
284
+ "scr_dir2_threshold_50": 0.1652543400252158,
285
+ "scr_dir1_threshold_100": 0.17112298953890875,
286
+ "scr_metric_threshold_100": 0.22033912003362105,
287
+ "scr_dir2_threshold_100": 0.22033912003362105,
288
+ "scr_dir1_threshold_500": 0.2513371199806205,
289
+ "scr_metric_threshold_500": 0.3093220124947467,
290
+ "scr_dir2_threshold_500": 0.3093220124947467
291
+ }
292
+ ],
293
+ "sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
294
+ "sae_lens_id": "layer_19/width_1m/average_l0_18",
295
+ "sae_lens_release_id": "gemma-scope-2b-pt-res",
296
+ "sae_lens_version": "4.4.5",
297
+ "eval_result_unstructured": null
298
+ }
gemma_scope/scr_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_29_eval_results.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_type_id": "scr",
3
+ "eval_config": {
4
+ "random_seed": 42,
5
+ "dataset_names": [
6
+ "LabHC/bias_in_bios_class_set1",
7
+ "canrager/amazon_reviews_mcauley_1and5"
8
+ ],
9
+ "perform_scr": true,
10
+ "early_stopping_patience": 20,
11
+ "train_set_size": 4000,
12
+ "test_set_size": 1000,
13
+ "context_length": 128,
14
+ "probe_train_batch_size": 16,
15
+ "probe_test_batch_size": 500,
16
+ "probe_epochs": 20,
17
+ "probe_lr": 0.001,
18
+ "probe_l1_penalty": 0.001,
19
+ "sae_batch_size": 5,
20
+ "llm_batch_size": 32,
21
+ "llm_dtype": "bfloat16",
22
+ "lower_vram_usage": true,
23
+ "model_name": "gemma-2-2b",
24
+ "n_values": [
25
+ 2,
26
+ 5,
27
+ 10,
28
+ 20,
29
+ 50,
30
+ 100,
31
+ 500
32
+ ],
33
+ "column1_vals_lookup": {
34
+ "LabHC/bias_in_bios_class_set1": [
35
+ [
36
+ "professor",
37
+ "nurse"
38
+ ],
39
+ [
40
+ "architect",
41
+ "journalist"
42
+ ],
43
+ [
44
+ "surgeon",
45
+ "psychologist"
46
+ ],
47
+ [
48
+ "attorney",
49
+ "teacher"
50
+ ]
51
+ ],
52
+ "canrager/amazon_reviews_mcauley_1and5": [
53
+ [
54
+ "Books",
55
+ "CDs_and_Vinyl"
56
+ ],
57
+ [
58
+ "Software",
59
+ "Electronics"
60
+ ],
61
+ [
62
+ "Pet_Supplies",
63
+ "Office_Products"
64
+ ],
65
+ [
66
+ "Industrial_and_Scientific",
67
+ "Toys_and_Games"
68
+ ]
69
+ ]
70
+ }
71
+ },
72
+ "eval_id": "0e8ec9a3-c6ea-4207-b4a9-46007cee4bfc",
73
+ "datetime_epoch_millis": 1732775210003,
74
+ "eval_result_metrics": {
75
+ "scr_metrics": {
76
+ "scr_dir1_threshold_2": 0.12643844925292913,
77
+ "scr_metric_threshold_2": 0.09377844569320483,
78
+ "scr_dir2_threshold_2": 0.09377844569320483,
79
+ "scr_dir1_threshold_5": 0.09083801465952376,
80
+ "scr_metric_threshold_5": 0.13084594762185095,
81
+ "scr_dir2_threshold_5": 0.13084594762185095,
82
+ "scr_dir1_threshold_10": 0.13487408593607592,
83
+ "scr_metric_threshold_10": 0.17308413895460062,
84
+ "scr_dir2_threshold_10": 0.17308413895460062,
85
+ "scr_dir1_threshold_20": 0.11530473055678761,
86
+ "scr_metric_threshold_20": 0.2176993274524422,
87
+ "scr_dir2_threshold_20": 0.2176993274524422,
88
+ "scr_dir1_threshold_50": 0.16190028497156714,
89
+ "scr_metric_threshold_50": 0.2937432254936019,
90
+ "scr_dir2_threshold_50": 0.2937432254936019,
91
+ "scr_dir1_threshold_100": 0.16580359382530607,
92
+ "scr_metric_threshold_100": 0.3760448666873901,
93
+ "scr_dir2_threshold_100": 0.3760448666873901,
94
+ "scr_dir1_threshold_500": -0.05063392903970242,
95
+ "scr_metric_threshold_500": 0.4739206866573681,
96
+ "scr_dir2_threshold_500": 0.4739206866573681
97
+ }
98
+ },
99
+ "eval_result_details": [
100
+ {
101
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
102
+ "scr_dir1_threshold_2": 0.2352945301358372,
103
+ "scr_metric_threshold_2": 0.06565661886956488,
104
+ "scr_dir2_threshold_2": 0.06565661886956488,
105
+ "scr_dir1_threshold_5": 0.2647059081334898,
106
+ "scr_metric_threshold_5": 0.047979929491924624,
107
+ "scr_dir2_threshold_5": 0.047979929491924624,
108
+ "scr_dir1_threshold_10": 0.2647059081334898,
109
+ "scr_metric_threshold_10": 0.06818183870683217,
110
+ "scr_dir2_threshold_10": 0.06818183870683217,
111
+ "scr_dir1_threshold_20": -0.014705688998826315,
112
+ "scr_metric_threshold_20": 0.08838389843850893,
113
+ "scr_dir2_threshold_20": 0.08838389843850893,
114
+ "scr_dir1_threshold_50": 0.029412254536306668,
115
+ "scr_metric_threshold_50": 0.11868698803602408,
116
+ "scr_dir2_threshold_50": 0.11868698803602408,
117
+ "scr_dir1_threshold_100": 0.044117943535132986,
118
+ "scr_metric_threshold_100": 0.1641415471739122,
119
+ "scr_dir2_threshold_100": 0.1641415471739122,
120
+ "scr_dir1_threshold_500": -1.1911757100620501,
121
+ "scr_metric_threshold_500": 0.3434344133714281,
122
+ "scr_dir2_threshold_500": 0.3434344133714281
123
+ },
124
+ {
125
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
126
+ "scr_dir1_threshold_2": 0.21428563825937388,
127
+ "scr_metric_threshold_2": 0.15451902187229455,
128
+ "scr_dir2_threshold_2": 0.15451902187229455,
129
+ "scr_dir1_threshold_5": 0.22321462783281748,
130
+ "scr_metric_threshold_5": 0.21574340882211163,
131
+ "scr_dir2_threshold_5": 0.21574340882211163,
132
+ "scr_dir1_threshold_10": 0.29464281912968693,
133
+ "scr_metric_threshold_10": 0.27988338344409136,
134
+ "scr_dir2_threshold_10": 0.27988338344409136,
135
+ "scr_dir1_threshold_20": 0.33035718087031307,
136
+ "scr_metric_threshold_20": 0.31778428543773607,
137
+ "scr_dir2_threshold_20": 0.31778428543773607,
138
+ "scr_dir1_threshold_50": 0.38392898957344357,
139
+ "scr_metric_threshold_50": 0.39941691722045686,
140
+ "scr_dir2_threshold_50": 0.39941691722045686,
141
+ "scr_dir1_threshold_100": 0.24107154261093916,
142
+ "scr_metric_threshold_100": 0.48979596447077167,
143
+ "scr_dir2_threshold_100": 0.48979596447077167,
144
+ "scr_dir1_threshold_500": 0.026785904351565262,
145
+ "scr_metric_threshold_500": 0.46064147794472093,
146
+ "scr_dir2_threshold_500": 0.46064147794472093
147
+ },
148
+ {
149
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
150
+ "scr_dir1_threshold_2": 0.31999938011184215,
151
+ "scr_metric_threshold_2": 0.026698974698841024,
152
+ "scr_dir2_threshold_2": 0.026698974698841024,
153
+ "scr_dir1_threshold_5": 0.2599989748003543,
154
+ "scr_metric_threshold_5": 0.036407732227001686,
155
+ "scr_dir2_threshold_5": 0.036407732227001686,
156
+ "scr_dir1_threshold_10": 0.31999938011184215,
157
+ "scr_metric_threshold_10": 0.04854364296933558,
158
+ "scr_dir2_threshold_10": 0.04854364296933558,
159
+ "scr_dir1_threshold_20": 0.2599989748003543,
160
+ "scr_metric_threshold_20": 0.0800970687679908,
161
+ "scr_dir2_threshold_20": 0.0800970687679908,
162
+ "scr_dir1_threshold_50": 0.36000004768370447,
163
+ "scr_metric_threshold_50": 0.20388347407697072,
164
+ "scr_dir2_threshold_50": 0.20388347407697072,
165
+ "scr_dir1_threshold_100": 0.3999995231629555,
166
+ "scr_metric_threshold_100": 0.30097090468710963,
167
+ "scr_dir2_threshold_100": 0.30097090468710963,
168
+ "scr_dir1_threshold_500": -0.18000002384185224,
169
+ "scr_metric_threshold_500": 0.24271835951814563,
170
+ "scr_dir2_threshold_500": 0.24271835951814563
171
+ },
172
+ {
173
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
174
+ "scr_dir1_threshold_2": 0.1666668171834812,
175
+ "scr_metric_threshold_2": 0.1398176170646801,
176
+ "scr_dir2_threshold_2": 0.1398176170646801,
177
+ "scr_dir1_threshold_5": 0.22727256307256596,
178
+ "scr_metric_threshold_5": 0.20060800350743382,
179
+ "scr_dir2_threshold_5": 0.20060800350743382,
180
+ "scr_dir1_threshold_10": 0.24242433820766984,
181
+ "scr_metric_threshold_10": 0.24316118343281026,
182
+ "scr_dir2_threshold_10": 0.24316118343281026,
183
+ "scr_dir1_threshold_20": 0.2651513235846603,
184
+ "scr_metric_threshold_20": 0.27659576009949804,
185
+ "scr_dir2_threshold_20": 0.27659576009949804,
186
+ "scr_dir1_threshold_50": 0.20454557769557552,
187
+ "scr_metric_threshold_50": 0.3556231718905268,
188
+ "scr_dir2_threshold_50": 0.3556231718905268,
189
+ "scr_dir1_threshold_100": 0.22727256307256596,
190
+ "scr_metric_threshold_100": 0.44072953174127966,
191
+ "scr_dir2_threshold_100": 0.44072953174127966,
192
+ "scr_dir1_threshold_500": 0.20454557769557552,
193
+ "scr_metric_threshold_500": 0.5440729169403075,
194
+ "scr_dir2_threshold_500": 0.5440729169403075
195
+ },
196
+ {
197
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
198
+ "scr_dir1_threshold_2": 0.005747051625836349,
199
+ "scr_metric_threshold_2": 0.10937499636202094,
200
+ "scr_dir2_threshold_2": 0.10937499636202094,
201
+ "scr_dir1_threshold_5": 0.022988549058726323,
202
+ "scr_metric_threshold_5": 0.1640624945430314,
203
+ "scr_dir2_threshold_5": 0.1640624945430314,
204
+ "scr_dir1_threshold_10": 0.045977098117452646,
205
+ "scr_metric_threshold_10": 0.1992187600044424,
206
+ "scr_dir2_threshold_10": 0.1992187600044424,
207
+ "scr_dir1_threshold_20": 0.051724149743289,
208
+ "scr_metric_threshold_20": 0.24609362539921728,
209
+ "scr_dir2_threshold_20": 0.24609362539921728,
210
+ "scr_dir1_threshold_50": 0.05747120136912535,
211
+ "scr_metric_threshold_50": 0.38671868724486125,
212
+ "scr_dir2_threshold_50": 0.38671868724486125,
213
+ "scr_dir1_threshold_100": 0.0632182529949617,
214
+ "scr_metric_threshold_100": 0.5195311163042696,
215
+ "scr_dir2_threshold_100": 0.5195311163042696,
216
+ "scr_dir1_threshold_500": 0.051724149743289,
217
+ "scr_metric_threshold_500": 0.7031250763975602,
218
+ "scr_dir2_threshold_500": 0.7031250763975602
219
+ },
220
+ {
221
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
222
+ "scr_dir1_threshold_2": 0.02150541424736498,
223
+ "scr_metric_threshold_2": 0.027777889470769958,
224
+ "scr_dir2_threshold_2": 0.027777889470769958,
225
+ "scr_dir1_threshold_5": 0.03225828159854488,
226
+ "scr_metric_threshold_5": 0.05555554241520354,
227
+ "scr_dir2_threshold_5": 0.05555554241520354,
228
+ "scr_dir1_threshold_10": 0.0591399692940024,
229
+ "scr_metric_threshold_10": 0.08333343188597349,
230
+ "scr_dir2_threshold_10": 0.08333343188597349,
231
+ "scr_dir1_threshold_20": 0.08602165698945992,
232
+ "scr_metric_threshold_20": 0.11507952370218591,
233
+ "scr_dir2_threshold_20": 0.11507952370218591,
234
+ "scr_dir1_threshold_50": 0.15053757927656003,
235
+ "scr_metric_threshold_50": 0.16666686377194698,
236
+ "scr_dir2_threshold_50": 0.16666686377194698,
237
+ "scr_dir1_threshold_100": 0.17204299352392502,
238
+ "scr_metric_threshold_100": 0.2738095097305752,
239
+ "scr_dir2_threshold_100": 0.2738095097305752,
240
+ "scr_dir1_threshold_500": 0.21505382201865497,
241
+ "scr_metric_threshold_500": 0.4682540264469558,
242
+ "scr_dir2_threshold_500": 0.4682540264469558
243
+ },
244
+ {
245
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
246
+ "scr_dir1_threshold_2": 0.037313396321503885,
247
+ "scr_metric_threshold_2": 0.19672131467909934,
248
+ "scr_dir2_threshold_2": 0.19672131467909934,
249
+ "scr_dir1_threshold_5": -0.335820566893535,
250
+ "scr_metric_threshold_5": 0.24590178991766937,
251
+ "scr_dir2_threshold_5": 0.24590178991766937,
252
+ "scr_dir1_threshold_10": -0.2388060033440874,
253
+ "scr_metric_threshold_10": 0.3606557761491922,
254
+ "scr_dir2_threshold_10": 0.3606557761491922,
255
+ "scr_dir1_threshold_20": -0.1791043913053728,
256
+ "scr_metric_threshold_20": 0.48196725027755244,
257
+ "scr_dir2_threshold_20": 0.48196725027755244,
258
+ "scr_dir1_threshold_50": -0.06716420234086116,
259
+ "scr_metric_threshold_50": 0.5409836642237883,
260
+ "scr_dir2_threshold_50": 0.5409836642237883,
261
+ "scr_dir1_threshold_100": -0.02985036120858645,
262
+ "scr_metric_threshold_100": 0.6032787244059127,
263
+ "scr_dir2_threshold_100": 0.6032787244059127,
264
+ "scr_dir1_threshold_500": 0.14179143979463976,
265
+ "scr_metric_threshold_500": 0.7409836251387762,
266
+ "scr_dir2_threshold_500": 0.7409836251387762
267
+ },
268
+ {
269
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
270
+ "scr_dir1_threshold_2": 0.010695366138193501,
271
+ "scr_metric_threshold_2": 0.029661132528367763,
272
+ "scr_dir2_threshold_2": 0.029661132528367763,
273
+ "scr_dir1_threshold_5": 0.032085779673226365,
274
+ "scr_metric_threshold_5": 0.08050868005043157,
275
+ "scr_dir2_threshold_5": 0.08050868005043157,
276
+ "scr_dir1_threshold_10": 0.09090917783855113,
277
+ "scr_metric_threshold_10": 0.10169509504412763,
278
+ "scr_dir2_threshold_10": 0.10169509504412763,
279
+ "scr_dir1_threshold_20": 0.12299463877042335,
280
+ "scr_metric_threshold_20": 0.13559320749684803,
281
+ "scr_dir2_threshold_20": 0.13559320749684803,
282
+ "scr_dir1_threshold_50": 0.17647083197868257,
283
+ "scr_metric_threshold_50": 0.17796603748424014,
284
+ "scr_dir2_threshold_50": 0.17796603748424014,
285
+ "scr_dir1_threshold_100": 0.2085562929105548,
286
+ "scr_metric_threshold_100": 0.21610163498529078,
287
+ "scr_dir2_threshold_100": 0.21610163498529078,
288
+ "scr_dir1_threshold_500": 0.3262034079825585,
289
+ "scr_metric_threshold_500": 0.28813559750105067,
290
+ "scr_dir2_threshold_500": 0.28813559750105067
291
+ }
292
+ ],
293
+ "sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
294
+ "sae_lens_id": "layer_19/width_1m/average_l0_29",
295
+ "sae_lens_release_id": "gemma-scope-2b-pt-res",
296
+ "sae_lens_version": "4.4.5",
297
+ "eval_result_unstructured": null
298
+ }
gemma_scope/scr_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_50_eval_results.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_type_id": "scr",
3
+ "eval_config": {
4
+ "random_seed": 42,
5
+ "dataset_names": [
6
+ "LabHC/bias_in_bios_class_set1",
7
+ "canrager/amazon_reviews_mcauley_1and5"
8
+ ],
9
+ "perform_scr": true,
10
+ "early_stopping_patience": 20,
11
+ "train_set_size": 4000,
12
+ "test_set_size": 1000,
13
+ "context_length": 128,
14
+ "probe_train_batch_size": 16,
15
+ "probe_test_batch_size": 500,
16
+ "probe_epochs": 20,
17
+ "probe_lr": 0.001,
18
+ "probe_l1_penalty": 0.001,
19
+ "sae_batch_size": 5,
20
+ "llm_batch_size": 32,
21
+ "llm_dtype": "bfloat16",
22
+ "lower_vram_usage": true,
23
+ "model_name": "gemma-2-2b",
24
+ "n_values": [
25
+ 2,
26
+ 5,
27
+ 10,
28
+ 20,
29
+ 50,
30
+ 100,
31
+ 500
32
+ ],
33
+ "column1_vals_lookup": {
34
+ "LabHC/bias_in_bios_class_set1": [
35
+ [
36
+ "professor",
37
+ "nurse"
38
+ ],
39
+ [
40
+ "architect",
41
+ "journalist"
42
+ ],
43
+ [
44
+ "surgeon",
45
+ "psychologist"
46
+ ],
47
+ [
48
+ "attorney",
49
+ "teacher"
50
+ ]
51
+ ],
52
+ "canrager/amazon_reviews_mcauley_1and5": [
53
+ [
54
+ "Books",
55
+ "CDs_and_Vinyl"
56
+ ],
57
+ [
58
+ "Software",
59
+ "Electronics"
60
+ ],
61
+ [
62
+ "Pet_Supplies",
63
+ "Office_Products"
64
+ ],
65
+ [
66
+ "Industrial_and_Scientific",
67
+ "Toys_and_Games"
68
+ ]
69
+ ]
70
+ }
71
+ },
72
+ "eval_id": "8fa32cd0-b56d-4237-b902-cc2d6ec6cbd6",
73
+ "datetime_epoch_millis": 1732787982072,
74
+ "eval_result_metrics": {
75
+ "scr_metrics": {
76
+ "scr_dir1_threshold_2": 0.08747594459933392,
77
+ "scr_metric_threshold_2": 0.15999384166712416,
78
+ "scr_dir2_threshold_2": 0.15999384166712416,
79
+ "scr_dir1_threshold_5": 0.08873316880165232,
80
+ "scr_metric_threshold_5": 0.20538973434744356,
81
+ "scr_dir2_threshold_5": 0.20538973434744356,
82
+ "scr_dir1_threshold_10": 0.1553595324107292,
83
+ "scr_metric_threshold_10": 0.23933960593557582,
84
+ "scr_dir2_threshold_10": 0.23933960593557582,
85
+ "scr_dir1_threshold_20": 0.20049255011780437,
86
+ "scr_metric_threshold_20": 0.2912222973945863,
87
+ "scr_dir2_threshold_20": 0.2912222973945863,
88
+ "scr_dir1_threshold_50": 0.20123883698141437,
89
+ "scr_metric_threshold_50": 0.36872842139860207,
90
+ "scr_dir2_threshold_50": 0.36872842139860207,
91
+ "scr_dir1_threshold_100": 0.10412677530300897,
92
+ "scr_metric_threshold_100": 0.4383315016334388,
93
+ "scr_dir2_threshold_100": 0.4383315016334388,
94
+ "scr_dir1_threshold_500": -0.12485906451240938,
95
+ "scr_metric_threshold_500": 0.5168769827267571,
96
+ "scr_dir2_threshold_500": 0.5168769827267571
97
+ }
98
+ },
99
+ "eval_result_details": [
100
+ {
101
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
102
+ "scr_dir1_threshold_2": 0.20588315213818456,
103
+ "scr_metric_threshold_2": 0.06565661886956488,
104
+ "scr_dir2_threshold_2": 0.06565661886956488,
105
+ "scr_dir1_threshold_5": 0.3088238516686228,
106
+ "scr_metric_threshold_5": 0.0606061791950303,
107
+ "scr_dir2_threshold_5": 0.0606061791950303,
108
+ "scr_dir1_threshold_10": 0.3088238516686228,
109
+ "scr_metric_threshold_10": 0.06565661886956488,
110
+ "scr_dir2_threshold_10": 0.06565661886956488,
111
+ "scr_dir1_threshold_20": 0.4117654277377151,
112
+ "scr_metric_threshold_20": 0.1010101481416146,
113
+ "scr_dir2_threshold_20": 0.1010101481416146,
114
+ "scr_dir1_threshold_50": 0.3529417952037558,
115
+ "scr_metric_threshold_50": 0.11363639784472028,
116
+ "scr_dir2_threshold_50": 0.11363639784472028,
117
+ "scr_dir1_threshold_100": 0.2500002191346635,
118
+ "scr_metric_threshold_100": 0.13131323773912976,
119
+ "scr_dir2_threshold_100": 0.13131323773912976,
120
+ "scr_dir1_threshold_500": -1.499999561730673,
121
+ "scr_metric_threshold_500": 0.24747485542111733,
122
+ "scr_dir2_threshold_500": 0.24747485542111733
123
+ },
124
+ {
125
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
126
+ "scr_dir1_threshold_2": 0.15178590435156525,
127
+ "scr_metric_threshold_2": 0.1486880203024162,
128
+ "scr_dir2_threshold_2": 0.1486880203024162,
129
+ "scr_dir1_threshold_5": 0.22321462783281748,
130
+ "scr_metric_threshold_5": 0.256559724713472,
131
+ "scr_dir2_threshold_5": 0.256559724713472,
132
+ "scr_dir1_threshold_10": 0.3392856382593739,
133
+ "scr_metric_threshold_10": 0.3352769425984771,
134
+ "scr_dir2_threshold_10": 0.3352769425984771,
135
+ "scr_dir1_threshold_20": 0.3928574469625044,
136
+ "scr_metric_threshold_20": 0.4081633326880509,
137
+ "scr_dir2_threshold_20": 0.4081633326880509,
138
+ "scr_dir1_threshold_50": 0.44642872348125223,
139
+ "scr_metric_threshold_50": 0.478134135105462,
140
+ "scr_dir2_threshold_50": 0.478134135105462,
141
+ "scr_dir1_threshold_100": 0.4285718087031305,
142
+ "scr_metric_threshold_100": 0.5597667668881827,
143
+ "scr_dir2_threshold_100": 0.5597667668881827,
144
+ "scr_dir1_threshold_500": 0.2678574469625044,
145
+ "scr_metric_threshold_500": 0.5247812787922537,
146
+ "scr_dir2_threshold_500": 0.5247812787922537
147
+ },
148
+ {
149
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
150
+ "scr_dir1_threshold_2": 0.09999988079073888,
151
+ "scr_metric_threshold_2": 0.03398057901282845,
152
+ "scr_dir2_threshold_2": 0.03398057901282845,
153
+ "scr_dir1_threshold_5": 0.2999996423722166,
154
+ "scr_metric_threshold_5": 0.05339794939768205,
155
+ "scr_dir2_threshold_5": 0.05339794939768205,
156
+ "scr_dir1_threshold_10": 0.3999995231629555,
157
+ "scr_metric_threshold_10": 0.06553400481148368,
158
+ "scr_dir2_threshold_10": 0.06553400481148368,
159
+ "scr_dir1_threshold_20": 0.419999260902581,
160
+ "scr_metric_threshold_20": 0.10679604346683183,
161
+ "scr_dir2_threshold_20": 0.10679604346683183,
162
+ "scr_dir1_threshold_50": 0.37999978542332996,
163
+ "scr_metric_threshold_50": 0.21116507839095816,
164
+ "scr_dir2_threshold_50": 0.21116507839095816,
165
+ "scr_dir1_threshold_100": 0.36000004768370447,
166
+ "scr_metric_threshold_100": 0.29854360680146863,
167
+ "scr_dir2_threshold_100": 0.29854360680146863,
168
+ "scr_dir1_threshold_500": 0.31999938011184215,
169
+ "scr_metric_threshold_500": 0.402912641725595,
170
+ "scr_dir2_threshold_500": 0.402912641725595
171
+ },
172
+ {
173
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
174
+ "scr_dir1_threshold_2": 0.13636371846371703,
175
+ "scr_metric_threshold_2": 0.1945288742786073,
176
+ "scr_dir2_threshold_2": 0.1945288742786073,
177
+ "scr_dir1_threshold_5": 0.2348486764153397,
178
+ "scr_metric_threshold_5": 0.28571436335818673,
179
+ "scr_dir2_threshold_5": 0.28571436335818673,
180
+ "scr_dir1_threshold_10": 0.34090929615929255,
181
+ "scr_metric_threshold_10": 0.33434649134328737,
182
+ "scr_dir2_threshold_10": 0.33434649134328737,
183
+ "scr_dir1_threshold_20": 0.3560606197439528,
184
+ "scr_metric_threshold_20": 0.3738601972388017,
185
+ "scr_dir2_threshold_20": 0.3738601972388017,
186
+ "scr_dir1_threshold_50": 0.17424247897581135,
187
+ "scr_metric_threshold_50": 0.46808516034824327,
188
+ "scr_dir2_threshold_50": 0.46808516034824327,
189
+ "scr_dir1_threshold_100": 0.1666668171834812,
190
+ "scr_metric_threshold_100": 0.5927052260945105,
191
+ "scr_dir2_threshold_100": 0.5927052260945105,
192
+ "scr_dir1_threshold_500": 0.12878760512094328,
193
+ "scr_metric_threshold_500": 0.6930091372636762,
194
+ "scr_dir2_threshold_500": 0.6930091372636762
195
+ },
196
+ {
197
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
198
+ "scr_dir1_threshold_2": 0.011494103251672699,
199
+ "scr_metric_threshold_2": 0.4453123854036597,
200
+ "scr_dir2_threshold_2": 0.4453123854036597,
201
+ "scr_dir1_threshold_5": 0.051724149743289,
202
+ "scr_metric_threshold_5": 0.5039063163931178,
203
+ "scr_dir2_threshold_5": 0.5039063163931178,
204
+ "scr_dir1_threshold_10": 0.07471269880201532,
205
+ "scr_metric_threshold_10": 0.5273437490905052,
206
+ "scr_dir2_threshold_10": 0.5273437490905052,
207
+ "scr_dir1_threshold_20": 0.10919535111241434,
208
+ "scr_metric_threshold_20": 0.5742188473159398,
209
+ "scr_dir2_threshold_20": 0.5742188473159398,
210
+ "scr_dir1_threshold_50": 0.12643684854530432,
211
+ "scr_metric_threshold_50": 0.6601561781499136,
212
+ "scr_dir2_threshold_50": 0.6601561781499136,
213
+ "scr_dir1_threshold_100": -0.6149424027382507,
214
+ "scr_metric_threshold_100": 0.6914062436335366,
215
+ "scr_dir2_threshold_100": 0.6914062436335366,
216
+ "scr_dir1_threshold_500": -0.5919538536795244,
217
+ "scr_metric_threshold_500": 0.6835938436779607,
218
+ "scr_dir2_threshold_500": 0.6835938436779607
219
+ },
220
+ {
221
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
222
+ "scr_dir1_threshold_2": 0.048387101942822495,
223
+ "scr_metric_threshold_2": 0.04761913772431862,
224
+ "scr_dir2_threshold_2": 0.04761913772431862,
225
+ "scr_dir1_threshold_5": 0.06451624274209494,
226
+ "scr_metric_threshold_5": 0.05952398128698237,
227
+ "scr_dir2_threshold_5": 0.05952398128698237,
228
+ "scr_dir1_threshold_10": 0.08064538354136737,
229
+ "scr_metric_threshold_10": 0.10317468013952215,
230
+ "scr_dir2_threshold_10": 0.10317468013952215,
231
+ "scr_dir1_threshold_20": 0.12365589158110252,
232
+ "scr_metric_threshold_20": 0.13888897430117703,
233
+ "scr_dir2_threshold_20": 0.13888897430117703,
234
+ "scr_dir1_threshold_50": 0.17204299352392502,
235
+ "scr_metric_threshold_50": 0.2341270132234779,
236
+ "scr_dir2_threshold_50": 0.2341270132234779,
237
+ "scr_dir1_threshold_100": 0.1344087589322824,
238
+ "scr_metric_threshold_100": 0.3333334910175576,
239
+ "scr_dir2_threshold_100": 0.3333334910175576,
240
+ "scr_dir1_threshold_500": 0.13978503238037496,
241
+ "scr_metric_threshold_500": 0.5119047252994956,
242
+ "scr_dir2_threshold_500": 0.5119047252994956
243
+ },
244
+ {
245
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
246
+ "scr_dir1_threshold_2": 0.029850806019357277,
247
+ "scr_metric_threshold_2": 0.27213115522983844,
248
+ "scr_dir2_threshold_2": 0.27213115522983844,
249
+ "scr_dir1_threshold_5": -0.4999997775946146,
250
+ "scr_metric_threshold_5": 0.3344262154119629,
251
+ "scr_dir2_threshold_5": 0.3344262154119629,
252
+ "scr_dir1_threshold_10": -0.36567137291289226,
253
+ "scr_metric_threshold_10": 0.3901639831223101,
254
+ "scr_dir2_threshold_10": 0.3901639831223101,
255
+ "scr_dir1_threshold_20": -0.3432836020064524,
256
+ "scr_metric_threshold_20": 0.5081968110147818,
257
+ "scr_dir2_threshold_20": 0.5081968110147818,
258
+ "scr_dir1_threshold_50": -0.18656698160751944,
259
+ "scr_metric_threshold_50": 0.6065573706418013,
260
+ "scr_dir2_threshold_50": 0.6065573706418013,
261
+ "scr_dir1_threshold_100": -0.089551973247301,
262
+ "scr_metric_threshold_100": 0.6622951383521485,
263
+ "scr_dir2_threshold_100": 0.6622951383521485,
264
+ "scr_dir1_threshold_500": -0.089551973247301,
265
+ "scr_metric_threshold_500": 0.7704918321118941,
266
+ "scr_dir2_threshold_500": 0.7704918321118941
267
+ },
268
+ {
269
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
270
+ "scr_dir1_threshold_2": 0.016042889836613183,
271
+ "scr_metric_threshold_2": 0.07203396251575987,
272
+ "scr_dir2_threshold_2": 0.07203396251575987,
273
+ "scr_dir1_threshold_5": 0.026737937233452542,
274
+ "scr_metric_threshold_5": 0.08898314502311447,
275
+ "scr_dir2_threshold_5": 0.08898314502311447,
276
+ "scr_dir1_threshold_10": 0.06417124060509859,
277
+ "scr_metric_threshold_10": 0.09322037750945592,
278
+ "scr_dir2_threshold_10": 0.09322037750945592,
279
+ "scr_dir1_threshold_20": 0.13369000490861685,
280
+ "scr_metric_threshold_20": 0.11864402498949342,
281
+ "scr_dir2_threshold_20": 0.11864402498949342,
282
+ "scr_dir1_threshold_50": 0.1443850523054562,
283
+ "scr_metric_threshold_50": 0.17796603748424014,
284
+ "scr_dir2_threshold_50": 0.17796603748424014,
285
+ "scr_dir1_threshold_100": 0.1978609267723613,
286
+ "scr_metric_threshold_100": 0.23728830254097566,
287
+ "scr_dir2_threshold_100": 0.23728830254097566,
288
+ "scr_dir1_threshold_500": 0.3262034079825585,
289
+ "scr_metric_threshold_500": 0.30084754752206383,
290
+ "scr_dir2_threshold_500": 0.30084754752206383
291
+ }
292
+ ],
293
+ "sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
294
+ "sae_lens_id": "layer_19/width_1m/average_l0_50",
295
+ "sae_lens_release_id": "gemma-scope-2b-pt-res",
296
+ "sae_lens_version": "4.4.5",
297
+ "eval_result_unstructured": null
298
+ }
gemma_scope/scr_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_88_eval_results.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_type_id": "scr",
3
+ "eval_config": {
4
+ "random_seed": 42,
5
+ "dataset_names": [
6
+ "LabHC/bias_in_bios_class_set1",
7
+ "canrager/amazon_reviews_mcauley_1and5"
8
+ ],
9
+ "perform_scr": true,
10
+ "early_stopping_patience": 20,
11
+ "train_set_size": 4000,
12
+ "test_set_size": 1000,
13
+ "context_length": 128,
14
+ "probe_train_batch_size": 16,
15
+ "probe_test_batch_size": 500,
16
+ "probe_epochs": 20,
17
+ "probe_lr": 0.001,
18
+ "probe_l1_penalty": 0.001,
19
+ "sae_batch_size": 5,
20
+ "llm_batch_size": 32,
21
+ "llm_dtype": "bfloat16",
22
+ "lower_vram_usage": true,
23
+ "model_name": "gemma-2-2b",
24
+ "n_values": [
25
+ 2,
26
+ 5,
27
+ 10,
28
+ 20,
29
+ 50,
30
+ 100,
31
+ 500
32
+ ],
33
+ "column1_vals_lookup": {
34
+ "LabHC/bias_in_bios_class_set1": [
35
+ [
36
+ "professor",
37
+ "nurse"
38
+ ],
39
+ [
40
+ "architect",
41
+ "journalist"
42
+ ],
43
+ [
44
+ "surgeon",
45
+ "psychologist"
46
+ ],
47
+ [
48
+ "attorney",
49
+ "teacher"
50
+ ]
51
+ ],
52
+ "canrager/amazon_reviews_mcauley_1and5": [
53
+ [
54
+ "Books",
55
+ "CDs_and_Vinyl"
56
+ ],
57
+ [
58
+ "Software",
59
+ "Electronics"
60
+ ],
61
+ [
62
+ "Pet_Supplies",
63
+ "Office_Products"
64
+ ],
65
+ [
66
+ "Industrial_and_Scientific",
67
+ "Toys_and_Games"
68
+ ]
69
+ ]
70
+ }
71
+ },
72
+ "eval_id": "e27fc6fb-c45d-4b4c-9da4-472a80f45bab",
73
+ "datetime_epoch_millis": 1732800750015,
74
+ "eval_result_metrics": {
75
+ "scr_metrics": {
76
+ "scr_dir1_threshold_2": 0.21264830382497119,
77
+ "scr_metric_threshold_2": 0.1818389025317222,
78
+ "scr_dir2_threshold_2": 0.1818389025317222,
79
+ "scr_dir1_threshold_5": 0.17133533645756982,
80
+ "scr_metric_threshold_5": 0.2450639539500707,
81
+ "scr_dir2_threshold_5": 0.2450639539500707,
82
+ "scr_dir1_threshold_10": 0.21522004701769837,
83
+ "scr_metric_threshold_10": 0.2881033586687831,
84
+ "scr_dir2_threshold_10": 0.2881033586687831,
85
+ "scr_dir1_threshold_20": 0.2230499282541247,
86
+ "scr_metric_threshold_20": 0.3358950334782536,
87
+ "scr_dir2_threshold_20": 0.3358950334782536,
88
+ "scr_dir1_threshold_50": 0.10437551693923294,
89
+ "scr_metric_threshold_50": 0.40656554322945876,
90
+ "scr_dir2_threshold_50": 0.40656554322945876,
91
+ "scr_dir1_threshold_100": -0.05099819322146019,
92
+ "scr_metric_threshold_100": 0.4631431256443061,
93
+ "scr_dir2_threshold_100": 0.4631431256443061,
94
+ "scr_dir1_threshold_500": -0.23269817062063827,
95
+ "scr_metric_threshold_500": 0.4942973245720607,
96
+ "scr_dir2_threshold_500": 0.4942973245720607
97
+ }
98
+ },
99
+ "eval_result_details": [
100
+ {
101
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
102
+ "scr_dir1_threshold_2": 0.2941181626697965,
103
+ "scr_metric_threshold_2": 0.047979929491924624,
104
+ "scr_dir2_threshold_2": 0.047979929491924624,
105
+ "scr_dir1_threshold_5": 0.38235317320140844,
106
+ "scr_metric_threshold_5": 0.09090911827577622,
107
+ "scr_dir2_threshold_5": 0.09090911827577622,
108
+ "scr_dir1_threshold_10": 0.4264711167365414,
109
+ "scr_metric_threshold_10": 0.1010101481416146,
110
+ "scr_dir2_threshold_10": 0.1010101481416146,
111
+ "scr_dir1_threshold_20": 0.4411768057353677,
112
+ "scr_metric_threshold_20": 0.11363639784472028,
113
+ "scr_dir2_threshold_20": 0.11363639784472028,
114
+ "scr_dir1_threshold_50": 0.47058818373302036,
115
+ "scr_metric_threshold_50": 0.13131323773912976,
116
+ "scr_dir2_threshold_50": 0.13131323773912976,
117
+ "scr_dir1_threshold_100": -0.823529102398122,
118
+ "scr_metric_threshold_100": 0.14646470727950273,
119
+ "scr_dir2_threshold_100": 0.14646470727950273,
120
+ "scr_dir1_threshold_500": -1.8970584239309076,
121
+ "scr_metric_threshold_500": 0.19191926641739082,
122
+ "scr_dir2_threshold_500": 0.19191926641739082
123
+ },
124
+ {
125
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
126
+ "scr_dir1_threshold_2": 0.32142872348125223,
127
+ "scr_metric_threshold_2": 0.16618085123760423,
128
+ "scr_dir2_threshold_2": 0.16618085123760423,
129
+ "scr_dir1_threshold_5": 0.40178590435156525,
130
+ "scr_metric_threshold_5": 0.265306140181066,
131
+ "scr_dir2_threshold_5": 0.265306140181066,
132
+ "scr_dir1_threshold_10": 0.45535718087031307,
133
+ "scr_metric_threshold_10": 0.37900867238755315,
134
+ "scr_dir2_threshold_10": 0.37900867238755315,
135
+ "scr_dir1_threshold_20": 0.5,
136
+ "scr_metric_threshold_20": 0.49271137836848733,
137
+ "scr_dir2_threshold_20": 0.49271137836848733,
138
+ "scr_dir1_threshold_50": 0.49107154261093916,
139
+ "scr_metric_threshold_50": 0.5801748379466394,
140
+ "scr_dir2_threshold_50": 0.5801748379466394,
141
+ "scr_dir1_threshold_100": 0.5178574469625045,
142
+ "scr_metric_threshold_100": 0.6647230574015229,
143
+ "scr_dir2_threshold_100": 0.6647230574015229,
144
+ "scr_dir1_threshold_500": 0.4375002660921914,
145
+ "scr_metric_threshold_500": 0.478134135105462,
146
+ "scr_dir2_threshold_500": 0.478134135105462
147
+ },
148
+ {
149
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
150
+ "scr_dir1_threshold_2": 0.33999911785146764,
151
+ "scr_metric_threshold_2": 0.07038831123983014,
152
+ "scr_dir2_threshold_2": 0.07038831123983014,
153
+ "scr_dir1_threshold_5": 0.419999260902581,
154
+ "scr_metric_threshold_5": 0.09466013272449793,
155
+ "scr_dir2_threshold_5": 0.09466013272449793,
156
+ "scr_dir1_threshold_10": 0.4599999284744433,
157
+ "scr_metric_threshold_10": 0.1359223160513138,
158
+ "scr_dir2_threshold_10": 0.1359223160513138,
159
+ "scr_dir1_threshold_20": 0.47999966621406887,
160
+ "scr_metric_threshold_20": 0.2087377805053172,
161
+ "scr_dir2_threshold_20": 0.2087377805053172,
162
+ "scr_dir1_threshold_50": 0.1399993562699899,
163
+ "scr_metric_threshold_50": 0.29126214715894894,
164
+ "scr_dir2_threshold_50": 0.29126214715894894,
165
+ "scr_dir1_threshold_100": 0.019999737739625517,
166
+ "scr_metric_threshold_100": 0.3665047648271256,
167
+ "scr_dir2_threshold_100": 0.3665047648271256,
168
+ "scr_dir1_threshold_500": -0.16000028610222672,
169
+ "scr_metric_threshold_500": 0.23543689987562594,
170
+ "scr_dir2_threshold_500": 0.23543689987562594
171
+ },
172
+ {
173
+ "dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
174
+ "scr_dir1_threshold_2": 0.32575752102418865,
175
+ "scr_metric_threshold_2": 0.20060800350743382,
176
+ "scr_dir2_threshold_2": 0.20060800350743382,
177
+ "scr_dir1_threshold_5": 0.38636371846371703,
178
+ "scr_metric_threshold_5": 0.3495440426617002,
179
+ "scr_dir2_threshold_5": 0.3495440426617002,
180
+ "scr_dir1_threshold_10": 0.42424247897581135,
181
+ "scr_metric_threshold_10": 0.4072947739054895,
182
+ "scr_dir2_threshold_10": 0.4072947739054895,
183
+ "scr_dir1_threshold_20": 0.32575752102418865,
184
+ "scr_metric_threshold_20": 0.41641337716417814,
185
+ "scr_dir2_threshold_20": 0.41641337716417814,
186
+ "scr_dir1_threshold_50": 0.28030309871976417,
187
+ "scr_metric_threshold_50": 0.5288753656218946,
188
+ "scr_dir2_threshold_50": 0.5288753656218946,
189
+ "scr_dir1_threshold_100": 0.3333331828165188,
190
+ "scr_metric_threshold_100": 0.5471123909701696,
191
+ "scr_dir2_threshold_100": 0.5471123909701696,
192
+ "scr_dir1_threshold_500": 0.007575661792330145,
193
+ "scr_metric_threshold_500": 0.6656535086567126,
194
+ "scr_dir2_threshold_500": 0.6656535086567126
195
+ },
196
+ {
197
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
198
+ "scr_dir1_threshold_2": 0.028735600684562674,
199
+ "scr_metric_threshold_2": 0.5156249163264817,
200
+ "scr_dir2_threshold_2": 0.5156249163264817,
201
+ "scr_dir1_threshold_5": 0.04022970393623537,
202
+ "scr_metric_threshold_5": 0.5781250472937277,
203
+ "scr_dir2_threshold_5": 0.5781250472937277,
204
+ "scr_dir1_threshold_10": 0.045977098117452646,
205
+ "scr_metric_threshold_10": 0.5898436472270916,
206
+ "scr_dir2_threshold_10": 0.5898436472270916,
207
+ "scr_dir1_threshold_20": 0.05747120136912535,
208
+ "scr_metric_threshold_20": 0.6445311454081021,
209
+ "scr_dir2_threshold_20": 0.6445311454081021,
210
+ "scr_dir1_threshold_50": -0.6264368485453043,
211
+ "scr_metric_threshold_50": 0.714843676330924,
212
+ "scr_dir2_threshold_50": 0.714843676330924,
213
+ "scr_dir1_threshold_100": -0.6091953511124143,
214
+ "scr_metric_threshold_100": 0.7226563091171597,
215
+ "scr_dir2_threshold_100": 0.7226563091171597,
216
+ "scr_dir1_threshold_500": -0.6781609982885933,
217
+ "scr_metric_threshold_500": 0.7265625090949477,
218
+ "scr_dir2_threshold_500": 0.7265625090949477
219
+ },
220
+ {
221
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
222
+ "scr_dir1_threshold_2": 0.06989251619018748,
223
+ "scr_metric_threshold_2": 0.0238096871253275,
224
+ "scr_dir2_threshold_2": 0.0238096871253275,
225
+ "scr_dir1_threshold_5": 0.10215047733373753,
226
+ "scr_metric_threshold_5": 0.04365093537887616,
227
+ "scr_dir2_threshold_5": 0.04365093537887616,
228
+ "scr_dir1_threshold_10": 0.12365589158110252,
229
+ "scr_metric_threshold_10": 0.08730163423141594,
230
+ "scr_dir2_threshold_10": 0.08730163423141594,
231
+ "scr_dir1_threshold_20": 0.10215047733373753,
232
+ "scr_metric_threshold_20": 0.15079381786384077,
233
+ "scr_dir2_threshold_20": 0.15079381786384077,
234
+ "scr_dir1_threshold_50": 0.12365589158110252,
235
+ "scr_metric_threshold_50": 0.23015881087803544,
236
+ "scr_dir2_threshold_50": 0.23015881087803544,
237
+ "scr_dir1_threshold_100": 0.16129044662773992,
238
+ "scr_metric_threshold_100": 0.36904778517921244,
239
+ "scr_dir2_threshold_100": 0.36904778517921244,
240
+ "scr_dir1_threshold_500": 0.21505382201865497,
241
+ "scr_metric_threshold_500": 0.5515874583329292,
242
+ "scr_dir2_threshold_500": 0.5515874583329292
243
+ },
244
+ {
245
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
246
+ "scr_dir1_threshold_2": 0.2089551973247301,
247
+ "scr_metric_threshold_2": 0.3114754963356824,
248
+ "scr_dir2_threshold_2": 0.3114754963356824,
249
+ "scr_dir1_threshold_5": -0.44776120066881747,
250
+ "scr_metric_threshold_5": 0.419672190095428,
251
+ "scr_dir2_threshold_5": 0.419672190095428,
252
+ "scr_dir1_threshold_10": -0.32089538628924175,
253
+ "scr_metric_threshold_10": 0.4688524699089378,
254
+ "scr_dir2_threshold_10": 0.4688524699089378,
255
+ "scr_dir1_threshold_20": -0.2611937742505272,
256
+ "scr_metric_threshold_20": 0.4950820306461671,
257
+ "scr_dir2_threshold_20": 0.4950820306461671,
258
+ "scr_dir1_threshold_50": -0.23134296823116993,
259
+ "scr_metric_threshold_50": 0.6360655776149192,
260
+ "scr_dir2_threshold_50": 0.6360655776149192,
261
+ "scr_dir1_threshold_100": -0.2537311839483806,
262
+ "scr_metric_threshold_100": 0.6852460528534893,
263
+ "scr_dir2_threshold_100": 0.6852460528534893,
264
+ "scr_dir1_threshold_500": -0.2089551973247301,
265
+ "scr_metric_threshold_500": 0.800000039085012,
266
+ "scr_dir2_threshold_500": 0.800000039085012
267
+ },
268
+ {
269
+ "dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
270
+ "scr_dir1_threshold_2": 0.11229959137358399,
271
+ "scr_metric_threshold_2": 0.11864402498949342,
272
+ "scr_dir2_threshold_2": 0.11864402498949342,
273
+ "scr_dir1_threshold_5": 0.08556165414013145,
274
+ "scr_metric_threshold_5": 0.11864402498949342,
275
+ "scr_dir2_threshold_5": 0.11864402498949342,
276
+ "scr_dir1_threshold_10": 0.10695206767516431,
277
+ "scr_metric_threshold_10": 0.13559320749684803,
278
+ "scr_dir2_threshold_10": 0.13559320749684803,
279
+ "scr_dir1_threshold_20": 0.13903752860703653,
280
+ "scr_metric_threshold_20": 0.1652543400252158,
281
+ "scr_dir2_threshold_20": 0.1652543400252158,
282
+ "scr_dir1_threshold_50": 0.18716587937552193,
283
+ "scr_metric_threshold_50": 0.1398306925451783,
284
+ "scr_dir2_threshold_50": 0.1398306925451783,
285
+ "scr_dir1_threshold_100": 0.2459892775408467,
286
+ "scr_metric_threshold_100": 0.20338993752626644,
287
+ "scr_dir2_threshold_100": 0.20338993752626644,
288
+ "scr_dir1_threshold_500": 0.42245979077817514,
289
+ "scr_metric_threshold_500": 0.3050847800084053,
290
+ "scr_dir2_threshold_500": 0.3050847800084053
291
+ }
292
+ ],
293
+ "sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
294
+ "sae_lens_id": "layer_19/width_1m/average_l0_88",
295
+ "sae_lens_release_id": "gemma-scope-2b-pt-res",
296
+ "sae_lens_version": "4.4.5",
297
+ "eval_result_unstructured": null
298
+ }