Upload folder using huggingface_hub
Browse files- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_107_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_19_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_207_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_26_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_58_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_73_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_157_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_16_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_18_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_29_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_50_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_88_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_114_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_13_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_21_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_36_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_63_eval_results.json +154 -0
- gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_9_eval_results.json +154 -0
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_107_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "271ac369-a8b4-4486-a86e-7fd9b8510241",
|
| 73 |
+
"datetime_epoch_millis": 1732802893750,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": 0.0013750046491622925,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.003700011968612671,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.0023250073194503785,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.003400002419948578,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.005000007152557374,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.0016000047326087952,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.003749993443489075,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.0064999997615814206,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.002750006318092346,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.00982501208782196,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.01420001983642578,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.004375007748603821,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.019775000214576722,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.024100005626678467,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.004325005412101746,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.05610000044107437,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.06300000548362732,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.006900005042552948,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.24575000554323195,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.25560001730918885,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.009850011765956878
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.0029000043869018555,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.004400014877319336,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": 0.0015000104904174805,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.002650028467178345,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.0042000293731689455,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": 0.0015500009059906006,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.004849985241889954,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.005999994277954101,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 0.001150009036064148,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.006650009751319885,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.00860002040863037,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": 0.0019500106573104858,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.016050001978874205,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.019200003147125243,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": 0.0031500011682510376,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.04840000569820404,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.05460001230239868,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.0062000066041946415,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.28369999825954434,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.2918000102043152,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.008100011944770813
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": -0.0001499950885772705,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.003000009059906006,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.0031500041484832765,
|
| 129 |
+
"tpp_threshold_5_total_metric": 0.004149976372718811,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.005799984931945801,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.0016500085592269897,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.0026500016450881963,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.007000005245208741,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.004350003600120544,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.013000014424324037,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.019800019264221192,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.006800004839897155,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.02349999845027924,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.02900000810623169,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.005500009655952454,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.0637999951839447,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.07139999866485595,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.0076000034809112545,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.20780001282691957,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.2194000244140625,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.011600011587142944
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_12/width_1m/average_l0_107",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_19_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "dbaf35ab-c1ca-4a1a-861f-4b2ac87c9295",
|
| 73 |
+
"datetime_epoch_millis": 1732807324236,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": 0.0012499898672103883,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.002999997138977051,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.0017500072717666626,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.0019749999046325684,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.004600006341934204,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.002625006437301636,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.004824995994567871,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.007300001382827759,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.0024750053882598875,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.013024991750717162,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.017899996042251586,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.004875004291534424,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.03225001096725464,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.04490001797676087,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.012650007009506225,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.05287500023841858,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.07230001091957092,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.019425010681152342,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.15495000779628754,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.17980001568794252,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.02485000789165497
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.0037999808788299564,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.004399991035461426,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": 0.0006000101566314697,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.0012999892234802244,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.0029999971389770507,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": 0.0017000079154968263,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.0034999936819076535,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.004200005531311035,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 0.0007000118494033813,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.014599984884262084,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.016999995708465575,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": 0.002400010824203491,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.031700012087821965,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.03660001754760742,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": 0.004900005459785461,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.0697500079870224,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.07740001678466797,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.007650008797645569,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.20540001392364501,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.2150000214576721,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.0096000075340271
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": -0.0013000011444091798,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.0016000032424926757,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.0029000043869018555,
|
| 129 |
+
"tpp_threshold_5_total_metric": 0.0026500105857849123,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.006200015544891357,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.003550004959106445,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.006149998307228089,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.010399997234344482,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.004249998927116394,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.01144999861717224,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.018799996376037596,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.0073499977588653564,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.032800009846687316,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.053200018405914304,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.02040000855922699,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.03599999248981476,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.06720000505447388,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.031200012564659117,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.10450000166893006,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.1446000099182129,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.04010000824928284
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_12/width_1m/average_l0_19",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_207_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "23d1ee1a-f1e0-4c9e-98b8-d940c929d764",
|
| 73 |
+
"datetime_epoch_millis": 1732811634933,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": 0.0022249951958656315,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.0039000034332275393,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.0016750082373619079,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.002974994480609894,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.0050999999046325685,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.0021250054240226747,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.00907500386238098,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.01230001449584961,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.0032250106334686278,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.014125008881092072,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.020200014114379883,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.0060750052332878115,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.04402499794960022,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.0502000093460083,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.006175011396408081,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.1209750071167946,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.13270001411437987,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.011725006997585295,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.355925016105175,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.376000028848648,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.020075012743473054
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.003949987888336182,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.004799997806549073,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": 0.0008500099182128906,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.004249989986419678,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.005799996852874756,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": 0.001550006866455078,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.010700005292892455,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.012000024318695068,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 0.0013000190258026123,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.018150004744529727,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.02040001153945923,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": 0.0022500067949295043,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.0487500011920929,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.05180001258850098,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": 0.003050011396408081,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.13575000762939454,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.14500001668930054,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.009250009059906006,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.41810002028942106,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.43100003004074094,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.012900009751319885
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": 0.0005000025033950807,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.003000009059906006,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.0025000065565109254,
|
| 129 |
+
"tpp_threshold_5_total_metric": 0.0016999989748001099,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.004400002956390381,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.002700003981590271,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.007450002431869507,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.01260000467300415,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.005150002241134643,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.010100013017654418,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.020000016689300536,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.009900003671646118,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.039299994707107544,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.04860000610351563,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.00930001139640808,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.10620000660419464,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.12040001153945923,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.014200004935264587,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.2937500119209289,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.32100002765655516,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.027250015735626222
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_12/width_1m/average_l0_207",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_26_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "cf318db4-f1c8-41f1-b0d9-63025b7cbaec",
|
| 73 |
+
"datetime_epoch_millis": 1732815981090,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": 0.00249999463558197,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.004100000858306885,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.0016000062227249144,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.00032500326633453367,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.002500009536743164,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.0021750062704086305,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.004599989950656891,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.007499998807907105,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.0029000088572502134,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.016175009310245514,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.021100014448165894,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.00492500513792038,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.02772500663995743,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.03370001316070557,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.005975006520748139,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.03942499905824661,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.04950000643730164,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.010075007379055024,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.14940000772476197,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.17930001616477964,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.0299000084400177
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.004249989986419678,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.004799997806549073,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": 0.0005500078201293946,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.0021500051021575927,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.0038000106811523437,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": 0.001650005578994751,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.0040499806404113775,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.004799997806549073,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 0.0007500171661376954,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.017499998211860657,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.0190000057220459,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": 0.0015000075101852417,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.02940000295639038,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.03420001268386841,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": 0.0048000097274780275,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.046800002455711365,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.05320000648498535,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.006400004029273987,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.19675000607967377,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.20720001459121704,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.010450008511543273
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": 0.0007499992847442628,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.003400003910064697,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.0026500046253204344,
|
| 129 |
+
"tpp_threshold_5_total_metric": -0.0014999985694885253,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.0012000083923339843,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.0027000069618225097,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.005149999260902405,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.010199999809265137,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.005050000548362732,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.014850020408630371,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.023200023174285888,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.008350002765655517,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.026050010323524476,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.03320001363754273,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.007150003314018249,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.03204999566078186,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.04580000638961792,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.01375001072883606,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.10205000936985015,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.15140001773834227,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.049350008368492126
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_12/width_1m/average_l0_26",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_58_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "2814ce7e-6088-4329-a8b2-16291a2e1ebd",
|
| 73 |
+
"datetime_epoch_millis": 1732820301277,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": 0.0025499999523162843,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.004100006818771362,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.001550006866455078,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.001600000262260437,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.0037000060081481935,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.0021000057458877563,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.020299994945526124,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.024500006437301637,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.004200011491775513,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.04507500231266022,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.05060001015663147,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.0055250078439712524,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.09865000545978547,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.10820001363754272,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.009550008177757262,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.1921500086784363,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.20770001411437988,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.015550005435943603,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.35172502100467684,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.3713000297546387,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.019575008749961854
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.0031499862670898438,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.0037999987602233888,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": 0.0006500124931335449,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.003349998593330383,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.005000007152557373,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": 0.0016500085592269897,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.013399982452392578,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.015799999237060547,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 0.0024000167846679687,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.046849998831748965,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.04960000514984131,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": 0.002750006318092346,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.11260000169277191,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.11860001087188721,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": 0.006000009179115295,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.21295000612735748,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.22120001316070556,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.008250007033348083,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.40855001211166386,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.41740002632141116,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.008850014209747315
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": 0.0019500136375427248,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.004400014877319336,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.002450001239776611,
|
| 129 |
+
"tpp_threshold_5_total_metric": -0.00014999806880950902,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.0024000048637390138,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.0025500029325485228,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.02720000743865967,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.03320001363754273,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.006000006198883056,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.043300005793571475,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.05160001516342163,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.008300009369850158,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.08470000922679902,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.09780001640319824,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.01310000717639923,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.17135001122951507,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.1942000150680542,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.022850003838539124,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.2949000298976898,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.3252000331878662,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.030300003290176392
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_12/width_1m/average_l0_58",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_73_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "d628e440-b8f9-4d08-82a3-69a8d55aafac",
|
| 73 |
+
"datetime_epoch_millis": 1732824644214,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": 0.0027750089764595033,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.004400014877319336,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.0016250059008598328,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.0014000117778778076,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.003600013256072998,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.0022000014781951905,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.004649992287158967,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.007700002193450928,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.0030500099062919614,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.009050004184246063,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.012400007247924803,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.0033500030636787414,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.03312500268220901,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.03870000839233398,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.005575005710124969,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.0787999987602234,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.08760001063346863,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.00880001187324524,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.2875000208616257,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.29920002818107605,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.011700007319450378
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.004050001502037048,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.004600012302398681,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": 0.0005500108003616333,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.0034000128507614136,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.005200016498565674,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": 0.0018000036478042602,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.005149996280670166,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.005800008773803711,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 0.0006500124931335449,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.0127000093460083,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.013800013065338134,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": 0.001100003719329834,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.031950002908706664,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.03520001173019409,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": 0.0032500088214874267,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.0886000007390976,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.09240001440048218,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.0038000136613845826,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.3336500197649002,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.3396000266075134,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.00595000684261322
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": 0.0015000164508819584,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.004200017452239991,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.0027000010013580322,
|
| 129 |
+
"tpp_threshold_5_total_metric": -0.0005999892950057984,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.002000010013580322,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.0025999993085861206,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.004149988293647767,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.009599995613098145,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.005450007319450378,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.0053999990224838255,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.011000001430511474,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.005600002408027649,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.03430000245571137,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.042200005054473876,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.007900002598762512,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.06899999678134919,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.08280000686645508,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.013800010085105896,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.24135002195835115,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.25880002975463867,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.017450007796287536
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_12/width_1m/average_l0_73",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_157_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "39ac12cf-d5af-4b21-910b-b266cf430acb",
|
| 73 |
+
"datetime_epoch_millis": 1732805595841,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": 0.014150001108646393,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.01700000762939453,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.0028500065207481386,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.023499999940395356,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.02820000648498535,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.004700006544589996,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.035924994945526124,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.04050000309944153,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.004575008153915405,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.05302499979734421,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.06069999933242798,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.007674999535083771,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.12385000586509703,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.13330000638961792,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.009450000524520875,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.21379999965429308,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.2249000072479248,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.011100007593631745,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.3955500215291977,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.41450002789497375,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.01895000636577606
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.01904999613761902,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.019400012493133546,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": 0.0003500163555145264,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.03189999461174011,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.03240001201629639,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": 0.0005000174045562744,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.047550007700920105,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.0476000189781189,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 5.00112771987915e-05,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.07140001058578492,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.07380001544952393,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": 0.0024000048637390138,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.1481000006198883,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.15220000743865966,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": 0.004100006818771362,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.2532499969005585,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.25720001459121705,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.0039500176906585695,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.43710002303123474,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.4452000379562378,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.008100014925003052
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": 0.009250006079673766,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.014600002765655517,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.005349996685981751,
|
| 129 |
+
"tpp_threshold_5_total_metric": 0.0151000052690506,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.024000000953674317,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.008899995684623718,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.024299982190132144,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.03339998722076416,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.009100005030632019,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.0346499890089035,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.04759998321533203,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.012949994206428528,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.0996000111103058,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.11440000534057618,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.014799994230270386,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.17435000240802764,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.19259999990463256,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.01824999749660492,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.35400002002716063,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.3838000178337097,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.029799997806549072
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_19/width_1m/average_l0_157",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_16_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "c91e3b8e-4b8c-45d8-b156-842f311e4497",
|
| 73 |
+
"datetime_epoch_millis": 1732810170175,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": 0.0011249989271163944,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.0033000051975250245,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.00217500627040863,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.0018500059843063357,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.0031000077724456787,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.0012500017881393432,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.006949996948242187,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.009299999475479125,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.0023500025272369385,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.007524999976158141,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.010600000619888306,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.003075000643730164,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.02012498825788498,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.023099994659423827,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.0029750064015388487,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.043374998867511755,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.04940000772476197,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.006025008857250214,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.14412499964237213,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.15420000553131102,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.010075005888938903
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.002499997615814209,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.0014000058174133301,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": -0.0010999917984008789,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.005600011348724366,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.004200017452239991,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": -0.001399993896484375,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.006749999523162842,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.006400012969970703,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": -0.0003499865531921387,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.011699995398521424,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.011800003051757813,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": 0.00010000765323638915,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.02924998104572296,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.029799997806549072,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": 0.0005500167608261108,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.05475001335144043,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.056800031661987306,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.002050018310546875,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.19165000617504122,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.19580001831054689,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.004150012135505676
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": -0.0002499997615814202,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.005200004577636719,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.005450004339218139,
|
| 129 |
+
"tpp_threshold_5_total_metric": -0.0018999993801116942,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.0019999980926513673,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.0038999974727630614,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.007149994373321532,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.012199985980987548,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.005049991607666016,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.00335000455379486,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.009399998188018798,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.006049993634223938,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.010999995470046998,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.016399991512298585,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.005399996042251587,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.03199998438358307,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.041999983787536624,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.009999999403953552,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.09659999310970306,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.11259999275207519,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.015999999642372132
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_19/width_1m/average_l0_16",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_18_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "e91e1736-be5e-4301-82a9-1b2aaddb213d",
|
| 73 |
+
"datetime_epoch_millis": 1732814609814,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": 0.0018749952316284182,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.004199999570846558,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.0023250043392181396,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.0027000039815902707,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.00440000295639038,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.0016999989748001099,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.007725001871585846,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.01050000786781311,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.0027750059962272644,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.00989999920129776,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.0125,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.00260000079870224,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.02392498552799225,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.026299989223480223,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.002375003695487976,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.04332500845193863,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.048300004005432134,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.0049749955534935,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.15995001047849655,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.1686000108718872,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.008650000393390655
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.0032499998807907103,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.002200007438659668,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": -0.0010499924421310425,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.005750009417533874,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.004600012302398681,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": -0.0011499971151351929,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.007550007104873657,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.007600021362304687,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 5.0014257431030275e-05,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.011550000309944153,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.01140000820159912,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": -0.00014999210834503173,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.028549984097480774,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.02799999713897705,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": -0.0005499869585037231,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.05665002167224884,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.05840002298355103,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.0017500013113021851,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.22364999651908873,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.227400004863739,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.0037500083446502684
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": 0.0004999905824661262,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.006199991703033448,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.005700001120567321,
|
| 129 |
+
"tpp_threshold_5_total_metric": -0.0003500014543533327,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.00419999361038208,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.004549995064735413,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.007899996638298035,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.013399994373321534,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.005499997735023498,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.008249998092651367,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.013599991798400879,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.005349993705749512,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.019299986958503726,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.0245999813079834,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.005299994349479675,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.02999999523162842,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.038199985027313234,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.008199989795684814,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.09625002443790437,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.1098000168800354,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.013549992442131042
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_19/width_1m/average_l0_18",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_29_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "74f6729e-0f7b-4c6e-b816-136226ed10e2",
|
| 73 |
+
"datetime_epoch_millis": 1732819145294,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": 0.003925006091594695,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.006600010395050048,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.002675004303455353,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.005200004577636719,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.00710000991821289,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.001900005340576172,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.01235000491142273,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.015400010347366334,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.0030500054359436034,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.01707499325275421,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.020099997520446777,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.003025004267692566,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.03347500860691071,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.03660001158714295,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.0031250029802322387,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.05972499251365662,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.06609999537467956,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.0063750028610229485,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.20580001473426818,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.21600001454353335,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.010199999809265137
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.006000009179115295,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.005200016498565674,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": -0.0007999926805496216,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.00845000147819519,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.007600009441375732,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": -0.000849992036819458,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.013000008463859559,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.013000023365020753,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 1.4901161193847656e-08,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.020499986410140992,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.020399999618530274,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": -9.998679161071777e-05,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.039050003886222845,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.03860001564025879,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": -0.0004499882459640503,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.06894998848438263,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.07120000123977661,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.0022500127553939818,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.260900017619133,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.2658000230789185,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.004900005459785461
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": 0.0018500030040740963,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.008000004291534423,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.006150001287460327,
|
| 129 |
+
"tpp_threshold_5_total_metric": 0.001950007677078247,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.006600010395050049,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.004650002717971802,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.011700001358985902,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.017799997329711915,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.006099995970726013,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.013650000095367432,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.01979999542236328,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.00614999532699585,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.027900013327598575,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.0346000075340271,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.006699994206428528,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.0504999965429306,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.06099998950958252,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.010499992966651916,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.15070001184940338,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.1662000060081482,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.015499994158744812
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_19/width_1m/average_l0_29",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_50_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "317b8b70-a5a1-489c-8e6a-6807435bd447",
|
| 73 |
+
"datetime_epoch_millis": 1732823642436,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": 0.006174992024898529,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.009199994802474975,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.0030250027775764465,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.009049998223781587,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.011500000953674316,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.002450002729892731,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.018500003218650817,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.021800005435943605,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.0033000022172927856,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.027625001966953278,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.03130000233650208,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.0036750003695487977,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.05174999833106994,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.05580000281333923,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.004050004482269287,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.08990000635385514,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.09760000705718994,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.007700000703334808,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.259050016105175,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.27120001912117003,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.012150003015995026
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.01159999668598175,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.010800004005432129,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": -0.0007999926805496216,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.017099994421005252,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.01640000343322754,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": -0.0006999909877777099,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.02355000376701355,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.023800015449523926,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 0.000250011682510376,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.037850001454353334,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.0380000114440918,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": 0.00015000998973846436,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.06305000483989714,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.06320002079010009,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": 0.0001500159502029419,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.11195001900196076,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.11480002403259278,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.002850005030632019,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.30735001564025877,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.311400032043457,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.004050016403198242
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": 0.0007499873638153083,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.007599985599517823,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.006849998235702514,
|
| 129 |
+
"tpp_threshold_5_total_metric": 0.0010000020265579224,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.006599998474121094,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.005599996447563172,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.013450002670288085,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.01979999542236328,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.006349992752075195,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.017400002479553225,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.024599993228912355,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.007199990749359131,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.04044999182224274,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.04839998483657837,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.007949993014335632,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.06784999370574951,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.08039999008178711,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.012549996376037598,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.21075001657009124,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.23100000619888306,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.02024998962879181
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_19/width_1m/average_l0_50",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_88_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "285f1a93-5e8f-466c-aaae-72b0f7659cc7",
|
| 73 |
+
"datetime_epoch_millis": 1732828187536,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": 0.0098750039935112,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.013500005006790161,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.0036250010132789614,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.014700004458427429,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.017800009250640868,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.0031000047922134398,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.026900011301040645,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.031000012159347532,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.004100000858306885,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.039100004732608794,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.04440000653266907,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.005300001800060273,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.07382499724626541,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.07970000505447387,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.005875007808208466,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.1286250099539757,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.1379000127315521,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.009275002777576447,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.3243000030517578,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.3394000113010407,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.015100008249282837
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.015050017833709718,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.014800024032592774,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": -0.00024999380111694335,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.022800010442733765,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.022600018978118898,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": -0.00019999146461486817,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.03480000495910644,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.035600018501281736,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 0.000800013542175293,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.051750007271766665,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.05240001678466797,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": 0.0006500095129013062,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.09155000746250153,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.09240002632141113,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": 0.000850018858909607,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.14890001118183135,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.15220001935958863,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.003300008177757263,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.3709500104188919,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.37580002546310426,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.004850015044212341
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": 0.004699990153312682,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.012199985980987548,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.007499995827674866,
|
| 129 |
+
"tpp_threshold_5_total_metric": 0.006599998474121093,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.012999999523162841,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.006400001049041748,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.01900001764297485,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.02640000581741333,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.007399988174438476,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.026450002193450926,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.036399996280670165,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.009949994087219239,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.056099987030029295,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.06699998378753662,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.010899996757507325,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.10835000872612,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.12360000610351562,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.01524999737739563,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.27764999568462373,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.3029999971389771,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.025350001454353333
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_19/width_1m/average_l0_88",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_114_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "f583dd17-2716-41f7-a813-4d19d384a362",
|
| 73 |
+
"datetime_epoch_millis": 1732740660080,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": 0.004625003039836884,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.006899994611740113,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.002274991571903229,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.004875001311302185,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.007299995422363282,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.0024249941110610963,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.008974997699260712,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.012899988889694215,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.0039249911904335015,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.016024994850158694,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.019799983501434325,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.003774988651275635,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.03919998854398728,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.045399981737136844,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.006199993193149567,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.07647501528263093,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.0827000081539154,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.006224992871284485,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.294900019466877,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.3055000066757202,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.010599987208843233
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.004500007629394532,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.00820000171661377,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": 0.003699994087219238,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.003850007057189941,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.007800006866455078,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": 0.003949999809265137,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.010449990630149841,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.014199984073638917,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 0.0037499934434890745,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.016349995136260988,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.020599985122680665,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": 0.004249989986419678,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.04004999101161957,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.04919998645782471,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": 0.00914999544620514,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.07160001695156097,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.08080000877380371,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.009199991822242737,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.32235002517700195,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.3348000168800354,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.012449991703033448
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": 0.004749998450279236,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.005599987506866455,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.0008499890565872193,
|
| 129 |
+
"tpp_threshold_5_total_metric": 0.005899995565414429,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.0067999839782714845,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.0008999884128570556,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.007500004768371583,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.011599993705749512,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.004099988937377929,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.015699994564056397,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.018999981880187988,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.0032999873161315917,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.03834998607635498,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.04159997701644898,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.0032499909400939943,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.08135001361370087,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.0846000075340271,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.003249993920326233,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.26745001375675204,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.27619999647140503,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.008749982714653015
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_5/width_1m/average_l0_114",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_13_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "974df560-d379-44a9-9132-0267eeb09208",
|
| 73 |
+
"datetime_epoch_millis": 1732748490561,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": -2.5005638599395405e-05,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.0029999911785125732,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.0030249968171119686,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.0020249903202056883,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.004499983787536621,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.0024749934673309326,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.0067499965429306035,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.01009998917579651,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.003349992632865906,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.00794999897480011,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.011199986934661866,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.0032499879598617554,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.013725003600120545,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.01669999361038208,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.002974990010261536,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.017924994230270386,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.022799980640411378,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.004874986410140991,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.06485000103712081,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.07039999365806579,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.005549992620944977
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.0001499891281127935,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.00399998426437378,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": 0.003849995136260986,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.0031999886035919186,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.006599986553192138,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": 0.0033999979496002197,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.004000002145767212,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.0075999975204467775,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 0.0035999953746795655,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.008750009536743163,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.013399994373321534,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": 0.004649984836578369,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.014950010180473329,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.020000004768371583,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": 0.005049994587898255,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.02194998562335968,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.027799975872039796,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.005849990248680115,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.06870000362396239,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.07419999837875366,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.00549999475479126
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": -0.0002000004053115843,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.0019999980926513673,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.0021999984979629516,
|
| 129 |
+
"tpp_threshold_5_total_metric": 0.000849992036819458,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.0023999810218811035,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.0015499889850616455,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.009499990940093996,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.01259998083114624,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.003099989891052246,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.007149988412857056,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.008999979496002198,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.0018499910831451416,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.012499997019767761,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.013399982452392578,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.0008999854326248169,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.013900002837181092,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.01779998540878296,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.0038999825716018676,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.06099999845027923,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.06659998893737792,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.005599990487098694
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_5/width_1m/average_l0_13",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_21_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "0f831b42-243e-4e17-bfd2-149f20f69e93",
|
| 73 |
+
"datetime_epoch_millis": 1732755201852,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": 0.00037499219179153447,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.0032999873161315917,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.002924995124340057,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.0019999951124191283,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.0045999884605407715,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.002599993348121643,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.008724993467330933,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.012099987268447876,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.0033749938011169434,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.011824998259544372,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.015399986505508423,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.00357498824596405,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.01889999508857727,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.02229998707771301,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.0033999919891357423,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.024949990212917328,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.030199980735778807,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.005249990522861481,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.09215000718832017,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.09779999852180482,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.00564999133348465
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.0007999926805496217,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.0045999884605407715,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": 0.00379999577999115,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.00359998345375061,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.00699998140335083,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": 0.0033999979496002197,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.005699992179870606,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.009199988842010499,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 0.0034999966621398924,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.016150012612342834,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.02120000123977661,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": 0.005049988627433777,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.019999995827674866,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.025599992275238036,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": 0.005599996447563172,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.029699987173080443,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.03579998016357422,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.006099992990493774,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.10000000298023225,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.1059999942779541,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.0059999912977218624
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": -5.0008296966552734e-05,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.001999986171722412,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.0020499944686889647,
|
| 129 |
+
"tpp_threshold_5_total_metric": 0.0004000067710876467,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.002199995517730713,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.0017999887466430664,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.01174999475479126,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.014999985694885254,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.0032499909400939943,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.00749998390674591,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.009599971771240234,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.0020999878644943236,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.017799994349479674,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.018999981880187988,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.001199987530708313,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.020199993252754213,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.0245999813079834,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.004399988055229187,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.08430001139640808,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.08960000276565552,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.005299991369247437
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_5/width_1m/average_l0_21",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_36_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "3df41e6d-e533-4791-b3d1-ca1eb2b3bcec",
|
| 73 |
+
"datetime_epoch_millis": 1732762041332,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": 0.0005249947309494019,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.003499990701675415,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.002974995970726013,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.003424997627735138,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.00629999041557312,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.002874992787837982,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.010299988090991974,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.013799983263015746,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.0034999951720237734,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.0132999986410141,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.017599982023239136,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.004299983382225037,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.023650002479553223,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.027599990367889404,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.003949987888336182,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.03372500389814377,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.03949999809265137,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.005774994194507599,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.13810000121593474,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.14399999380111694,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.00589999258518219
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.0009999901056289675,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.0045999884605407715,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": 0.003599998354911804,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.005050000548362732,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.008599996566772461,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": 0.003549996018409729,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.006549990177154541,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.009999990463256836,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 0.003450000286102295,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.017250004410743716,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.023199987411499024,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": 0.00594998300075531,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.024850019812583925,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.031400012969970706,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": 0.0065499931573867794,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.03709999322891236,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.04419999122619629,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.007099997997283935,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.15629999637603761,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.1625999927520752,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.006299996376037597
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": 4.999935626983634e-05,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.0023999929428100584,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.002349993586540222,
|
| 129 |
+
"tpp_threshold_5_total_metric": 0.0017999947071075445,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.00399998426437378,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.002199989557266235,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.014049986004829405,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.017599976062774657,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.0035499900579452513,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.009349992871284485,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.011999976634979249,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.002649983763694763,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.022449985146522522,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.023799967765808106,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.0013499826192855835,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.03035001456737518,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.03480000495910644,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.004449990391731262,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.1199000060558319,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.12539999485015868,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.005499988794326782
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_5/width_1m/average_l0_36",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_63_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "8cd7b0b0-8e74-4c53-946d-881ac1127796",
|
| 73 |
+
"datetime_epoch_millis": 1732768990817,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": 4.999488592147814e-05,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.0029999911785125732,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.0029499962925910953,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.0044749930500984185,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.007399988174438476,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.0029249951243400575,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.011475001275539399,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.015099996328353883,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.0036249950528144835,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.01495000422000885,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.019899988174438478,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.004949983954429626,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.02775000035762787,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.032699990272521975,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.004949989914894104,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.04887499958276749,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.05499998927116394,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.006124989688396454,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.1950250118970871,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.20260000228881836,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.007574990391731262
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 9.998977184295628e-05,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.0037999868392944334,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": 0.003699997067451477,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.005249992012977599,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.008999991416931152,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": 0.0037499994039535524,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.006950005888938904,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.010600006580352784,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 0.0036500006914138793,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.019800007343292236,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.02659999132156372,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": 0.0067999839782714845,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.02834999561309814,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.03579999208450317,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": 0.007449996471405029,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.046100002527236936,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.05439999103546143,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.008299988508224488,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.21875001192092897,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.2266000032424927,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.00784999132156372
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": 0.0,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.002199995517730713,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.002199995517730713,
|
| 129 |
+
"tpp_threshold_5_total_metric": 0.0036999940872192386,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.005799984931945801,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.0020999908447265625,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.015999996662139894,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.01959998607635498,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.003599989414215088,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.010100001096725464,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.013199985027313232,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.0030999839305877685,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.027150005102157593,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.029599988460540773,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.002449983358383179,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.051649996638298036,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.055599987506866455,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.003949990868568421,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.17130001187324523,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.17860000133514403,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.007299989461898804
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_5/width_1m/average_l0_63",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|
gemma_scope/tpp_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_9_eval_results.json
ADDED
|
@@ -0,0 +1,154 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "tpp",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
| 8 |
+
],
|
| 9 |
+
"perform_scr": false,
|
| 10 |
+
"early_stopping_patience": 20,
|
| 11 |
+
"train_set_size": 4000,
|
| 12 |
+
"test_set_size": 1000,
|
| 13 |
+
"context_length": 128,
|
| 14 |
+
"probe_train_batch_size": 16,
|
| 15 |
+
"probe_test_batch_size": 500,
|
| 16 |
+
"probe_epochs": 20,
|
| 17 |
+
"probe_lr": 0.001,
|
| 18 |
+
"probe_l1_penalty": 0.001,
|
| 19 |
+
"sae_batch_size": 5,
|
| 20 |
+
"llm_batch_size": 32,
|
| 21 |
+
"llm_dtype": "bfloat16",
|
| 22 |
+
"lower_vram_usage": true,
|
| 23 |
+
"model_name": "gemma-2-2b",
|
| 24 |
+
"n_values": [
|
| 25 |
+
2,
|
| 26 |
+
5,
|
| 27 |
+
10,
|
| 28 |
+
20,
|
| 29 |
+
50,
|
| 30 |
+
100,
|
| 31 |
+
500
|
| 32 |
+
],
|
| 33 |
+
"column1_vals_lookup": {
|
| 34 |
+
"LabHC/bias_in_bios_class_set1": [
|
| 35 |
+
[
|
| 36 |
+
"professor",
|
| 37 |
+
"nurse"
|
| 38 |
+
],
|
| 39 |
+
[
|
| 40 |
+
"architect",
|
| 41 |
+
"journalist"
|
| 42 |
+
],
|
| 43 |
+
[
|
| 44 |
+
"surgeon",
|
| 45 |
+
"psychologist"
|
| 46 |
+
],
|
| 47 |
+
[
|
| 48 |
+
"attorney",
|
| 49 |
+
"teacher"
|
| 50 |
+
]
|
| 51 |
+
],
|
| 52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
| 53 |
+
[
|
| 54 |
+
"Books",
|
| 55 |
+
"CDs_and_Vinyl"
|
| 56 |
+
],
|
| 57 |
+
[
|
| 58 |
+
"Software",
|
| 59 |
+
"Electronics"
|
| 60 |
+
],
|
| 61 |
+
[
|
| 62 |
+
"Pet_Supplies",
|
| 63 |
+
"Office_Products"
|
| 64 |
+
],
|
| 65 |
+
[
|
| 66 |
+
"Industrial_and_Scientific",
|
| 67 |
+
"Toys_and_Games"
|
| 68 |
+
]
|
| 69 |
+
]
|
| 70 |
+
}
|
| 71 |
+
},
|
| 72 |
+
"eval_id": "1e487729-3364-47c2-a6ce-685bbf26552d",
|
| 73 |
+
"datetime_epoch_millis": 1732775650303,
|
| 74 |
+
"eval_result_metrics": {
|
| 75 |
+
"tpp_metrics": {
|
| 76 |
+
"tpp_threshold_2_total_metric": -0.00012500137090683,
|
| 77 |
+
"tpp_threshold_2_intended_diff_only": 0.0028999924659729006,
|
| 78 |
+
"tpp_threshold_2_unintended_diff_only": 0.0030249938368797306,
|
| 79 |
+
"tpp_threshold_5_total_metric": 0.0013999953866004947,
|
| 80 |
+
"tpp_threshold_5_intended_diff_only": 0.003899991512298584,
|
| 81 |
+
"tpp_threshold_5_unintended_diff_only": 0.0024999961256980895,
|
| 82 |
+
"tpp_threshold_10_total_metric": 0.008075000345706941,
|
| 83 |
+
"tpp_threshold_10_intended_diff_only": 0.011299991607666017,
|
| 84 |
+
"tpp_threshold_10_unintended_diff_only": 0.003224991261959076,
|
| 85 |
+
"tpp_threshold_20_total_metric": 0.006974992156028748,
|
| 86 |
+
"tpp_threshold_20_intended_diff_only": 0.010299980640411377,
|
| 87 |
+
"tpp_threshold_20_unintended_diff_only": 0.003324988484382629,
|
| 88 |
+
"tpp_threshold_50_total_metric": 0.01184999793767929,
|
| 89 |
+
"tpp_threshold_50_intended_diff_only": 0.014799988269805907,
|
| 90 |
+
"tpp_threshold_50_unintended_diff_only": 0.0029499903321266174,
|
| 91 |
+
"tpp_threshold_100_total_metric": 0.017200005054473878,
|
| 92 |
+
"tpp_threshold_100_intended_diff_only": 0.022199994325637816,
|
| 93 |
+
"tpp_threshold_100_unintended_diff_only": 0.00499998927116394,
|
| 94 |
+
"tpp_threshold_500_total_metric": 0.05437500625848771,
|
| 95 |
+
"tpp_threshold_500_intended_diff_only": 0.05979999303817749,
|
| 96 |
+
"tpp_threshold_500_unintended_diff_only": 0.005424986779689788
|
| 97 |
+
}
|
| 98 |
+
},
|
| 99 |
+
"eval_result_details": [
|
| 100 |
+
{
|
| 101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_tpp_results",
|
| 102 |
+
"tpp_threshold_2_total_metric": 0.0007499963045120238,
|
| 103 |
+
"tpp_threshold_2_intended_diff_only": 0.0045999884605407715,
|
| 104 |
+
"tpp_threshold_2_unintended_diff_only": 0.0038499921560287476,
|
| 105 |
+
"tpp_threshold_5_total_metric": 0.0031000018119812017,
|
| 106 |
+
"tpp_threshold_5_intended_diff_only": 0.006599998474121094,
|
| 107 |
+
"tpp_threshold_5_unintended_diff_only": 0.0034999966621398924,
|
| 108 |
+
"tpp_threshold_10_total_metric": 0.0056499958038330085,
|
| 109 |
+
"tpp_threshold_10_intended_diff_only": 0.009199988842010499,
|
| 110 |
+
"tpp_threshold_10_unintended_diff_only": 0.0035499930381774902,
|
| 111 |
+
"tpp_threshold_20_total_metric": 0.009649994969367982,
|
| 112 |
+
"tpp_threshold_20_intended_diff_only": 0.014199984073638917,
|
| 113 |
+
"tpp_threshold_20_unintended_diff_only": 0.004549989104270935,
|
| 114 |
+
"tpp_threshold_50_total_metric": 0.012649989128112792,
|
| 115 |
+
"tpp_threshold_50_intended_diff_only": 0.017599987983703613,
|
| 116 |
+
"tpp_threshold_50_unintended_diff_only": 0.00494999885559082,
|
| 117 |
+
"tpp_threshold_100_total_metric": 0.021200004220008853,
|
| 118 |
+
"tpp_threshold_100_intended_diff_only": 0.027199995517730714,
|
| 119 |
+
"tpp_threshold_100_unintended_diff_only": 0.0059999912977218624,
|
| 120 |
+
"tpp_threshold_500_total_metric": 0.06080001294612885,
|
| 121 |
+
"tpp_threshold_500_intended_diff_only": 0.06660000085830689,
|
| 122 |
+
"tpp_threshold_500_unintended_diff_only": 0.005799987912178039
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_tpp_results",
|
| 126 |
+
"tpp_threshold_2_total_metric": -0.0009999990463256839,
|
| 127 |
+
"tpp_threshold_2_intended_diff_only": 0.0011999964714050292,
|
| 128 |
+
"tpp_threshold_2_unintended_diff_only": 0.002199995517730713,
|
| 129 |
+
"tpp_threshold_5_total_metric": -0.0003000110387802123,
|
| 130 |
+
"tpp_threshold_5_intended_diff_only": 0.0011999845504760743,
|
| 131 |
+
"tpp_threshold_5_unintended_diff_only": 0.0014999955892562866,
|
| 132 |
+
"tpp_threshold_10_total_metric": 0.010500004887580872,
|
| 133 |
+
"tpp_threshold_10_intended_diff_only": 0.013399994373321534,
|
| 134 |
+
"tpp_threshold_10_unintended_diff_only": 0.0028999894857406616,
|
| 135 |
+
"tpp_threshold_20_total_metric": 0.0042999893426895145,
|
| 136 |
+
"tpp_threshold_20_intended_diff_only": 0.006399977207183838,
|
| 137 |
+
"tpp_threshold_20_unintended_diff_only": 0.0020999878644943236,
|
| 138 |
+
"tpp_threshold_50_total_metric": 0.011050006747245789,
|
| 139 |
+
"tpp_threshold_50_intended_diff_only": 0.011999988555908203,
|
| 140 |
+
"tpp_threshold_50_unintended_diff_only": 0.0009499818086624145,
|
| 141 |
+
"tpp_threshold_100_total_metric": 0.013200005888938904,
|
| 142 |
+
"tpp_threshold_100_intended_diff_only": 0.017199993133544922,
|
| 143 |
+
"tpp_threshold_100_unintended_diff_only": 0.003999987244606018,
|
| 144 |
+
"tpp_threshold_500_total_metric": 0.04794999957084656,
|
| 145 |
+
"tpp_threshold_500_intended_diff_only": 0.05299998521804809,
|
| 146 |
+
"tpp_threshold_500_unintended_diff_only": 0.005049985647201538
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 150 |
+
"sae_lens_id": "layer_5/width_1m/average_l0_9",
|
| 151 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 152 |
+
"sae_lens_version": "4.4.5",
|
| 153 |
+
"eval_result_unstructured": null
|
| 154 |
+
}
|