Upload folder using huggingface_hub
Browse files- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_107_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_19_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_207_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_26_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_58_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_73_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_157_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_16_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_18_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_29_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_50_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_88_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_114_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_13_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_21_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_36_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_63_eval_results.json +215 -0
- gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_9_eval_results.json +215 -0
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_107_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "2152cc35-f62f-4bc2-b5dd-82a1fa957350",
|
| 33 |
+
"datetime_epoch_millis": 1732829183765,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.95139375,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6537937500000001,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.7223437500000001,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.7772000000000001,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.83160625,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.8772187499999999,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.9226875000000001,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.6482749999999999,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.6911437500000001,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.7850062500000001,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.8371875,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.89884375,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.92945625,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9598000000000001,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.6428,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.6910000000000001,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.7916000000000001,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.8336,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.8968,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9378,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.6537999999999999,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.7007999999999999,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.8507999999999999,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.8928,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.9202,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.9398,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9446,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.671,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7164,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7598,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8008,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8652,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.9052000000000001,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.6756,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.6824,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.7806,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.8193999999999999,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.8922000000000001,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.9282,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.9111999999999998,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6921999999999999,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7430000000000001,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7534000000000001,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.7969999999999999,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.842,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.8894,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.6447999999999999,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.6923999999999999,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.8011999999999999,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.8368,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.8722,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.8994,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.9012,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.6056,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.64,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.6658,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.7574,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.8156000000000001,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8622,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.6095999999999999,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.6405999999999998,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.752,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.7856000000000001,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.8256,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.8779999999999999,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.981,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.673,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.724,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.764,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.827,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.848,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.933,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.582,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.6,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.7,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.757,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.923,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.969,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9693999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.662,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.7014,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.7582,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.8030000000000002,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.8614,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9270000000000002,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.6295999999999999,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.6688000000000001,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.7148,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.8106,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.8882,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.9292,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.94475,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.6387499999999999,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.7837500000000001,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.8280000000000001,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.8732500000000001,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.89775,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9295,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.599,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.6357499999999999,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.70825,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.7985,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.87075,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.89325,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9992000000000001,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.6450000000000001,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.7792,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.8968,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9608000000000001,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.991,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9974000000000001,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.7918,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.9084,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.9724,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.9968,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.9985999999999999,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.9987999999999999,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_12/width_1m/average_l0_107",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_19_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "b30d3f76-8187-49e4-a377-1b4d4794bcda",
|
| 33 |
+
"datetime_epoch_millis": 1732834005832,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.95139375,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6537937500000001,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.7223437500000001,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.7772000000000001,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.83160625,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.8772187499999999,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.9226875000000001,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.627475,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.67053125,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.73448125,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.8145062500000001,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.86846875,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.9058187500000001,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9598000000000001,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.6428,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.6910000000000001,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.7916000000000001,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.8336,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.8968,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9378,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.616,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.6772,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.7574,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.8455999999999999,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.9017999999999999,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.9262,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9446,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.671,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7164,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7598,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8008,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8652,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.9052000000000001,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.6734,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.7030000000000001,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.7122,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.8358000000000001,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.8936,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.915,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.9111999999999998,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6921999999999999,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7430000000000001,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7534000000000001,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.7969999999999999,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.842,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.8894,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.6138,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.662,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.6732,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.789,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.8502000000000001,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.8762000000000001,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.9012,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.6056,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.64,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.6658,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.7574,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.8156000000000001,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8622,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.5494000000000001,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.6326,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.6664,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.7365999999999999,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.7872,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.8506,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.981,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.673,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.724,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.764,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.827,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.848,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.933,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.649,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.674,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.715,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.773,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.8634999999999999,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.9305000000000001,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9693999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.662,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.7014,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.7582,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.8030000000000002,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.8614,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9270000000000002,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.6254000000000001,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.6452,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.7048,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.8206,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.86,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.8969999999999999,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.94475,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.6387499999999999,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.7837500000000001,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.8280000000000001,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.8732500000000001,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.89775,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9295,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.619,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.65325,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.70025,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.7372500000000001,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.8002499999999999,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.8532500000000001,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9992000000000001,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.6450000000000001,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.7792,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.8968,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9608000000000001,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.991,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9974000000000001,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.6738,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.717,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.9465999999999999,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.9782,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.9912000000000001,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.9978,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_12/width_1m/average_l0_19",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_207_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "b152193d-2e86-4f7c-ab37-f2c49a5e43ab",
|
| 33 |
+
"datetime_epoch_millis": 1732839196988,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.95139375,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6537937500000001,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.7223437500000001,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.7772000000000001,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.83160625,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.8772187499999999,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.9226875000000001,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.6875687500000001,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.73154375,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.809075,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.85929375,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.9051187500000001,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.9346062499999999,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9598000000000001,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.6428,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.6910000000000001,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.7916000000000001,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.8336,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.8968,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9378,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.6634,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.6708,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.8081999999999999,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.8905999999999998,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.9262,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.942,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9446,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.671,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7164,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7598,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8008,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8652,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.9052000000000001,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.6818,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.712,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.8039999999999999,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.8628,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.9086000000000001,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.9298000000000002,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.9111999999999998,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6921999999999999,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7430000000000001,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7534000000000001,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.7969999999999999,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.842,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.8894,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.6488,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.7594000000000001,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.8002,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.859,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.873,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.9075999999999999,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.9012,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.6056,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.64,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.6658,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.7574,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.8156000000000001,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8622,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.5774,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.615,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.7425999999999999,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.8093999999999999,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.8326,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.8794000000000001,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.981,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.673,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.724,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.764,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.827,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.848,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.933,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.641,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.678,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.77,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.811,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.9175,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.9755,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9693999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.662,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.7014,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.7582,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.8030000000000002,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.8614,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9270000000000002,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.653,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.6692,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.7214,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.7719999999999999,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.8864000000000001,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.9258000000000001,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.94475,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.6387499999999999,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.7837500000000001,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.8280000000000001,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.8732500000000001,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.89775,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9295,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.76675,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.81175,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.832,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.87175,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.89825,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.91775,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9992000000000001,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.6450000000000001,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.7792,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.8968,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9608000000000001,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.991,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9974000000000001,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.8684,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.9362,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.9942,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.9978,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.9984,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.999,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_12/width_1m/average_l0_207",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_26_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "50888268-cb08-44fa-beb2-0bd7ca5d5826",
|
| 33 |
+
"datetime_epoch_millis": 1732843873915,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.95139375,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6537937500000001,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.7223437500000001,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.7772000000000001,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.83160625,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.8772187499999999,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.9226875000000001,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.62875,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.6551812499999999,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.73898125,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.817175,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.8776999999999999,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.91249375,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9598000000000001,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.6428,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.6910000000000001,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.7916000000000001,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.8336,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.8968,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9378,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.6358,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.6552,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.726,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.8792,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.9007999999999999,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.9296000000000001,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9446,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.671,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7164,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7598,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8008,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8652,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.9052000000000001,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.6990000000000001,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.7076,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.728,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.7896000000000001,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.8968,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.9164000000000001,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.9111999999999998,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6921999999999999,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7430000000000001,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7534000000000001,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.7969999999999999,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.842,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.8894,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.6058,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.6328,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.739,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.7948,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.8522000000000001,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.8812000000000001,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.9012,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.6056,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.64,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.6658,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.7574,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.8156000000000001,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8622,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.5574000000000001,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.6234,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.6936,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.7559999999999999,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.7936,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.8523999999999999,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.981,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.673,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.724,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.764,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.827,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.848,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.933,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.558,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.635,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.711,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.796,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.896,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.95,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9693999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.662,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.7014,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.7582,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.8030000000000002,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.8614,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9270000000000002,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.6348,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.639,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.7087999999999999,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.8238,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.8702,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.9057999999999999,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.94475,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.6387499999999999,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.7837500000000001,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.8280000000000001,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.8732500000000001,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.89775,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9295,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.639,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.65325,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.6652499999999999,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.7280000000000001,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.818,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.86675,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9992000000000001,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.6450000000000001,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.7792,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.8968,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9608000000000001,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.991,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9974000000000001,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.7002,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.6952,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.9401999999999999,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.97,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.994,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.9978000000000001,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_12/width_1m/average_l0_26",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_58_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "820c150c-cd60-430a-8926-2abee617bb27",
|
| 33 |
+
"datetime_epoch_millis": 1732848444855,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.95139375,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6537937500000001,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.7223437500000001,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.7772000000000001,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.83160625,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.8772187499999999,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.9226875000000001,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.66771875,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.7066062499999999,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.7661187500000001,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.8053125000000001,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.8574499999999999,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.909,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9598000000000001,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.6428,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.6910000000000001,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.7916000000000001,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.8336,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.8968,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9378,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.6896000000000001,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.7247999999999999,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.7602000000000001,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.8138,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.8939999999999999,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.9353999999999999,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9446,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.671,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7164,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7598,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8008,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8652,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.9052000000000001,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.6896,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.6998,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.7337999999999999,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.7734,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.8450000000000001,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.914,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.9111999999999998,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6921999999999999,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7430000000000001,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7534000000000001,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.7969999999999999,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.842,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.8894,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.6148,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.6784,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.7488,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.7846,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.8274000000000001,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.8766,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.9012,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.6056,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.64,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.6658,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.7574,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.8156000000000001,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8622,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.6586000000000001,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.6676000000000001,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.69,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.7268,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.7798,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.8476000000000001,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.981,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.673,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.724,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.764,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.827,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.848,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.933,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.651,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.746,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.753,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.766,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.844,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.928,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9693999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.662,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.7014,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.7582,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.8030000000000002,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.8614,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9270000000000002,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.644,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.6848,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.7494,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.8214,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.8462,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.8958,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.94475,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.6387499999999999,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.7837500000000001,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.8280000000000001,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.8732500000000001,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.89775,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9295,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.67675,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.7002499999999999,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.7197500000000001,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.7605000000000001,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.8260000000000001,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.8759999999999999,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9992000000000001,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.6450000000000001,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.7792,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.8968,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9608000000000001,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.991,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9974000000000001,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.7173999999999999,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.7512,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.974,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.9960000000000001,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.9972,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.9986,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_12/width_1m/average_l0_58",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_12_width_1m_average_l0_73_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "65890e24-6a95-4d11-ae3e-8dde56f2a20a",
|
| 33 |
+
"datetime_epoch_millis": 1732853022963,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.95139375,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6537937500000001,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.7223437500000001,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.7772000000000001,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.83160625,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.8772187499999999,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.9226875000000001,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.6329749999999998,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.6819125,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.7779874999999999,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.83498125,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.86626875,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.9184249999999999,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9598000000000001,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.6428,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.6910000000000001,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.7916000000000001,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.8336,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.8968,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9378,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.6364,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.6538,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.807,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.8482,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.8901999999999999,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.9384,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9446,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.671,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7164,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7598,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8008,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8652,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.9052000000000001,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.6831999999999999,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.6798,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.7372,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.7908,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.8408000000000001,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.9088,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.9111999999999998,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6921999999999999,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7430000000000001,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7534000000000001,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.7969999999999999,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.842,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.8894,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.5846,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.6799999999999999,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.7696000000000001,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.8244,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.845,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.883,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.9012,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.6056,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.64,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.6658,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.7574,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.8156000000000001,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8622,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.6584000000000001,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.6689999999999999,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.7132,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.7718,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.7928,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.8628,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.981,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.673,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.724,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.764,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.827,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.848,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.933,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.587,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.671,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.744,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.83,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.8454999999999999,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.954,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9693999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.662,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.7014,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.7582,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.8030000000000002,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.8614,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9270000000000002,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.6136,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.6394,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.7642,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.8612,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.8912000000000001,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.9178000000000001,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.94475,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.6387499999999999,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.7837500000000001,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.8280000000000001,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.8732500000000001,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.89775,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9295,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.597,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.6405,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.7105,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.75725,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.82725,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.884,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9992000000000001,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.6450000000000001,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.7792,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.8968,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9608000000000001,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.991,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9974000000000001,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.7036,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.8218,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.9782,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.9962,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.9974000000000001,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.9986,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_12/width_1m/average_l0_73",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_157_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "a0f7e8d4-4104-4345-ab49-a49e5fc52f80",
|
| 33 |
+
"datetime_epoch_millis": 1732833864431,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.9516937499999999,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6981875,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.75983125,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.8185562500000001,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.8697250000000001,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.90351875,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.9333812499999998,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.8296812500000001,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.8718250000000001,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.90261875,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.92265,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.9360562499999999,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.9496875,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9606,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.658,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.7243999999999999,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.805,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.868,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.9146000000000001,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9518000000000001,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.8672000000000001,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.8992000000000001,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.9145999999999999,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.9380000000000001,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.9537999999999999,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.9650000000000001,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9471999999999999,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.6699999999999999,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7036,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7646,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8288,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8764,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.9196,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.7575999999999999,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.8234,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.898,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.9202,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.9384,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.95,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.907,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6832,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7338,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7846,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.8426,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.8772,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.9033999999999999,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.7668,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.837,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.859,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.8842000000000001,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.908,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.9221999999999999,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.9162000000000001,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.6135999999999999,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.7056,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.7588000000000001,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.8166,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.8615999999999999,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8804000000000001,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.7752000000000001,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.7988,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.8326,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.8626000000000001,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.8937999999999999,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.9098,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.9704999999999999,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.697,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.743,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.79,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.86,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.88,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.942,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.884,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.914,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.915,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.919,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.922,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.963,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9693999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.6354,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.6962000000000002,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.8066000000000001,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.867,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.9179999999999999,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9391999999999999,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.7533999999999998,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.8600000000000001,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.9221999999999999,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.9414,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.9527999999999999,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.9581999999999999,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.9432500000000001,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.6905,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.78625,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.8462500000000001,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.876,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.90075,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9312499999999999,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.84325,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.853,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.8807499999999999,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.917,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.92025,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.9305000000000001,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9994,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.9378,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.9858,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.9926,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9987999999999999,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.9996,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9994,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.9899999999999999,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.9892,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.9987999999999999,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.9987999999999999,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.9994,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.9987999999999999,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_19/width_1m/average_l0_157",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_16_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "bc79fe88-fbc8-400e-8888-fdd2694a20d5",
|
| 33 |
+
"datetime_epoch_millis": 1732839049411,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.9516937499999999,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6981875,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.75983125,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.8185562500000001,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.8697250000000001,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.90351875,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.9333812499999998,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.6864812499999999,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.74515625,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.8041625,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.8599312500000001,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.89705,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.92021875,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9606,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.658,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.7243999999999999,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.805,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.868,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.9146000000000001,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9518000000000001,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.6801999999999999,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.7605999999999999,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.8501999999999998,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.8928,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.9294,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.9443999999999999,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9471999999999999,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.6699999999999999,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7036,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7646,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8288,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8764,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.9196,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.6921999999999999,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.6892,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.7651999999999999,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.8548000000000002,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.9039999999999999,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.9269999999999999,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.907,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6832,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7338,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7846,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.8426,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.8772,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.9033999999999999,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.6818000000000002,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.7163999999999999,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.7664,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.8272,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.8741999999999999,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.8939999999999999,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.9162000000000001,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.6135999999999999,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.7056,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.7588000000000001,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.8166,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.8615999999999999,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8804000000000001,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.6416000000000001,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.658,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.6986000000000001,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.7554000000000001,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.8156000000000001,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.8682000000000001,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.9704999999999999,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.697,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.743,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.79,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.86,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.88,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.942,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.585,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.843,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.885,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.896,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.9115,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.934,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9693999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.6354,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.6962000000000002,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.8066000000000001,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.867,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.9179999999999999,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9391999999999999,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.5835999999999999,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.6614,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.7656,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.8603999999999999,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.9099999999999999,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.923,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.9432500000000001,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.6905,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.78625,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.8462500000000001,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.876,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.90075,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9312499999999999,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.64525,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.64825,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.7115,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.79525,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.8335000000000001,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.8727499999999999,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9994,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.9378,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.9858,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.9926,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9987999999999999,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.9996,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9994,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.9822,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.9843999999999999,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.9907999999999999,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.9975999999999999,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.9982,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.9984,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_19/width_1m/average_l0_16",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_18_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "13b2d7c7-909d-4ba7-bab6-36ba48eb617a",
|
| 33 |
+
"datetime_epoch_millis": 1732843634197,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.9516937499999999,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6981875,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.75983125,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.8185562500000001,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.8697250000000001,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.90351875,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.9333812499999998,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.73201875,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.7801374999999999,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.8549937499999999,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.88908125,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.91265625,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.9333875,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9606,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.658,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.7243999999999999,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.805,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.868,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.9146000000000001,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9518000000000001,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.7756000000000001,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.8044,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.867,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.8926000000000001,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.9373999999999999,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.9507999999999999,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9471999999999999,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.6699999999999999,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7036,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7646,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8288,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8764,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.9196,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.6831999999999999,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.7766,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.8854,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.9072000000000001,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.9278000000000001,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.9339999999999999,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.907,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6832,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7338,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7846,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.8426,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.8772,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.9033999999999999,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.704,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.7362,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.8134,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.8583999999999999,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.8792,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.8986000000000001,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.9162000000000001,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.6135999999999999,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.7056,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.7588000000000001,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.8166,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.8615999999999999,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8804000000000001,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.6970000000000001,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.7036,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.7773999999999999,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.8321999999999999,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.865,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.8987999999999999,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.9704999999999999,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.697,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.743,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.79,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.86,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.88,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.942,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.82,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.886,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.904,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.906,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.917,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.955,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9693999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.6354,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.6962000000000002,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.8066000000000001,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.867,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.9179999999999999,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9391999999999999,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.5846,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.6918,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.8532,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.8960000000000001,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.9148,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.9358000000000001,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.9432500000000001,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.6905,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.78625,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.8462500000000001,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.876,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.90075,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9312499999999999,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.66875,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.6745,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.7417499999999999,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.82225,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.86125,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.8955,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9994,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.9378,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.9858,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.9926,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9987999999999999,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.9996,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9994,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.9229999999999998,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.968,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.9978,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.998,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.9987999999999999,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.9986,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_19/width_1m/average_l0_18",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_29_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "2268e811-51e5-4e6c-b331-a5e8437b327e",
|
| 33 |
+
"datetime_epoch_millis": 1732848202890,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.9516937499999999,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6981875,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.75983125,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.8185562500000001,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.8697250000000001,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.90351875,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.9333812499999998,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.7340749999999999,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.78225,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.84896875,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.89109375,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.923075,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.94041875,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9606,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.658,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.7243999999999999,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.805,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.868,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.9146000000000001,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9518000000000001,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.7854,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.8324,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.8870000000000001,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.9067999999999999,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.9461999999999999,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.9566000000000001,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9471999999999999,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.6699999999999999,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7036,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7646,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8288,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8764,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.9196,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.6828000000000001,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.7853999999999999,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.8699999999999999,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.9190000000000002,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.9358000000000001,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.943,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.907,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6832,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7338,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7846,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.8426,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.8772,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.9033999999999999,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.7110000000000001,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.7348,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.8061999999999999,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.869,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.8862,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.9057999999999999,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.9162000000000001,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.6135999999999999,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.7056,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.7588000000000001,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.8166,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.8615999999999999,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8804000000000001,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.712,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.741,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.8145999999999999,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.8350000000000002,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.8804000000000001,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.909,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.9704999999999999,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.697,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.743,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.79,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.86,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.88,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.942,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.651,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.71,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.824,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.874,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.924,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.955,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9693999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.6354,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.6962000000000002,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.8066000000000001,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.867,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.9179999999999999,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9391999999999999,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.6516,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.7534000000000001,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.8448,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.9046000000000001,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.9364000000000001,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.9461999999999999,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.9432500000000001,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.6905,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.78625,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.8462500000000001,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.876,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.90075,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9312499999999999,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.696,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.718,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.7557499999999999,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.82275,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.877,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.9087500000000001,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9994,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.9378,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.9858,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.9926,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9987999999999999,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.9996,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9994,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.9827999999999999,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.983,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.9894000000000001,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.9975999999999999,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.9986,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.999,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_19/width_1m/average_l0_29",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_50_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "bff5b485-3fb1-4f7a-a457-0b85a7a95907",
|
| 33 |
+
"datetime_epoch_millis": 1732852771682,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.9516937499999999,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6981875,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.75983125,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.8185562500000001,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.8697250000000001,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.90351875,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.9333812499999998,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.7520562500000001,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.7858437499999998,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.8401437500000001,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.90544375,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.92734375,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.94395625,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9606,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.658,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.7243999999999999,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.805,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.868,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.9146000000000001,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9518000000000001,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.8109999999999999,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.8532,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.9039999999999999,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.929,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.9490000000000001,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.9632,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9471999999999999,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.6699999999999999,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7036,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7646,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8288,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8764,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.9196,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.6864000000000001,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.7882,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.8964000000000001,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.9228,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.9426,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.943,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.907,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6832,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7338,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7846,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.8426,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.8772,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.9033999999999999,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.7056000000000001,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.7405999999999999,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.8200000000000001,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.8782,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.8986000000000001,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.9172,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.9162000000000001,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.6135999999999999,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.7056,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.7588000000000001,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.8166,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.8615999999999999,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8804000000000001,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.7278,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.7718,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.8114000000000001,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.8616000000000001,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.8896000000000001,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.9141999999999999,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.9704999999999999,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.697,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.743,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.79,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.86,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.88,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.942,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.584,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.591,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.661,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.887,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.9215,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.953,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9693999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.6354,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.6962000000000002,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.8066000000000001,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.867,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.9179999999999999,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9391999999999999,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.7981999999999999,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.8134,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.8737999999999999,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.9258000000000001,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.9381999999999999,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.9516,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.9432500000000001,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.6905,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.78625,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.8462500000000001,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.876,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.90075,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9312499999999999,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.71625,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.74175,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.76475,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.84075,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.88025,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.91025,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9994,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.9378,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.9858,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.9926,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9987999999999999,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.9996,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9994,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.9872,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.9867999999999999,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.9898,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.9984,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.999,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.9992000000000001,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_19/width_1m/average_l0_50",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_19_width_1m_average_l0_88_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "fb9a6b9b-9594-4e25-86f7-c128409fbb8d",
|
| 33 |
+
"datetime_epoch_millis": 1732856564628,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.9516937499999999,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6981875,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.75983125,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.8185562500000001,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.8697250000000001,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.90351875,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.9333812499999998,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.795825,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.8353687500000001,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.88688125,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.9121375,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.9344312499999999,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.9482125,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9606,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.658,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.7243999999999999,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.805,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.868,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.9146000000000001,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9518000000000001,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.8158,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.8602000000000001,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.9062000000000001,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.9358000000000001,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.9541999999999999,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.961,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9471999999999999,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.6699999999999999,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7036,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7646,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8288,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8764,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.9196,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.6886,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.7384000000000001,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.898,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.9274000000000001,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.9422,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.943,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.907,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6832,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7338,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7846,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.8426,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.8772,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.9033999999999999,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.7384000000000001,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.8106,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.8615999999999999,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.8918000000000001,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.9024000000000001,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.923,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.9162000000000001,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.6135999999999999,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.7056,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.7588000000000001,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.8166,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.8615999999999999,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8804000000000001,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.7586,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.7902,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.8343999999999999,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.859,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.8931999999999999,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.915,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.9704999999999999,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.697,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.743,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.79,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.86,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.88,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.942,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.882,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.885,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.886,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.896,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.95,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.965,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9693999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.6354,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.6962000000000002,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.8066000000000001,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.867,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.9179999999999999,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9391999999999999,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.7538,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.8208,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.8884000000000001,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.9277999999999998,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.9405999999999999,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.9570000000000001,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.9432500000000001,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.6905,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.78625,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.8462500000000001,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.876,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.90075,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9312499999999999,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.74,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.78775,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.82925,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.8605,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.8932499999999999,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.9225,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9994,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.9378,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.9858,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.9926,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9987999999999999,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.9996,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9994,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.9894000000000001,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.9899999999999999,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.9911999999999999,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.9987999999999999,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.9996,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.9992000000000001,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_19/width_1m/average_l0_88",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_114_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "d072f69e-c31b-4695-af06-66ac05b6e278",
|
| 33 |
+
"datetime_epoch_millis": 1732785615378,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.9402125,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6785437499999999,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.727125,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.7773812499999999,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.82091875,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.8598187500000002,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.90058125,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.6874375,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.74446875,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.8248375000000001,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.8701937500000001,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.8994624999999999,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.919625,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9581999999999999,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.6644,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.6841999999999999,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.7525999999999999,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.8286000000000001,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.8592000000000001,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9119999999999999,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.6806000000000001,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.765,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.8618,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.908,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.9283999999999999,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.943,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9416,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.6638,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7262000000000001,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7604,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8001999999999999,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8406,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.8902000000000001,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.6592,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.7036,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.8036,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.8573999999999999,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.8867999999999998,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.9164,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.9036,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6784000000000001,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7112,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7512000000000001,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.7872,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.8202,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.8676,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.6022000000000001,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.7451999999999999,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.7792000000000001,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.8428000000000001,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.8677999999999999,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.8855999999999999,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.889,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.614,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.6384,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.6812,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.7150000000000001,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.7776,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8318,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.6442,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.6744,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.7736000000000001,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.8109999999999999,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.8472000000000002,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.874,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.9255,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.63,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.684,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.736,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.767,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.8,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.854,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.603,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.669,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.821,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.871,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.894,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.9,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9743999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.6656000000000001,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.7036,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.779,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.8356,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.8981999999999999,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9346,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.6841999999999999,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.7065999999999999,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.7918000000000001,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.837,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.906,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.9416,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.9299999999999999,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.70875,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.774,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.80725,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.85275,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.8897499999999999,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9152500000000001,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.6925,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.71875,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.7705,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.8367499999999999,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.8664999999999999,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.897,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9994,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.8034000000000001,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.8954000000000001,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.9513999999999999,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9810000000000001,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.993,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9992000000000001,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.9336,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.9732,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.9972,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.9975999999999999,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.999,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.9994,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_5/width_1m/average_l0_114",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_13_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "73aed515-217b-4b3a-af7c-46ba9afd8884",
|
| 33 |
+
"datetime_epoch_millis": 1732795337429,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.9402125,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6785437499999999,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.727125,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.7773812499999999,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.82091875,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.8598187500000002,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.90058125,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.6987250000000002,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.746175,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.8057937500000001,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.83911875,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.86868125,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.897175,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9581999999999999,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.6644,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.6841999999999999,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.7525999999999999,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.8286000000000001,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.8592000000000001,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9119999999999999,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.7686,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.8552,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.8695999999999999,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.892,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.9108,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.9335999999999999,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9416,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.6638,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7262000000000001,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7604,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8001999999999999,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8406,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.8902000000000001,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.7542,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.7574,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.8140000000000001,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.8388,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.8800000000000001,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.9052,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.9036,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6784000000000001,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7112,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7512000000000001,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.7872,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.8202,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.8676,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.6852,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.7366,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.7956,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.8220000000000001,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.849,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.875,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.889,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.614,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.6384,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.6812,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.7150000000000001,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.7776,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8318,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.6344000000000001,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.6642,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.7368,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.7624,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.7986000000000001,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.8462,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.9255,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.63,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.684,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.736,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.767,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.8,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.854,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.586,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.659,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.73,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.755,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.786,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.83,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9743999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.6656000000000001,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.7036,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.779,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.8356,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.8981999999999999,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9346,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.6896,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.7030000000000001,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.8301999999999999,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.8586,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.8997999999999999,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.9216,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.9299999999999999,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.70875,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.774,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.80725,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.85275,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.8897499999999999,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9152500000000001,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.6080000000000001,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.675,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.71375,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.7957500000000001,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.83125,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.869,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9994,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.8034000000000001,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.8954000000000001,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.9513999999999999,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9810000000000001,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.993,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9992000000000001,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.8638,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.9189999999999999,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.9564,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.9884000000000001,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.994,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.9968,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_5/width_1m/average_l0_13",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_21_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "322874d6-e1ba-467e-90d3-fcbd8fcd7dd1",
|
| 33 |
+
"datetime_epoch_millis": 1732802971709,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.9402125,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6785437499999999,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.727125,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.7773812499999999,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.82091875,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.8598187500000002,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.90058125,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.6845125,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.736275,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.8109937500000001,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.8581,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.88205625,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.90725,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9581999999999999,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.6644,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.6841999999999999,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.7525999999999999,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.8286000000000001,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.8592000000000001,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9119999999999999,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.7074,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.8036,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.8714000000000001,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.8966000000000001,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.9094,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.9347999999999999,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9416,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.6638,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7262000000000001,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7604,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8001999999999999,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8406,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.8902000000000001,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.6718,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.6986,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.8295999999999999,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.869,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.8878,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.9138000000000002,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.9036,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6784000000000001,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7112,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7512000000000001,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.7872,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.8202,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.8676,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.6227999999999999,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.6536,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.8024000000000001,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.8346,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.8475999999999999,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.8752000000000001,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.889,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.614,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.6384,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.6812,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.7150000000000001,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.7776,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8318,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.6497999999999999,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.6796,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.727,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.7822,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.8214,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.8534,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.9255,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.63,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.684,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.736,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.767,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.8,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.854,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.614,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.706,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.727,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.798,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.85,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.876,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9743999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.6656000000000001,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.7036,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.779,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.8356,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.8981999999999999,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9346,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.6910000000000001,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.7106,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.8257999999999999,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.8728,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.89,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.9212,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.9299999999999999,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.70875,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.774,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.80725,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.85275,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.8897499999999999,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9152500000000001,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.6184999999999999,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.671,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.71175,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.8160000000000001,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.85325,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.885,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9994,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.8034000000000001,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.8954000000000001,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.9513999999999999,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9810000000000001,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.993,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9992000000000001,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.9007999999999999,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.9672000000000001,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.993,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.9955999999999999,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.9969999999999999,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.9986,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_5/width_1m/average_l0_21",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_36_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "5fcfca99-5508-4b9e-a8a9-a629485f7cef",
|
| 33 |
+
"datetime_epoch_millis": 1732808876689,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.9402125,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6785437499999999,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.727125,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.7773812499999999,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.82091875,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.8598187500000002,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.90058125,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.69280625,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.7446687500000001,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.820975,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.862225,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.8913125,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.9145375000000001,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9581999999999999,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.6644,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.6841999999999999,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.7525999999999999,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.8286000000000001,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.8592000000000001,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9119999999999999,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.7060000000000001,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.7510000000000001,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.8821999999999999,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.9032,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.9224,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.9412,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9416,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.6638,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7262000000000001,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7604,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8001999999999999,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8406,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.8902000000000001,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.6722000000000001,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.6986000000000001,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.8071999999999999,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.8694,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.885,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.916,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.9036,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6784000000000001,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7112,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7512000000000001,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.7872,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.8202,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.8676,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.6312,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.7078000000000001,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.8074,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.841,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.8535999999999999,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.8884000000000001,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.889,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.614,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.6384,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.6812,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.7150000000000001,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.7776,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8318,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.6666,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.6814,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.748,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.7706,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.8379999999999999,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.8692,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.9255,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.63,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.684,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.736,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.767,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.8,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.854,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.651,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.738,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.751,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.8474999999999999,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.88,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.884,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9743999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.6656000000000001,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.7036,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.779,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.8356,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.8981999999999999,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9346,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.6914,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.711,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.8301999999999999,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.868,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.8949999999999999,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.9326000000000001,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.9299999999999999,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.70875,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.774,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.80725,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.85275,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.8897499999999999,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9152500000000001,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.62925,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.67875,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.748,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.8025,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.8585,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.8865,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9994,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.8034000000000001,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.8954000000000001,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.9513999999999999,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9810000000000001,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.993,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9992000000000001,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.8948,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.9907999999999999,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.9938,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.9955999999999999,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.998,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.9984,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_5/width_1m/average_l0_36",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_63_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "2c8650d9-ebd8-42c4-bf1b-489f45624046",
|
| 33 |
+
"datetime_epoch_millis": 1732814735171,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.9402125,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6785437499999999,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.727125,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.7773812499999999,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.82091875,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.8598187500000002,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.90058125,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.6963124999999999,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.7402500000000001,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.8132187499999999,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.8723937500000001,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.8967499999999999,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.9198375,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9581999999999999,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.6644,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.6841999999999999,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.7525999999999999,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.8286000000000001,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.8592000000000001,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9119999999999999,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.6898,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.7596,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.8661999999999999,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.906,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.9244,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.942,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9416,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.6638,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7262000000000001,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7604,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8001999999999999,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8406,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.8902000000000001,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.6824,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.6802,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.7718,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.8596,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.9008,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.915,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.9036,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6784000000000001,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7112,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7512000000000001,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.7872,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.8202,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.8676,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.6622,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.7438,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.8058,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.8472,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.8565999999999999,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.8924000000000001,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.889,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.614,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.6384,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.6812,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.7150000000000001,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.7776,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8318,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.6517999999999999,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.6876000000000001,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.751,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.805,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.8486,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.8742000000000001,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.9255,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.63,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.684,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.736,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.767,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.8,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.854,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.589,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.653,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.755,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.856,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.894,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.907,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9743999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.6656000000000001,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.7036,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.779,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.8356,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.8981999999999999,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9346,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.6969999999999998,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.702,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.7989999999999999,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.8757999999999999,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.8873999999999999,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.9356,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.9299999999999999,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.70875,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.774,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.80725,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.85275,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.8897499999999999,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9152500000000001,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.6335,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.716,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.7627499999999999,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.83275,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.8639999999999999,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.8935,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9994,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.8034000000000001,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.8954000000000001,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.9513999999999999,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9810000000000001,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.993,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9992000000000001,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.9648,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.9798,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.9942,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.9968,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.9982000000000001,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.999,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_5/width_1m/average_l0_63",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|
gemma_scope/sparse_probing_1M/gemma-scope-2b-pt-res_layer_5_width_1m_average_l0_9_eval_results.json
ADDED
|
@@ -0,0 +1,215 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eval_type_id": "sparse_probing",
|
| 3 |
+
"eval_config": {
|
| 4 |
+
"random_seed": 42,
|
| 5 |
+
"dataset_names": [
|
| 6 |
+
"LabHC/bias_in_bios_class_set1",
|
| 7 |
+
"LabHC/bias_in_bios_class_set2",
|
| 8 |
+
"LabHC/bias_in_bios_class_set3",
|
| 9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
| 10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
| 11 |
+
"codeparrot/github-code",
|
| 12 |
+
"fancyzhx/ag_news",
|
| 13 |
+
"Helsinki-NLP/europarl"
|
| 14 |
+
],
|
| 15 |
+
"probe_train_set_size": 4000,
|
| 16 |
+
"probe_test_set_size": 1000,
|
| 17 |
+
"context_length": 128,
|
| 18 |
+
"sae_batch_size": 5,
|
| 19 |
+
"llm_batch_size": 32,
|
| 20 |
+
"llm_dtype": "bfloat16",
|
| 21 |
+
"model_name": "gemma-2-2b",
|
| 22 |
+
"k_values": [
|
| 23 |
+
1,
|
| 24 |
+
2,
|
| 25 |
+
5,
|
| 26 |
+
10,
|
| 27 |
+
20,
|
| 28 |
+
50
|
| 29 |
+
],
|
| 30 |
+
"lower_vram_usage": true
|
| 31 |
+
},
|
| 32 |
+
"eval_id": "fd646c98-922d-4d93-a9c7-5755ad21bde0",
|
| 33 |
+
"datetime_epoch_millis": 1732819979852,
|
| 34 |
+
"eval_result_metrics": {
|
| 35 |
+
"llm": {
|
| 36 |
+
"llm_test_accuracy": 0.9402125,
|
| 37 |
+
"llm_top_1_test_accuracy": 0.6785437499999999,
|
| 38 |
+
"llm_top_2_test_accuracy": 0.727125,
|
| 39 |
+
"llm_top_5_test_accuracy": 0.7773812499999999,
|
| 40 |
+
"llm_top_10_test_accuracy": 0.82091875,
|
| 41 |
+
"llm_top_20_test_accuracy": 0.8598187500000002,
|
| 42 |
+
"llm_top_50_test_accuracy": 0.90058125,
|
| 43 |
+
"llm_top_100_test_accuracy": null
|
| 44 |
+
},
|
| 45 |
+
"sae": {
|
| 46 |
+
"sae_test_accuracy": -1.0,
|
| 47 |
+
"sae_top_1_test_accuracy": 0.67448125,
|
| 48 |
+
"sae_top_2_test_accuracy": 0.7074374999999999,
|
| 49 |
+
"sae_top_5_test_accuracy": 0.7915562499999999,
|
| 50 |
+
"sae_top_10_test_accuracy": 0.8273875,
|
| 51 |
+
"sae_top_20_test_accuracy": 0.86016875,
|
| 52 |
+
"sae_top_50_test_accuracy": 0.890275,
|
| 53 |
+
"sae_top_100_test_accuracy": null
|
| 54 |
+
}
|
| 55 |
+
},
|
| 56 |
+
"eval_result_details": [
|
| 57 |
+
{
|
| 58 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
| 59 |
+
"llm_test_accuracy": 0.9581999999999999,
|
| 60 |
+
"llm_top_1_test_accuracy": 0.6644,
|
| 61 |
+
"llm_top_2_test_accuracy": 0.6841999999999999,
|
| 62 |
+
"llm_top_5_test_accuracy": 0.7525999999999999,
|
| 63 |
+
"llm_top_10_test_accuracy": 0.8286000000000001,
|
| 64 |
+
"llm_top_20_test_accuracy": 0.8592000000000001,
|
| 65 |
+
"llm_top_50_test_accuracy": 0.9119999999999999,
|
| 66 |
+
"llm_top_100_test_accuracy": null,
|
| 67 |
+
"sae_test_accuracy": -1.0,
|
| 68 |
+
"sae_top_1_test_accuracy": 0.7386,
|
| 69 |
+
"sae_top_2_test_accuracy": 0.8231999999999999,
|
| 70 |
+
"sae_top_5_test_accuracy": 0.8521999999999998,
|
| 71 |
+
"sae_top_10_test_accuracy": 0.8858,
|
| 72 |
+
"sae_top_20_test_accuracy": 0.9118,
|
| 73 |
+
"sae_top_50_test_accuracy": 0.9308,
|
| 74 |
+
"sae_top_100_test_accuracy": null
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
| 78 |
+
"llm_test_accuracy": 0.9416,
|
| 79 |
+
"llm_top_1_test_accuracy": 0.6638,
|
| 80 |
+
"llm_top_2_test_accuracy": 0.7262000000000001,
|
| 81 |
+
"llm_top_5_test_accuracy": 0.7604,
|
| 82 |
+
"llm_top_10_test_accuracy": 0.8001999999999999,
|
| 83 |
+
"llm_top_20_test_accuracy": 0.8406,
|
| 84 |
+
"llm_top_50_test_accuracy": 0.8902000000000001,
|
| 85 |
+
"llm_top_100_test_accuracy": null,
|
| 86 |
+
"sae_test_accuracy": -1.0,
|
| 87 |
+
"sae_top_1_test_accuracy": 0.6796,
|
| 88 |
+
"sae_top_2_test_accuracy": 0.711,
|
| 89 |
+
"sae_top_5_test_accuracy": 0.8144,
|
| 90 |
+
"sae_top_10_test_accuracy": 0.8238,
|
| 91 |
+
"sae_top_20_test_accuracy": 0.8667999999999999,
|
| 92 |
+
"sae_top_50_test_accuracy": 0.9019999999999999,
|
| 93 |
+
"sae_top_100_test_accuracy": null
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
| 97 |
+
"llm_test_accuracy": 0.9036,
|
| 98 |
+
"llm_top_1_test_accuracy": 0.6784000000000001,
|
| 99 |
+
"llm_top_2_test_accuracy": 0.7112,
|
| 100 |
+
"llm_top_5_test_accuracy": 0.7512000000000001,
|
| 101 |
+
"llm_top_10_test_accuracy": 0.7872,
|
| 102 |
+
"llm_top_20_test_accuracy": 0.8202,
|
| 103 |
+
"llm_top_50_test_accuracy": 0.8676,
|
| 104 |
+
"llm_top_100_test_accuracy": null,
|
| 105 |
+
"sae_test_accuracy": -1.0,
|
| 106 |
+
"sae_top_1_test_accuracy": 0.6648,
|
| 107 |
+
"sae_top_2_test_accuracy": 0.7464000000000001,
|
| 108 |
+
"sae_top_5_test_accuracy": 0.7956,
|
| 109 |
+
"sae_top_10_test_accuracy": 0.8394,
|
| 110 |
+
"sae_top_20_test_accuracy": 0.8539999999999999,
|
| 111 |
+
"sae_top_50_test_accuracy": 0.8784000000000001,
|
| 112 |
+
"sae_top_100_test_accuracy": null
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
| 116 |
+
"llm_test_accuracy": 0.889,
|
| 117 |
+
"llm_top_1_test_accuracy": 0.614,
|
| 118 |
+
"llm_top_2_test_accuracy": 0.6384,
|
| 119 |
+
"llm_top_5_test_accuracy": 0.6812,
|
| 120 |
+
"llm_top_10_test_accuracy": 0.7150000000000001,
|
| 121 |
+
"llm_top_20_test_accuracy": 0.7776,
|
| 122 |
+
"llm_top_50_test_accuracy": 0.8318,
|
| 123 |
+
"llm_top_100_test_accuracy": null,
|
| 124 |
+
"sae_test_accuracy": -1.0,
|
| 125 |
+
"sae_top_1_test_accuracy": 0.6334000000000001,
|
| 126 |
+
"sae_top_2_test_accuracy": 0.6494,
|
| 127 |
+
"sae_top_5_test_accuracy": 0.7124,
|
| 128 |
+
"sae_top_10_test_accuracy": 0.7562000000000001,
|
| 129 |
+
"sae_top_20_test_accuracy": 0.789,
|
| 130 |
+
"sae_top_50_test_accuracy": 0.8291999999999999,
|
| 131 |
+
"sae_top_100_test_accuracy": null
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
| 135 |
+
"llm_test_accuracy": 0.9255,
|
| 136 |
+
"llm_top_1_test_accuracy": 0.63,
|
| 137 |
+
"llm_top_2_test_accuracy": 0.684,
|
| 138 |
+
"llm_top_5_test_accuracy": 0.736,
|
| 139 |
+
"llm_top_10_test_accuracy": 0.767,
|
| 140 |
+
"llm_top_20_test_accuracy": 0.8,
|
| 141 |
+
"llm_top_50_test_accuracy": 0.854,
|
| 142 |
+
"llm_top_100_test_accuracy": null,
|
| 143 |
+
"sae_test_accuracy": -1.0,
|
| 144 |
+
"sae_top_1_test_accuracy": 0.588,
|
| 145 |
+
"sae_top_2_test_accuracy": 0.574,
|
| 146 |
+
"sae_top_5_test_accuracy": 0.678,
|
| 147 |
+
"sae_top_10_test_accuracy": 0.714,
|
| 148 |
+
"sae_top_20_test_accuracy": 0.7585,
|
| 149 |
+
"sae_top_50_test_accuracy": 0.81,
|
| 150 |
+
"sae_top_100_test_accuracy": null
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"dataset_name": "codeparrot/github-code_results",
|
| 154 |
+
"llm_test_accuracy": 0.9743999999999999,
|
| 155 |
+
"llm_top_1_test_accuracy": 0.6656000000000001,
|
| 156 |
+
"llm_top_2_test_accuracy": 0.7036,
|
| 157 |
+
"llm_top_5_test_accuracy": 0.779,
|
| 158 |
+
"llm_top_10_test_accuracy": 0.8356,
|
| 159 |
+
"llm_top_20_test_accuracy": 0.8981999999999999,
|
| 160 |
+
"llm_top_50_test_accuracy": 0.9346,
|
| 161 |
+
"llm_top_100_test_accuracy": null,
|
| 162 |
+
"sae_test_accuracy": -1.0,
|
| 163 |
+
"sae_top_1_test_accuracy": 0.6942,
|
| 164 |
+
"sae_top_2_test_accuracy": 0.7076,
|
| 165 |
+
"sae_top_5_test_accuracy": 0.8305999999999999,
|
| 166 |
+
"sae_top_10_test_accuracy": 0.8603999999999999,
|
| 167 |
+
"sae_top_20_test_accuracy": 0.8969999999999999,
|
| 168 |
+
"sae_top_50_test_accuracy": 0.9174,
|
| 169 |
+
"sae_top_100_test_accuracy": null
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
| 173 |
+
"llm_test_accuracy": 0.9299999999999999,
|
| 174 |
+
"llm_top_1_test_accuracy": 0.70875,
|
| 175 |
+
"llm_top_2_test_accuracy": 0.774,
|
| 176 |
+
"llm_top_5_test_accuracy": 0.80725,
|
| 177 |
+
"llm_top_10_test_accuracy": 0.85275,
|
| 178 |
+
"llm_top_20_test_accuracy": 0.8897499999999999,
|
| 179 |
+
"llm_top_50_test_accuracy": 0.9152500000000001,
|
| 180 |
+
"llm_top_100_test_accuracy": null,
|
| 181 |
+
"sae_test_accuracy": -1.0,
|
| 182 |
+
"sae_top_1_test_accuracy": 0.5992500000000001,
|
| 183 |
+
"sae_top_2_test_accuracy": 0.6094999999999999,
|
| 184 |
+
"sae_top_5_test_accuracy": 0.70625,
|
| 185 |
+
"sae_top_10_test_accuracy": 0.7655,
|
| 186 |
+
"sae_top_20_test_accuracy": 0.8152500000000001,
|
| 187 |
+
"sae_top_50_test_accuracy": 0.859,
|
| 188 |
+
"sae_top_100_test_accuracy": null
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
| 192 |
+
"llm_test_accuracy": 0.9994,
|
| 193 |
+
"llm_top_1_test_accuracy": 0.8034000000000001,
|
| 194 |
+
"llm_top_2_test_accuracy": 0.8954000000000001,
|
| 195 |
+
"llm_top_5_test_accuracy": 0.9513999999999999,
|
| 196 |
+
"llm_top_10_test_accuracy": 0.9810000000000001,
|
| 197 |
+
"llm_top_20_test_accuracy": 0.993,
|
| 198 |
+
"llm_top_50_test_accuracy": 0.9992000000000001,
|
| 199 |
+
"llm_top_100_test_accuracy": null,
|
| 200 |
+
"sae_test_accuracy": -1.0,
|
| 201 |
+
"sae_top_1_test_accuracy": 0.798,
|
| 202 |
+
"sae_top_2_test_accuracy": 0.8384,
|
| 203 |
+
"sae_top_5_test_accuracy": 0.943,
|
| 204 |
+
"sae_top_10_test_accuracy": 0.974,
|
| 205 |
+
"sae_top_20_test_accuracy": 0.9890000000000001,
|
| 206 |
+
"sae_top_50_test_accuracy": 0.9954000000000001,
|
| 207 |
+
"sae_top_100_test_accuracy": null
|
| 208 |
+
}
|
| 209 |
+
],
|
| 210 |
+
"sae_bench_commit_hash": "6f2e38f6481933249b70185f9d3b68737eac44a1",
|
| 211 |
+
"sae_lens_id": "layer_5/width_1m/average_l0_9",
|
| 212 |
+
"sae_lens_release_id": "gemma-scope-2b-pt-res",
|
| 213 |
+
"sae_lens_version": "4.4.5",
|
| 214 |
+
"eval_result_unstructured": null
|
| 215 |
+
}
|