gpt2 SAEs
Collection
13 items
•
Updated
TopK SAE trained bewteen each layer for GPT-2
gpt2_defaults = {
"data_dir": "data/fineweb_edu_10b",
"eval_interval": 250,
"eval_steps": 100,
"batch_size": 1,
"gradient_accumulation_steps": 32 // 1,
"learning_rate": 5e-4,
"warmup_steps": 750,
"max_steps": 7500,
"decay_lr": True,
"min_lr": 1e-4,
}
SAEConfig("topk.tblock.gpt2",
gpt_config = gpt_options['gpt2'],
n_features=tuple(768 * n for n in (32,)*13),
sae_variant=SAEVariant.TOPK,
top_k = (32,) * 13,
sae_keys=gen_sae_keys(n_features=13, loc="standard"),
),