File size: 1,617 Bytes
d45879a
 
 
 
 
 
 
 
927fd2f
d45879a
81e6d46
 
d45879a
 
 
927fd2f
d45879a
 
 
927fd2f
 
 
 
d45879a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
---
license: mit
library_name: transformers
pipeline_tag: text-generation
---

# Training Language Models To Explain Their Own Computations

This is a **Llama-3.1-8B-Instruct** explainer model fine-tuned for the **input ablations** task for the **Llama-3.1-8B-Instruct** target model, as described in [this paper](https://arxiv.org/abs/2511.08579). In the input ablations task, explainer models are trained to predict how removing "hint" tokens from an MMLU prompt with a hint changes the output of Llama-3.1-8B-Instruct. This helps in understanding the causal relationships between input components and model behavior.

[Repository](https://github.com/TransluceAI/introspective-interp) | 
[Paper](https://arxiv.org/abs/2511.08579)

## Sample Usage

To evaluate the explainer model on the input ablation task, you can use the evaluation script provided in the GitHub repository.

```bash
uv run --env-file .env evaluate.py \
  --config config/input_ablation/instruct_instruct_hint.yaml \
  --target_model_path meta-llama/Llama-3.1-8B-Instruct \
  --task hint_attribution \
  --model_path Transluce/input_ablation_llama3.1_8b_instruct_llama3.1_8b_instruct \
  --output_dir /PATH/TO/RESULTS/ \
  --batch_size 64
```

## Citation

```bibtex
@misc{li2025traininglanguagemodelsexplain,
      title={Training Language Models to Explain Their Own Computations}, 
      author={Belinda Z. Li and Zifan Carl Guo and Vincent Huang and Jacob Steinhardt and Jacob Andreas},
      year={2025},
      eprint={2511.08579},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2511.08579}, 
}
```