Add pipeline tag and GitHub link
Browse filesHi, I'm Niels from the community science team at Hugging Face.
This PR adds the `text-generation` pipeline tag to the model metadata to make it more discoverable on the Hub. I've also added explicit links to the paper and the official GitHub repository. Providing the repository link is particularly important for this model, as it requires the custom `ContinuousLlama` class provided there to correctly handle continuous token embeddings.
Best,
Niels
README.md
CHANGED
|
@@ -1,9 +1,10 @@
|
|
| 1 |
---
|
| 2 |
-
license: mit
|
| 3 |
-
language:
|
| 4 |
-
- en
|
| 5 |
base_model:
|
| 6 |
- meta-llama/Llama-3.1-8B
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
---
|
| 8 |
|
| 9 |
# Model Card
|
|
@@ -12,7 +13,8 @@ This is a Llama-3.1-8B base model fine-tuned to explain continuous features from
|
|
| 12 |
This model was trained to map SAE features from Llama-3.1-8B's residual stream to their explanations derived from Neuronpedia.
|
| 13 |
It generalizes to explaining any arbitrary continuous feature from Llama-3.1-8B's residual stream.
|
| 14 |
|
| 15 |
-
|
|
|
|
| 16 |
|
| 17 |
## Usage
|
| 18 |
|
|
@@ -25,7 +27,7 @@ import torch
|
|
| 25 |
import numpy as np
|
| 26 |
from transformers import AutoTokenizer
|
| 27 |
|
| 28 |
-
# Load the continuous model class
|
| 29 |
from model.continuous_llama import ContinuousLlama
|
| 30 |
|
| 31 |
# Load the model and tokenizer
|
|
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
| 2 |
base_model:
|
| 3 |
- meta-llama/Llama-3.1-8B
|
| 4 |
+
language:
|
| 5 |
+
- en
|
| 6 |
+
license: mit
|
| 7 |
+
pipeline_tag: text-generation
|
| 8 |
---
|
| 9 |
|
| 10 |
# Model Card
|
|
|
|
| 13 |
This model was trained to map SAE features from Llama-3.1-8B's residual stream to their explanations derived from Neuronpedia.
|
| 14 |
It generalizes to explaining any arbitrary continuous feature from Llama-3.1-8B's residual stream.
|
| 15 |
|
| 16 |
+
- **Paper:** [Training Language Models to Explain Their Own Computations](https://arxiv.org/abs/2511.08579)
|
| 17 |
+
- **Repository:** [https://github.com/TransluceAI/introspective-interp](https://github.com/TransluceAI/introspective-interp)
|
| 18 |
|
| 19 |
## Usage
|
| 20 |
|
|
|
|
| 27 |
import numpy as np
|
| 28 |
from transformers import AutoTokenizer
|
| 29 |
|
| 30 |
+
# Load the continuous model class (requires code from the GitHub repo)
|
| 31 |
from model.continuous_llama import ContinuousLlama
|
| 32 |
|
| 33 |
# Load the model and tokenizer
|