belindazli nielsr HF Staff commited on
Commit
b8a900f
·
verified ·
1 Parent(s): 3bcdcee

Add pipeline tag and repository link (#1)

Browse files

- Add pipeline tag and repository link (46c629f89b25c13be59ee17ec906e255e8aa8523)


Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +7 -5
README.md CHANGED
@@ -1,9 +1,10 @@
1
  ---
2
- license: mit
3
- language:
4
- - en
5
  base_model:
6
  - meta-llama/Llama-3-8B
 
 
 
 
7
  ---
8
 
9
  # Model Card
@@ -12,13 +13,14 @@ This is a Llama-3-8B base model fine-tuned to explain continuous features from L
12
  This model was trained to map SAE features from Llama-3.1-8B's residual stream to their explanations derived from Neuronpedia.
13
  It generalizes to explaining any arbitrary continuous feature from Llama-3.1-8B's residual stream.
14
 
15
- See [paper](https://arxiv.org/abs/2511.08579) for more details.
 
16
 
17
  ## Usage
18
 
19
  Use the code below to get started with the model.
20
 
21
- **Note**: This model requires custom handling of continuous tokens. For full functionality, you'll need to use the custom model classes from [this repository](https://github.com/TransluceAI/introspective-interp/tree/main) that can properly embed feature vectors at the `<|reserved_special_token_12|>` tokens. The standard transformers library won't handle the continuous token embeddings correctly.
22
 
23
  ```python
24
  import torch
 
1
  ---
 
 
 
2
  base_model:
3
  - meta-llama/Llama-3-8B
4
+ language:
5
+ - en
6
+ license: mit
7
+ pipeline_tag: text-generation
8
  ---
9
 
10
  # Model Card
 
13
  This model was trained to map SAE features from Llama-3.1-8B's residual stream to their explanations derived from Neuronpedia.
14
  It generalizes to explaining any arbitrary continuous feature from Llama-3.1-8B's residual stream.
15
 
16
+ - **Paper:** [Training Language Models to Explain Their Own Computations](https://arxiv.org/abs/2511.08579)
17
+ - **Repository:** [https://github.com/TransluceAI/introspective-interp](https://github.com/TransluceAI/introspective-interp)
18
 
19
  ## Usage
20
 
21
  Use the code below to get started with the model.
22
 
23
+ **Note**: This model requires custom handling of continuous tokens. For full functionality, you'll need to use the custom model classes from [the GitHub repository](https://github.com/TransluceAI/introspective-interp/tree/main) that can properly embed feature vectors at the `<|reserved_special_token_12|>` tokens. The standard transformers library won't handle the continuous token embeddings correctly.
24
 
25
  ```python
26
  import torch