nielsr HF Staff commited on
Commit
4f5ad11
·
verified ·
1 Parent(s): f2785ff

Add pipeline tag and GitHub link

Browse files

Hi, I'm Niels from the community science team at Hugging Face.

This PR adds the `text-generation` pipeline tag to the model metadata to make it more discoverable on the Hub. I've also added explicit links to the paper and the official GitHub repository. Providing the repository link is particularly important for this model, as it requires the custom `ContinuousLlama` class provided there to correctly handle continuous token embeddings.

Best,
Niels

Files changed (1) hide show
  1. README.md +7 -5
README.md CHANGED
@@ -1,9 +1,10 @@
1
  ---
2
- license: mit
3
- language:
4
- - en
5
  base_model:
6
  - meta-llama/Llama-3.1-8B
 
 
 
 
7
  ---
8
 
9
  # Model Card
@@ -12,7 +13,8 @@ This is a Llama-3.1-8B base model fine-tuned to explain continuous features from
12
  This model was trained to map SAE features from Llama-3.1-8B's residual stream to their explanations derived from Neuronpedia.
13
  It generalizes to explaining any arbitrary continuous feature from Llama-3.1-8B's residual stream.
14
 
15
- See [paper](https://arxiv.org/abs/2511.08579) for more details.
 
16
 
17
  ## Usage
18
 
@@ -25,7 +27,7 @@ import torch
25
  import numpy as np
26
  from transformers import AutoTokenizer
27
 
28
- # Load the continuous model class
29
  from model.continuous_llama import ContinuousLlama
30
 
31
  # Load the model and tokenizer
 
1
  ---
 
 
 
2
  base_model:
3
  - meta-llama/Llama-3.1-8B
4
+ language:
5
+ - en
6
+ license: mit
7
+ pipeline_tag: text-generation
8
  ---
9
 
10
  # Model Card
 
13
  This model was trained to map SAE features from Llama-3.1-8B's residual stream to their explanations derived from Neuronpedia.
14
  It generalizes to explaining any arbitrary continuous feature from Llama-3.1-8B's residual stream.
15
 
16
+ - **Paper:** [Training Language Models to Explain Their Own Computations](https://arxiv.org/abs/2511.08579)
17
+ - **Repository:** [https://github.com/TransluceAI/introspective-interp](https://github.com/TransluceAI/introspective-interp)
18
 
19
  ## Usage
20
 
 
27
  import numpy as np
28
  from transformers import AutoTokenizer
29
 
30
+ # Load the continuous model class (requires code from the GitHub repo)
31
  from model.continuous_llama import ContinuousLlama
32
 
33
  # Load the model and tokenizer