100 million
Browse files
README.md
CHANGED
|
@@ -36,7 +36,7 @@ Training logs are available [via wandb](https://wandb.ai/lewington/ViT-L-14-laio
|
|
| 36 |
- Training Dataset: [Laion-2b](https://huggingface.co/datasets/laion/laion2B-multi-joined-translated-to-en)
|
| 37 |
- SAE Architecture: topk with k=32
|
| 38 |
- Layer Location: always the residual stream
|
| 39 |
-
- Training Checkpoints: every ~
|
| 40 |
- Number of features per autoencoder: 65536 (expansion factor 16)
|
| 41 |
|
| 42 |
## Usage
|
|
|
|
| 36 |
- Training Dataset: [Laion-2b](https://huggingface.co/datasets/laion/laion2B-multi-joined-translated-to-en)
|
| 37 |
- SAE Architecture: topk with k=32
|
| 38 |
- Layer Location: always the residual stream
|
| 39 |
+
- Training Checkpoints: every ~100 million tokens
|
| 40 |
- Number of features per autoencoder: 65536 (expansion factor 16)
|
| 41 |
|
| 42 |
## Usage
|