| --- |
| language: en |
| license: apache-2.0 |
| source_code: https://github.com/pwesp/sail |
| tags: |
| - sparse-autoencoder |
| - matryoshka |
| - ct |
| - mri |
| --- |
| |
| # SAIL — Pretrained SAE Weights |
|
|
| Pretrained Matryoshka Sparse Autoencoder (SAE) weights for the [SAIL](https://github.com/pwesp/sail) repository. See the project page for the full pipeline and usage instructions. |
|
|
| Two checkpoints are provided, one for each foundation model (FM) embedding space: |
|
|
| | File | Foundation model | Input dim | Dictionary sizes | k values | |
| |------|-----------------|-----------|-----------------|----------| |
| | `biomedparse_sae.ckpt` | BiomedParse | 1536 | 128, 512, 2048, 8192 | 20, 40, 80, 160 | |
| | `dinov3_sae.ckpt` | DINOv3 | 1024 | 128, 512, 2048, 8192 | 5, 10, 20, 40 | |
|
|
| Both SAEs were trained on CT and MRI embeddings from the [TotalSegmentator](https://github.com/wasserth/TotalSegmentator) dataset. |
|
|
| ## Usage |
|
|
| To download these weights and place them in the expected directory structure, run from the repo root: |
|
|
| ```bash |
| bash pretrained/download_weights.sh |
| ``` |
|
|
| ## Citation |
|
|
| If you find this work useful, please cite our paper: |
|
|
| ```bibtex |
| @misc{sail2026, |
| title = {Sparse Autoencoders for Interpretable Medical Image Representation Learning}, |
| author = {Wesp, Philipp and Holland, Robbie and Sideri-Lampretsa, Vasiliki and Gatidis, Sergios}, |
| year = 2026, |
| journal = {arXiv.org}, |
| howpublished = {https://arxiv.org/abs/2603.23794v1} |
| } |
| ``` |