| --- |
| license: mit |
| tags: |
| - kornia |
| - image-classification |
| - backbone |
| --- |
| |
| # kornia/tiny_vit |
| |
| Pretrained weights for **TinyViT**, |
| used as the encoder backbone in |
| [`kornia.models.SegmentAnything`](https://kornia.readthedocs.io/en/latest/models.html) |
| (MobileSAM) and available via |
| [`kornia.models.TinyViT`](https://kornia.readthedocs.io/en/latest/models.html). |
| |
| TinyViT is a small Vision Transformer trained with knowledge distillation from large |
| teacher models on ImageNet-22K. ECCV 2022. |
| |
| **Original repo:** [microsoft/Cream/TinyViT](https://github.com/microsoft/Cream/tree/main/TinyViT) |
| |
| ## Weights |
| |
| | File | Params | Pre-training | Fine-tuning | |
| |------|--------|-------------|-------------| |
| | `tiny_vit_5m_22k_distill.pth` | 5M | ImageNet-22K | — | |
| | `tiny_vit_5m_22kto1k_distill.pth` | 5M | ImageNet-22K | ImageNet-1K 224 | |
| | `tiny_vit_11m_22k_distill.pth` | 11M | ImageNet-22K | — | |
| | `tiny_vit_11m_22kto1k_distill.pth` | 11M | ImageNet-22K | ImageNet-1K 224 | |
| | `tiny_vit_21m_22k_distill.pth` | 21M | ImageNet-22K | — | |
| | `tiny_vit_21m_22kto1k_distill.pth` | 21M | ImageNet-22K | ImageNet-1K 224 | |
| | `tiny_vit_21m_22kto1k_384_distill.pth` | 21M | ImageNet-22K | ImageNet-1K 384 | |
| | `tiny_vit_21m_22kto1k_512_distill.pth` | 21M | ImageNet-22K | ImageNet-1K 512 | |
|
|
| ## Citation |
|
|
| ```bibtex |
| @inproceedings{wu2022tinyvit, |
| title = {{TinyViT}: Fast Pretraining Distillation for Small Vision Transformers}, |
| author = {Wu, Kan and Zhang, Jinnian and Peng, Houwen and Liu, Mengchen |
| and Xiao, Bin and Fu, Jianlong and Yuan, Lu}, |
| booktitle = {ECCV}, |
| year = {2022} |
| } |
| ``` |
|
|