|
|
--- |
|
|
license: apache-2.0 |
|
|
datasets: |
|
|
- ILSVRC/imagenet-1k |
|
|
metrics: |
|
|
- accuracy |
|
|
language: |
|
|
- en |
|
|
tags: |
|
|
- vision |
|
|
- image-classification |
|
|
- pytorch_model_hub_mixin |
|
|
pipeline_tag: image-classification |
|
|
library_name: PyTorch |
|
|
model_index: |
|
|
- name: SpaRTAN-S |
|
|
results: |
|
|
- task: |
|
|
type: image-classification |
|
|
dataset: |
|
|
type: ILSVRC/imagenet-1k |
|
|
name: ImageNet-1k |
|
|
metrics: |
|
|
- name: top-1 accuracy |
|
|
type: accuracy |
|
|
value: 82.35 |
|
|
- name: top-5 accuracy |
|
|
type: accuracy |
|
|
value: 96.14 |
|
|
--- |
|
|
|
|
|
# SpaRTAN-S |
|
|
|
|
|
SpaRTAN is a lightweight architectural design which shows consistent efficiency and competitive performance when benchmarked against ImageNet and COCO dataset. It was introduced in the paper [SpaRTAN](https://arxiv.org/abs/2507.10999) and released in this [repo](https://github.com/henry-pay/SpaRTAN). SpaRTAN-S is a scaled-up version of SpaRTAN-T. |
|
|
|
|
|
# Model Description |
|
|
|
|
|
SpaRTAN-S shares the same configurations as SpaRTAN-T presented in the paper, [SpaRTAN](https://arxiv.org/abs/2507.10999), except the number of channels at each stage, as outlined below. |
|
|
|
|
|
| Stage | Channel | |
|
|
|:---:|:---:| |
|
|
| S1 | 64 | |
|
|
| S2 | 128 | |
|
|
| S3 | 320 | |
|
|
| S4 | 512 | |
|
|
|
|
|
# Intended Uses & Limitations |
|
|
|
|
|
You can use the raw model for image classification. Using as a feature extractor, SpaRTAN-S can be fine-tuned on various downstream tasks including object detection. |
|
|
|
|
|
# Training Procedure |
|
|
|
|
|
Same training procedure as outlined in the paper, [SpaRTAN](https://arxiv.org/abs/2507.10999), is used to train this model. |
|
|
|
|
|
# Evaluation Result |
|
|
|
|
|
| Model | Resolution | Params (M) | FLOPs (G) | Top-1 (%) | top-5 (%) | |
|
|
|:---:|:---:|:---:|:---:|:---:|:---:| |
|
|
| SpaRTAN-S | 224x224 | 18.51 | 3.86 | 82.35 | 96.14 | |
|
|
|
|
|
# Implementation |
|
|
|
|
|
Please refer to this [repo](https://github.com/henry-pay/SpaRTAN) for full implementation. |
|
|
|
|
|
# Citation |
|
|
|
|
|
```bibtex |
|
|
@inproceedings{ |
|
|
title={SpaRTAN: Spatial Reinforcement Token-based Aggregation Network for Visual Recognition}, |
|
|
author={Pay, Quan Bi and Baskaran, Vishnu Monn and Loo, Junn Yong and Wong, KokSheik and See, Simon}, |
|
|
booktitle={2025 International Joint Conference on Neural Networks (IJCNN)}, |
|
|
pages={to appear}, |
|
|
year={2025}, |
|
|
organization={IEEE}, |
|
|
note={Accepted} |
|
|
} |
|
|
``` |
|
|
|