henry-pay commited on
Commit
5a6b483
·
1 Parent(s): ae86706

Add model

Browse files
Files changed (3) hide show
  1. README.md +77 -0
  2. config.json +23 -0
  3. model.safetensors +3 -0
README.md ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - ILSVRC/imagenet-1k
5
+ metrics:
6
+ - accuracy
7
+ language:
8
+ - en
9
+ tags:
10
+ - vision
11
+ - image-classification
12
+ - pytorch_model_hub_mixin
13
+ pipeline_tag: image-classification
14
+ library_name: PyTorch
15
+ model_index:
16
+ - name: SpaRTAN-S
17
+ results:
18
+ - task:
19
+ type: image-classification
20
+ dataset:
21
+ type: ILSVRC/imagenet-1k
22
+ name: ImageNet-1k
23
+ metrics:
24
+ - name: top-1 accuracy
25
+ type: accuracy
26
+ value: 82.35
27
+ - name: top-5 accuracy
28
+ type: accuracy
29
+ value: 96.14
30
+ ---
31
+
32
+ # SpaRTAN-S
33
+
34
+ SpaRTAN is a lightweight architectural design which shows consistent efficiency and competitive performance when benchmarked against ImageNet and COCO dataset. It was introduced in the paper [SpaRTAN](https://arxiv.org/abs/2507.10999) and released in this [repo](https://github.com/henry-pay/SpaRTAN). SpaRTAN-S is a scaled-up version of SpaRTAN-T.
35
+
36
+ # Model Description
37
+
38
+ SpaRTAN-S shares the same configurations as SpaRTAN-T presented in the paper, [SpaRTAN](https://arxiv.org/abs/2507.10999), except the number of channels at each stage, as outlined below.
39
+
40
+ | Stage | Channel |
41
+ |:---:|:---:|
42
+ | S1 | 64 |
43
+ | S2 | 128 |
44
+ | S3 | 320 |
45
+ | S4 | 512 |
46
+
47
+ # Intended Uses & Limitations
48
+
49
+ You can use the raw model for image classification. Using as a feature extractor, SpaRTAN-S can be fine-tuned on various downstream tasks including object detection.
50
+
51
+ # Training Procedure
52
+
53
+ Same training procedure as outlined in the paper, [SpaRTAN](https://arxiv.org/abs/2507.10999), is used to train this model.
54
+
55
+ # Evaluation Result
56
+
57
+ | Model | Resolution | Params (M) | FLOPs (G) | Top-1 (%) | top-5 (%) |
58
+ |:---:|:---:|:---:|:---:|:---:|:---:|
59
+ | SpaRTAN-S | 224x224 | 18.51 | 3.86 | 82.35 | 96.14 |
60
+
61
+ # Implementation
62
+
63
+ Please refer to this [repo](https://github.com/henry-pay/SpaRTAN) for full implementation.
64
+
65
+ # Citation
66
+
67
+ ```bibtex
68
+ @inproceedings{
69
+ title={SpaRTAN: Spatial Reinforcement Token-based Aggregation Network for Visual Recognition},
70
+ author={Pay, Quan Bi and Baskaran, Vishnu Monn and Loo, Junn Yong and Wong, KokSheik and See, Simon},
71
+ booktitle={2025 International Joint Conference on Neural Networks (IJCNN)},
72
+ pages={to appear},
73
+ year={2025},
74
+ organization={IEEE},
75
+ note={Accepted}
76
+ }
77
+ ```
config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dims": [
3
+ 128,
4
+ 320,
5
+ 512
6
+ ],
7
+ "dropout": 0.1,
8
+ "expand_ratios": [
9
+ 4,
10
+ 4,
11
+ 2,
12
+ 2
13
+ ],
14
+ "init_dim": 64,
15
+ "layer_depths": [
16
+ 3,
17
+ 3,
18
+ 12,
19
+ 2
20
+ ],
21
+ "num_classes": 1000,
22
+ "num_layer": 4
23
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:834c5949b4472448006e3a8500953f0df1ab8dd8ec7653214f0044e4e4fe9e14
3
+ size 74291312