Improve model card: Add PTQ4VM paper, pipeline tag, and library

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +14 -3
README.md CHANGED
@@ -1,13 +1,15 @@
1
  ---
2
  license: apache-2.0
 
 
3
  ---
4
 
5
-
6
-
7
  <br>
8
 
9
  # Vim Model Card
10
 
 
 
11
  ## Model Details
12
 
13
  Vision Mamba (Vim) is a generic backbone trained on the ImageNet-1K dataset for vision tasks.
@@ -16,6 +18,8 @@ Vision Mamba (Vim) is a generic backbone trained on the ImageNet-1K dataset for
16
  - **Model type:** A generic vision backbone based on the bidirectional state space model (SSM) architecture.
17
  - **License:** Non-commercial license
18
 
 
 
19
 
20
  ### Model Sources
21
 
@@ -55,4 +59,11 @@ Vim-tiny is evaluated on ImageNet-1K val set, and achieves 76.1% Top-1 Acc. By f
55
  }
56
  ```
57
 
58
-
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ pipeline_tag: image-classification
4
+ library_name: pytorch
5
  ---
6
 
 
 
7
  <br>
8
 
9
  # Vim Model Card
10
 
11
+ This repository contains the model based on the paper [PTQ4VM: Post-Training Quantization for Visual Mamba](https://huggingface.co/papers/2412.20386).
12
+
13
  ## Model Details
14
 
15
  Vision Mamba (Vim) is a generic backbone trained on the ImageNet-1K dataset for vision tasks.
 
18
  - **Model type:** A generic vision backbone based on the bidirectional state space model (SSM) architecture.
19
  - **License:** Non-commercial license
20
 
21
+ ### Github repository:
22
+ https://github.com/YoungHyun197/ptq4vm
23
 
24
  ### Model Sources
25
 
 
59
  }
60
  ```
61
 
62
+ ```
63
+ @article{cho2024ptq4vm,
64
+ title={PTQ4VM: Post-Training Quantization for Visual Mamba},
65
+ author={Cho, Younghyun and Lee, Changhun and Kim, Seonggon and Park, Eunhyeok},
66
+ journal={arXiv preprint arXiv:2412.20386},
67
+ year={2024}
68
+ }
69
+ ```