narugo1992 commited on
Commit
0793b4d
·
verified ·
1 Parent(s): c620f85

Export model 'vit_mediumd_patch16_reg4_gap_256.sbb_in12k_ft_in1k', on 2025-01-20 07:22:31 UTC

Browse files
README.md CHANGED
@@ -53,7 +53,6 @@ base_model:
53
  - timm/nextvit_large.bd_in1k
54
  - timm/nextvit_small.bd_in1k_384
55
  - timm/nextvit_small.bd_ssld_6m_in1k
56
- - timm/repghostnet_058.in1k
57
  - timm/repghostnet_080.in1k
58
  - timm/repghostnet_111.in1k
59
  - timm/repghostnet_150.in1k
@@ -99,6 +98,7 @@ base_model:
99
  - timm/vit_medium_patch16_reg4_gap_256.sbb_in12k_ft_in1k
100
  - timm/vit_mediumd_patch16_reg4_gap_256.sbb2_e200_in12k
101
  - timm/vit_mediumd_patch16_reg4_gap_256.sbb2_e200_in12k_ft_in1k
 
102
  - timm/vit_mediumd_patch16_reg4_gap_384.sbb2_e200_in12k_ft_in1k
103
  - timm/vit_small_patch14_reg4_dinov2.lvd142m
104
  language:
@@ -114,7 +114,7 @@ ONNX export version from [TIMM](https://huggingface.co/timm).
114
 
115
  # Models
116
 
117
- 485 models exported from TIMM in total.
118
 
119
  ## Beit
120
 
@@ -879,7 +879,7 @@ ONNX export version from [TIMM](https://huggingface.co/timm).
879
 
880
  ## VisionTransformer
881
 
882
- 62 models with model class `VisionTransformer`.
883
 
884
  | Name | Params | Flops | Input Size | Can Classify | Features | Classes | Dataset | Model | Architecture | Created At |
885
  |:-------------------------------------------------------------------------------------------------------------------------------------------------|:---------|:--------|-------------:|:---------------|-----------:|----------:|:------------------|:------------------|:---------------------------------|:-------------|
@@ -913,6 +913,7 @@ ONNX export version from [TIMM](https://huggingface.co/timm).
913
  | [vit_base_patch16_siglip_gap_224.webli](https://huggingface.co/timm/vit_base_patch16_siglip_gap_224.webli) | 85.6M | 16.8G | 224 | False | 768 | 768 | | VisionTransformer | vit_base_patch16_siglip_gap_224 | 2024-12-24 |
914
  | [vit_mediumd_patch16_reg4_gap_256.sbb2_e200_in12k](https://huggingface.co/timm/vit_mediumd_patch16_reg4_gap_256.sbb2_e200_in12k) | 69.5M | 16.5G | 256 | True | 512 | 11821 | imagenet-12k | VisionTransformer | vit_mediumd_patch16_reg4_gap_256 | 2024-08-14 |
915
  | [vit_mediumd_patch16_reg4_gap_256.sbb2_e200_in12k_ft_in1k](https://huggingface.co/timm/vit_mediumd_patch16_reg4_gap_256.sbb2_e200_in12k_ft_in1k) | 64.0M | 16.5G | 256 | True | 512 | 1000 | imagenet-1k | VisionTransformer | vit_mediumd_patch16_reg4_gap_256 | 2024-08-21 |
 
916
  | [vit_betwixt_patch16_reg4_gap_256.sbb_in12k_ft_in1k](https://huggingface.co/timm/vit_betwixt_patch16_reg4_gap_256.sbb_in12k_ft_in1k) | 60.2M | 15.5G | 256 | True | 640 | 1000 | imagenet-1k | VisionTransformer | vit_betwixt_patch16_reg4_gap_256 | 2024-05-10 |
917
  | [vit_betwixt_patch16_reg4_gap_256.sbb_in1k](https://huggingface.co/timm/vit_betwixt_patch16_reg4_gap_256.sbb_in1k) | 60.2M | 15.5G | 256 | True | 640 | 1000 | imagenet-1k | VisionTransformer | vit_betwixt_patch16_reg4_gap_256 | 2024-05-10 |
918
  | [vit_betwixt_patch16_reg1_gap_256.sbb_in1k](https://huggingface.co/timm/vit_betwixt_patch16_reg1_gap_256.sbb_in1k) | 60.2M | 15.3G | 256 | True | 640 | 1000 | imagenet-1k | VisionTransformer | vit_betwixt_patch16_reg1_gap_256 | 2024-05-10 |
 
53
  - timm/nextvit_large.bd_in1k
54
  - timm/nextvit_small.bd_in1k_384
55
  - timm/nextvit_small.bd_ssld_6m_in1k
 
56
  - timm/repghostnet_080.in1k
57
  - timm/repghostnet_111.in1k
58
  - timm/repghostnet_150.in1k
 
98
  - timm/vit_medium_patch16_reg4_gap_256.sbb_in12k_ft_in1k
99
  - timm/vit_mediumd_patch16_reg4_gap_256.sbb2_e200_in12k
100
  - timm/vit_mediumd_patch16_reg4_gap_256.sbb2_e200_in12k_ft_in1k
101
+ - timm/vit_mediumd_patch16_reg4_gap_256.sbb_in12k_ft_in1k
102
  - timm/vit_mediumd_patch16_reg4_gap_384.sbb2_e200_in12k_ft_in1k
103
  - timm/vit_small_patch14_reg4_dinov2.lvd142m
104
  language:
 
114
 
115
  # Models
116
 
117
+ 486 models exported from TIMM in total.
118
 
119
  ## Beit
120
 
 
879
 
880
  ## VisionTransformer
881
 
882
+ 63 models with model class `VisionTransformer`.
883
 
884
  | Name | Params | Flops | Input Size | Can Classify | Features | Classes | Dataset | Model | Architecture | Created At |
885
  |:-------------------------------------------------------------------------------------------------------------------------------------------------|:---------|:--------|-------------:|:---------------|-----------:|----------:|:------------------|:------------------|:---------------------------------|:-------------|
 
913
  | [vit_base_patch16_siglip_gap_224.webli](https://huggingface.co/timm/vit_base_patch16_siglip_gap_224.webli) | 85.6M | 16.8G | 224 | False | 768 | 768 | | VisionTransformer | vit_base_patch16_siglip_gap_224 | 2024-12-24 |
914
  | [vit_mediumd_patch16_reg4_gap_256.sbb2_e200_in12k](https://huggingface.co/timm/vit_mediumd_patch16_reg4_gap_256.sbb2_e200_in12k) | 69.5M | 16.5G | 256 | True | 512 | 11821 | imagenet-12k | VisionTransformer | vit_mediumd_patch16_reg4_gap_256 | 2024-08-14 |
915
  | [vit_mediumd_patch16_reg4_gap_256.sbb2_e200_in12k_ft_in1k](https://huggingface.co/timm/vit_mediumd_patch16_reg4_gap_256.sbb2_e200_in12k_ft_in1k) | 64.0M | 16.5G | 256 | True | 512 | 1000 | imagenet-1k | VisionTransformer | vit_mediumd_patch16_reg4_gap_256 | 2024-08-21 |
916
+ | [vit_mediumd_patch16_reg4_gap_256.sbb_in12k_ft_in1k](https://huggingface.co/timm/vit_mediumd_patch16_reg4_gap_256.sbb_in12k_ft_in1k) | 64.0M | 16.5G | 256 | True | 512 | 1000 | imagenet-1k | VisionTransformer | vit_mediumd_patch16_reg4_gap_256 | 2024-05-10 |
917
  | [vit_betwixt_patch16_reg4_gap_256.sbb_in12k_ft_in1k](https://huggingface.co/timm/vit_betwixt_patch16_reg4_gap_256.sbb_in12k_ft_in1k) | 60.2M | 15.5G | 256 | True | 640 | 1000 | imagenet-1k | VisionTransformer | vit_betwixt_patch16_reg4_gap_256 | 2024-05-10 |
918
  | [vit_betwixt_patch16_reg4_gap_256.sbb_in1k](https://huggingface.co/timm/vit_betwixt_patch16_reg4_gap_256.sbb_in1k) | 60.2M | 15.5G | 256 | True | 640 | 1000 | imagenet-1k | VisionTransformer | vit_betwixt_patch16_reg4_gap_256 | 2024-05-10 |
919
  | [vit_betwixt_patch16_reg1_gap_256.sbb_in1k](https://huggingface.co/timm/vit_betwixt_patch16_reg1_gap_256.sbb_in1k) | 60.2M | 15.3G | 256 | True | 640 | 1000 | imagenet-1k | VisionTransformer | vit_betwixt_patch16_reg1_gap_256 | 2024-05-10 |
models.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2536f00e902540de3e4d1eccb32dfe9f8d31a57e62a2cd4a08c0410a9f9967d9
3
- size 35108
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:181abec6ab39e4b7cb91bb9d310eae8478cd77d898a20083a4de8f309ce19482
3
+ size 35133
vit_mediumd_patch16_reg4_gap_256.sbb_in12k_ft_in1k/meta.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b718f90a0d5962f52331a2a80d927c3b1e7df462b63cf4bb1b976ec0fdf7863
3
+ size 169887
vit_mediumd_patch16_reg4_gap_256.sbb_in12k_ft_in1k/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92b535309329f20ba3c828ba1a10223a3b1458228f2b43ae5efe091974610797
3
+ size 256732827
vit_mediumd_patch16_reg4_gap_256.sbb_in12k_ft_in1k/preprocess.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4c3ceed99ae49aa023c8c7cbc224be5ba023f409004ac0091862f1e398d784e
3
+ size 642