narugo1992
commited on
Export model 'vit_small_patch16_384.augreg_in1k', on 2025-01-20 04:27:01 UTC
Browse files
README.md
CHANGED
|
@@ -114,7 +114,7 @@ ONNX export version from [TIMM](https://huggingface.co/timm).
|
|
| 114 |
|
| 115 |
# Models
|
| 116 |
|
| 117 |
-
|
| 118 |
|
| 119 |
## ByobNet
|
| 120 |
|
|
@@ -453,7 +453,7 @@ ONNX export version from [TIMM](https://huggingface.co/timm).
|
|
| 453 |
|
| 454 |
## VisionTransformer
|
| 455 |
|
| 456 |
-
|
| 457 |
|
| 458 |
| Name | Params | Flops | Input Size | Can Classify | Features | Classes | Dataset | Model | Architecture | Created At |
|
| 459 |
|:-------------------------------------------------------------------------------------------------------------------------------------------------|:---------|:--------|-------------:|:---------------|-----------:|----------:|:------------|:------------------|:---------------------------------|:-------------|
|
|
@@ -467,6 +467,7 @@ ONNX export version from [TIMM](https://huggingface.co/timm).
|
|
| 467 |
| [vit_base_patch16_rpn_224.sw_in1k](https://huggingface.co/timm/vit_base_patch16_rpn_224.sw_in1k) | 86.4M | 16.8G | 224 | True | 768 | 1000 | imagenet-1k | VisionTransformer | vit_base_patch16_rpn_224 | 2022-12-22 |
|
| 468 |
| [vit_betwixt_patch16_reg1_gap_256.sbb_in1k](https://huggingface.co/timm/vit_betwixt_patch16_reg1_gap_256.sbb_in1k) | 60.2M | 15.3G | 256 | True | 640 | 1000 | imagenet-1k | VisionTransformer | vit_betwixt_patch16_reg1_gap_256 | 2024-05-10 |
|
| 469 |
| [vit_base_patch32_clip_384.laion2b_ft_in12k_in1k](https://huggingface.co/timm/vit_base_patch32_clip_384.laion2b_ft_in12k_in1k) | 88.2M | 12.7G | 384 | True | 768 | 1000 | imagenet-1k | VisionTransformer | vit_base_patch32_clip_384 | 2022-11-05 |
|
|
|
|
| 470 |
| [vit_little_patch16_reg4_gap_256.sbb_in1k](https://huggingface.co/timm/vit_little_patch16_reg4_gap_256.sbb_in1k) | 22.4M | 5.7G | 256 | True | 320 | 1000 | imagenet-1k | VisionTransformer | vit_little_patch16_reg4_gap_256 | 2024-05-10 |
|
| 471 |
| [vit_little_patch16_reg1_gap_256.sbb_in12k_ft_in1k](https://huggingface.co/timm/vit_little_patch16_reg1_gap_256.sbb_in12k_ft_in1k) | 22.4M | 5.7G | 256 | True | 320 | 1000 | imagenet-1k | VisionTransformer | vit_little_patch16_reg1_gap_256 | 2024-05-27 |
|
| 472 |
| [flexivit_small.1200ep_in1k](https://huggingface.co/timm/flexivit_small.1200ep_in1k) | 22.0M | 4.9G | 240 | True | 384 | 1000 | imagenet-1k | VisionTransformer | flexivit_small | 2022-12-22 |
|
|
|
|
| 114 |
|
| 115 |
# Models
|
| 116 |
|
| 117 |
+
131 models exported from TIMM in total.
|
| 118 |
|
| 119 |
## ByobNet
|
| 120 |
|
|
|
|
| 453 |
|
| 454 |
## VisionTransformer
|
| 455 |
|
| 456 |
+
15 models with model class `VisionTransformer`.
|
| 457 |
|
| 458 |
| Name | Params | Flops | Input Size | Can Classify | Features | Classes | Dataset | Model | Architecture | Created At |
|
| 459 |
|:-------------------------------------------------------------------------------------------------------------------------------------------------|:---------|:--------|-------------:|:---------------|-----------:|----------:|:------------|:------------------|:---------------------------------|:-------------|
|
|
|
|
| 467 |
| [vit_base_patch16_rpn_224.sw_in1k](https://huggingface.co/timm/vit_base_patch16_rpn_224.sw_in1k) | 86.4M | 16.8G | 224 | True | 768 | 1000 | imagenet-1k | VisionTransformer | vit_base_patch16_rpn_224 | 2022-12-22 |
|
| 468 |
| [vit_betwixt_patch16_reg1_gap_256.sbb_in1k](https://huggingface.co/timm/vit_betwixt_patch16_reg1_gap_256.sbb_in1k) | 60.2M | 15.3G | 256 | True | 640 | 1000 | imagenet-1k | VisionTransformer | vit_betwixt_patch16_reg1_gap_256 | 2024-05-10 |
|
| 469 |
| [vit_base_patch32_clip_384.laion2b_ft_in12k_in1k](https://huggingface.co/timm/vit_base_patch32_clip_384.laion2b_ft_in12k_in1k) | 88.2M | 12.7G | 384 | True | 768 | 1000 | imagenet-1k | VisionTransformer | vit_base_patch32_clip_384 | 2022-11-05 |
|
| 470 |
+
| [vit_small_patch16_384.augreg_in1k](https://huggingface.co/timm/vit_small_patch16_384.augreg_in1k) | 22.0M | 12.4G | 384 | True | 384 | 1000 | imagenet-1k | VisionTransformer | vit_small_patch16_384 | 2022-12-22 |
|
| 471 |
| [vit_little_patch16_reg4_gap_256.sbb_in1k](https://huggingface.co/timm/vit_little_patch16_reg4_gap_256.sbb_in1k) | 22.4M | 5.7G | 256 | True | 320 | 1000 | imagenet-1k | VisionTransformer | vit_little_patch16_reg4_gap_256 | 2024-05-10 |
|
| 472 |
| [vit_little_patch16_reg1_gap_256.sbb_in12k_ft_in1k](https://huggingface.co/timm/vit_little_patch16_reg1_gap_256.sbb_in12k_ft_in1k) | 22.4M | 5.7G | 256 | True | 320 | 1000 | imagenet-1k | VisionTransformer | vit_little_patch16_reg1_gap_256 | 2024-05-27 |
|
| 473 |
| [flexivit_small.1200ep_in1k](https://huggingface.co/timm/flexivit_small.1200ep_in1k) | 22.0M | 4.9G | 240 | True | 384 | 1000 | imagenet-1k | VisionTransformer | flexivit_small | 2022-12-22 |
|
models.parquet
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:301ad14fd1a1ea1fafc524fdf55f589e7e02b8b26b7f233632f88800e51fc0a6
|
| 3 |
+
size 17296
|
vit_small_patch16_384.augreg_in1k/meta.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:422c706aadfd1494adb2b5429653c237f8871d4e244cfc8c37c6df3929a0ece5
|
| 3 |
+
size 169842
|
vit_small_patch16_384.augreg_in1k/model.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:901f2e83d413b07f9e170451b5f42693911b30422ec7fb521fb4879fbd45e8b2
|
| 3 |
+
size 88958535
|
vit_small_patch16_384.augreg_in1k/preprocess.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9a08b6782724d3a3274363ac60488e44bc3f5df57b1c3d6ae9caad44d03e625e
|
| 3 |
+
size 642
|