Upload folder using huggingface_hub
Browse files- README.md +23 -0
- config.json +3 -2
- onnx/model.onnx +3 -0
- onnx/model_bnb4.onnx +3 -0
- onnx/model_fp16.onnx +3 -0
- onnx/model_int8.onnx +3 -0
- onnx/model_q4.onnx +3 -0
- onnx/model_q4f16.onnx +3 -0
- onnx/model_quantized.onnx +3 -0
- onnx/model_uint8.onnx +3 -0
- preprocessor_config.json +1 -1
- quantize_config.json +18 -0
README.md
CHANGED
|
@@ -1,3 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
|
| 2 |
---
|
| 3 |
tags:
|
|
|
|
| 1 |
+
---
|
| 2 |
+
library_name: transformers.js
|
| 3 |
+
base_model:
|
| 4 |
+
- Sioood/autotrain-vehicle-classification-test
|
| 5 |
+
pipeline_tag: image-classification
|
| 6 |
+
---
|
| 7 |
+
|
| 8 |
+
|
| 9 |
+
|
| 10 |
+
# autotrain-vehicle-classification-test (ONNX)
|
| 11 |
+
|
| 12 |
+
|
| 13 |
+
This is an ONNX version of [Sioood/autotrain-vehicle-classification-test](https://huggingface.co/Sioood/autotrain-vehicle-classification-test). It was automatically converted and uploaded using [this Hugging Face Space](https://huggingface.co/spaces/onnx-community/convert-to-onnx).
|
| 14 |
+
|
| 15 |
+
|
| 16 |
+
## Usage with Transformers.js
|
| 17 |
+
|
| 18 |
+
|
| 19 |
+
See the pipeline documentation for `image-classification`: https://huggingface.co/docs/transformers.js/api/pipelines#module_pipelines.ImageClassificationPipeline
|
| 20 |
+
|
| 21 |
+
|
| 22 |
+
---
|
| 23 |
+
|
| 24 |
|
| 25 |
---
|
| 26 |
tags:
|
config.json
CHANGED
|
@@ -1,5 +1,6 @@
|
|
| 1 |
{
|
| 2 |
-
"
|
|
|
|
| 3 |
"_num_labels": 4,
|
| 4 |
"architectures": [
|
| 5 |
"ViTForImageClassification"
|
|
@@ -33,5 +34,5 @@
|
|
| 33 |
"problem_type": "single_label_classification",
|
| 34 |
"qkv_bias": true,
|
| 35 |
"torch_dtype": "float32",
|
| 36 |
-
"transformers_version": "4.
|
| 37 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"_attn_implementation_autoset": true,
|
| 3 |
+
"_name_or_path": "Sioood/autotrain-vehicle-classification-test",
|
| 4 |
"_num_labels": 4,
|
| 5 |
"architectures": [
|
| 6 |
"ViTForImageClassification"
|
|
|
|
| 34 |
"problem_type": "single_label_classification",
|
| 35 |
"qkv_bias": true,
|
| 36 |
"torch_dtype": "float32",
|
| 37 |
+
"transformers_version": "4.49.0"
|
| 38 |
}
|
onnx/model.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d068b0318396f64023aeeb8b0d07618b81f4cbd4bef0bf34db89f375de5b95d7
|
| 3 |
+
size 343407840
|
onnx/model_bnb4.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8b0fb03c4125070f5752d2aaabafa0af0bfe89eb0f70aa2d4e4971a8f3adeed4
|
| 3 |
+
size 51456162
|
onnx/model_fp16.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fe084deb84f4438247684217ffbd4b143d4cfc922319cacaf0cca803b94bc202
|
| 3 |
+
size 171804458
|
onnx/model_int8.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5df1602608ca68b9037cc64094402cc7ff55ff998107a59673ba1611cf80975c
|
| 3 |
+
size 87335183
|
onnx/model_q4.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:55d07baeee13d8c1f28bbb4d45322f71da496b51353c2236d493f0b703761cc5
|
| 3 |
+
size 56764050
|
onnx/model_q4f16.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dbf39e6b8a65037f8007bccc0f763414150311f1525b0375c1f9d7908903fc13
|
| 3 |
+
size 49721661
|
onnx/model_quantized.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c41cf6c7abdaba56a85d7e37731ead69bcc9d9dd2af1a6f6956b1912192834e
|
| 3 |
+
size 87335183
|
onnx/model_uint8.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c41cf6c7abdaba56a85d7e37731ead69bcc9d9dd2af1a6f6956b1912192834e
|
| 3 |
+
size 87335183
|
preprocessor_config.json
CHANGED
|
@@ -8,7 +8,7 @@
|
|
| 8 |
0.5,
|
| 9 |
0.5
|
| 10 |
],
|
| 11 |
-
"image_processor_type": "
|
| 12 |
"image_std": [
|
| 13 |
0.5,
|
| 14 |
0.5,
|
|
|
|
| 8 |
0.5,
|
| 9 |
0.5
|
| 10 |
],
|
| 11 |
+
"image_processor_type": "ViTFeatureExtractor",
|
| 12 |
"image_std": [
|
| 13 |
0.5,
|
| 14 |
0.5,
|
quantize_config.json
ADDED
|
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"modes": [
|
| 3 |
+
"fp16",
|
| 4 |
+
"q8",
|
| 5 |
+
"int8",
|
| 6 |
+
"uint8",
|
| 7 |
+
"q4",
|
| 8 |
+
"q4f16",
|
| 9 |
+
"bnb4"
|
| 10 |
+
],
|
| 11 |
+
"per_channel": true,
|
| 12 |
+
"reduce_range": true,
|
| 13 |
+
"block_size": null,
|
| 14 |
+
"is_symmetric": true,
|
| 15 |
+
"accuracy_level": null,
|
| 16 |
+
"quant_type": 1,
|
| 17 |
+
"op_block_list": null
|
| 18 |
+
}
|