Omdano commited on
Commit
8eada2c
·
verified ·
1 Parent(s): 8bf2e5c

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +49 -0
README.md ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # DinoV3 Vision Transformer Huge (INT8 Quantized)
2
+
3
+ INT8 quantized version of `facebook/dinov3-vith16plus-pretrain-lvd1689m` using BitsAndBytes.
4
+
5
+ ## Model Details
6
+
7
+ - **Base Model**: DinoV3 Vision Transformer Huge (840M parameters)
8
+ - **Quantization**: INT8 weight-only quantization via BitsAndBytes
9
+ - **Size**: ~845MB (from ~1.7GB original)
10
+ - **Compression**: ~2x size reduction
11
+ - **Accuracy Loss**: <1% typical
12
+
13
+ ## Usage
14
+
15
+ ```python
16
+ from transformers import AutoModel, BitsAndBytesConfig
17
+
18
+ # Load the INT8 quantized model
19
+ model = AutoModel.from_pretrained(
20
+ "Omdano/INT8-H16P",
21
+ trust_remote_code=True,
22
+ quantization_config=BitsAndBytesConfig(load_in_8bit=True),
23
+ device_map="auto"
24
+ )
25
+
26
+ # Use for feature extraction or classification
27
+ outputs = model(pixel_values=inputs)
28
+ ```
29
+
30
+ ## Benefits
31
+
32
+ - **2x smaller** than full precision model
33
+ - **Faster inference** on GPU
34
+ - **Same API** as original DinoV3
35
+ - **Minimal accuracy loss** (<1%)
36
+
37
+ ## Requirements
38
+
39
+ ```bash
40
+ pip install transformers bitsandbytes torch
41
+ ```
42
+
43
+ ## Original Model
44
+
45
+ Based on [facebook/dinov3-vith16plus-pretrain-lvd1689m](https://huggingface.co/facebook/dinov3-vith16plus-pretrain-lvd1689m)
46
+
47
+ ## License
48
+
49
+ Apache 2.0 (same as original DinoV3)