EricRollei commited on
Commit
4fdac1c
·
verified ·
1 Parent(s): cdff52b

initial model card

Browse files
Files changed (1) hide show
  1. README.md +124 -3
README.md CHANGED
@@ -1,3 +1,124 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: tencent-hunyuan-community
4
+ license_link: https://huggingface.co/tencent/HunyuanImage-3.0/blob/main/LICENSE.txt
5
+ tags:
6
+ - text-to-image
7
+ - hunyuan
8
+ - quantization
9
+ - int8
10
+ - comfyui
11
+ pipeline_tag: text-to-image
12
+ ---
13
+
14
+ # Hunyuan Image 3.0 - INT8 Quantized
15
+
16
+ This is an **INT8 quantized version** of Tencent's [HunyuanImage-3.0](https://huggingface.co/tencent/HunyuanImage-3.0) model, optimized for high-end GPU workflows without CPU offloading.
17
+
18
+ ## Model Description
19
+
20
+ INT8 quantization of the Hunyuan Image 3.0 text-to-image diffusion transformer, providing a balance between the full BF16 precision and more aggressive NF4 quantization. This version maintains excellent image quality while reducing memory requirements.
21
+
22
+ **Key Features:**
23
+ - 🎯 High quality output comparable to BF16
24
+ - 💾 ~80GB VRAM required (fits RTX 6000 Ada/Blackwell)
25
+ - ⚡ ~3.5 minutes generation time at base resolution
26
+ - 🔧 Designed for ComfyUI workflows
27
+
28
+ ## VRAM Requirements
29
+
30
+ | Phase | VRAM Usage |
31
+ |-------|------------|
32
+ | Weight Loading | ~80 GB |
33
+ | Inference (additional) | ~12-20 GB |
34
+ | **Total** | **~92-100 GB** |
35
+
36
+ **Recommended Hardware:**
37
+ - NVIDIA RTX 6000 Ada (48GB) - requires model split/offload
38
+ - NVIDIA RTX 6000 Blackwell (96GB) - fits entirely in VRAM ✅
39
+ - Multi-GPU setups with 80GB+ combined VRAM
40
+
41
+ ## Usage
42
+
43
+ ### ComfyUI (Recommended)
44
+
45
+ This model is designed to work with the [Comfy_HunyuanImage3](https://github.com/EricRollei/Comfy_HunyuanImage3) custom nodes:
46
+ ```bash
47
+ cd ComfyUI/custom_nodes
48
+ git clone https://github.com/EricRollei/Comfy_HunyuanImage3
49
+ ```
50
+
51
+ Install the nodes and download this model to your ComfyUI models directory. The nodes handle INT8 loading automatically.
52
+
53
+ ### Direct Usage
54
+ ```python
55
+ # INT8 weights can be loaded with standard torch quantization
56
+ # See the ComfyUI nodes for reference implementation
57
+ ```
58
+
59
+ ## Performance
60
+
61
+ - **Generation Time**: ~3.5 minutes for base resolution (1024x1024)
62
+ - **Weight Loading**: ~60 seconds (one-time per session)
63
+ - **Quality**: Excellent - minimal degradation from BF16
64
+ - **Speed**: Faster inference than BF16 due to reduced memory bandwidth
65
+
66
+ ## Quantization Details
67
+
68
+ - **Method**: INT8 per-channel quantization
69
+ - **Target**: Hunyuan Image 3.0 transformer backbone
70
+ - **Precision Loss**: Minimal - image quality remains high
71
+ - **Trade-off**: Middle ground between NF4 (lower quality) and BF16 (highest VRAM)
72
+
73
+ ## Original Model
74
+
75
+ This is a quantized derivative of [Tencent's HunyuanImage-3.0](https://huggingface.co/tencent/HunyuanImage-3.0).
76
+
77
+ **Original Model Details:**
78
+ - Architecture: Diffusion Transformer
79
+ - Resolution: Up to 2048x2048
80
+ - Language Support: English and Chinese prompts
81
+ - License: [Tencent Hunyuan Community License](https://huggingface.co/tencent/HunyuanImage-3.0/blob/main/LICENSE.txt)
82
+
83
+ Please review the original model card and license for full details on capabilities and restrictions.
84
+
85
+ ## Limitations
86
+
87
+ - Requires high-end professional GPU (80GB+ VRAM)
88
+ - Not suitable for consumer GPUs (4090, 5090) without further optimization
89
+ - INT8 quantization may introduce minor quality differences in edge cases
90
+ - Loading time adds ~1 minute overhead to first generation
91
+
92
+ ## Credits
93
+
94
+ **Original Model**: [Tencent Hunyuan Team](https://huggingface.co/tencent)
95
+ **Quantization**: Eric Rollei
96
+ **ComfyUI Integration**: [Comfy_HunyuanImage3](https://github.com/EricRollei/Comfy_HunyuanImage3)
97
+
98
+ ## License
99
+
100
+ This model inherits the license from the original Hunyuan Image 3.0 model:
101
+ - **License**: [Tencent Hunyuan Community License](https://huggingface.co/tencent/HunyuanImage-3.0/blob/main/LICENSE.txt)
102
+ - Please review the original license for commercial use restrictions and requirements
103
+
104
+ ## Citation
105
+ ```bibtex
106
+ @misc{hunyuan-image-3-int8,
107
+ author = {Rollei, Eric},
108
+ title = {Hunyuan Image 3.0 INT8 Quantized},
109
+ year = {2024},
110
+ publisher = {Hugging Face},
111
+ howpublished = {\url{https://huggingface.co/[YOUR_USERNAME]/[MODEL_NAME]}}
112
+ }
113
+ ```
114
+
115
+ Original model citation:
116
+ ```bibtex
117
+ @misc{tencent2024hunyuan,
118
+ title={Hunyuan Image 3.0},
119
+ author={Tencent Hunyuan Team},
120
+ year={2024},
121
+ publisher={Hugging Face},
122
+ howpublished={\url{https://huggingface.co/tencent/HunyuanImage-3.0}}
123
+ }
124
+ ```