EricRollei's picture
update minor changes
ef3f1f8 verified
---
license: other
license_name: tencent-hunyuan-community
license_link: https://huggingface.co/tencent/HunyuanImage-3.0/blob/main/LICENSE.txt
tags:
- text-to-image
- hunyuan
- quantization
- int8
- comfyui
- custom nodes
- autoregressive
- Dit
- Hunyuan-Image-3
pipeline_tag: text-to-image
---
# Hunyuan Image 3.0 - INT8 Quantized
This is an **INT8 quantized version** of Tencent's [HunyuanImage-3.0](https://huggingface.co/tencent/HunyuanImage-3.0) model, optimized for high-end GPU workflows without CPU offloading.
## Model Description
INT8 quantization of the Hunyuan Image 3.0 text-to-image diffusion transformer, providing a balance between the full BF16 precision and more aggressive NF4 quantization. This version maintains excellent image quality while reducing memory requirements.
**Key Features:**
- 🎯 High quality output comparable to BF16
- 💾 ~80GB VRAM required (fits RTX 6000 Ada/Blackwell)
- âš¡ ~3.5 minutes generation time at base resolution
- 🔧 Designed for ComfyUI workflows
## VRAM Requirements
| Phase | VRAM Usage |
|-------|------------|
| Weight Loading | ~80 GB |
| Inference (additional) | ~12-20 GB |
| **Total** | **~92-100 GB** |
**Recommended Hardware:**
- NVIDIA RTX 6000 Ada (48GB) - requires model split/offload
- NVIDIA RTX 6000 Blackwell (96GB) - fits entirely in VRAM ✅ Workflows on the github page
- Multi-GPU setups with 80GB+ combined VRAM
## Usage
### ComfyUI (Recommended)
This model is designed to work with the [Comfy_HunyuanImage3](https://github.com/EricRollei/Comfy_HunyuanImage3) custom nodes:
```bash
cd ComfyUI/custom_nodes
git clone https://github.com/EricRollei/Comfy_HunyuanImage3
```
Install the nodes and download this model to your ComfyUI models directory. The nodes handle INT8 loading automatically.
### Direct Usage
```python
# INT8 weights can be loaded with standard torch quantization
# See the ComfyUI nodes for reference implementation
```
## Performance
- **Generation Time**: ~3.5 minutes for base resolution (1024x1024)
- **Weight Loading**: ~60 seconds (one-time per session)
- **Quality**: Excellent - minimal degradation from BF16
- **Speed**: Faster inference than BF16 due to reduced memory bandwidth
## Quantization Details
- **Method**: INT8 per-channel quantization
- **Target**: Hunyuan Image 3.0 transformer backbone
- **Precision Loss**: Minimal - image quality remains high
- **Trade-off**: Middle ground between NF4 (lower quality) and BF16 (highest VRAM)
## Original Model
This is a quantized derivative of [Tencent's HunyuanImage-3.0](https://huggingface.co/tencent/HunyuanImage-3.0).
**Original Model Details:**
- Architecture: Diffusion Transformer
- Resolution: Up to 2048x2048
- Language Support: English and Chinese prompts
- License: [Tencent Hunyuan Community License](https://huggingface.co/tencent/HunyuanImage-3.0/blob/main/LICENSE.txt)
Please review the original model card and license for full details on capabilities and restrictions.
## Limitations
- Requires high-end professional GPU (80GB+ VRAM)
- Not suitable for consumer GPUs (4090, 5090) without further optimization
- INT8 quantization may introduce minor quality differences in edge cases
- Loading time adds ~1 minute overhead to first generation
## Credits
**Original Model**: [Tencent Hunyuan Team](https://huggingface.co/tencent)
**Quantization**: Eric Rollei
**ComfyUI Integration**: [Comfy_HunyuanImage3](https://github.com/EricRollei/Comfy_HunyuanImage3)
## License
This model inherits the license from the original Hunyuan Image 3.0 model:
- **License**: [Tencent Hunyuan Community License](https://huggingface.co/tencent/HunyuanImage-3.0/blob/main/LICENSE.txt)
- Please review the original license for commercial use restrictions and requirements
## Citation
```bibtex
@misc{hunyuan-image-3-int8,
author = {Rollei, Eric},
title = {Hunyuan Image 3.0 INT8 Quantized},
year = {2024},
publisher = {Hugging Face},
howpublished = {\url{https://huggingface.co/[YOUR_USERNAME]/[MODEL_NAME]}}
}
```
Original model citation:
```bibtex
@misc{tencent2024hunyuan,
title={Hunyuan Image 3.0},
author={Tencent Hunyuan Team},
year={2024},
publisher={Hugging Face},
howpublished={\url{https://huggingface.co/tencent/HunyuanImage-3.0}}
}
```