--- license: other license_name: tencent-hunyuan-community license_link: https://huggingface.co/tencent/HunyuanImage-3.0/blob/main/LICENSE.txt tags: - text-to-image - hunyuan - quantization - int8 - comfyui - custom nodes - autoregressive - Dit - Hunyuan-Image-3 pipeline_tag: text-to-image --- # Hunyuan Image 3.0 - INT8 Quantized This is an **INT8 quantized version** of Tencent's [HunyuanImage-3.0](https://huggingface.co/tencent/HunyuanImage-3.0) model, optimized for high-end GPU workflows without CPU offloading. ## Model Description INT8 quantization of the Hunyuan Image 3.0 text-to-image diffusion transformer, providing a balance between the full BF16 precision and more aggressive NF4 quantization. This version maintains excellent image quality while reducing memory requirements. **Key Features:** - 🎯 High quality output comparable to BF16 - 💾 ~80GB VRAM required (fits RTX 6000 Ada/Blackwell) - ⚡ ~3.5 minutes generation time at base resolution - 🔧 Designed for ComfyUI workflows ## VRAM Requirements | Phase | VRAM Usage | |-------|------------| | Weight Loading | ~80 GB | | Inference (additional) | ~12-20 GB | | **Total** | **~92-100 GB** | **Recommended Hardware:** - NVIDIA RTX 6000 Ada (48GB) - requires model split/offload - NVIDIA RTX 6000 Blackwell (96GB) - fits entirely in VRAM ✅ Workflows on the github page - Multi-GPU setups with 80GB+ combined VRAM ## Usage ### ComfyUI (Recommended) This model is designed to work with the [Comfy_HunyuanImage3](https://github.com/EricRollei/Comfy_HunyuanImage3) custom nodes: ```bash cd ComfyUI/custom_nodes git clone https://github.com/EricRollei/Comfy_HunyuanImage3 ``` Install the nodes and download this model to your ComfyUI models directory. The nodes handle INT8 loading automatically. ### Direct Usage ```python # INT8 weights can be loaded with standard torch quantization # See the ComfyUI nodes for reference implementation ``` ## Performance - **Generation Time**: ~3.5 minutes for base resolution (1024x1024) - **Weight Loading**: ~60 seconds (one-time per session) - **Quality**: Excellent - minimal degradation from BF16 - **Speed**: Faster inference than BF16 due to reduced memory bandwidth ## Quantization Details - **Method**: INT8 per-channel quantization - **Target**: Hunyuan Image 3.0 transformer backbone - **Precision Loss**: Minimal - image quality remains high - **Trade-off**: Middle ground between NF4 (lower quality) and BF16 (highest VRAM) ## Original Model This is a quantized derivative of [Tencent's HunyuanImage-3.0](https://huggingface.co/tencent/HunyuanImage-3.0). **Original Model Details:** - Architecture: Diffusion Transformer - Resolution: Up to 2048x2048 - Language Support: English and Chinese prompts - License: [Tencent Hunyuan Community License](https://huggingface.co/tencent/HunyuanImage-3.0/blob/main/LICENSE.txt) Please review the original model card and license for full details on capabilities and restrictions. ## Limitations - Requires high-end professional GPU (80GB+ VRAM) - Not suitable for consumer GPUs (4090, 5090) without further optimization - INT8 quantization may introduce minor quality differences in edge cases - Loading time adds ~1 minute overhead to first generation ## Credits **Original Model**: [Tencent Hunyuan Team](https://huggingface.co/tencent) **Quantization**: Eric Rollei **ComfyUI Integration**: [Comfy_HunyuanImage3](https://github.com/EricRollei/Comfy_HunyuanImage3) ## License This model inherits the license from the original Hunyuan Image 3.0 model: - **License**: [Tencent Hunyuan Community License](https://huggingface.co/tencent/HunyuanImage-3.0/blob/main/LICENSE.txt) - Please review the original license for commercial use restrictions and requirements ## Citation ```bibtex @misc{hunyuan-image-3-int8, author = {Rollei, Eric}, title = {Hunyuan Image 3.0 INT8 Quantized}, year = {2024}, publisher = {Hugging Face}, howpublished = {\url{https://huggingface.co/[YOUR_USERNAME]/[MODEL_NAME]}} } ``` Original model citation: ```bibtex @misc{tencent2024hunyuan, title={Hunyuan Image 3.0}, author={Tencent Hunyuan Team}, year={2024}, publisher={Hugging Face}, howpublished={\url{https://huggingface.co/tencent/HunyuanImage-3.0}} } ```