Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
|
@@ -19,7 +19,7 @@ base_model:
|
|
| 19 |
|
| 20 |
# FLUX.1-dev-ControlNet-Union-Pro-2.0 (fp8)
|
| 21 |
|
| 22 |
-
This repository contains an unified ControlNet for FLUX.1-dev model released by [Shakker Labs](https://huggingface.co/Shakker-Labs). This
|
| 23 |
|
| 24 |
# FP8 Quantization
|
| 25 |
This model has been quantized from the original BFloat16 format to FP8 format. The benefits include:
|
|
@@ -27,6 +27,8 @@ This model has been quantized from the original BFloat16 format to FP8 format. T
|
|
| 27 |
- **Faster Inference**: Potential speed improvements, especially on hardware with FP8 support
|
| 28 |
- **Minimal Quality Loss**: Carefully calibrated quantization process to preserve output quality
|
| 29 |
|
|
|
|
|
|
|
| 30 |
# Keynotes
|
| 31 |
In comparison with [Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro](https://huggingface.co/Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro),
|
| 32 |
- Remove mode embedding, has smaller model size.
|
|
@@ -157,6 +159,15 @@ pipe.to("cuda")
|
|
| 157 |
|
| 158 |
See `fp8_inference_example.py` for a complete example.
|
| 159 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 160 |
# Resources
|
| 161 |
- [InstantX/FLUX.1-dev-IP-Adapter](https://huggingface.co/InstantX/FLUX.1-dev-IP-Adapter)
|
| 162 |
- [InstantX/FLUX.1-dev-Controlnet-Canny](https://huggingface.co/InstantX/FLUX.1-dev-Controlnet-Canny)
|
|
|
|
| 19 |
|
| 20 |
# FLUX.1-dev-ControlNet-Union-Pro-2.0 (fp8)
|
| 21 |
|
| 22 |
+
This repository contains an unified ControlNet for FLUX.1-dev model released by [Shakker Labs](https://huggingface.co/Shakker-Labs). This is a direct quantization of the original model to FP8 format for optimized inference performance (not a fine-tuned version). We provide an [online demo](https://huggingface.co/spaces/Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro-2.0).
|
| 23 |
|
| 24 |
# FP8 Quantization
|
| 25 |
This model has been quantized from the original BFloat16 format to FP8 format. The benefits include:
|
|
|
|
| 27 |
- **Faster Inference**: Potential speed improvements, especially on hardware with FP8 support
|
| 28 |
- **Minimal Quality Loss**: Carefully calibrated quantization process to preserve output quality
|
| 29 |
|
| 30 |
+
Note: This is a direct quantization of [Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro-2.0](https://huggingface.co/Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro-2.0) and preserves all the functionality of the original model.
|
| 31 |
+
|
| 32 |
# Keynotes
|
| 33 |
In comparison with [Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro](https://huggingface.co/Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro),
|
| 34 |
- Remove mode embedding, has smaller model size.
|
|
|
|
| 159 |
|
| 160 |
See `fp8_inference_example.py` for a complete example.
|
| 161 |
|
| 162 |
+
# Pushing Model to Hugging Face Hub
|
| 163 |
+
To push your FP8 quantized model to the Hugging Face Hub, use the included script:
|
| 164 |
+
|
| 165 |
+
```bash
|
| 166 |
+
python push_model_to_hub.py --repo_id "ABDALLALSWAITI/FLUX.1-dev-ControlNet-Union-Pro-2.0-fp8"
|
| 167 |
+
```
|
| 168 |
+
|
| 169 |
+
You will need to have the `huggingface_hub` library installed and be logged in with your Hugging Face credentials.
|
| 170 |
+
|
| 171 |
# Resources
|
| 172 |
- [InstantX/FLUX.1-dev-IP-Adapter](https://huggingface.co/InstantX/FLUX.1-dev-IP-Adapter)
|
| 173 |
- [InstantX/FLUX.1-dev-Controlnet-Canny](https://huggingface.co/InstantX/FLUX.1-dev-Controlnet-Canny)
|