Update README.md
Browse files
README.md
CHANGED
|
@@ -18,7 +18,9 @@ This repository contains merged models, built upon the base models:
|
|
| 18 |
- [Freepik's Flux.1-Lite-8B-alpha](https://huggingface.co/Freepik/flux.1-lite-8B-alpha)
|
| 19 |
- [Flux-dev-de-distill](https://huggingface.co/nyanko7/flux-dev-de-distill)
|
| 20 |
|
| 21 |
-
Detail enhancement and acceleration techniques have been applied, particularly optimized for NVIDIA 4XXX cards (maybe 3XXX too). The goal is to have high efficiency accelerated models with lower overhead.
|
|
|
|
|
|
|
| 22 |
|
| 23 |
**====================================**
|
| 24 |
|
|
@@ -49,12 +51,12 @@ Detail enhancement and acceleration techniques have been applied, particularly o
|
|
| 49 |
- **Best of Flux: Style Enhancing LoRA** (Weight: 0.06) ([Model Link](https://civitai.com/models/821668))
|
| 50 |
|
| 51 |
**Distillation Used:**
|
| 52 |
-
- **Flux distilled lora** (Weight: Hyper -
|
| 53 |
|
| 54 |
3. **GGUF Quantized Models (Q8_0)**:
|
| 55 |
-
- [flux.1-lite-8B-alpha-Hyper-8.Steps-Detail.Plus-Distilled-Q8_0](https://huggingface.co/mdmachine/FLUX.Model.Merge-Detail.Enhancement.and.Acceleration/blob/main/Lite-8B-Plus/Distilled/GGUF/Q8_0/flux.1-lite-8B-alpha-Hyper-8.Steps-Detail.Plus-Distilled-Q8_0.gguf)
|
| 56 |
-
- [flux.1-lite-8B-alpha-Hyper-16.Steps-Detail.Plus-Distilled-Q8_0](https://huggingface.co/mdmachine/FLUX.Model.Merge-Detail.Enhancement.and.Acceleration/blob/main/Lite-8B-Plus/Distilled/GGUF/Q8_0/flux.1-lite-8B-alpha-Hyper-16.Steps-Detail.Plus-Distilled-Q8_0.gguf)
|
| 57 |
-
- [flux.1-lite-8B-alpha-Turbo-8.Steps-Detail.Plus-Distilled-Q8_0](https://huggingface.co/mdmachine/FLUX.Model.Merge-Detail.Enhancement.and.Acceleration/blob/main/Lite-8B-Plus/Distilled/GGUF/Q8_0/flux.1-lite-8B-alpha-Turbo-8.Steps-Detail.Plus-Distilled-Q8_0.gguf)
|
| 58 |
|
| 59 |
**====================================**
|
| 60 |
|
|
|
|
| 18 |
- [Freepik's Flux.1-Lite-8B-alpha](https://huggingface.co/Freepik/flux.1-lite-8B-alpha)
|
| 19 |
- [Flux-dev-de-distill](https://huggingface.co/nyanko7/flux-dev-de-distill)
|
| 20 |
|
| 21 |
+
Detail enhancement and acceleration techniques have been applied, particularly optimized for NVIDIA 4XXX cards (maybe 3XXX too). The goal is to have high efficiency accelerated models with lower overhead.
|
| 22 |
+
|
| 23 |
+
The (de-re-destill | Distilled) models can be used with CFG remaining at 1-ish. Also the baked-in accelerators work (mostly) as intended. This is a result of workflow optimization by-products. Things can be added/removed or changed at any time.
|
| 24 |
|
| 25 |
**====================================**
|
| 26 |
|
|
|
|
| 51 |
- **Best of Flux: Style Enhancing LoRA** (Weight: 0.06) ([Model Link](https://civitai.com/models/821668))
|
| 52 |
|
| 53 |
**Distillation Used:**
|
| 54 |
+
- **Flux distilled lora** (Weight: Hyper - .65 | Turbo - 0.50) ([Model Link](https://civitai.com/models/977247/flux-distilled-lora))
|
| 55 |
|
| 56 |
3. **GGUF Quantized Models (Q8_0)**:
|
| 57 |
+
- [flux.1-lite-8B-alpha-Hyper-8.Steps-Detail.Plus-Distilled-V2-Q8_0](https://huggingface.co/mdmachine/FLUX.Model.Merge-Detail.Enhancement.and.Acceleration/blob/main/Lite-8B-Plus/Distilled/GGUF/Q8_0/flux.1-lite-8B-alpha-Hyper-8.Steps-Detail.Plus-Distilled-V2-Q8_0.gguf)
|
| 58 |
+
- [flux.1-lite-8B-alpha-Hyper-16.Steps-Detail.Plus-Distilled-V2-Q8_0](https://huggingface.co/mdmachine/FLUX.Model.Merge-Detail.Enhancement.and.Acceleration/blob/main/Lite-8B-Plus/Distilled/GGUF/Q8_0/flux.1-lite-8B-alpha-Hyper-16.Steps-Detail.Plus-Distilled-V2-Q8_0.gguf)
|
| 59 |
+
- [flux.1-lite-8B-alpha-Turbo-8.Steps-Detail.Plus-Distilled-V2-Q8_0](https://huggingface.co/mdmachine/FLUX.Model.Merge-Detail.Enhancement.and.Acceleration/blob/main/Lite-8B-Plus/Distilled/GGUF/Q8_0/flux.1-lite-8B-alpha-Turbo-8.Steps-Detail.Plus-Distilled-V2-Q8_0.gguf)
|
| 60 |
|
| 61 |
**====================================**
|
| 62 |
|