Flex.2-preview / README.md
Eviation's picture
Update README.md
22b2c27 verified
|
raw
history blame
7.29 kB
---
license: apache-2.0
pipeline_tag: text-to-image
library_name: gguf
tags:
- flex
- flux
- gguf
- safetensors
base_model:
- ostris/Flex.2-preview
---
# Info
Various quantizations for [hf:ostris/Flex.2-preview](https://huggingface.co/ostris/Flex.2-preview/).
# Safetensors
| Filename | Quant Type | File Size | Description | Example Image |
| -------- | ---------- | --------- | ----------- | ------------- |
| [Flex.2-preview-fp8_e4m3fn_scaled.safetensors](https://huggingface.co/Eviation/Flex.2-preview/blob/main/Flex.2-preview-fp8_e4m3fn_scaled.safetensors) | F8_E4M3FN | 8.17GB | Scale per weight tensor | - |
| [Flex.2-preview-fp8_e5m2_scaled.safetensors](https://huggingface.co/Eviation/Flex.2-preview/blob/main/Flex.2-preview-fp8_e5m2_scaled.safetensors) | F8_E5M2 | 8.17GB | Scale per weight tensor | - |
| [Flex.2-preview-fp8_e4m3fn.safetensors](https://huggingface.co/Eviation/Flex.2-preview/blob/main/Flex.2-preview-fp8_e4m3fn.safetensors) | F8_E4M3FN | 8.16GB | - | - |
| [Flex.2-preview-fp8_e5m2.safetensors](https://huggingface.co/Eviation/Flex.2-preview/blob/main/Flex.2-preview-fp8_e5m2.safetensors) | F8_E5M2 | 8.16GB | - | - |
# Pure GGUF
- pure, conversion from safetensors BF16 via F32 gguf
- architecture: flex.2 (as not all tensor shapes match to flux)
- no imatrix was used to quantize
- biases and norms: F32
- img_in.weight: BF16 (due to tensor shape and block sizes)
- everything else according to file type
| Filename | Quant Type | File Size | Description / L2 Loss Step 25 | Example Image |
| -------- | ---------- | --------- | ----------------------------- | ------------- |
| [Flex.2-preview-BF16.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/Flex.2-preview-BF16.gguf) | BF16 | 16.3GB | - | - |
| [Flex.2-preview-Q8_0.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/pure/Flex.2-preview-Q8_0.gguf) | Q8_0 | 8.68GB | TBC | - |
| [Flex.2-preview-Q6_K.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/pure/Flex.2-preview-Q6_K.gguf) | Q6_K | 6.70GB | TBC | - |
| [Flex.2-preview-Q5_1.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/pure/Flex.2-preview-Q5_1.gguf) | Q5_1 | 6.13GB | TBC | - |
| [Flex.2-preview-Q5_0.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/pure/Flex.2-preview-Q5_0.gguf) | Q5_0 | 5.62GB | TBC | - |
| [Flex.2-preview-Q4_1.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/pure/Flex.2-preview-Q4_1.gguf) | Q4_1 | 5.11GB | TBC | - |
| [Flex.2-preview-IQ4_NL.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/pure/Flex.2-preview-IQ4_NL.gguf) | IQ4_NL | 4.60GB | TBC | - |
| [Flex.2-preview-Q4_0.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/pure/Flex.2-preview-Q4_0.gguf) | Q4_0 | 4.60GB | TBC | - |
| [Flex.2-preview-Q3_K_S.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/pure/Flex.2-preview-Q3_K_S.gguf) | Q3_K_S | 3.52GB | TBC | - |
# Fluxified GGUF
- conversion from safetensors BF16 via F32 gguf
- truncated img_in.weight tensor to first 16 latent channels
- lost ability to do inpainting and process control image
- should be a drop-in replacement for FLUX
- architecture: flux
- dynamic quantization?
| Filename | Quant type | File Size | Description / L2 Loss Step 25 | Example Image |
| -------- | ---------- | --------- | ----------------------------- | ------------- |
| [Flex.2-preview-fluxified-Q8_0.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified/Flex.2-preview-fluxified-Q8_0.gguf) | Q8_0 | 8.39GB | TBC | - |
| [Flex.2-preview-fluxified-Q6_K.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified/Flex.2-preview-fluxified-Q6_K.gguf) | Q6_K | 6.74GB | TBC | - |
| [Flex.2-preview-fluxified-Q5_1.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified/Flex.2-preview-fluxified-Q5_1.gguf) | Q5_1 | 6.19GB | TBC | - |
| [Flex.2-preview-fluxified-Q5_0.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified/Flex.2-preview-fluxified-Q5_0.gguf) | Q5_0 | 5.70GB | TBC | - |
| [Flex.2-preview-fluxified-Q5_K_S.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified/Flex.2-preview-fluxified-Q5_K_S.gguf) | Q5_K_S | 5.67GB | TBC | - |
| [Flex.2-preview-fluxified-Q4_1.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified/Flex.2-preview-fluxified-Q4_1.gguf) | Q4_1 | 5.22GB | TBC | - |
| [Flex.2-preview-fluxified-Q4_0.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified/Flex.2-preview-fluxified-Q4_0.gguf) | Q4_0 | 4.72GB | TBC | - |
| [Flex.2-preview-fluxified-Q4_K_S.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified/Flex.2-preview-fluxified-Q4_K_S.gguf) | Q4_K_S | 4.58GB | TBC | - |
| [Flex.2-preview-fluxified-Q3_K_S.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified/Flex.2-preview-fluxified-Q3_K_S.gguf) | Q3_K_S | 3.52GB | TBC | - |
# Fluxified GGUF Imatrix
- Fluxified GGUF + Importance Matrix
| Filename | Quant type | File Size | Description / L2 Loss Step 25 | Example Image |
| -------- | ---------- | --------- | ----------------------------- | ------------- |
| [Flex.2-preview-fluxified-Q8_0.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified-imat/Flex.2-preview-fluxified-Q8_0_gguf) | Q8_0 | 8.39GB | TBC | - |
| [Flex.2-preview-fluxified-Q6_K.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified-imat/Flex.2-preview-fluxified-Q6_K_gguf) | Q6_K | 6.74GB | TBC | - |
| [Flex.2-preview-fluxified-Q5_1.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified-imat/Flex.2-preview-fluxified-Q5_1_gguf) | Q5_1 | 6.19GB | TBC | - |
| [Flex.2-preview-fluxified-Q5_0.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified-imat/Flex.2-preview-fluxified-Q5_0_gguf) | Q5_0 | 5.70GB | TBC | - |
| [Flex.2-preview-fluxified-Q5_K_S.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified-imat/Flex.2-preview-fluxified-Q5_K_S.gguf) | Q5_K_S | 5.67GB | TBC
| [Flex.2-preview-fluxified-Q4_1.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified-imat/Flex.2-preview-fluxified-Q4_1_gguf) | Q4_1 | 5.22GB | TBC | - || - |
| [Flex.2-preview-fluxified-Q4_0.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified-imat/Flex.2-preview-fluxified-Q4_0_gguf) | Q4_0 | 4.72GB | TBC | - || - |
| [Flex.2-preview-fluxified-IQ4_NL.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified-imat/Flex.2-preview-fluxified-IQ4_NL.gguf) | IQ4_NL | 4.58GB | TBC | - |
| [Flex.2-preview-fluxified-Q4_K_S.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified-imat/Flex.2-preview-fluxified-Q4_K_S.gguf) | Q4_K_S | 4.58GB | TBC | -
| [Flex.2-preview-fluxified-IQ4_XS.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified-imat/Flex.2-preview-fluxified-IQ4_XS.gguf) | IQ4_XS | 4.37GB | TBC | - ||
| [Flex.2-preview-fluxified-Q3_K_S.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified-imat/Flex.2-preview-fluxified-Q3_K_S.gguf) | Q3_K_S | 3.52GB | TBC | - |
| [Flex.2-preview-fluxified-Q2_K.gguf](https://huggingface.co/Eviation/Flex.2-preview/blob/main/fluxified-imat/Flex.2-preview-fluxified-Q2_K.gguf) | Q2_K | 2.82GB | TBC | - |