docs: update model card — 5000-step run, scale=3.0 recommendation
Browse files
README.md
CHANGED
|
@@ -34,7 +34,7 @@ A **PEFT LoRA adapter** trained on top of [Tongyi-MAI/Z-Image-Turbo](https://hug
|
|
| 34 |
| Target modules | `to_q`, `to_k`, `to_v`, `w1`, `w2`, `w3` |
|
| 35 |
| Trainable params | ~39 M |
|
| 36 |
| Adapter size | ~271 MB |
|
| 37 |
-
| Training steps | 3 000 |
|
| 38 |
| Training resolution | 512 × 512 |
|
| 39 |
| Dataset | [DownFlow/fuliji](https://huggingface.co/datasets/DownFlow/fuliji) (8 artists, ~200 images) |
|
| 40 |
|
|
@@ -89,10 +89,10 @@ PEFT exposes a scaling multiplier per adapter. Increase it to push the style har
|
|
| 89 |
# After PeftModel.from_pretrained ...
|
| 90 |
for module in pipe.transformer.modules():
|
| 91 |
if hasattr(module, "scaling"):
|
| 92 |
-
module.scaling = {k: v *
|
| 93 |
```
|
| 94 |
|
| 95 |
-
Recommended
|
| 96 |
|
| 97 |
---
|
| 98 |
|
|
@@ -224,8 +224,8 @@ Prepend `by <artist>, ` at the start of your prompt.
|
|
| 224 |
- **Base model**: `Tongyi-MAI/Z-Image-Turbo` (8-step flow matching, CFG-free)
|
| 225 |
- **Method**: PEFT LoRA, rank=32, alpha=32, dropout=0.05
|
| 226 |
- **Dataset**: `DownFlow/fuliji` filtered to artists with ≥ 21 images
|
| 227 |
-
- **Steps**: 3 000
|
| 228 |
-
- **Optimizer**: AdamW, lr=1e-4, warmup=100 steps
|
| 229 |
- **Batch**: 1 × 4 gradient accumulation = effective batch 4
|
| 230 |
- **Augmentation**: horizontal flip, caption dropout 5%, timestep bias 1.2
|
| 231 |
- **Regularisation**: 25% of batches sample from a 277-image generic dataset
|
|
|
|
| 34 |
| Target modules | `to_q`, `to_k`, `to_v`, `w1`, `w2`, `w3` |
|
| 35 |
| Trainable params | ~39 M |
|
| 36 |
| Adapter size | ~271 MB |
|
| 37 |
+
| Training steps | **5 000** (3 000 at lr=1e-4 + 2 000 continued at lr=5e-5, EMA) |
|
| 38 |
| Training resolution | 512 × 512 |
|
| 39 |
| Dataset | [DownFlow/fuliji](https://huggingface.co/datasets/DownFlow/fuliji) (8 artists, ~200 images) |
|
| 40 |
|
|
|
|
| 89 |
# After PeftModel.from_pretrained ...
|
| 90 |
for module in pipe.transformer.modules():
|
| 91 |
if hasattr(module, "scaling"):
|
| 92 |
+
module.scaling = {k: v * 3.0 for k, v in module.scaling.items()}
|
| 93 |
```
|
| 94 |
|
| 95 |
+
Recommended value: **3.0** (step-5000 EMA, strong identity with no colour artefacts on 8-step inference). Lighter alternative: 1.2. Values above 5 may saturate style.
|
| 96 |
|
| 97 |
---
|
| 98 |
|
|
|
|
| 224 |
- **Base model**: `Tongyi-MAI/Z-Image-Turbo` (8-step flow matching, CFG-free)
|
| 225 |
- **Method**: PEFT LoRA, rank=32, alpha=32, dropout=0.05
|
| 226 |
- **Dataset**: `DownFlow/fuliji` filtered to artists with ≥ 21 images
|
| 227 |
+
- **Steps**: **5 000** — 3 000 initial (lr=1e-4) + 2 000 continuation (lr=5e-5, resumed from step 3000 EMA)
|
| 228 |
+
- **Optimizer**: AdamW, lr=1e-4→5e-5, warmup=100 steps each phase
|
| 229 |
- **Batch**: 1 × 4 gradient accumulation = effective batch 4
|
| 230 |
- **Augmentation**: horizontal flip, caption dropout 5%, timestep bias 1.2
|
| 231 |
- **Regularisation**: 25% of batches sample from a 277-image generic dataset
|