GCStream commited on
Commit
b1216f2
·
verified ·
1 Parent(s): f545af6

docs: update model card — 5000-step run, scale=3.0 recommendation

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -34,7 +34,7 @@ A **PEFT LoRA adapter** trained on top of [Tongyi-MAI/Z-Image-Turbo](https://hug
34
  | Target modules | `to_q`, `to_k`, `to_v`, `w1`, `w2`, `w3` |
35
  | Trainable params | ~39 M |
36
  | Adapter size | ~271 MB |
37
- | Training steps | 3 000 |
38
  | Training resolution | 512 × 512 |
39
  | Dataset | [DownFlow/fuliji](https://huggingface.co/datasets/DownFlow/fuliji) (8 artists, ~200 images) |
40
 
@@ -89,10 +89,10 @@ PEFT exposes a scaling multiplier per adapter. Increase it to push the style har
89
  # After PeftModel.from_pretrained ...
90
  for module in pipe.transformer.modules():
91
  if hasattr(module, "scaling"):
92
- module.scaling = {k: v * 1.5 for k, v in module.scaling.items()}
93
  ```
94
 
95
- Recommended range: **1.0 2.0**. Values above 3.0 may cause colour artefacts.
96
 
97
  ---
98
 
@@ -224,8 +224,8 @@ Prepend `by <artist>, ` at the start of your prompt.
224
  - **Base model**: `Tongyi-MAI/Z-Image-Turbo` (8-step flow matching, CFG-free)
225
  - **Method**: PEFT LoRA, rank=32, alpha=32, dropout=0.05
226
  - **Dataset**: `DownFlow/fuliji` filtered to artists with ≥ 21 images
227
- - **Steps**: 3 000 with EMA (decay=0.9999)
228
- - **Optimizer**: AdamW, lr=1e-4, warmup=100 steps
229
  - **Batch**: 1 × 4 gradient accumulation = effective batch 4
230
  - **Augmentation**: horizontal flip, caption dropout 5%, timestep bias 1.2
231
  - **Regularisation**: 25% of batches sample from a 277-image generic dataset
 
34
  | Target modules | `to_q`, `to_k`, `to_v`, `w1`, `w2`, `w3` |
35
  | Trainable params | ~39 M |
36
  | Adapter size | ~271 MB |
37
+ | Training steps | **5 000** (3 000 at lr=1e-4 + 2 000 continued at lr=5e-5, EMA) |
38
  | Training resolution | 512 × 512 |
39
  | Dataset | [DownFlow/fuliji](https://huggingface.co/datasets/DownFlow/fuliji) (8 artists, ~200 images) |
40
 
 
89
  # After PeftModel.from_pretrained ...
90
  for module in pipe.transformer.modules():
91
  if hasattr(module, "scaling"):
92
+ module.scaling = {k: v * 3.0 for k, v in module.scaling.items()}
93
  ```
94
 
95
+ Recommended value: **3.0** (step-5000 EMA, strong identity with no colour artefacts on 8-step inference). Lighter alternative: 1.2. Values above 5 may saturate style.
96
 
97
  ---
98
 
 
224
  - **Base model**: `Tongyi-MAI/Z-Image-Turbo` (8-step flow matching, CFG-free)
225
  - **Method**: PEFT LoRA, rank=32, alpha=32, dropout=0.05
226
  - **Dataset**: `DownFlow/fuliji` filtered to artists with ≥ 21 images
227
+ - **Steps**: **5 000** — 3 000 initial (lr=1e-4) + 2 000 continuation (lr=5e-5, resumed from step 3000 EMA)
228
+ - **Optimizer**: AdamW, lr=1e-4→5e-5, warmup=100 steps each phase
229
  - **Batch**: 1 × 4 gradient accumulation = effective batch 4
230
  - **Augmentation**: horizontal flip, caption dropout 5%, timestep bias 1.2
231
  - **Regularisation**: 25% of batches sample from a 277-image generic dataset