|
|
--- |
|
|
tags: |
|
|
- text-to-image |
|
|
- lora |
|
|
- diffusers |
|
|
- template:diffusion-lora |
|
|
license: apache-2.0 |
|
|
base_model: |
|
|
- Tongyi-MAI/Z-Image-Turbo |
|
|
pipeline_tag: text-to-image |
|
|
library_name: diffusers |
|
|
--- |
|
|
|
|
|
# zimage_lora — Z-Image Turbo LoRA experiments |
|
|
|
|
|
This repository contains training artifacts and provenance for a LoRA adapter trained on top of Tongyi-MAI/Z-Image-Turbo (Z-Image Turbo). The main experiment in this repo is the `16/` folder, which holds a LoRA tuned with a base rank of 16 for linear and conv adapters. |
|
|
|
|
|
Purpose |
|
|
- Collect, preserve and document a training run so it can be uploaded to Hugging Face or reproduced locally. |
|
|
|
|
|
Contents |
|
|
- `16/` — experiment folder with model artifacts, training config and logs. |
|
|
|
|
|
Quick summary |
|
|
- Base model: Tongyi-MAI/Z-Image-Turbo (arch: zimage:turbo) |
|
|
- LoRA type: UNet LoRA with rank 16 (linear and conv set to 16, alpha 16) |
|
|
- Training steps: 3000 (checkpoint saves every 250 steps) |
|
|
- Output format: diffusers safetensors |
|
|
|
|
|
Usage notes |
|
|
- Files in `16/` are training artifacts. To use the resulting LoRA, merge or load the safetensors using a compatible Z-Image-Turbo pipeline (Diffusers or a compatible loader). |
|
|
|
|
|
For maintainers (see AGENTS.md) — how to review, test, and publish artifacts is documented in `AGENTS.md`. |
|
|
|
|
|
If you need a short, experiment-focused README (example usage and provenance) look in `16/README.md`. |
|
|
|