Upload 32 files
Browse files- checkpoint-3200/README.md +129 -0
- checkpoint-3200/optimizer.bin +3 -0
- checkpoint-3200/pytorch_lora_weights.safetensors +3 -0
- checkpoint-3200/random_states_0.pkl +3 -0
- checkpoint-3200/scheduler.bin +3 -0
- checkpoint-3200/simpletuner_config.json +387 -0
- checkpoint-3200/training_state-training-images.json +1 -0
- checkpoint-3200/training_state.json +1 -0
- checkpoint-3400/README.md +129 -0
- checkpoint-3400/optimizer.bin +3 -0
- checkpoint-3400/pytorch_lora_weights.safetensors +3 -0
- checkpoint-3400/random_states_0.pkl +3 -0
- checkpoint-3400/scheduler.bin +3 -0
- checkpoint-3400/simpletuner_config.json +387 -0
- checkpoint-3400/training_state-training-images.json +1 -0
- checkpoint-3400/training_state.json +1 -0
- checkpoint-3600/README.md +129 -0
- checkpoint-3600/optimizer.bin +3 -0
- checkpoint-3600/pytorch_lora_weights.safetensors +3 -0
- checkpoint-3600/random_states_0.pkl +3 -0
- checkpoint-3600/scheduler.bin +3 -0
- checkpoint-3600/simpletuner_config.json +387 -0
- checkpoint-3600/training_state-training-images.json +1 -0
- checkpoint-3600/training_state.json +1 -0
- checkpoint-4000/README.md +129 -0
- checkpoint-4000/optimizer.bin +3 -0
- checkpoint-4000/pytorch_lora_weights.safetensors +3 -0
- checkpoint-4000/random_states_0.pkl +3 -0
- checkpoint-4000/scheduler.bin +3 -0
- checkpoint-4000/simpletuner_config.json +387 -0
- checkpoint-4000/training_state-training-images.json +1 -0
- checkpoint-4000/training_state.json +1 -0
checkpoint-3200/README.md
ADDED
|
@@ -0,0 +1,129 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: other
|
| 3 |
+
base_model: "black-forest-labs/FLUX.2-dev"
|
| 4 |
+
tags:
|
| 5 |
+
- flux2
|
| 6 |
+
- flux2-diffusers
|
| 7 |
+
- text-to-image
|
| 8 |
+
- image-to-image
|
| 9 |
+
- diffusers
|
| 10 |
+
- simpletuner
|
| 11 |
+
- not-for-all-audiences
|
| 12 |
+
- lora
|
| 13 |
+
|
| 14 |
+
- template:sd-lora
|
| 15 |
+
- standard
|
| 16 |
+
pipeline_tag: text-to-image
|
| 17 |
+
inference: true
|
| 18 |
+
|
| 19 |
+
---
|
| 20 |
+
|
| 21 |
+
# quzo/fl2
|
| 22 |
+
|
| 23 |
+
This is a PEFT LoRA derived from [black-forest-labs/FLUX.2-dev](https://huggingface.co/black-forest-labs/FLUX.2-dev).
|
| 24 |
+
|
| 25 |
+
The main validation prompt used during training was:
|
| 26 |
+
```
|
| 27 |
+
bm82 man
|
| 28 |
+
```
|
| 29 |
+
|
| 30 |
+
|
| 31 |
+
## Validation settings
|
| 32 |
+
- CFG: `7.5`
|
| 33 |
+
- CFG Rescale: `0.0`
|
| 34 |
+
- Steps: `20`
|
| 35 |
+
- Sampler: `FlowMatchEulerDiscreteScheduler`
|
| 36 |
+
- Seed: `None`
|
| 37 |
+
- Resolution: `1024x1024`
|
| 38 |
+
|
| 39 |
+
|
| 40 |
+
Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
|
| 41 |
+
|
| 42 |
+
|
| 43 |
+
|
| 44 |
+
|
| 45 |
+
<Gallery />
|
| 46 |
+
|
| 47 |
+
The text encoder **was not** trained.
|
| 48 |
+
You may reuse the base model text encoder for inference.
|
| 49 |
+
|
| 50 |
+
|
| 51 |
+
## Training settings
|
| 52 |
+
|
| 53 |
+
- Training epochs: 533
|
| 54 |
+
- Training steps: 3200
|
| 55 |
+
- Learning rate: 0.0001
|
| 56 |
+
- Learning rate schedule: constant_with_warmup
|
| 57 |
+
- Warmup steps: 0
|
| 58 |
+
- Max grad value: 2.0
|
| 59 |
+
- Effective batch size: 2
|
| 60 |
+
- Micro-batch size: 2
|
| 61 |
+
- Gradient accumulation steps: 1
|
| 62 |
+
- Number of GPUs: 1
|
| 63 |
+
- Gradient checkpointing: True
|
| 64 |
+
- Prediction type: flow_matching[]
|
| 65 |
+
- Optimizer: adamw_bf16
|
| 66 |
+
- Trainable parameter precision: Pure BF16
|
| 67 |
+
- Base model precision: `no_change`
|
| 68 |
+
- Caption dropout probability: 0.1%
|
| 69 |
+
|
| 70 |
+
|
| 71 |
+
|
| 72 |
+
- LoRA Rank: 16
|
| 73 |
+
- LoRA Alpha: 16.0
|
| 74 |
+
- LoRA Dropout: 0.1
|
| 75 |
+
- LoRA initialisation style: default
|
| 76 |
+
- LoRA mode: Standard
|
| 77 |
+
|
| 78 |
+
|
| 79 |
+
## Datasets
|
| 80 |
+
|
| 81 |
+
### training-images
|
| 82 |
+
- Repeats: 0
|
| 83 |
+
- Total number of images: 12
|
| 84 |
+
- Total number of aspect buckets: 2
|
| 85 |
+
- Resolution: 1.048576 megapixels
|
| 86 |
+
- Cropped: False
|
| 87 |
+
- Crop style: None
|
| 88 |
+
- Crop aspect: None
|
| 89 |
+
- Used for regularisation data: No
|
| 90 |
+
|
| 91 |
+
|
| 92 |
+
## Inference
|
| 93 |
+
|
| 94 |
+
|
| 95 |
+
```python
|
| 96 |
+
import torch
|
| 97 |
+
from diffusers import DiffusionPipeline
|
| 98 |
+
|
| 99 |
+
model_id = 'black-forest-labs/FLUX.2-dev'
|
| 100 |
+
adapter_id = 'quzo/fl2'
|
| 101 |
+
pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
|
| 102 |
+
pipeline.load_lora_weights(adapter_id)
|
| 103 |
+
|
| 104 |
+
prompt = "bm82 man"
|
| 105 |
+
negative_prompt = 'blurry, cropped, ugly'
|
| 106 |
+
|
| 107 |
+
## Optional: quantise the model to save on vram.
|
| 108 |
+
## Note: The model was not quantised during training, so it is not necessary to quantise it during inference time.
|
| 109 |
+
#from optimum.quanto import quantize, freeze, qint8
|
| 110 |
+
#quantize(pipeline.transformer, weights=qint8)
|
| 111 |
+
#freeze(pipeline.transformer)
|
| 112 |
+
|
| 113 |
+
pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
|
| 114 |
+
model_output = pipeline(
|
| 115 |
+
prompt=prompt,
|
| 116 |
+
negative_prompt=negative_prompt,
|
| 117 |
+
num_inference_steps=20,
|
| 118 |
+
generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
|
| 119 |
+
width=1024,
|
| 120 |
+
height=1024,
|
| 121 |
+
guidance_scale=7.5,
|
| 122 |
+
).images[0]
|
| 123 |
+
|
| 124 |
+
model_output.save("output.png", format="PNG")
|
| 125 |
+
|
| 126 |
+
```
|
| 127 |
+
|
| 128 |
+
|
| 129 |
+
|
checkpoint-3200/optimizer.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bcd3a3a7c184a0031b7eef26313fd7020b0b6fcca1949725feca4d9993b0f896
|
| 3 |
+
size 321008843
|
checkpoint-3200/pytorch_lora_weights.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:88ade5a67008a9aba49b0b6aabf1dd60c57d072029937bf5fb092c49a212413f
|
| 3 |
+
size 106979136
|
checkpoint-3200/random_states_0.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:05c0092c4c618520653c8345b12f78775995d65375428b89ab85aaa646f7482d
|
| 3 |
+
size 14757
|
checkpoint-3200/scheduler.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ce701db56c2af726940feb1b070907060c446d2e48225a0937aba9bfefcf32e6
|
| 3 |
+
size 1401
|
checkpoint-3200/simpletuner_config.json
ADDED
|
@@ -0,0 +1,387 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"model_family": "flux2",
|
| 3 |
+
"model_flavour": "dev",
|
| 4 |
+
"controlnet": false,
|
| 5 |
+
"pretrained_model_name_or_path": "black-forest-labs/FLUX.2-dev",
|
| 6 |
+
"output_dir": "/home/Ubuntu/.simpletuner/output",
|
| 7 |
+
"logging_dir": "/home/Ubuntu/.simpletuner/output/logs",
|
| 8 |
+
"model_type": "lora",
|
| 9 |
+
"seed": null,
|
| 10 |
+
"resolution": 1024,
|
| 11 |
+
"resume_from_checkpoint": null,
|
| 12 |
+
"prediction_type": null,
|
| 13 |
+
"pretrained_vae_model_name_or_path": "black-forest-labs/FLUX.2-dev",
|
| 14 |
+
"vae_dtype": "bf16",
|
| 15 |
+
"vae_cache_ondemand": false,
|
| 16 |
+
"vae_cache_disable": false,
|
| 17 |
+
"accelerator_cache_clear_interval": null,
|
| 18 |
+
"aspect_bucket_rounding": 2,
|
| 19 |
+
"base_model_precision": "no_change",
|
| 20 |
+
"text_encoder_1_precision": "no_change",
|
| 21 |
+
"text_encoder_2_precision": "no_change",
|
| 22 |
+
"text_encoder_3_precision": "no_change",
|
| 23 |
+
"text_encoder_4_precision": "no_change",
|
| 24 |
+
"gradient_checkpointing_interval": null,
|
| 25 |
+
"offload_during_startup": false,
|
| 26 |
+
"quantize_via": "accelerator",
|
| 27 |
+
"wan_force_2_1_time_embedding": false,
|
| 28 |
+
"fuse_qkv_projections": false,
|
| 29 |
+
"rescale_betas_zero_snr": false,
|
| 30 |
+
"control": false,
|
| 31 |
+
"controlnet_custom_config": null,
|
| 32 |
+
"controlnet_model_name_or_path": null,
|
| 33 |
+
"tread_config": null,
|
| 34 |
+
"pretrained_transformer_model_name_or_path": null,
|
| 35 |
+
"pretrained_transformer_subfolder": "transformer",
|
| 36 |
+
"pretrained_unet_model_name_or_path": null,
|
| 37 |
+
"pretrained_unet_subfolder": "unet",
|
| 38 |
+
"pretrained_t5_model_name_or_path": null,
|
| 39 |
+
"revision": null,
|
| 40 |
+
"variant": null,
|
| 41 |
+
"base_model_default_dtype": "bf16",
|
| 42 |
+
"unet_attention_slice": false,
|
| 43 |
+
"num_train_epochs": 834,
|
| 44 |
+
"max_train_steps": 5000,
|
| 45 |
+
"train_batch_size": 2,
|
| 46 |
+
"learning_rate": 0.0001,
|
| 47 |
+
"optimizer": "adamw_bf16",
|
| 48 |
+
"lr_scheduler": "constant_with_warmup",
|
| 49 |
+
"gradient_accumulation_steps": 1,
|
| 50 |
+
"lr_warmup_steps": 0,
|
| 51 |
+
"checkpoints_total_limit": 5,
|
| 52 |
+
"gradient_checkpointing": true,
|
| 53 |
+
"enable_group_offload": false,
|
| 54 |
+
"ramtorch": false,
|
| 55 |
+
"ramtorch_target_modules": null,
|
| 56 |
+
"ramtorch_text_encoder": false,
|
| 57 |
+
"ramtorch_vae": false,
|
| 58 |
+
"ramtorch_controlnet": false,
|
| 59 |
+
"group_offload_type": "block_level",
|
| 60 |
+
"group_offload_blocks_per_group": 1,
|
| 61 |
+
"group_offload_use_stream": false,
|
| 62 |
+
"group_offload_to_disk_path": "",
|
| 63 |
+
"group_offload_text_encoder": false,
|
| 64 |
+
"group_offload_vae": false,
|
| 65 |
+
"offload_during_save": false,
|
| 66 |
+
"enable_chunked_feed_forward": false,
|
| 67 |
+
"feed_forward_chunk_size": null,
|
| 68 |
+
"train_text_encoder": false,
|
| 69 |
+
"text_encoder_lr": null,
|
| 70 |
+
"lyrics_embedder_train": false,
|
| 71 |
+
"lyrics_embedder_optimizer": null,
|
| 72 |
+
"lyrics_embedder_lr": null,
|
| 73 |
+
"lyrics_embedder_lr_scheduler": null,
|
| 74 |
+
"lr_num_cycles": 1,
|
| 75 |
+
"lr_power": 0.8,
|
| 76 |
+
"use_soft_min_snr": false,
|
| 77 |
+
"use_ema": false,
|
| 78 |
+
"ema_device": "cpu",
|
| 79 |
+
"ema_cpu_only": false,
|
| 80 |
+
"ema_update_interval": 1,
|
| 81 |
+
"ema_foreach_disable": false,
|
| 82 |
+
"ema_decay": 0.995,
|
| 83 |
+
"lora_rank": 16,
|
| 84 |
+
"lora_alpha": 16.0,
|
| 85 |
+
"lora_type": "standard",
|
| 86 |
+
"lora_dropout": 0.1,
|
| 87 |
+
"lora_init_type": "default",
|
| 88 |
+
"peft_lora_mode": "standard",
|
| 89 |
+
"singlora_ramp_up_steps": 0,
|
| 90 |
+
"init_lora": null,
|
| 91 |
+
"lycoris_config": "/home/Ubuntu/.simpletuner/configs/config/lycoris_config.json",
|
| 92 |
+
"init_lokr_norm": null,
|
| 93 |
+
"flux_lora_target": "all",
|
| 94 |
+
"acestep_lora_target": "attn_qkv+linear_qkv",
|
| 95 |
+
"use_dora": false,
|
| 96 |
+
"resolution_type": "pixel_area",
|
| 97 |
+
"data_backend_config": "/home/Ubuntu/.simpletuner/configs/111/multidatabackend-bartek_mizak.json",
|
| 98 |
+
"caption_strategy": "filename",
|
| 99 |
+
"conditioning_multidataset_sampling": "random",
|
| 100 |
+
"instance_prompt": null,
|
| 101 |
+
"parquet_caption_column": null,
|
| 102 |
+
"parquet_filename_column": null,
|
| 103 |
+
"ignore_missing_files": false,
|
| 104 |
+
"vae_cache_scan_behaviour": "recreate",
|
| 105 |
+
"vae_enable_slicing": false,
|
| 106 |
+
"vae_enable_tiling": false,
|
| 107 |
+
"vae_batch_size": 4,
|
| 108 |
+
"max_upscale_threshold": null,
|
| 109 |
+
"caption_dropout_probability": 0.1,
|
| 110 |
+
"tokenizer_max_length": 512,
|
| 111 |
+
"audio_max_duration_seconds": null,
|
| 112 |
+
"audio_min_duration_seconds": null,
|
| 113 |
+
"audio_channels": 1,
|
| 114 |
+
"audio_duration_interval": 3.0,
|
| 115 |
+
"audio_truncation_mode": "beginning",
|
| 116 |
+
"validation_step_interval": 100,
|
| 117 |
+
"validation_epoch_interval": null,
|
| 118 |
+
"disable_benchmark": false,
|
| 119 |
+
"validation_preview": false,
|
| 120 |
+
"validation_preview_steps": 1,
|
| 121 |
+
"validation_prompt": "bm82 man",
|
| 122 |
+
"validation_lyrics": null,
|
| 123 |
+
"validation_audio_duration": 30.0,
|
| 124 |
+
"num_validation_images": 1,
|
| 125 |
+
"num_eval_images": 4,
|
| 126 |
+
"eval_steps_interval": null,
|
| 127 |
+
"eval_timesteps": 28,
|
| 128 |
+
"eval_dataset_pooling": false,
|
| 129 |
+
"evaluation_type": "none",
|
| 130 |
+
"pretrained_evaluation_model_name_or_path": "openai/clip-vit-large-patch14-336",
|
| 131 |
+
"validation_guidance": 7.5,
|
| 132 |
+
"validation_num_inference_steps": 20,
|
| 133 |
+
"validation_on_startup": false,
|
| 134 |
+
"validation_method": "simpletuner-local",
|
| 135 |
+
"validation_external_script": null,
|
| 136 |
+
"validation_external_background": false,
|
| 137 |
+
"validation_using_datasets": false,
|
| 138 |
+
"validation_torch_compile": false,
|
| 139 |
+
"validation_guidance_real": 1.0,
|
| 140 |
+
"validation_no_cfg_until_timestep": 2,
|
| 141 |
+
"validation_negative_prompt": "blurry, cropped, ugly",
|
| 142 |
+
"validation_randomize": false,
|
| 143 |
+
"validation_seed": null,
|
| 144 |
+
"validation_multigpu": "batch-parallel",
|
| 145 |
+
"validation_disable": false,
|
| 146 |
+
"validation_prompt_library": false,
|
| 147 |
+
"user_prompt_library": null,
|
| 148 |
+
"eval_dataset_id": null,
|
| 149 |
+
"validation_stitch_input_location": "left",
|
| 150 |
+
"validation_guidance_rescale": 0.0,
|
| 151 |
+
"validation_disable_unconditional": false,
|
| 152 |
+
"validation_guidance_skip_layers": null,
|
| 153 |
+
"validation_guidance_skip_layers_start": 0.01,
|
| 154 |
+
"validation_guidance_skip_layers_stop": 0.2,
|
| 155 |
+
"validation_guidance_skip_scale": 2.8,
|
| 156 |
+
"validation_lycoris_strength": 1.0,
|
| 157 |
+
"validation_noise_scheduler": null,
|
| 158 |
+
"validation_num_video_frames": null,
|
| 159 |
+
"validation_resolution": "1024x1024",
|
| 160 |
+
"validation_seed_source": "cpu",
|
| 161 |
+
"validation_torch_compile_mode": "max-autotune",
|
| 162 |
+
"validation_adapter_path": null,
|
| 163 |
+
"validation_adapter_name": null,
|
| 164 |
+
"validation_adapter_strength": 1.0,
|
| 165 |
+
"validation_adapter_mode": "adapter_only",
|
| 166 |
+
"validation_adapter_config": null,
|
| 167 |
+
"i_know_what_i_am_doing": false,
|
| 168 |
+
"flow_sigmoid_scale": 1.0,
|
| 169 |
+
"flux_fast_schedule": false,
|
| 170 |
+
"flow_use_uniform_schedule": false,
|
| 171 |
+
"flow_use_beta_schedule": false,
|
| 172 |
+
"flow_beta_schedule_alpha": 2.0,
|
| 173 |
+
"flow_beta_schedule_beta": 2.0,
|
| 174 |
+
"flow_schedule_shift": 3.0,
|
| 175 |
+
"flow_schedule_auto_shift": false,
|
| 176 |
+
"flux_guidance_mode": "constant",
|
| 177 |
+
"flux_attention_masked_training": false,
|
| 178 |
+
"flux_guidance_value": 1.0,
|
| 179 |
+
"flux_guidance_min": 0.0,
|
| 180 |
+
"flux_guidance_max": 4.0,
|
| 181 |
+
"t5_padding": "unmodified",
|
| 182 |
+
"sd3_clip_uncond_behaviour": "empty_string",
|
| 183 |
+
"sd3_t5_uncond_behaviour": null,
|
| 184 |
+
"soft_min_snr_sigma_data": null,
|
| 185 |
+
"mixed_precision": "bf16",
|
| 186 |
+
"attention_mechanism": "diffusers",
|
| 187 |
+
"sla_config": null,
|
| 188 |
+
"sageattention_usage": {
|
| 189 |
+
"_value_": "inference",
|
| 190 |
+
"_name_": "INFERENCE",
|
| 191 |
+
"__objclass__": "{'_generate_next_value_': <staticmethod(<function Enum._generate_next_value_ at 0x7649ca9cdf80>)>, '__module__': 'simpletuner.helpers.training.attention_backend', 'allows_training': <property object at 0x764872f93880>, 'allows_inference': <property object at 0x764872f93790>, 'from_raw': <classmethod(<function AttentionBackendMode.from_raw at 0x764872fe85e0>)>, '_new_member_': <built-in method __new__ of type object at 0xa234e0>, '_use_args_': True, '_member_names_': ['TRAINING', 'INFERENCE', 'TRAINING_AND_INFERENCE'], '_member_map_': {'TRAINING': <AttentionBackendMode.TRAINING: 'training'>, 'INFERENCE': <AttentionBackendMode.INFERENCE: 'inference'>, 'TRAINING_AND_INFERENCE': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>}, '_value2member_map_': {'training': <AttentionBackendMode.TRAINING: 'training'>, 'inference': <AttentionBackendMode.INFERENCE: 'inference'>, 'training+inference': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>}, '_unhashable_values_': [], '_member_type_': <class 'str'>, '_value_repr_': <slot wrapper '__repr__' of 'str' objects>, '__dict__': <attribute '__dict__' of 'AttentionBackendMode' objects>, '__weakref__': <attribute '__weakref__' of 'AttentionBackendMode' objects>, '__doc__': None, 'TRAINING': <AttentionBackendMode.TRAINING: 'training'>, 'INFERENCE': <AttentionBackendMode.INFERENCE: 'inference'>, 'TRAINING_AND_INFERENCE': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>, '__repr__': <function Enum.__repr__ at 0x7649ca9ce0c0>, '__str__': <function Enum.__str__ at 0x7649ca9ce160>, '__format__': <function Enum.__format__ at 0x7649ca9ce2a0>, '__new__': <function Enum.__new__ at 0x7649ca9cde40>}",
|
| 192 |
+
"_sort_order_": 1
|
| 193 |
+
},
|
| 194 |
+
"disable_tf32": false,
|
| 195 |
+
"set_grads_to_none": false,
|
| 196 |
+
"noise_offset": 0.1,
|
| 197 |
+
"noise_offset_probability": 0.25,
|
| 198 |
+
"input_perturbation": 0.0,
|
| 199 |
+
"input_perturbation_steps": 0,
|
| 200 |
+
"lr_end": "4e-7",
|
| 201 |
+
"lr_scale": false,
|
| 202 |
+
"lr_scale_sqrt": false,
|
| 203 |
+
"ignore_final_epochs": false,
|
| 204 |
+
"freeze_encoder_before": 12,
|
| 205 |
+
"freeze_encoder_after": 17,
|
| 206 |
+
"freeze_encoder_strategy": "after",
|
| 207 |
+
"layer_freeze_strategy": null,
|
| 208 |
+
"fully_unload_text_encoder": false,
|
| 209 |
+
"save_text_encoder": false,
|
| 210 |
+
"text_encoder_limit": 100,
|
| 211 |
+
"prepend_instance_prompt": false,
|
| 212 |
+
"only_instance_prompt": false,
|
| 213 |
+
"data_aesthetic_score": 7.0,
|
| 214 |
+
"delete_unwanted_images": false,
|
| 215 |
+
"delete_problematic_images": false,
|
| 216 |
+
"disable_bucket_pruning": false,
|
| 217 |
+
"allow_dataset_oversubscription": false,
|
| 218 |
+
"disable_segmented_timestep_sampling": false,
|
| 219 |
+
"preserve_data_backend_cache": false,
|
| 220 |
+
"override_dataset_config": false,
|
| 221 |
+
"cache_dir": "/home/Ubuntu/.simpletuner/output/cache",
|
| 222 |
+
"cache_dir_text": "cache",
|
| 223 |
+
"cache_dir_vae": "",
|
| 224 |
+
"compress_disk_cache": true,
|
| 225 |
+
"aspect_bucket_disable_rebuild": false,
|
| 226 |
+
"keep_vae_loaded": false,
|
| 227 |
+
"skip_file_discovery": "",
|
| 228 |
+
"data_backend_sampling": "auto-weighting",
|
| 229 |
+
"image_processing_batch_size": 32,
|
| 230 |
+
"write_batch_size": 128,
|
| 231 |
+
"read_batch_size": 25,
|
| 232 |
+
"enable_multiprocessing": false,
|
| 233 |
+
"accelerate_config": null,
|
| 234 |
+
"deepspeed_config": null,
|
| 235 |
+
"fsdp_enable": false,
|
| 236 |
+
"fsdp_version": 2,
|
| 237 |
+
"fsdp_reshard_after_forward": false,
|
| 238 |
+
"fsdp_state_dict_type": "SHARDED_STATE_DICT",
|
| 239 |
+
"fsdp_cpu_ram_efficient_loading": false,
|
| 240 |
+
"fsdp_auto_wrap_policy": "TRANSFORMER_BASED_WRAP",
|
| 241 |
+
"fsdp_limit_all_gathers": false,
|
| 242 |
+
"fsdp_cpu_offload": false,
|
| 243 |
+
"fsdp_activation_checkpointing": false,
|
| 244 |
+
"fsdp_transformer_layer_cls_to_wrap": null,
|
| 245 |
+
"context_parallel_size": 1,
|
| 246 |
+
"context_parallel_comm_strategy": "allgather",
|
| 247 |
+
"num_processes": 1,
|
| 248 |
+
"num_machines": 1,
|
| 249 |
+
"accelerate_extra_args": null,
|
| 250 |
+
"main_process_ip": "127.0.0.1",
|
| 251 |
+
"main_process_port": 29500,
|
| 252 |
+
"machine_rank": 0,
|
| 253 |
+
"same_network": true,
|
| 254 |
+
"dynamo_backend": "no",
|
| 255 |
+
"dynamo_mode": "",
|
| 256 |
+
"dynamo_fullgraph": false,
|
| 257 |
+
"dynamo_dynamic": false,
|
| 258 |
+
"dynamo_use_regional_compilation": false,
|
| 259 |
+
"max_workers": 32,
|
| 260 |
+
"aws_max_pool_connections": 128,
|
| 261 |
+
"torch_num_threads": 8,
|
| 262 |
+
"dataloader_prefetch": false,
|
| 263 |
+
"dataloader_prefetch_qlen": 10,
|
| 264 |
+
"aspect_bucket_worker_count": 12,
|
| 265 |
+
"aspect_bucket_alignment": 16,
|
| 266 |
+
"minimum_image_size": null,
|
| 267 |
+
"maximum_image_size": null,
|
| 268 |
+
"target_downsample_size": null,
|
| 269 |
+
"metadata_update_interval": 3600,
|
| 270 |
+
"debug_aspect_buckets": false,
|
| 271 |
+
"debug_dataset_loader": false,
|
| 272 |
+
"print_filenames": false,
|
| 273 |
+
"print_sampler_statistics": false,
|
| 274 |
+
"timestep_bias_strategy": null,
|
| 275 |
+
"timestep_bias_begin": 0,
|
| 276 |
+
"timestep_bias_end": 1000,
|
| 277 |
+
"timestep_bias_multiplier": 1.0,
|
| 278 |
+
"timestep_bias_portion": 0.25,
|
| 279 |
+
"training_scheduler_timestep_spacing": "trailing",
|
| 280 |
+
"inference_scheduler_timestep_spacing": "trailing",
|
| 281 |
+
"loss_type": "l2",
|
| 282 |
+
"huber_schedule": "snr",
|
| 283 |
+
"huber_c": 0.1,
|
| 284 |
+
"snr_gamma": null,
|
| 285 |
+
"masked_loss_probability": 1.0,
|
| 286 |
+
"hidream_use_load_balancing_loss": false,
|
| 287 |
+
"hidream_load_balancing_loss_weight": null,
|
| 288 |
+
"adam_beta1": 0.9,
|
| 289 |
+
"adam_beta2": 0.999,
|
| 290 |
+
"optimizer_beta1": null,
|
| 291 |
+
"optimizer_beta2": null,
|
| 292 |
+
"optimizer_cpu_offload_method": null,
|
| 293 |
+
"gradient_precision": null,
|
| 294 |
+
"adam_weight_decay": 0.01,
|
| 295 |
+
"adam_epsilon": 1e-08,
|
| 296 |
+
"prodigy_steps": null,
|
| 297 |
+
"max_grad_norm": 2.0,
|
| 298 |
+
"optimizer_config": null,
|
| 299 |
+
"grad_clip_method": "value",
|
| 300 |
+
"optimizer_offload_gradients": false,
|
| 301 |
+
"fuse_optimizer": false,
|
| 302 |
+
"optimizer_release_gradients": false,
|
| 303 |
+
"push_to_hub": false,
|
| 304 |
+
"publishing_config": null,
|
| 305 |
+
"post_upload_script": null,
|
| 306 |
+
"push_checkpoints_to_hub": false,
|
| 307 |
+
"push_to_hub_background": false,
|
| 308 |
+
"hub_model_id": "quzo/fl2",
|
| 309 |
+
"model_card_private": false,
|
| 310 |
+
"model_card_safe_for_work": false,
|
| 311 |
+
"model_card_note": null,
|
| 312 |
+
"report_to": "none",
|
| 313 |
+
"custom_tracker": null,
|
| 314 |
+
"checkpoint_step_interval": 200,
|
| 315 |
+
"checkpoint_epoch_interval": null,
|
| 316 |
+
"checkpointing_rolling_steps": 0,
|
| 317 |
+
"checkpointing_use_tempdir": false,
|
| 318 |
+
"checkpoints_rolling_total_limit": 1,
|
| 319 |
+
"tracker_run_name": "simpletuner-testing",
|
| 320 |
+
"tracker_project_name": "simpletuner",
|
| 321 |
+
"tracker_image_layout": "gallery",
|
| 322 |
+
"enable_watermark": false,
|
| 323 |
+
"framerate": null,
|
| 324 |
+
"seed_for_each_device": true,
|
| 325 |
+
"snr_weight": 1.0,
|
| 326 |
+
"webhook_config": [
|
| 327 |
+
{
|
| 328 |
+
"webhook_type": "raw",
|
| 329 |
+
"callback_url": "http://0.0.0.0:8001/callback",
|
| 330 |
+
"log_level": "info",
|
| 331 |
+
"ssl_no_verify": false
|
| 332 |
+
}
|
| 333 |
+
],
|
| 334 |
+
"webhook_reporting_interval": null,
|
| 335 |
+
"distillation_method": null,
|
| 336 |
+
"distillation_config": null,
|
| 337 |
+
"ema_validation": "comparison",
|
| 338 |
+
"local_rank": -1,
|
| 339 |
+
"ltx_train_mode": "i2v",
|
| 340 |
+
"ltx_i2v_prob": 0.1,
|
| 341 |
+
"ltx_partial_noise_fraction": 0.05,
|
| 342 |
+
"ltx_protect_first_frame": false,
|
| 343 |
+
"offload_param_path": null,
|
| 344 |
+
"offset_noise": false,
|
| 345 |
+
"quantize_activations": false,
|
| 346 |
+
"refiner_training": false,
|
| 347 |
+
"refiner_training_invert_schedule": false,
|
| 348 |
+
"refiner_training_strength": 0.2,
|
| 349 |
+
"sdxl_refiner_uses_full_range": false,
|
| 350 |
+
"sana_complex_human_instruction": "Given a user prompt, generate an 'Enhanced prompt' that provides detailed visual descriptions suitable for image generation. Evaluate the level of detail in the user prompt:\n- If the prompt is simple, focus on adding specifics about colors, shapes, sizes, textures, and spatial relationships to create vivid and concrete scenes.\n- Do not add detail that is not in the original prompt! Do not mention colors if the User Prompt does not!\n- If the prompt is already detailed, refine and enhance the existing details slightly without overcomplicating.\n- If the prompt contains explicit material, do not remove this information from the output, which would damage the performance of downstream classifiers.\nHere are examples of how to transform or refine prompts:\n- User Prompt: A cat sleeping -> Enhanced: A cat sleeping peacefully, showcasing the joy of pet ownership. Cute floof kitty cat gatto.\n- User Prompt: A busy city street -> Enhanced: A bustling city street scene featuring a crowd of people.\nPlease generate only the enhanced description for the prompt below and avoid including any additional commentary or evaluations:\nUser Prompt: ",
|
| 351 |
+
"vae_path": "black-forest-labs/FLUX.2-dev",
|
| 352 |
+
"accelerator_project_config": {
|
| 353 |
+
"project_dir": "/home/Ubuntu/.simpletuner/output",
|
| 354 |
+
"logging_dir": "/home/Ubuntu/.simpletuner/output/logs",
|
| 355 |
+
"automatic_checkpoint_naming": false,
|
| 356 |
+
"total_limit": null,
|
| 357 |
+
"iteration": 16,
|
| 358 |
+
"save_on_each_node": false
|
| 359 |
+
},
|
| 360 |
+
"process_group_kwargs": {
|
| 361 |
+
"backend": "nccl",
|
| 362 |
+
"init_method": null,
|
| 363 |
+
"timeout": "1:30:00"
|
| 364 |
+
},
|
| 365 |
+
"is_quantized": false,
|
| 366 |
+
"weight_dtype": "torch.bfloat16",
|
| 367 |
+
"disable_accelerator": false,
|
| 368 |
+
"lora_initialisation_style": true,
|
| 369 |
+
"checkpointing_steps": 200,
|
| 370 |
+
"use_fsdp": false,
|
| 371 |
+
"model_type_label": "Flux.2",
|
| 372 |
+
"use_deepspeed_optimizer": false,
|
| 373 |
+
"use_deepspeed_scheduler": false,
|
| 374 |
+
"base_weight_dtype": "torch.bfloat16",
|
| 375 |
+
"is_quanto": false,
|
| 376 |
+
"is_torchao": false,
|
| 377 |
+
"is_bnb": false,
|
| 378 |
+
"flow_matching": true,
|
| 379 |
+
"overrode_max_train_steps": false,
|
| 380 |
+
"total_num_batches": 6,
|
| 381 |
+
"num_update_steps_per_epoch": 6,
|
| 382 |
+
"total_batch_size": 2,
|
| 383 |
+
"enable_adamw_bf16": true,
|
| 384 |
+
"is_schedulefree": false,
|
| 385 |
+
"is_lr_scheduler_disabled": false,
|
| 386 |
+
"total_steps_remaining_at_start": 5000
|
| 387 |
+
}
|
checkpoint-3200/training_state-training-images.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"aspect_ratio_bucket_indices": {"1.0": ["/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im008.png", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im021.png"], "0.74": ["/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0014.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9743.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9827.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9935.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0003.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9796.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9890.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9865.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9260.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0090.jpg"]}, "buckets": ["1.0", "0.74"], "exhausted_buckets": [], "batch_size": 2, "current_bucket": 1, "seen_images": {"/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im008.png": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im021.png": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9743.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9796.jpg": true}, "current_epoch": 534}
|
checkpoint-3200/training_state.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"global_step": 3200, "epoch_step": 3733, "epoch": 534, "exhausted_backends": [], "repeats": {"training-images": 0}}
|
checkpoint-3400/README.md
ADDED
|
@@ -0,0 +1,129 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: other
|
| 3 |
+
base_model: "black-forest-labs/FLUX.2-dev"
|
| 4 |
+
tags:
|
| 5 |
+
- flux2
|
| 6 |
+
- flux2-diffusers
|
| 7 |
+
- text-to-image
|
| 8 |
+
- image-to-image
|
| 9 |
+
- diffusers
|
| 10 |
+
- simpletuner
|
| 11 |
+
- not-for-all-audiences
|
| 12 |
+
- lora
|
| 13 |
+
|
| 14 |
+
- template:sd-lora
|
| 15 |
+
- standard
|
| 16 |
+
pipeline_tag: text-to-image
|
| 17 |
+
inference: true
|
| 18 |
+
|
| 19 |
+
---
|
| 20 |
+
|
| 21 |
+
# quzo/fl2
|
| 22 |
+
|
| 23 |
+
This is a PEFT LoRA derived from [black-forest-labs/FLUX.2-dev](https://huggingface.co/black-forest-labs/FLUX.2-dev).
|
| 24 |
+
|
| 25 |
+
The main validation prompt used during training was:
|
| 26 |
+
```
|
| 27 |
+
bm82 man
|
| 28 |
+
```
|
| 29 |
+
|
| 30 |
+
|
| 31 |
+
## Validation settings
|
| 32 |
+
- CFG: `7.5`
|
| 33 |
+
- CFG Rescale: `0.0`
|
| 34 |
+
- Steps: `20`
|
| 35 |
+
- Sampler: `FlowMatchEulerDiscreteScheduler`
|
| 36 |
+
- Seed: `None`
|
| 37 |
+
- Resolution: `1024x1024`
|
| 38 |
+
|
| 39 |
+
|
| 40 |
+
Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
|
| 41 |
+
|
| 42 |
+
|
| 43 |
+
|
| 44 |
+
|
| 45 |
+
<Gallery />
|
| 46 |
+
|
| 47 |
+
The text encoder **was not** trained.
|
| 48 |
+
You may reuse the base model text encoder for inference.
|
| 49 |
+
|
| 50 |
+
|
| 51 |
+
## Training settings
|
| 52 |
+
|
| 53 |
+
- Training epochs: 566
|
| 54 |
+
- Training steps: 3400
|
| 55 |
+
- Learning rate: 0.0001
|
| 56 |
+
- Learning rate schedule: constant_with_warmup
|
| 57 |
+
- Warmup steps: 0
|
| 58 |
+
- Max grad value: 2.0
|
| 59 |
+
- Effective batch size: 2
|
| 60 |
+
- Micro-batch size: 2
|
| 61 |
+
- Gradient accumulation steps: 1
|
| 62 |
+
- Number of GPUs: 1
|
| 63 |
+
- Gradient checkpointing: True
|
| 64 |
+
- Prediction type: flow_matching[]
|
| 65 |
+
- Optimizer: adamw_bf16
|
| 66 |
+
- Trainable parameter precision: Pure BF16
|
| 67 |
+
- Base model precision: `no_change`
|
| 68 |
+
- Caption dropout probability: 0.1%
|
| 69 |
+
|
| 70 |
+
|
| 71 |
+
|
| 72 |
+
- LoRA Rank: 16
|
| 73 |
+
- LoRA Alpha: 16.0
|
| 74 |
+
- LoRA Dropout: 0.1
|
| 75 |
+
- LoRA initialisation style: default
|
| 76 |
+
- LoRA mode: Standard
|
| 77 |
+
|
| 78 |
+
|
| 79 |
+
## Datasets
|
| 80 |
+
|
| 81 |
+
### training-images
|
| 82 |
+
- Repeats: 0
|
| 83 |
+
- Total number of images: 12
|
| 84 |
+
- Total number of aspect buckets: 1
|
| 85 |
+
- Resolution: 1.048576 megapixels
|
| 86 |
+
- Cropped: False
|
| 87 |
+
- Crop style: None
|
| 88 |
+
- Crop aspect: None
|
| 89 |
+
- Used for regularisation data: No
|
| 90 |
+
|
| 91 |
+
|
| 92 |
+
## Inference
|
| 93 |
+
|
| 94 |
+
|
| 95 |
+
```python
|
| 96 |
+
import torch
|
| 97 |
+
from diffusers import DiffusionPipeline
|
| 98 |
+
|
| 99 |
+
model_id = 'black-forest-labs/FLUX.2-dev'
|
| 100 |
+
adapter_id = 'quzo/fl2'
|
| 101 |
+
pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
|
| 102 |
+
pipeline.load_lora_weights(adapter_id)
|
| 103 |
+
|
| 104 |
+
prompt = "bm82 man"
|
| 105 |
+
negative_prompt = 'blurry, cropped, ugly'
|
| 106 |
+
|
| 107 |
+
## Optional: quantise the model to save on vram.
|
| 108 |
+
## Note: The model was not quantised during training, so it is not necessary to quantise it during inference time.
|
| 109 |
+
#from optimum.quanto import quantize, freeze, qint8
|
| 110 |
+
#quantize(pipeline.transformer, weights=qint8)
|
| 111 |
+
#freeze(pipeline.transformer)
|
| 112 |
+
|
| 113 |
+
pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
|
| 114 |
+
model_output = pipeline(
|
| 115 |
+
prompt=prompt,
|
| 116 |
+
negative_prompt=negative_prompt,
|
| 117 |
+
num_inference_steps=20,
|
| 118 |
+
generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
|
| 119 |
+
width=1024,
|
| 120 |
+
height=1024,
|
| 121 |
+
guidance_scale=7.5,
|
| 122 |
+
).images[0]
|
| 123 |
+
|
| 124 |
+
model_output.save("output.png", format="PNG")
|
| 125 |
+
|
| 126 |
+
```
|
| 127 |
+
|
| 128 |
+
|
| 129 |
+
|
checkpoint-3400/optimizer.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5a1cf4adf651ea45be3c0204a409ab22a3913476ee69840d1ffce3c8b45697c1
|
| 3 |
+
size 321008843
|
checkpoint-3400/pytorch_lora_weights.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:414a5aff8b7277e413b9f104c7edd4bbcd12ab0d261beb22649b90a62d304fc4
|
| 3 |
+
size 106979136
|
checkpoint-3400/random_states_0.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:01b906f76c1ba1117bc03af59e66750f2205f8a2f01c5e6240e20dc436024aae
|
| 3 |
+
size 14885
|
checkpoint-3400/scheduler.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f02eb14703c391ff6b0683ceeea335d568ec18788de76c6f5aa899ec9eb225ae
|
| 3 |
+
size 1401
|
checkpoint-3400/simpletuner_config.json
ADDED
|
@@ -0,0 +1,387 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"model_family": "flux2",
|
| 3 |
+
"model_flavour": "dev",
|
| 4 |
+
"controlnet": false,
|
| 5 |
+
"pretrained_model_name_or_path": "black-forest-labs/FLUX.2-dev",
|
| 6 |
+
"output_dir": "/home/Ubuntu/.simpletuner/output",
|
| 7 |
+
"logging_dir": "/home/Ubuntu/.simpletuner/output/logs",
|
| 8 |
+
"model_type": "lora",
|
| 9 |
+
"seed": null,
|
| 10 |
+
"resolution": 1024,
|
| 11 |
+
"resume_from_checkpoint": null,
|
| 12 |
+
"prediction_type": null,
|
| 13 |
+
"pretrained_vae_model_name_or_path": "black-forest-labs/FLUX.2-dev",
|
| 14 |
+
"vae_dtype": "bf16",
|
| 15 |
+
"vae_cache_ondemand": false,
|
| 16 |
+
"vae_cache_disable": false,
|
| 17 |
+
"accelerator_cache_clear_interval": null,
|
| 18 |
+
"aspect_bucket_rounding": 2,
|
| 19 |
+
"base_model_precision": "no_change",
|
| 20 |
+
"text_encoder_1_precision": "no_change",
|
| 21 |
+
"text_encoder_2_precision": "no_change",
|
| 22 |
+
"text_encoder_3_precision": "no_change",
|
| 23 |
+
"text_encoder_4_precision": "no_change",
|
| 24 |
+
"gradient_checkpointing_interval": null,
|
| 25 |
+
"offload_during_startup": false,
|
| 26 |
+
"quantize_via": "accelerator",
|
| 27 |
+
"wan_force_2_1_time_embedding": false,
|
| 28 |
+
"fuse_qkv_projections": false,
|
| 29 |
+
"rescale_betas_zero_snr": false,
|
| 30 |
+
"control": false,
|
| 31 |
+
"controlnet_custom_config": null,
|
| 32 |
+
"controlnet_model_name_or_path": null,
|
| 33 |
+
"tread_config": null,
|
| 34 |
+
"pretrained_transformer_model_name_or_path": null,
|
| 35 |
+
"pretrained_transformer_subfolder": "transformer",
|
| 36 |
+
"pretrained_unet_model_name_or_path": null,
|
| 37 |
+
"pretrained_unet_subfolder": "unet",
|
| 38 |
+
"pretrained_t5_model_name_or_path": null,
|
| 39 |
+
"revision": null,
|
| 40 |
+
"variant": null,
|
| 41 |
+
"base_model_default_dtype": "bf16",
|
| 42 |
+
"unet_attention_slice": false,
|
| 43 |
+
"num_train_epochs": 834,
|
| 44 |
+
"max_train_steps": 5000,
|
| 45 |
+
"train_batch_size": 2,
|
| 46 |
+
"learning_rate": 0.0001,
|
| 47 |
+
"optimizer": "adamw_bf16",
|
| 48 |
+
"lr_scheduler": "constant_with_warmup",
|
| 49 |
+
"gradient_accumulation_steps": 1,
|
| 50 |
+
"lr_warmup_steps": 0,
|
| 51 |
+
"checkpoints_total_limit": 5,
|
| 52 |
+
"gradient_checkpointing": true,
|
| 53 |
+
"enable_group_offload": false,
|
| 54 |
+
"ramtorch": false,
|
| 55 |
+
"ramtorch_target_modules": null,
|
| 56 |
+
"ramtorch_text_encoder": false,
|
| 57 |
+
"ramtorch_vae": false,
|
| 58 |
+
"ramtorch_controlnet": false,
|
| 59 |
+
"group_offload_type": "block_level",
|
| 60 |
+
"group_offload_blocks_per_group": 1,
|
| 61 |
+
"group_offload_use_stream": false,
|
| 62 |
+
"group_offload_to_disk_path": "",
|
| 63 |
+
"group_offload_text_encoder": false,
|
| 64 |
+
"group_offload_vae": false,
|
| 65 |
+
"offload_during_save": false,
|
| 66 |
+
"enable_chunked_feed_forward": false,
|
| 67 |
+
"feed_forward_chunk_size": null,
|
| 68 |
+
"train_text_encoder": false,
|
| 69 |
+
"text_encoder_lr": null,
|
| 70 |
+
"lyrics_embedder_train": false,
|
| 71 |
+
"lyrics_embedder_optimizer": null,
|
| 72 |
+
"lyrics_embedder_lr": null,
|
| 73 |
+
"lyrics_embedder_lr_scheduler": null,
|
| 74 |
+
"lr_num_cycles": 1,
|
| 75 |
+
"lr_power": 0.8,
|
| 76 |
+
"use_soft_min_snr": false,
|
| 77 |
+
"use_ema": false,
|
| 78 |
+
"ema_device": "cpu",
|
| 79 |
+
"ema_cpu_only": false,
|
| 80 |
+
"ema_update_interval": 1,
|
| 81 |
+
"ema_foreach_disable": false,
|
| 82 |
+
"ema_decay": 0.995,
|
| 83 |
+
"lora_rank": 16,
|
| 84 |
+
"lora_alpha": 16.0,
|
| 85 |
+
"lora_type": "standard",
|
| 86 |
+
"lora_dropout": 0.1,
|
| 87 |
+
"lora_init_type": "default",
|
| 88 |
+
"peft_lora_mode": "standard",
|
| 89 |
+
"singlora_ramp_up_steps": 0,
|
| 90 |
+
"init_lora": null,
|
| 91 |
+
"lycoris_config": "/home/Ubuntu/.simpletuner/configs/config/lycoris_config.json",
|
| 92 |
+
"init_lokr_norm": null,
|
| 93 |
+
"flux_lora_target": "all",
|
| 94 |
+
"acestep_lora_target": "attn_qkv+linear_qkv",
|
| 95 |
+
"use_dora": false,
|
| 96 |
+
"resolution_type": "pixel_area",
|
| 97 |
+
"data_backend_config": "/home/Ubuntu/.simpletuner/configs/111/multidatabackend-bartek_mizak.json",
|
| 98 |
+
"caption_strategy": "filename",
|
| 99 |
+
"conditioning_multidataset_sampling": "random",
|
| 100 |
+
"instance_prompt": null,
|
| 101 |
+
"parquet_caption_column": null,
|
| 102 |
+
"parquet_filename_column": null,
|
| 103 |
+
"ignore_missing_files": false,
|
| 104 |
+
"vae_cache_scan_behaviour": "recreate",
|
| 105 |
+
"vae_enable_slicing": false,
|
| 106 |
+
"vae_enable_tiling": false,
|
| 107 |
+
"vae_batch_size": 4,
|
| 108 |
+
"max_upscale_threshold": null,
|
| 109 |
+
"caption_dropout_probability": 0.1,
|
| 110 |
+
"tokenizer_max_length": 512,
|
| 111 |
+
"audio_max_duration_seconds": null,
|
| 112 |
+
"audio_min_duration_seconds": null,
|
| 113 |
+
"audio_channels": 1,
|
| 114 |
+
"audio_duration_interval": 3.0,
|
| 115 |
+
"audio_truncation_mode": "beginning",
|
| 116 |
+
"validation_step_interval": 100,
|
| 117 |
+
"validation_epoch_interval": null,
|
| 118 |
+
"disable_benchmark": false,
|
| 119 |
+
"validation_preview": false,
|
| 120 |
+
"validation_preview_steps": 1,
|
| 121 |
+
"validation_prompt": "bm82 man",
|
| 122 |
+
"validation_lyrics": null,
|
| 123 |
+
"validation_audio_duration": 30.0,
|
| 124 |
+
"num_validation_images": 1,
|
| 125 |
+
"num_eval_images": 4,
|
| 126 |
+
"eval_steps_interval": null,
|
| 127 |
+
"eval_timesteps": 28,
|
| 128 |
+
"eval_dataset_pooling": false,
|
| 129 |
+
"evaluation_type": "none",
|
| 130 |
+
"pretrained_evaluation_model_name_or_path": "openai/clip-vit-large-patch14-336",
|
| 131 |
+
"validation_guidance": 7.5,
|
| 132 |
+
"validation_num_inference_steps": 20,
|
| 133 |
+
"validation_on_startup": false,
|
| 134 |
+
"validation_method": "simpletuner-local",
|
| 135 |
+
"validation_external_script": null,
|
| 136 |
+
"validation_external_background": false,
|
| 137 |
+
"validation_using_datasets": false,
|
| 138 |
+
"validation_torch_compile": false,
|
| 139 |
+
"validation_guidance_real": 1.0,
|
| 140 |
+
"validation_no_cfg_until_timestep": 2,
|
| 141 |
+
"validation_negative_prompt": "blurry, cropped, ugly",
|
| 142 |
+
"validation_randomize": false,
|
| 143 |
+
"validation_seed": null,
|
| 144 |
+
"validation_multigpu": "batch-parallel",
|
| 145 |
+
"validation_disable": false,
|
| 146 |
+
"validation_prompt_library": false,
|
| 147 |
+
"user_prompt_library": null,
|
| 148 |
+
"eval_dataset_id": null,
|
| 149 |
+
"validation_stitch_input_location": "left",
|
| 150 |
+
"validation_guidance_rescale": 0.0,
|
| 151 |
+
"validation_disable_unconditional": false,
|
| 152 |
+
"validation_guidance_skip_layers": null,
|
| 153 |
+
"validation_guidance_skip_layers_start": 0.01,
|
| 154 |
+
"validation_guidance_skip_layers_stop": 0.2,
|
| 155 |
+
"validation_guidance_skip_scale": 2.8,
|
| 156 |
+
"validation_lycoris_strength": 1.0,
|
| 157 |
+
"validation_noise_scheduler": null,
|
| 158 |
+
"validation_num_video_frames": null,
|
| 159 |
+
"validation_resolution": "1024x1024",
|
| 160 |
+
"validation_seed_source": "cpu",
|
| 161 |
+
"validation_torch_compile_mode": "max-autotune",
|
| 162 |
+
"validation_adapter_path": null,
|
| 163 |
+
"validation_adapter_name": null,
|
| 164 |
+
"validation_adapter_strength": 1.0,
|
| 165 |
+
"validation_adapter_mode": "adapter_only",
|
| 166 |
+
"validation_adapter_config": null,
|
| 167 |
+
"i_know_what_i_am_doing": false,
|
| 168 |
+
"flow_sigmoid_scale": 1.0,
|
| 169 |
+
"flux_fast_schedule": false,
|
| 170 |
+
"flow_use_uniform_schedule": false,
|
| 171 |
+
"flow_use_beta_schedule": false,
|
| 172 |
+
"flow_beta_schedule_alpha": 2.0,
|
| 173 |
+
"flow_beta_schedule_beta": 2.0,
|
| 174 |
+
"flow_schedule_shift": 3.0,
|
| 175 |
+
"flow_schedule_auto_shift": false,
|
| 176 |
+
"flux_guidance_mode": "constant",
|
| 177 |
+
"flux_attention_masked_training": false,
|
| 178 |
+
"flux_guidance_value": 1.0,
|
| 179 |
+
"flux_guidance_min": 0.0,
|
| 180 |
+
"flux_guidance_max": 4.0,
|
| 181 |
+
"t5_padding": "unmodified",
|
| 182 |
+
"sd3_clip_uncond_behaviour": "empty_string",
|
| 183 |
+
"sd3_t5_uncond_behaviour": null,
|
| 184 |
+
"soft_min_snr_sigma_data": null,
|
| 185 |
+
"mixed_precision": "bf16",
|
| 186 |
+
"attention_mechanism": "diffusers",
|
| 187 |
+
"sla_config": null,
|
| 188 |
+
"sageattention_usage": {
|
| 189 |
+
"_value_": "inference",
|
| 190 |
+
"_name_": "INFERENCE",
|
| 191 |
+
"__objclass__": "{'_generate_next_value_': <staticmethod(<function Enum._generate_next_value_ at 0x7649ca9cdf80>)>, '__module__': 'simpletuner.helpers.training.attention_backend', 'allows_training': <property object at 0x764872f93880>, 'allows_inference': <property object at 0x764872f93790>, 'from_raw': <classmethod(<function AttentionBackendMode.from_raw at 0x764872fe85e0>)>, '_new_member_': <built-in method __new__ of type object at 0xa234e0>, '_use_args_': True, '_member_names_': ['TRAINING', 'INFERENCE', 'TRAINING_AND_INFERENCE'], '_member_map_': {'TRAINING': <AttentionBackendMode.TRAINING: 'training'>, 'INFERENCE': <AttentionBackendMode.INFERENCE: 'inference'>, 'TRAINING_AND_INFERENCE': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>}, '_value2member_map_': {'training': <AttentionBackendMode.TRAINING: 'training'>, 'inference': <AttentionBackendMode.INFERENCE: 'inference'>, 'training+inference': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>}, '_unhashable_values_': [], '_member_type_': <class 'str'>, '_value_repr_': <slot wrapper '__repr__' of 'str' objects>, '__dict__': <attribute '__dict__' of 'AttentionBackendMode' objects>, '__weakref__': <attribute '__weakref__' of 'AttentionBackendMode' objects>, '__doc__': None, 'TRAINING': <AttentionBackendMode.TRAINING: 'training'>, 'INFERENCE': <AttentionBackendMode.INFERENCE: 'inference'>, 'TRAINING_AND_INFERENCE': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>, '__repr__': <function Enum.__repr__ at 0x7649ca9ce0c0>, '__str__': <function Enum.__str__ at 0x7649ca9ce160>, '__format__': <function Enum.__format__ at 0x7649ca9ce2a0>, '__new__': <function Enum.__new__ at 0x7649ca9cde40>}",
|
| 192 |
+
"_sort_order_": 1
|
| 193 |
+
},
|
| 194 |
+
"disable_tf32": false,
|
| 195 |
+
"set_grads_to_none": false,
|
| 196 |
+
"noise_offset": 0.1,
|
| 197 |
+
"noise_offset_probability": 0.25,
|
| 198 |
+
"input_perturbation": 0.0,
|
| 199 |
+
"input_perturbation_steps": 0,
|
| 200 |
+
"lr_end": "4e-7",
|
| 201 |
+
"lr_scale": false,
|
| 202 |
+
"lr_scale_sqrt": false,
|
| 203 |
+
"ignore_final_epochs": false,
|
| 204 |
+
"freeze_encoder_before": 12,
|
| 205 |
+
"freeze_encoder_after": 17,
|
| 206 |
+
"freeze_encoder_strategy": "after",
|
| 207 |
+
"layer_freeze_strategy": null,
|
| 208 |
+
"fully_unload_text_encoder": false,
|
| 209 |
+
"save_text_encoder": false,
|
| 210 |
+
"text_encoder_limit": 100,
|
| 211 |
+
"prepend_instance_prompt": false,
|
| 212 |
+
"only_instance_prompt": false,
|
| 213 |
+
"data_aesthetic_score": 7.0,
|
| 214 |
+
"delete_unwanted_images": false,
|
| 215 |
+
"delete_problematic_images": false,
|
| 216 |
+
"disable_bucket_pruning": false,
|
| 217 |
+
"allow_dataset_oversubscription": false,
|
| 218 |
+
"disable_segmented_timestep_sampling": false,
|
| 219 |
+
"preserve_data_backend_cache": false,
|
| 220 |
+
"override_dataset_config": false,
|
| 221 |
+
"cache_dir": "/home/Ubuntu/.simpletuner/output/cache",
|
| 222 |
+
"cache_dir_text": "cache",
|
| 223 |
+
"cache_dir_vae": "",
|
| 224 |
+
"compress_disk_cache": true,
|
| 225 |
+
"aspect_bucket_disable_rebuild": false,
|
| 226 |
+
"keep_vae_loaded": false,
|
| 227 |
+
"skip_file_discovery": "",
|
| 228 |
+
"data_backend_sampling": "auto-weighting",
|
| 229 |
+
"image_processing_batch_size": 32,
|
| 230 |
+
"write_batch_size": 128,
|
| 231 |
+
"read_batch_size": 25,
|
| 232 |
+
"enable_multiprocessing": false,
|
| 233 |
+
"accelerate_config": null,
|
| 234 |
+
"deepspeed_config": null,
|
| 235 |
+
"fsdp_enable": false,
|
| 236 |
+
"fsdp_version": 2,
|
| 237 |
+
"fsdp_reshard_after_forward": false,
|
| 238 |
+
"fsdp_state_dict_type": "SHARDED_STATE_DICT",
|
| 239 |
+
"fsdp_cpu_ram_efficient_loading": false,
|
| 240 |
+
"fsdp_auto_wrap_policy": "TRANSFORMER_BASED_WRAP",
|
| 241 |
+
"fsdp_limit_all_gathers": false,
|
| 242 |
+
"fsdp_cpu_offload": false,
|
| 243 |
+
"fsdp_activation_checkpointing": false,
|
| 244 |
+
"fsdp_transformer_layer_cls_to_wrap": null,
|
| 245 |
+
"context_parallel_size": 1,
|
| 246 |
+
"context_parallel_comm_strategy": "allgather",
|
| 247 |
+
"num_processes": 1,
|
| 248 |
+
"num_machines": 1,
|
| 249 |
+
"accelerate_extra_args": null,
|
| 250 |
+
"main_process_ip": "127.0.0.1",
|
| 251 |
+
"main_process_port": 29500,
|
| 252 |
+
"machine_rank": 0,
|
| 253 |
+
"same_network": true,
|
| 254 |
+
"dynamo_backend": "no",
|
| 255 |
+
"dynamo_mode": "",
|
| 256 |
+
"dynamo_fullgraph": false,
|
| 257 |
+
"dynamo_dynamic": false,
|
| 258 |
+
"dynamo_use_regional_compilation": false,
|
| 259 |
+
"max_workers": 32,
|
| 260 |
+
"aws_max_pool_connections": 128,
|
| 261 |
+
"torch_num_threads": 8,
|
| 262 |
+
"dataloader_prefetch": false,
|
| 263 |
+
"dataloader_prefetch_qlen": 10,
|
| 264 |
+
"aspect_bucket_worker_count": 12,
|
| 265 |
+
"aspect_bucket_alignment": 16,
|
| 266 |
+
"minimum_image_size": null,
|
| 267 |
+
"maximum_image_size": null,
|
| 268 |
+
"target_downsample_size": null,
|
| 269 |
+
"metadata_update_interval": 3600,
|
| 270 |
+
"debug_aspect_buckets": false,
|
| 271 |
+
"debug_dataset_loader": false,
|
| 272 |
+
"print_filenames": false,
|
| 273 |
+
"print_sampler_statistics": false,
|
| 274 |
+
"timestep_bias_strategy": null,
|
| 275 |
+
"timestep_bias_begin": 0,
|
| 276 |
+
"timestep_bias_end": 1000,
|
| 277 |
+
"timestep_bias_multiplier": 1.0,
|
| 278 |
+
"timestep_bias_portion": 0.25,
|
| 279 |
+
"training_scheduler_timestep_spacing": "trailing",
|
| 280 |
+
"inference_scheduler_timestep_spacing": "trailing",
|
| 281 |
+
"loss_type": "l2",
|
| 282 |
+
"huber_schedule": "snr",
|
| 283 |
+
"huber_c": 0.1,
|
| 284 |
+
"snr_gamma": null,
|
| 285 |
+
"masked_loss_probability": 1.0,
|
| 286 |
+
"hidream_use_load_balancing_loss": false,
|
| 287 |
+
"hidream_load_balancing_loss_weight": null,
|
| 288 |
+
"adam_beta1": 0.9,
|
| 289 |
+
"adam_beta2": 0.999,
|
| 290 |
+
"optimizer_beta1": null,
|
| 291 |
+
"optimizer_beta2": null,
|
| 292 |
+
"optimizer_cpu_offload_method": null,
|
| 293 |
+
"gradient_precision": null,
|
| 294 |
+
"adam_weight_decay": 0.01,
|
| 295 |
+
"adam_epsilon": 1e-08,
|
| 296 |
+
"prodigy_steps": null,
|
| 297 |
+
"max_grad_norm": 2.0,
|
| 298 |
+
"optimizer_config": null,
|
| 299 |
+
"grad_clip_method": "value",
|
| 300 |
+
"optimizer_offload_gradients": false,
|
| 301 |
+
"fuse_optimizer": false,
|
| 302 |
+
"optimizer_release_gradients": false,
|
| 303 |
+
"push_to_hub": false,
|
| 304 |
+
"publishing_config": null,
|
| 305 |
+
"post_upload_script": null,
|
| 306 |
+
"push_checkpoints_to_hub": false,
|
| 307 |
+
"push_to_hub_background": false,
|
| 308 |
+
"hub_model_id": "quzo/fl2",
|
| 309 |
+
"model_card_private": false,
|
| 310 |
+
"model_card_safe_for_work": false,
|
| 311 |
+
"model_card_note": null,
|
| 312 |
+
"report_to": "none",
|
| 313 |
+
"custom_tracker": null,
|
| 314 |
+
"checkpoint_step_interval": 200,
|
| 315 |
+
"checkpoint_epoch_interval": null,
|
| 316 |
+
"checkpointing_rolling_steps": 0,
|
| 317 |
+
"checkpointing_use_tempdir": false,
|
| 318 |
+
"checkpoints_rolling_total_limit": 1,
|
| 319 |
+
"tracker_run_name": "simpletuner-testing",
|
| 320 |
+
"tracker_project_name": "simpletuner",
|
| 321 |
+
"tracker_image_layout": "gallery",
|
| 322 |
+
"enable_watermark": false,
|
| 323 |
+
"framerate": null,
|
| 324 |
+
"seed_for_each_device": true,
|
| 325 |
+
"snr_weight": 1.0,
|
| 326 |
+
"webhook_config": [
|
| 327 |
+
{
|
| 328 |
+
"webhook_type": "raw",
|
| 329 |
+
"callback_url": "http://0.0.0.0:8001/callback",
|
| 330 |
+
"log_level": "info",
|
| 331 |
+
"ssl_no_verify": false
|
| 332 |
+
}
|
| 333 |
+
],
|
| 334 |
+
"webhook_reporting_interval": null,
|
| 335 |
+
"distillation_method": null,
|
| 336 |
+
"distillation_config": null,
|
| 337 |
+
"ema_validation": "comparison",
|
| 338 |
+
"local_rank": -1,
|
| 339 |
+
"ltx_train_mode": "i2v",
|
| 340 |
+
"ltx_i2v_prob": 0.1,
|
| 341 |
+
"ltx_partial_noise_fraction": 0.05,
|
| 342 |
+
"ltx_protect_first_frame": false,
|
| 343 |
+
"offload_param_path": null,
|
| 344 |
+
"offset_noise": false,
|
| 345 |
+
"quantize_activations": false,
|
| 346 |
+
"refiner_training": false,
|
| 347 |
+
"refiner_training_invert_schedule": false,
|
| 348 |
+
"refiner_training_strength": 0.2,
|
| 349 |
+
"sdxl_refiner_uses_full_range": false,
|
| 350 |
+
"sana_complex_human_instruction": "Given a user prompt, generate an 'Enhanced prompt' that provides detailed visual descriptions suitable for image generation. Evaluate the level of detail in the user prompt:\n- If the prompt is simple, focus on adding specifics about colors, shapes, sizes, textures, and spatial relationships to create vivid and concrete scenes.\n- Do not add detail that is not in the original prompt! Do not mention colors if the User Prompt does not!\n- If the prompt is already detailed, refine and enhance the existing details slightly without overcomplicating.\n- If the prompt contains explicit material, do not remove this information from the output, which would damage the performance of downstream classifiers.\nHere are examples of how to transform or refine prompts:\n- User Prompt: A cat sleeping -> Enhanced: A cat sleeping peacefully, showcasing the joy of pet ownership. Cute floof kitty cat gatto.\n- User Prompt: A busy city street -> Enhanced: A bustling city street scene featuring a crowd of people.\nPlease generate only the enhanced description for the prompt below and avoid including any additional commentary or evaluations:\nUser Prompt: ",
|
| 351 |
+
"vae_path": "black-forest-labs/FLUX.2-dev",
|
| 352 |
+
"accelerator_project_config": {
|
| 353 |
+
"project_dir": "/home/Ubuntu/.simpletuner/output",
|
| 354 |
+
"logging_dir": "/home/Ubuntu/.simpletuner/output/logs",
|
| 355 |
+
"automatic_checkpoint_naming": false,
|
| 356 |
+
"total_limit": null,
|
| 357 |
+
"iteration": 17,
|
| 358 |
+
"save_on_each_node": false
|
| 359 |
+
},
|
| 360 |
+
"process_group_kwargs": {
|
| 361 |
+
"backend": "nccl",
|
| 362 |
+
"init_method": null,
|
| 363 |
+
"timeout": "1:30:00"
|
| 364 |
+
},
|
| 365 |
+
"is_quantized": false,
|
| 366 |
+
"weight_dtype": "torch.bfloat16",
|
| 367 |
+
"disable_accelerator": false,
|
| 368 |
+
"lora_initialisation_style": true,
|
| 369 |
+
"checkpointing_steps": 200,
|
| 370 |
+
"use_fsdp": false,
|
| 371 |
+
"model_type_label": "Flux.2",
|
| 372 |
+
"use_deepspeed_optimizer": false,
|
| 373 |
+
"use_deepspeed_scheduler": false,
|
| 374 |
+
"base_weight_dtype": "torch.bfloat16",
|
| 375 |
+
"is_quanto": false,
|
| 376 |
+
"is_torchao": false,
|
| 377 |
+
"is_bnb": false,
|
| 378 |
+
"flow_matching": true,
|
| 379 |
+
"overrode_max_train_steps": false,
|
| 380 |
+
"total_num_batches": 6,
|
| 381 |
+
"num_update_steps_per_epoch": 6,
|
| 382 |
+
"total_batch_size": 2,
|
| 383 |
+
"enable_adamw_bf16": true,
|
| 384 |
+
"is_schedulefree": false,
|
| 385 |
+
"is_lr_scheduler_disabled": false,
|
| 386 |
+
"total_steps_remaining_at_start": 5000
|
| 387 |
+
}
|
checkpoint-3400/training_state-training-images.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"aspect_ratio_bucket_indices": {"1.0": ["/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im008.png", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im021.png"], "0.74": ["/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0014.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9743.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9827.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9935.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0003.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9796.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9890.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9865.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9260.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0090.jpg"]}, "buckets": ["0.74"], "exhausted_buckets": ["1.0"], "batch_size": 2, "current_bucket": 0, "seen_images": {"/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im021.png": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im008.png": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9935.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9890.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0003.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9743.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9827.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9260.jpg": true}, "current_epoch": 567}
|
checkpoint-3400/training_state.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"global_step": 3400, "epoch_step": 3966, "epoch": 567, "exhausted_backends": [], "repeats": {"training-images": 0}}
|
checkpoint-3600/README.md
ADDED
|
@@ -0,0 +1,129 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: other
|
| 3 |
+
base_model: "black-forest-labs/FLUX.2-dev"
|
| 4 |
+
tags:
|
| 5 |
+
- flux2
|
| 6 |
+
- flux2-diffusers
|
| 7 |
+
- text-to-image
|
| 8 |
+
- image-to-image
|
| 9 |
+
- diffusers
|
| 10 |
+
- simpletuner
|
| 11 |
+
- not-for-all-audiences
|
| 12 |
+
- lora
|
| 13 |
+
|
| 14 |
+
- template:sd-lora
|
| 15 |
+
- standard
|
| 16 |
+
pipeline_tag: text-to-image
|
| 17 |
+
inference: true
|
| 18 |
+
|
| 19 |
+
---
|
| 20 |
+
|
| 21 |
+
# quzo/fl2
|
| 22 |
+
|
| 23 |
+
This is a PEFT LoRA derived from [black-forest-labs/FLUX.2-dev](https://huggingface.co/black-forest-labs/FLUX.2-dev).
|
| 24 |
+
|
| 25 |
+
The main validation prompt used during training was:
|
| 26 |
+
```
|
| 27 |
+
bm82 man
|
| 28 |
+
```
|
| 29 |
+
|
| 30 |
+
|
| 31 |
+
## Validation settings
|
| 32 |
+
- CFG: `7.5`
|
| 33 |
+
- CFG Rescale: `0.0`
|
| 34 |
+
- Steps: `20`
|
| 35 |
+
- Sampler: `FlowMatchEulerDiscreteScheduler`
|
| 36 |
+
- Seed: `None`
|
| 37 |
+
- Resolution: `1024x1024`
|
| 38 |
+
|
| 39 |
+
|
| 40 |
+
Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
|
| 41 |
+
|
| 42 |
+
|
| 43 |
+
|
| 44 |
+
|
| 45 |
+
<Gallery />
|
| 46 |
+
|
| 47 |
+
The text encoder **was not** trained.
|
| 48 |
+
You may reuse the base model text encoder for inference.
|
| 49 |
+
|
| 50 |
+
|
| 51 |
+
## Training settings
|
| 52 |
+
|
| 53 |
+
- Training epochs: 599
|
| 54 |
+
- Training steps: 3600
|
| 55 |
+
- Learning rate: 0.0001
|
| 56 |
+
- Learning rate schedule: constant_with_warmup
|
| 57 |
+
- Warmup steps: 0
|
| 58 |
+
- Max grad value: 2.0
|
| 59 |
+
- Effective batch size: 2
|
| 60 |
+
- Micro-batch size: 2
|
| 61 |
+
- Gradient accumulation steps: 1
|
| 62 |
+
- Number of GPUs: 1
|
| 63 |
+
- Gradient checkpointing: True
|
| 64 |
+
- Prediction type: flow_matching[]
|
| 65 |
+
- Optimizer: adamw_bf16
|
| 66 |
+
- Trainable parameter precision: Pure BF16
|
| 67 |
+
- Base model precision: `no_change`
|
| 68 |
+
- Caption dropout probability: 0.1%
|
| 69 |
+
|
| 70 |
+
|
| 71 |
+
|
| 72 |
+
- LoRA Rank: 16
|
| 73 |
+
- LoRA Alpha: 16.0
|
| 74 |
+
- LoRA Dropout: 0.1
|
| 75 |
+
- LoRA initialisation style: default
|
| 76 |
+
- LoRA mode: Standard
|
| 77 |
+
|
| 78 |
+
|
| 79 |
+
## Datasets
|
| 80 |
+
|
| 81 |
+
### training-images
|
| 82 |
+
- Repeats: 0
|
| 83 |
+
- Total number of images: 12
|
| 84 |
+
- Total number of aspect buckets: 1
|
| 85 |
+
- Resolution: 1.048576 megapixels
|
| 86 |
+
- Cropped: False
|
| 87 |
+
- Crop style: None
|
| 88 |
+
- Crop aspect: None
|
| 89 |
+
- Used for regularisation data: No
|
| 90 |
+
|
| 91 |
+
|
| 92 |
+
## Inference
|
| 93 |
+
|
| 94 |
+
|
| 95 |
+
```python
|
| 96 |
+
import torch
|
| 97 |
+
from diffusers import DiffusionPipeline
|
| 98 |
+
|
| 99 |
+
model_id = 'black-forest-labs/FLUX.2-dev'
|
| 100 |
+
adapter_id = 'quzo/fl2'
|
| 101 |
+
pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
|
| 102 |
+
pipeline.load_lora_weights(adapter_id)
|
| 103 |
+
|
| 104 |
+
prompt = "bm82 man"
|
| 105 |
+
negative_prompt = 'blurry, cropped, ugly'
|
| 106 |
+
|
| 107 |
+
## Optional: quantise the model to save on vram.
|
| 108 |
+
## Note: The model was not quantised during training, so it is not necessary to quantise it during inference time.
|
| 109 |
+
#from optimum.quanto import quantize, freeze, qint8
|
| 110 |
+
#quantize(pipeline.transformer, weights=qint8)
|
| 111 |
+
#freeze(pipeline.transformer)
|
| 112 |
+
|
| 113 |
+
pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
|
| 114 |
+
model_output = pipeline(
|
| 115 |
+
prompt=prompt,
|
| 116 |
+
negative_prompt=negative_prompt,
|
| 117 |
+
num_inference_steps=20,
|
| 118 |
+
generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
|
| 119 |
+
width=1024,
|
| 120 |
+
height=1024,
|
| 121 |
+
guidance_scale=7.5,
|
| 122 |
+
).images[0]
|
| 123 |
+
|
| 124 |
+
model_output.save("output.png", format="PNG")
|
| 125 |
+
|
| 126 |
+
```
|
| 127 |
+
|
| 128 |
+
|
| 129 |
+
|
checkpoint-3600/optimizer.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f020ed312a541740949018b668e757324f91d58a2c4e8cf7be9b56b4b7ce1cb9
|
| 3 |
+
size 321008843
|
checkpoint-3600/pytorch_lora_weights.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f5daaa3075a66bf8aad8942dfd83df8683047f13a9fbdd8cbd35e3090e8789f7
|
| 3 |
+
size 106979136
|
checkpoint-3600/random_states_0.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2890a0caf51b59c66e1dac8a4761f5003fbe092a9bd8d548920ac59abfae98c5
|
| 3 |
+
size 14693
|
checkpoint-3600/scheduler.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6322143ad2efc56c6c09c3066b83311a4b8987351a4fe633c0f1c92f289e9d4f
|
| 3 |
+
size 1401
|
checkpoint-3600/simpletuner_config.json
ADDED
|
@@ -0,0 +1,387 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"model_family": "flux2",
|
| 3 |
+
"model_flavour": "dev",
|
| 4 |
+
"controlnet": false,
|
| 5 |
+
"pretrained_model_name_or_path": "black-forest-labs/FLUX.2-dev",
|
| 6 |
+
"output_dir": "/home/Ubuntu/.simpletuner/output",
|
| 7 |
+
"logging_dir": "/home/Ubuntu/.simpletuner/output/logs",
|
| 8 |
+
"model_type": "lora",
|
| 9 |
+
"seed": null,
|
| 10 |
+
"resolution": 1024,
|
| 11 |
+
"resume_from_checkpoint": null,
|
| 12 |
+
"prediction_type": null,
|
| 13 |
+
"pretrained_vae_model_name_or_path": "black-forest-labs/FLUX.2-dev",
|
| 14 |
+
"vae_dtype": "bf16",
|
| 15 |
+
"vae_cache_ondemand": false,
|
| 16 |
+
"vae_cache_disable": false,
|
| 17 |
+
"accelerator_cache_clear_interval": null,
|
| 18 |
+
"aspect_bucket_rounding": 2,
|
| 19 |
+
"base_model_precision": "no_change",
|
| 20 |
+
"text_encoder_1_precision": "no_change",
|
| 21 |
+
"text_encoder_2_precision": "no_change",
|
| 22 |
+
"text_encoder_3_precision": "no_change",
|
| 23 |
+
"text_encoder_4_precision": "no_change",
|
| 24 |
+
"gradient_checkpointing_interval": null,
|
| 25 |
+
"offload_during_startup": false,
|
| 26 |
+
"quantize_via": "accelerator",
|
| 27 |
+
"wan_force_2_1_time_embedding": false,
|
| 28 |
+
"fuse_qkv_projections": false,
|
| 29 |
+
"rescale_betas_zero_snr": false,
|
| 30 |
+
"control": false,
|
| 31 |
+
"controlnet_custom_config": null,
|
| 32 |
+
"controlnet_model_name_or_path": null,
|
| 33 |
+
"tread_config": null,
|
| 34 |
+
"pretrained_transformer_model_name_or_path": null,
|
| 35 |
+
"pretrained_transformer_subfolder": "transformer",
|
| 36 |
+
"pretrained_unet_model_name_or_path": null,
|
| 37 |
+
"pretrained_unet_subfolder": "unet",
|
| 38 |
+
"pretrained_t5_model_name_or_path": null,
|
| 39 |
+
"revision": null,
|
| 40 |
+
"variant": null,
|
| 41 |
+
"base_model_default_dtype": "bf16",
|
| 42 |
+
"unet_attention_slice": false,
|
| 43 |
+
"num_train_epochs": 834,
|
| 44 |
+
"max_train_steps": 5000,
|
| 45 |
+
"train_batch_size": 2,
|
| 46 |
+
"learning_rate": 0.0001,
|
| 47 |
+
"optimizer": "adamw_bf16",
|
| 48 |
+
"lr_scheduler": "constant_with_warmup",
|
| 49 |
+
"gradient_accumulation_steps": 1,
|
| 50 |
+
"lr_warmup_steps": 0,
|
| 51 |
+
"checkpoints_total_limit": 5,
|
| 52 |
+
"gradient_checkpointing": true,
|
| 53 |
+
"enable_group_offload": false,
|
| 54 |
+
"ramtorch": false,
|
| 55 |
+
"ramtorch_target_modules": null,
|
| 56 |
+
"ramtorch_text_encoder": false,
|
| 57 |
+
"ramtorch_vae": false,
|
| 58 |
+
"ramtorch_controlnet": false,
|
| 59 |
+
"group_offload_type": "block_level",
|
| 60 |
+
"group_offload_blocks_per_group": 1,
|
| 61 |
+
"group_offload_use_stream": false,
|
| 62 |
+
"group_offload_to_disk_path": "",
|
| 63 |
+
"group_offload_text_encoder": false,
|
| 64 |
+
"group_offload_vae": false,
|
| 65 |
+
"offload_during_save": false,
|
| 66 |
+
"enable_chunked_feed_forward": false,
|
| 67 |
+
"feed_forward_chunk_size": null,
|
| 68 |
+
"train_text_encoder": false,
|
| 69 |
+
"text_encoder_lr": null,
|
| 70 |
+
"lyrics_embedder_train": false,
|
| 71 |
+
"lyrics_embedder_optimizer": null,
|
| 72 |
+
"lyrics_embedder_lr": null,
|
| 73 |
+
"lyrics_embedder_lr_scheduler": null,
|
| 74 |
+
"lr_num_cycles": 1,
|
| 75 |
+
"lr_power": 0.8,
|
| 76 |
+
"use_soft_min_snr": false,
|
| 77 |
+
"use_ema": false,
|
| 78 |
+
"ema_device": "cpu",
|
| 79 |
+
"ema_cpu_only": false,
|
| 80 |
+
"ema_update_interval": 1,
|
| 81 |
+
"ema_foreach_disable": false,
|
| 82 |
+
"ema_decay": 0.995,
|
| 83 |
+
"lora_rank": 16,
|
| 84 |
+
"lora_alpha": 16.0,
|
| 85 |
+
"lora_type": "standard",
|
| 86 |
+
"lora_dropout": 0.1,
|
| 87 |
+
"lora_init_type": "default",
|
| 88 |
+
"peft_lora_mode": "standard",
|
| 89 |
+
"singlora_ramp_up_steps": 0,
|
| 90 |
+
"init_lora": null,
|
| 91 |
+
"lycoris_config": "/home/Ubuntu/.simpletuner/configs/config/lycoris_config.json",
|
| 92 |
+
"init_lokr_norm": null,
|
| 93 |
+
"flux_lora_target": "all",
|
| 94 |
+
"acestep_lora_target": "attn_qkv+linear_qkv",
|
| 95 |
+
"use_dora": false,
|
| 96 |
+
"resolution_type": "pixel_area",
|
| 97 |
+
"data_backend_config": "/home/Ubuntu/.simpletuner/configs/111/multidatabackend-bartek_mizak.json",
|
| 98 |
+
"caption_strategy": "filename",
|
| 99 |
+
"conditioning_multidataset_sampling": "random",
|
| 100 |
+
"instance_prompt": null,
|
| 101 |
+
"parquet_caption_column": null,
|
| 102 |
+
"parquet_filename_column": null,
|
| 103 |
+
"ignore_missing_files": false,
|
| 104 |
+
"vae_cache_scan_behaviour": "recreate",
|
| 105 |
+
"vae_enable_slicing": false,
|
| 106 |
+
"vae_enable_tiling": false,
|
| 107 |
+
"vae_batch_size": 4,
|
| 108 |
+
"max_upscale_threshold": null,
|
| 109 |
+
"caption_dropout_probability": 0.1,
|
| 110 |
+
"tokenizer_max_length": 512,
|
| 111 |
+
"audio_max_duration_seconds": null,
|
| 112 |
+
"audio_min_duration_seconds": null,
|
| 113 |
+
"audio_channels": 1,
|
| 114 |
+
"audio_duration_interval": 3.0,
|
| 115 |
+
"audio_truncation_mode": "beginning",
|
| 116 |
+
"validation_step_interval": 100,
|
| 117 |
+
"validation_epoch_interval": null,
|
| 118 |
+
"disable_benchmark": false,
|
| 119 |
+
"validation_preview": false,
|
| 120 |
+
"validation_preview_steps": 1,
|
| 121 |
+
"validation_prompt": "bm82 man",
|
| 122 |
+
"validation_lyrics": null,
|
| 123 |
+
"validation_audio_duration": 30.0,
|
| 124 |
+
"num_validation_images": 1,
|
| 125 |
+
"num_eval_images": 4,
|
| 126 |
+
"eval_steps_interval": null,
|
| 127 |
+
"eval_timesteps": 28,
|
| 128 |
+
"eval_dataset_pooling": false,
|
| 129 |
+
"evaluation_type": "none",
|
| 130 |
+
"pretrained_evaluation_model_name_or_path": "openai/clip-vit-large-patch14-336",
|
| 131 |
+
"validation_guidance": 7.5,
|
| 132 |
+
"validation_num_inference_steps": 20,
|
| 133 |
+
"validation_on_startup": false,
|
| 134 |
+
"validation_method": "simpletuner-local",
|
| 135 |
+
"validation_external_script": null,
|
| 136 |
+
"validation_external_background": false,
|
| 137 |
+
"validation_using_datasets": false,
|
| 138 |
+
"validation_torch_compile": false,
|
| 139 |
+
"validation_guidance_real": 1.0,
|
| 140 |
+
"validation_no_cfg_until_timestep": 2,
|
| 141 |
+
"validation_negative_prompt": "blurry, cropped, ugly",
|
| 142 |
+
"validation_randomize": false,
|
| 143 |
+
"validation_seed": null,
|
| 144 |
+
"validation_multigpu": "batch-parallel",
|
| 145 |
+
"validation_disable": false,
|
| 146 |
+
"validation_prompt_library": false,
|
| 147 |
+
"user_prompt_library": null,
|
| 148 |
+
"eval_dataset_id": null,
|
| 149 |
+
"validation_stitch_input_location": "left",
|
| 150 |
+
"validation_guidance_rescale": 0.0,
|
| 151 |
+
"validation_disable_unconditional": false,
|
| 152 |
+
"validation_guidance_skip_layers": null,
|
| 153 |
+
"validation_guidance_skip_layers_start": 0.01,
|
| 154 |
+
"validation_guidance_skip_layers_stop": 0.2,
|
| 155 |
+
"validation_guidance_skip_scale": 2.8,
|
| 156 |
+
"validation_lycoris_strength": 1.0,
|
| 157 |
+
"validation_noise_scheduler": null,
|
| 158 |
+
"validation_num_video_frames": null,
|
| 159 |
+
"validation_resolution": "1024x1024",
|
| 160 |
+
"validation_seed_source": "cpu",
|
| 161 |
+
"validation_torch_compile_mode": "max-autotune",
|
| 162 |
+
"validation_adapter_path": null,
|
| 163 |
+
"validation_adapter_name": null,
|
| 164 |
+
"validation_adapter_strength": 1.0,
|
| 165 |
+
"validation_adapter_mode": "adapter_only",
|
| 166 |
+
"validation_adapter_config": null,
|
| 167 |
+
"i_know_what_i_am_doing": false,
|
| 168 |
+
"flow_sigmoid_scale": 1.0,
|
| 169 |
+
"flux_fast_schedule": false,
|
| 170 |
+
"flow_use_uniform_schedule": false,
|
| 171 |
+
"flow_use_beta_schedule": false,
|
| 172 |
+
"flow_beta_schedule_alpha": 2.0,
|
| 173 |
+
"flow_beta_schedule_beta": 2.0,
|
| 174 |
+
"flow_schedule_shift": 3.0,
|
| 175 |
+
"flow_schedule_auto_shift": false,
|
| 176 |
+
"flux_guidance_mode": "constant",
|
| 177 |
+
"flux_attention_masked_training": false,
|
| 178 |
+
"flux_guidance_value": 1.0,
|
| 179 |
+
"flux_guidance_min": 0.0,
|
| 180 |
+
"flux_guidance_max": 4.0,
|
| 181 |
+
"t5_padding": "unmodified",
|
| 182 |
+
"sd3_clip_uncond_behaviour": "empty_string",
|
| 183 |
+
"sd3_t5_uncond_behaviour": null,
|
| 184 |
+
"soft_min_snr_sigma_data": null,
|
| 185 |
+
"mixed_precision": "bf16",
|
| 186 |
+
"attention_mechanism": "diffusers",
|
| 187 |
+
"sla_config": null,
|
| 188 |
+
"sageattention_usage": {
|
| 189 |
+
"_value_": "inference",
|
| 190 |
+
"_name_": "INFERENCE",
|
| 191 |
+
"__objclass__": "{'_generate_next_value_': <staticmethod(<function Enum._generate_next_value_ at 0x7649ca9cdf80>)>, '__module__': 'simpletuner.helpers.training.attention_backend', 'allows_training': <property object at 0x764872f93880>, 'allows_inference': <property object at 0x764872f93790>, 'from_raw': <classmethod(<function AttentionBackendMode.from_raw at 0x764872fe85e0>)>, '_new_member_': <built-in method __new__ of type object at 0xa234e0>, '_use_args_': True, '_member_names_': ['TRAINING', 'INFERENCE', 'TRAINING_AND_INFERENCE'], '_member_map_': {'TRAINING': <AttentionBackendMode.TRAINING: 'training'>, 'INFERENCE': <AttentionBackendMode.INFERENCE: 'inference'>, 'TRAINING_AND_INFERENCE': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>}, '_value2member_map_': {'training': <AttentionBackendMode.TRAINING: 'training'>, 'inference': <AttentionBackendMode.INFERENCE: 'inference'>, 'training+inference': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>}, '_unhashable_values_': [], '_member_type_': <class 'str'>, '_value_repr_': <slot wrapper '__repr__' of 'str' objects>, '__dict__': <attribute '__dict__' of 'AttentionBackendMode' objects>, '__weakref__': <attribute '__weakref__' of 'AttentionBackendMode' objects>, '__doc__': None, 'TRAINING': <AttentionBackendMode.TRAINING: 'training'>, 'INFERENCE': <AttentionBackendMode.INFERENCE: 'inference'>, 'TRAINING_AND_INFERENCE': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>, '__repr__': <function Enum.__repr__ at 0x7649ca9ce0c0>, '__str__': <function Enum.__str__ at 0x7649ca9ce160>, '__format__': <function Enum.__format__ at 0x7649ca9ce2a0>, '__new__': <function Enum.__new__ at 0x7649ca9cde40>}",
|
| 192 |
+
"_sort_order_": 1
|
| 193 |
+
},
|
| 194 |
+
"disable_tf32": false,
|
| 195 |
+
"set_grads_to_none": false,
|
| 196 |
+
"noise_offset": 0.1,
|
| 197 |
+
"noise_offset_probability": 0.25,
|
| 198 |
+
"input_perturbation": 0.0,
|
| 199 |
+
"input_perturbation_steps": 0,
|
| 200 |
+
"lr_end": "4e-7",
|
| 201 |
+
"lr_scale": false,
|
| 202 |
+
"lr_scale_sqrt": false,
|
| 203 |
+
"ignore_final_epochs": false,
|
| 204 |
+
"freeze_encoder_before": 12,
|
| 205 |
+
"freeze_encoder_after": 17,
|
| 206 |
+
"freeze_encoder_strategy": "after",
|
| 207 |
+
"layer_freeze_strategy": null,
|
| 208 |
+
"fully_unload_text_encoder": false,
|
| 209 |
+
"save_text_encoder": false,
|
| 210 |
+
"text_encoder_limit": 100,
|
| 211 |
+
"prepend_instance_prompt": false,
|
| 212 |
+
"only_instance_prompt": false,
|
| 213 |
+
"data_aesthetic_score": 7.0,
|
| 214 |
+
"delete_unwanted_images": false,
|
| 215 |
+
"delete_problematic_images": false,
|
| 216 |
+
"disable_bucket_pruning": false,
|
| 217 |
+
"allow_dataset_oversubscription": false,
|
| 218 |
+
"disable_segmented_timestep_sampling": false,
|
| 219 |
+
"preserve_data_backend_cache": false,
|
| 220 |
+
"override_dataset_config": false,
|
| 221 |
+
"cache_dir": "/home/Ubuntu/.simpletuner/output/cache",
|
| 222 |
+
"cache_dir_text": "cache",
|
| 223 |
+
"cache_dir_vae": "",
|
| 224 |
+
"compress_disk_cache": true,
|
| 225 |
+
"aspect_bucket_disable_rebuild": false,
|
| 226 |
+
"keep_vae_loaded": false,
|
| 227 |
+
"skip_file_discovery": "",
|
| 228 |
+
"data_backend_sampling": "auto-weighting",
|
| 229 |
+
"image_processing_batch_size": 32,
|
| 230 |
+
"write_batch_size": 128,
|
| 231 |
+
"read_batch_size": 25,
|
| 232 |
+
"enable_multiprocessing": false,
|
| 233 |
+
"accelerate_config": null,
|
| 234 |
+
"deepspeed_config": null,
|
| 235 |
+
"fsdp_enable": false,
|
| 236 |
+
"fsdp_version": 2,
|
| 237 |
+
"fsdp_reshard_after_forward": false,
|
| 238 |
+
"fsdp_state_dict_type": "SHARDED_STATE_DICT",
|
| 239 |
+
"fsdp_cpu_ram_efficient_loading": false,
|
| 240 |
+
"fsdp_auto_wrap_policy": "TRANSFORMER_BASED_WRAP",
|
| 241 |
+
"fsdp_limit_all_gathers": false,
|
| 242 |
+
"fsdp_cpu_offload": false,
|
| 243 |
+
"fsdp_activation_checkpointing": false,
|
| 244 |
+
"fsdp_transformer_layer_cls_to_wrap": null,
|
| 245 |
+
"context_parallel_size": 1,
|
| 246 |
+
"context_parallel_comm_strategy": "allgather",
|
| 247 |
+
"num_processes": 1,
|
| 248 |
+
"num_machines": 1,
|
| 249 |
+
"accelerate_extra_args": null,
|
| 250 |
+
"main_process_ip": "127.0.0.1",
|
| 251 |
+
"main_process_port": 29500,
|
| 252 |
+
"machine_rank": 0,
|
| 253 |
+
"same_network": true,
|
| 254 |
+
"dynamo_backend": "no",
|
| 255 |
+
"dynamo_mode": "",
|
| 256 |
+
"dynamo_fullgraph": false,
|
| 257 |
+
"dynamo_dynamic": false,
|
| 258 |
+
"dynamo_use_regional_compilation": false,
|
| 259 |
+
"max_workers": 32,
|
| 260 |
+
"aws_max_pool_connections": 128,
|
| 261 |
+
"torch_num_threads": 8,
|
| 262 |
+
"dataloader_prefetch": false,
|
| 263 |
+
"dataloader_prefetch_qlen": 10,
|
| 264 |
+
"aspect_bucket_worker_count": 12,
|
| 265 |
+
"aspect_bucket_alignment": 16,
|
| 266 |
+
"minimum_image_size": null,
|
| 267 |
+
"maximum_image_size": null,
|
| 268 |
+
"target_downsample_size": null,
|
| 269 |
+
"metadata_update_interval": 3600,
|
| 270 |
+
"debug_aspect_buckets": false,
|
| 271 |
+
"debug_dataset_loader": false,
|
| 272 |
+
"print_filenames": false,
|
| 273 |
+
"print_sampler_statistics": false,
|
| 274 |
+
"timestep_bias_strategy": null,
|
| 275 |
+
"timestep_bias_begin": 0,
|
| 276 |
+
"timestep_bias_end": 1000,
|
| 277 |
+
"timestep_bias_multiplier": 1.0,
|
| 278 |
+
"timestep_bias_portion": 0.25,
|
| 279 |
+
"training_scheduler_timestep_spacing": "trailing",
|
| 280 |
+
"inference_scheduler_timestep_spacing": "trailing",
|
| 281 |
+
"loss_type": "l2",
|
| 282 |
+
"huber_schedule": "snr",
|
| 283 |
+
"huber_c": 0.1,
|
| 284 |
+
"snr_gamma": null,
|
| 285 |
+
"masked_loss_probability": 1.0,
|
| 286 |
+
"hidream_use_load_balancing_loss": false,
|
| 287 |
+
"hidream_load_balancing_loss_weight": null,
|
| 288 |
+
"adam_beta1": 0.9,
|
| 289 |
+
"adam_beta2": 0.999,
|
| 290 |
+
"optimizer_beta1": null,
|
| 291 |
+
"optimizer_beta2": null,
|
| 292 |
+
"optimizer_cpu_offload_method": null,
|
| 293 |
+
"gradient_precision": null,
|
| 294 |
+
"adam_weight_decay": 0.01,
|
| 295 |
+
"adam_epsilon": 1e-08,
|
| 296 |
+
"prodigy_steps": null,
|
| 297 |
+
"max_grad_norm": 2.0,
|
| 298 |
+
"optimizer_config": null,
|
| 299 |
+
"grad_clip_method": "value",
|
| 300 |
+
"optimizer_offload_gradients": false,
|
| 301 |
+
"fuse_optimizer": false,
|
| 302 |
+
"optimizer_release_gradients": false,
|
| 303 |
+
"push_to_hub": false,
|
| 304 |
+
"publishing_config": null,
|
| 305 |
+
"post_upload_script": null,
|
| 306 |
+
"push_checkpoints_to_hub": false,
|
| 307 |
+
"push_to_hub_background": false,
|
| 308 |
+
"hub_model_id": "quzo/fl2",
|
| 309 |
+
"model_card_private": false,
|
| 310 |
+
"model_card_safe_for_work": false,
|
| 311 |
+
"model_card_note": null,
|
| 312 |
+
"report_to": "none",
|
| 313 |
+
"custom_tracker": null,
|
| 314 |
+
"checkpoint_step_interval": 200,
|
| 315 |
+
"checkpoint_epoch_interval": null,
|
| 316 |
+
"checkpointing_rolling_steps": 0,
|
| 317 |
+
"checkpointing_use_tempdir": false,
|
| 318 |
+
"checkpoints_rolling_total_limit": 1,
|
| 319 |
+
"tracker_run_name": "simpletuner-testing",
|
| 320 |
+
"tracker_project_name": "simpletuner",
|
| 321 |
+
"tracker_image_layout": "gallery",
|
| 322 |
+
"enable_watermark": false,
|
| 323 |
+
"framerate": null,
|
| 324 |
+
"seed_for_each_device": true,
|
| 325 |
+
"snr_weight": 1.0,
|
| 326 |
+
"webhook_config": [
|
| 327 |
+
{
|
| 328 |
+
"webhook_type": "raw",
|
| 329 |
+
"callback_url": "http://0.0.0.0:8001/callback",
|
| 330 |
+
"log_level": "info",
|
| 331 |
+
"ssl_no_verify": false
|
| 332 |
+
}
|
| 333 |
+
],
|
| 334 |
+
"webhook_reporting_interval": null,
|
| 335 |
+
"distillation_method": null,
|
| 336 |
+
"distillation_config": null,
|
| 337 |
+
"ema_validation": "comparison",
|
| 338 |
+
"local_rank": -1,
|
| 339 |
+
"ltx_train_mode": "i2v",
|
| 340 |
+
"ltx_i2v_prob": 0.1,
|
| 341 |
+
"ltx_partial_noise_fraction": 0.05,
|
| 342 |
+
"ltx_protect_first_frame": false,
|
| 343 |
+
"offload_param_path": null,
|
| 344 |
+
"offset_noise": false,
|
| 345 |
+
"quantize_activations": false,
|
| 346 |
+
"refiner_training": false,
|
| 347 |
+
"refiner_training_invert_schedule": false,
|
| 348 |
+
"refiner_training_strength": 0.2,
|
| 349 |
+
"sdxl_refiner_uses_full_range": false,
|
| 350 |
+
"sana_complex_human_instruction": "Given a user prompt, generate an 'Enhanced prompt' that provides detailed visual descriptions suitable for image generation. Evaluate the level of detail in the user prompt:\n- If the prompt is simple, focus on adding specifics about colors, shapes, sizes, textures, and spatial relationships to create vivid and concrete scenes.\n- Do not add detail that is not in the original prompt! Do not mention colors if the User Prompt does not!\n- If the prompt is already detailed, refine and enhance the existing details slightly without overcomplicating.\n- If the prompt contains explicit material, do not remove this information from the output, which would damage the performance of downstream classifiers.\nHere are examples of how to transform or refine prompts:\n- User Prompt: A cat sleeping -> Enhanced: A cat sleeping peacefully, showcasing the joy of pet ownership. Cute floof kitty cat gatto.\n- User Prompt: A busy city street -> Enhanced: A bustling city street scene featuring a crowd of people.\nPlease generate only the enhanced description for the prompt below and avoid including any additional commentary or evaluations:\nUser Prompt: ",
|
| 351 |
+
"vae_path": "black-forest-labs/FLUX.2-dev",
|
| 352 |
+
"accelerator_project_config": {
|
| 353 |
+
"project_dir": "/home/Ubuntu/.simpletuner/output",
|
| 354 |
+
"logging_dir": "/home/Ubuntu/.simpletuner/output/logs",
|
| 355 |
+
"automatic_checkpoint_naming": false,
|
| 356 |
+
"total_limit": null,
|
| 357 |
+
"iteration": 18,
|
| 358 |
+
"save_on_each_node": false
|
| 359 |
+
},
|
| 360 |
+
"process_group_kwargs": {
|
| 361 |
+
"backend": "nccl",
|
| 362 |
+
"init_method": null,
|
| 363 |
+
"timeout": "1:30:00"
|
| 364 |
+
},
|
| 365 |
+
"is_quantized": false,
|
| 366 |
+
"weight_dtype": "torch.bfloat16",
|
| 367 |
+
"disable_accelerator": false,
|
| 368 |
+
"lora_initialisation_style": true,
|
| 369 |
+
"checkpointing_steps": 200,
|
| 370 |
+
"use_fsdp": false,
|
| 371 |
+
"model_type_label": "Flux.2",
|
| 372 |
+
"use_deepspeed_optimizer": false,
|
| 373 |
+
"use_deepspeed_scheduler": false,
|
| 374 |
+
"base_weight_dtype": "torch.bfloat16",
|
| 375 |
+
"is_quanto": false,
|
| 376 |
+
"is_torchao": false,
|
| 377 |
+
"is_bnb": false,
|
| 378 |
+
"flow_matching": true,
|
| 379 |
+
"overrode_max_train_steps": false,
|
| 380 |
+
"total_num_batches": 6,
|
| 381 |
+
"num_update_steps_per_epoch": 6,
|
| 382 |
+
"total_batch_size": 2,
|
| 383 |
+
"enable_adamw_bf16": true,
|
| 384 |
+
"is_schedulefree": false,
|
| 385 |
+
"is_lr_scheduler_disabled": false,
|
| 386 |
+
"total_steps_remaining_at_start": 5000
|
| 387 |
+
}
|
checkpoint-3600/training_state-training-images.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"aspect_ratio_bucket_indices": {"1.0": ["/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im008.png", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im021.png"], "0.74": ["/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0014.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9743.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9827.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9935.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0003.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9796.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9890.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9865.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9260.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0090.jpg"]}, "buckets": ["0.74"], "exhausted_buckets": ["1.0"], "batch_size": 2, "current_bucket": 0, "seen_images": {"/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im008.png": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im021.png": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0014.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0003.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9827.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9260.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0090.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9890.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9743.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9865.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9796.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9935.jpg": true}, "current_epoch": 600}
|
checkpoint-3600/training_state.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"global_step": 3600, "epoch_step": 4199, "epoch": 600, "exhausted_backends": [], "repeats": {"training-images": 0}}
|
checkpoint-4000/README.md
ADDED
|
@@ -0,0 +1,129 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: other
|
| 3 |
+
base_model: "black-forest-labs/FLUX.2-dev"
|
| 4 |
+
tags:
|
| 5 |
+
- flux2
|
| 6 |
+
- flux2-diffusers
|
| 7 |
+
- text-to-image
|
| 8 |
+
- image-to-image
|
| 9 |
+
- diffusers
|
| 10 |
+
- simpletuner
|
| 11 |
+
- not-for-all-audiences
|
| 12 |
+
- lora
|
| 13 |
+
|
| 14 |
+
- template:sd-lora
|
| 15 |
+
- standard
|
| 16 |
+
pipeline_tag: text-to-image
|
| 17 |
+
inference: true
|
| 18 |
+
|
| 19 |
+
---
|
| 20 |
+
|
| 21 |
+
# quzo/fl2
|
| 22 |
+
|
| 23 |
+
This is a PEFT LoRA derived from [black-forest-labs/FLUX.2-dev](https://huggingface.co/black-forest-labs/FLUX.2-dev).
|
| 24 |
+
|
| 25 |
+
The main validation prompt used during training was:
|
| 26 |
+
```
|
| 27 |
+
bm82 man
|
| 28 |
+
```
|
| 29 |
+
|
| 30 |
+
|
| 31 |
+
## Validation settings
|
| 32 |
+
- CFG: `7.5`
|
| 33 |
+
- CFG Rescale: `0.0`
|
| 34 |
+
- Steps: `20`
|
| 35 |
+
- Sampler: `FlowMatchEulerDiscreteScheduler`
|
| 36 |
+
- Seed: `None`
|
| 37 |
+
- Resolution: `1024x1024`
|
| 38 |
+
|
| 39 |
+
|
| 40 |
+
Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
|
| 41 |
+
|
| 42 |
+
|
| 43 |
+
|
| 44 |
+
|
| 45 |
+
<Gallery />
|
| 46 |
+
|
| 47 |
+
The text encoder **was not** trained.
|
| 48 |
+
You may reuse the base model text encoder for inference.
|
| 49 |
+
|
| 50 |
+
|
| 51 |
+
## Training settings
|
| 52 |
+
|
| 53 |
+
- Training epochs: 666
|
| 54 |
+
- Training steps: 4000
|
| 55 |
+
- Learning rate: 0.0001
|
| 56 |
+
- Learning rate schedule: constant_with_warmup
|
| 57 |
+
- Warmup steps: 0
|
| 58 |
+
- Max grad value: 2.0
|
| 59 |
+
- Effective batch size: 2
|
| 60 |
+
- Micro-batch size: 2
|
| 61 |
+
- Gradient accumulation steps: 1
|
| 62 |
+
- Number of GPUs: 1
|
| 63 |
+
- Gradient checkpointing: True
|
| 64 |
+
- Prediction type: flow_matching[]
|
| 65 |
+
- Optimizer: adamw_bf16
|
| 66 |
+
- Trainable parameter precision: Pure BF16
|
| 67 |
+
- Base model precision: `no_change`
|
| 68 |
+
- Caption dropout probability: 0.1%
|
| 69 |
+
|
| 70 |
+
|
| 71 |
+
|
| 72 |
+
- LoRA Rank: 16
|
| 73 |
+
- LoRA Alpha: 16.0
|
| 74 |
+
- LoRA Dropout: 0.1
|
| 75 |
+
- LoRA initialisation style: default
|
| 76 |
+
- LoRA mode: Standard
|
| 77 |
+
|
| 78 |
+
|
| 79 |
+
## Datasets
|
| 80 |
+
|
| 81 |
+
### training-images
|
| 82 |
+
- Repeats: 0
|
| 83 |
+
- Total number of images: 12
|
| 84 |
+
- Total number of aspect buckets: 1
|
| 85 |
+
- Resolution: 1.048576 megapixels
|
| 86 |
+
- Cropped: False
|
| 87 |
+
- Crop style: None
|
| 88 |
+
- Crop aspect: None
|
| 89 |
+
- Used for regularisation data: No
|
| 90 |
+
|
| 91 |
+
|
| 92 |
+
## Inference
|
| 93 |
+
|
| 94 |
+
|
| 95 |
+
```python
|
| 96 |
+
import torch
|
| 97 |
+
from diffusers import DiffusionPipeline
|
| 98 |
+
|
| 99 |
+
model_id = 'black-forest-labs/FLUX.2-dev'
|
| 100 |
+
adapter_id = 'quzo/fl2'
|
| 101 |
+
pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
|
| 102 |
+
pipeline.load_lora_weights(adapter_id)
|
| 103 |
+
|
| 104 |
+
prompt = "bm82 man"
|
| 105 |
+
negative_prompt = 'blurry, cropped, ugly'
|
| 106 |
+
|
| 107 |
+
## Optional: quantise the model to save on vram.
|
| 108 |
+
## Note: The model was not quantised during training, so it is not necessary to quantise it during inference time.
|
| 109 |
+
#from optimum.quanto import quantize, freeze, qint8
|
| 110 |
+
#quantize(pipeline.transformer, weights=qint8)
|
| 111 |
+
#freeze(pipeline.transformer)
|
| 112 |
+
|
| 113 |
+
pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
|
| 114 |
+
model_output = pipeline(
|
| 115 |
+
prompt=prompt,
|
| 116 |
+
negative_prompt=negative_prompt,
|
| 117 |
+
num_inference_steps=20,
|
| 118 |
+
generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
|
| 119 |
+
width=1024,
|
| 120 |
+
height=1024,
|
| 121 |
+
guidance_scale=7.5,
|
| 122 |
+
).images[0]
|
| 123 |
+
|
| 124 |
+
model_output.save("output.png", format="PNG")
|
| 125 |
+
|
| 126 |
+
```
|
| 127 |
+
|
| 128 |
+
|
| 129 |
+
|
checkpoint-4000/optimizer.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fb5e08c3dd2f66e213afe2a25b1326f19804647e293077e94c7081620e01658e
|
| 3 |
+
size 321008843
|
checkpoint-4000/pytorch_lora_weights.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:263d549f7e01a872a1f228806b722556cb672a1e30876f93eb1601f703299642
|
| 3 |
+
size 106979136
|
checkpoint-4000/random_states_0.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:042666dd1082a430bd12fc8536141c84d36683e4390d2ea7c2f635ce006f9880
|
| 3 |
+
size 14757
|
checkpoint-4000/scheduler.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8ceee27b48e7dcd059bd1aa420fb9860574384cf52a7228f1287f54609342ffe
|
| 3 |
+
size 1401
|
checkpoint-4000/simpletuner_config.json
ADDED
|
@@ -0,0 +1,387 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"model_family": "flux2",
|
| 3 |
+
"model_flavour": "dev",
|
| 4 |
+
"controlnet": false,
|
| 5 |
+
"pretrained_model_name_or_path": "black-forest-labs/FLUX.2-dev",
|
| 6 |
+
"output_dir": "/home/Ubuntu/.simpletuner/output",
|
| 7 |
+
"logging_dir": "/home/Ubuntu/.simpletuner/output/logs",
|
| 8 |
+
"model_type": "lora",
|
| 9 |
+
"seed": null,
|
| 10 |
+
"resolution": 1024,
|
| 11 |
+
"resume_from_checkpoint": null,
|
| 12 |
+
"prediction_type": null,
|
| 13 |
+
"pretrained_vae_model_name_or_path": "black-forest-labs/FLUX.2-dev",
|
| 14 |
+
"vae_dtype": "bf16",
|
| 15 |
+
"vae_cache_ondemand": false,
|
| 16 |
+
"vae_cache_disable": false,
|
| 17 |
+
"accelerator_cache_clear_interval": null,
|
| 18 |
+
"aspect_bucket_rounding": 2,
|
| 19 |
+
"base_model_precision": "no_change",
|
| 20 |
+
"text_encoder_1_precision": "no_change",
|
| 21 |
+
"text_encoder_2_precision": "no_change",
|
| 22 |
+
"text_encoder_3_precision": "no_change",
|
| 23 |
+
"text_encoder_4_precision": "no_change",
|
| 24 |
+
"gradient_checkpointing_interval": null,
|
| 25 |
+
"offload_during_startup": false,
|
| 26 |
+
"quantize_via": "accelerator",
|
| 27 |
+
"wan_force_2_1_time_embedding": false,
|
| 28 |
+
"fuse_qkv_projections": false,
|
| 29 |
+
"rescale_betas_zero_snr": false,
|
| 30 |
+
"control": false,
|
| 31 |
+
"controlnet_custom_config": null,
|
| 32 |
+
"controlnet_model_name_or_path": null,
|
| 33 |
+
"tread_config": null,
|
| 34 |
+
"pretrained_transformer_model_name_or_path": null,
|
| 35 |
+
"pretrained_transformer_subfolder": "transformer",
|
| 36 |
+
"pretrained_unet_model_name_or_path": null,
|
| 37 |
+
"pretrained_unet_subfolder": "unet",
|
| 38 |
+
"pretrained_t5_model_name_or_path": null,
|
| 39 |
+
"revision": null,
|
| 40 |
+
"variant": null,
|
| 41 |
+
"base_model_default_dtype": "bf16",
|
| 42 |
+
"unet_attention_slice": false,
|
| 43 |
+
"num_train_epochs": 834,
|
| 44 |
+
"max_train_steps": 5000,
|
| 45 |
+
"train_batch_size": 2,
|
| 46 |
+
"learning_rate": 0.0001,
|
| 47 |
+
"optimizer": "adamw_bf16",
|
| 48 |
+
"lr_scheduler": "constant_with_warmup",
|
| 49 |
+
"gradient_accumulation_steps": 1,
|
| 50 |
+
"lr_warmup_steps": 0,
|
| 51 |
+
"checkpoints_total_limit": 5,
|
| 52 |
+
"gradient_checkpointing": true,
|
| 53 |
+
"enable_group_offload": false,
|
| 54 |
+
"ramtorch": false,
|
| 55 |
+
"ramtorch_target_modules": null,
|
| 56 |
+
"ramtorch_text_encoder": false,
|
| 57 |
+
"ramtorch_vae": false,
|
| 58 |
+
"ramtorch_controlnet": false,
|
| 59 |
+
"group_offload_type": "block_level",
|
| 60 |
+
"group_offload_blocks_per_group": 1,
|
| 61 |
+
"group_offload_use_stream": false,
|
| 62 |
+
"group_offload_to_disk_path": "",
|
| 63 |
+
"group_offload_text_encoder": false,
|
| 64 |
+
"group_offload_vae": false,
|
| 65 |
+
"offload_during_save": false,
|
| 66 |
+
"enable_chunked_feed_forward": false,
|
| 67 |
+
"feed_forward_chunk_size": null,
|
| 68 |
+
"train_text_encoder": false,
|
| 69 |
+
"text_encoder_lr": null,
|
| 70 |
+
"lyrics_embedder_train": false,
|
| 71 |
+
"lyrics_embedder_optimizer": null,
|
| 72 |
+
"lyrics_embedder_lr": null,
|
| 73 |
+
"lyrics_embedder_lr_scheduler": null,
|
| 74 |
+
"lr_num_cycles": 1,
|
| 75 |
+
"lr_power": 0.8,
|
| 76 |
+
"use_soft_min_snr": false,
|
| 77 |
+
"use_ema": false,
|
| 78 |
+
"ema_device": "cpu",
|
| 79 |
+
"ema_cpu_only": false,
|
| 80 |
+
"ema_update_interval": 1,
|
| 81 |
+
"ema_foreach_disable": false,
|
| 82 |
+
"ema_decay": 0.995,
|
| 83 |
+
"lora_rank": 16,
|
| 84 |
+
"lora_alpha": 16.0,
|
| 85 |
+
"lora_type": "standard",
|
| 86 |
+
"lora_dropout": 0.1,
|
| 87 |
+
"lora_init_type": "default",
|
| 88 |
+
"peft_lora_mode": "standard",
|
| 89 |
+
"singlora_ramp_up_steps": 0,
|
| 90 |
+
"init_lora": null,
|
| 91 |
+
"lycoris_config": "/home/Ubuntu/.simpletuner/configs/config/lycoris_config.json",
|
| 92 |
+
"init_lokr_norm": null,
|
| 93 |
+
"flux_lora_target": "all",
|
| 94 |
+
"acestep_lora_target": "attn_qkv+linear_qkv",
|
| 95 |
+
"use_dora": false,
|
| 96 |
+
"resolution_type": "pixel_area",
|
| 97 |
+
"data_backend_config": "/home/Ubuntu/.simpletuner/configs/111/multidatabackend-bartek_mizak.json",
|
| 98 |
+
"caption_strategy": "filename",
|
| 99 |
+
"conditioning_multidataset_sampling": "random",
|
| 100 |
+
"instance_prompt": null,
|
| 101 |
+
"parquet_caption_column": null,
|
| 102 |
+
"parquet_filename_column": null,
|
| 103 |
+
"ignore_missing_files": false,
|
| 104 |
+
"vae_cache_scan_behaviour": "recreate",
|
| 105 |
+
"vae_enable_slicing": false,
|
| 106 |
+
"vae_enable_tiling": false,
|
| 107 |
+
"vae_batch_size": 4,
|
| 108 |
+
"max_upscale_threshold": null,
|
| 109 |
+
"caption_dropout_probability": 0.1,
|
| 110 |
+
"tokenizer_max_length": 512,
|
| 111 |
+
"audio_max_duration_seconds": null,
|
| 112 |
+
"audio_min_duration_seconds": null,
|
| 113 |
+
"audio_channels": 1,
|
| 114 |
+
"audio_duration_interval": 3.0,
|
| 115 |
+
"audio_truncation_mode": "beginning",
|
| 116 |
+
"validation_step_interval": 100,
|
| 117 |
+
"validation_epoch_interval": null,
|
| 118 |
+
"disable_benchmark": false,
|
| 119 |
+
"validation_preview": false,
|
| 120 |
+
"validation_preview_steps": 1,
|
| 121 |
+
"validation_prompt": "bm82 man",
|
| 122 |
+
"validation_lyrics": null,
|
| 123 |
+
"validation_audio_duration": 30.0,
|
| 124 |
+
"num_validation_images": 1,
|
| 125 |
+
"num_eval_images": 4,
|
| 126 |
+
"eval_steps_interval": null,
|
| 127 |
+
"eval_timesteps": 28,
|
| 128 |
+
"eval_dataset_pooling": false,
|
| 129 |
+
"evaluation_type": "none",
|
| 130 |
+
"pretrained_evaluation_model_name_or_path": "openai/clip-vit-large-patch14-336",
|
| 131 |
+
"validation_guidance": 7.5,
|
| 132 |
+
"validation_num_inference_steps": 20,
|
| 133 |
+
"validation_on_startup": false,
|
| 134 |
+
"validation_method": "simpletuner-local",
|
| 135 |
+
"validation_external_script": null,
|
| 136 |
+
"validation_external_background": false,
|
| 137 |
+
"validation_using_datasets": false,
|
| 138 |
+
"validation_torch_compile": false,
|
| 139 |
+
"validation_guidance_real": 1.0,
|
| 140 |
+
"validation_no_cfg_until_timestep": 2,
|
| 141 |
+
"validation_negative_prompt": "blurry, cropped, ugly",
|
| 142 |
+
"validation_randomize": false,
|
| 143 |
+
"validation_seed": null,
|
| 144 |
+
"validation_multigpu": "batch-parallel",
|
| 145 |
+
"validation_disable": false,
|
| 146 |
+
"validation_prompt_library": false,
|
| 147 |
+
"user_prompt_library": null,
|
| 148 |
+
"eval_dataset_id": null,
|
| 149 |
+
"validation_stitch_input_location": "left",
|
| 150 |
+
"validation_guidance_rescale": 0.0,
|
| 151 |
+
"validation_disable_unconditional": false,
|
| 152 |
+
"validation_guidance_skip_layers": null,
|
| 153 |
+
"validation_guidance_skip_layers_start": 0.01,
|
| 154 |
+
"validation_guidance_skip_layers_stop": 0.2,
|
| 155 |
+
"validation_guidance_skip_scale": 2.8,
|
| 156 |
+
"validation_lycoris_strength": 1.0,
|
| 157 |
+
"validation_noise_scheduler": null,
|
| 158 |
+
"validation_num_video_frames": null,
|
| 159 |
+
"validation_resolution": "1024x1024",
|
| 160 |
+
"validation_seed_source": "cpu",
|
| 161 |
+
"validation_torch_compile_mode": "max-autotune",
|
| 162 |
+
"validation_adapter_path": null,
|
| 163 |
+
"validation_adapter_name": null,
|
| 164 |
+
"validation_adapter_strength": 1.0,
|
| 165 |
+
"validation_adapter_mode": "adapter_only",
|
| 166 |
+
"validation_adapter_config": null,
|
| 167 |
+
"i_know_what_i_am_doing": false,
|
| 168 |
+
"flow_sigmoid_scale": 1.0,
|
| 169 |
+
"flux_fast_schedule": false,
|
| 170 |
+
"flow_use_uniform_schedule": false,
|
| 171 |
+
"flow_use_beta_schedule": false,
|
| 172 |
+
"flow_beta_schedule_alpha": 2.0,
|
| 173 |
+
"flow_beta_schedule_beta": 2.0,
|
| 174 |
+
"flow_schedule_shift": 3.0,
|
| 175 |
+
"flow_schedule_auto_shift": false,
|
| 176 |
+
"flux_guidance_mode": "constant",
|
| 177 |
+
"flux_attention_masked_training": false,
|
| 178 |
+
"flux_guidance_value": 1.0,
|
| 179 |
+
"flux_guidance_min": 0.0,
|
| 180 |
+
"flux_guidance_max": 4.0,
|
| 181 |
+
"t5_padding": "unmodified",
|
| 182 |
+
"sd3_clip_uncond_behaviour": "empty_string",
|
| 183 |
+
"sd3_t5_uncond_behaviour": null,
|
| 184 |
+
"soft_min_snr_sigma_data": null,
|
| 185 |
+
"mixed_precision": "bf16",
|
| 186 |
+
"attention_mechanism": "diffusers",
|
| 187 |
+
"sla_config": null,
|
| 188 |
+
"sageattention_usage": {
|
| 189 |
+
"_value_": "inference",
|
| 190 |
+
"_name_": "INFERENCE",
|
| 191 |
+
"__objclass__": "{'_generate_next_value_': <staticmethod(<function Enum._generate_next_value_ at 0x7649ca9cdf80>)>, '__module__': 'simpletuner.helpers.training.attention_backend', 'allows_training': <property object at 0x764872f93880>, 'allows_inference': <property object at 0x764872f93790>, 'from_raw': <classmethod(<function AttentionBackendMode.from_raw at 0x764872fe85e0>)>, '_new_member_': <built-in method __new__ of type object at 0xa234e0>, '_use_args_': True, '_member_names_': ['TRAINING', 'INFERENCE', 'TRAINING_AND_INFERENCE'], '_member_map_': {'TRAINING': <AttentionBackendMode.TRAINING: 'training'>, 'INFERENCE': <AttentionBackendMode.INFERENCE: 'inference'>, 'TRAINING_AND_INFERENCE': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>}, '_value2member_map_': {'training': <AttentionBackendMode.TRAINING: 'training'>, 'inference': <AttentionBackendMode.INFERENCE: 'inference'>, 'training+inference': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>}, '_unhashable_values_': [], '_member_type_': <class 'str'>, '_value_repr_': <slot wrapper '__repr__' of 'str' objects>, '__dict__': <attribute '__dict__' of 'AttentionBackendMode' objects>, '__weakref__': <attribute '__weakref__' of 'AttentionBackendMode' objects>, '__doc__': None, 'TRAINING': <AttentionBackendMode.TRAINING: 'training'>, 'INFERENCE': <AttentionBackendMode.INFERENCE: 'inference'>, 'TRAINING_AND_INFERENCE': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>, '__repr__': <function Enum.__repr__ at 0x7649ca9ce0c0>, '__str__': <function Enum.__str__ at 0x7649ca9ce160>, '__format__': <function Enum.__format__ at 0x7649ca9ce2a0>, '__new__': <function Enum.__new__ at 0x7649ca9cde40>}",
|
| 192 |
+
"_sort_order_": 1
|
| 193 |
+
},
|
| 194 |
+
"disable_tf32": false,
|
| 195 |
+
"set_grads_to_none": false,
|
| 196 |
+
"noise_offset": 0.1,
|
| 197 |
+
"noise_offset_probability": 0.25,
|
| 198 |
+
"input_perturbation": 0.0,
|
| 199 |
+
"input_perturbation_steps": 0,
|
| 200 |
+
"lr_end": "4e-7",
|
| 201 |
+
"lr_scale": false,
|
| 202 |
+
"lr_scale_sqrt": false,
|
| 203 |
+
"ignore_final_epochs": false,
|
| 204 |
+
"freeze_encoder_before": 12,
|
| 205 |
+
"freeze_encoder_after": 17,
|
| 206 |
+
"freeze_encoder_strategy": "after",
|
| 207 |
+
"layer_freeze_strategy": null,
|
| 208 |
+
"fully_unload_text_encoder": false,
|
| 209 |
+
"save_text_encoder": false,
|
| 210 |
+
"text_encoder_limit": 100,
|
| 211 |
+
"prepend_instance_prompt": false,
|
| 212 |
+
"only_instance_prompt": false,
|
| 213 |
+
"data_aesthetic_score": 7.0,
|
| 214 |
+
"delete_unwanted_images": false,
|
| 215 |
+
"delete_problematic_images": false,
|
| 216 |
+
"disable_bucket_pruning": false,
|
| 217 |
+
"allow_dataset_oversubscription": false,
|
| 218 |
+
"disable_segmented_timestep_sampling": false,
|
| 219 |
+
"preserve_data_backend_cache": false,
|
| 220 |
+
"override_dataset_config": false,
|
| 221 |
+
"cache_dir": "/home/Ubuntu/.simpletuner/output/cache",
|
| 222 |
+
"cache_dir_text": "cache",
|
| 223 |
+
"cache_dir_vae": "",
|
| 224 |
+
"compress_disk_cache": true,
|
| 225 |
+
"aspect_bucket_disable_rebuild": false,
|
| 226 |
+
"keep_vae_loaded": false,
|
| 227 |
+
"skip_file_discovery": "",
|
| 228 |
+
"data_backend_sampling": "auto-weighting",
|
| 229 |
+
"image_processing_batch_size": 32,
|
| 230 |
+
"write_batch_size": 128,
|
| 231 |
+
"read_batch_size": 25,
|
| 232 |
+
"enable_multiprocessing": false,
|
| 233 |
+
"accelerate_config": null,
|
| 234 |
+
"deepspeed_config": null,
|
| 235 |
+
"fsdp_enable": false,
|
| 236 |
+
"fsdp_version": 2,
|
| 237 |
+
"fsdp_reshard_after_forward": false,
|
| 238 |
+
"fsdp_state_dict_type": "SHARDED_STATE_DICT",
|
| 239 |
+
"fsdp_cpu_ram_efficient_loading": false,
|
| 240 |
+
"fsdp_auto_wrap_policy": "TRANSFORMER_BASED_WRAP",
|
| 241 |
+
"fsdp_limit_all_gathers": false,
|
| 242 |
+
"fsdp_cpu_offload": false,
|
| 243 |
+
"fsdp_activation_checkpointing": false,
|
| 244 |
+
"fsdp_transformer_layer_cls_to_wrap": null,
|
| 245 |
+
"context_parallel_size": 1,
|
| 246 |
+
"context_parallel_comm_strategy": "allgather",
|
| 247 |
+
"num_processes": 1,
|
| 248 |
+
"num_machines": 1,
|
| 249 |
+
"accelerate_extra_args": null,
|
| 250 |
+
"main_process_ip": "127.0.0.1",
|
| 251 |
+
"main_process_port": 29500,
|
| 252 |
+
"machine_rank": 0,
|
| 253 |
+
"same_network": true,
|
| 254 |
+
"dynamo_backend": "no",
|
| 255 |
+
"dynamo_mode": "",
|
| 256 |
+
"dynamo_fullgraph": false,
|
| 257 |
+
"dynamo_dynamic": false,
|
| 258 |
+
"dynamo_use_regional_compilation": false,
|
| 259 |
+
"max_workers": 32,
|
| 260 |
+
"aws_max_pool_connections": 128,
|
| 261 |
+
"torch_num_threads": 8,
|
| 262 |
+
"dataloader_prefetch": false,
|
| 263 |
+
"dataloader_prefetch_qlen": 10,
|
| 264 |
+
"aspect_bucket_worker_count": 12,
|
| 265 |
+
"aspect_bucket_alignment": 16,
|
| 266 |
+
"minimum_image_size": null,
|
| 267 |
+
"maximum_image_size": null,
|
| 268 |
+
"target_downsample_size": null,
|
| 269 |
+
"metadata_update_interval": 3600,
|
| 270 |
+
"debug_aspect_buckets": false,
|
| 271 |
+
"debug_dataset_loader": false,
|
| 272 |
+
"print_filenames": false,
|
| 273 |
+
"print_sampler_statistics": false,
|
| 274 |
+
"timestep_bias_strategy": null,
|
| 275 |
+
"timestep_bias_begin": 0,
|
| 276 |
+
"timestep_bias_end": 1000,
|
| 277 |
+
"timestep_bias_multiplier": 1.0,
|
| 278 |
+
"timestep_bias_portion": 0.25,
|
| 279 |
+
"training_scheduler_timestep_spacing": "trailing",
|
| 280 |
+
"inference_scheduler_timestep_spacing": "trailing",
|
| 281 |
+
"loss_type": "l2",
|
| 282 |
+
"huber_schedule": "snr",
|
| 283 |
+
"huber_c": 0.1,
|
| 284 |
+
"snr_gamma": null,
|
| 285 |
+
"masked_loss_probability": 1.0,
|
| 286 |
+
"hidream_use_load_balancing_loss": false,
|
| 287 |
+
"hidream_load_balancing_loss_weight": null,
|
| 288 |
+
"adam_beta1": 0.9,
|
| 289 |
+
"adam_beta2": 0.999,
|
| 290 |
+
"optimizer_beta1": null,
|
| 291 |
+
"optimizer_beta2": null,
|
| 292 |
+
"optimizer_cpu_offload_method": null,
|
| 293 |
+
"gradient_precision": null,
|
| 294 |
+
"adam_weight_decay": 0.01,
|
| 295 |
+
"adam_epsilon": 1e-08,
|
| 296 |
+
"prodigy_steps": null,
|
| 297 |
+
"max_grad_norm": 2.0,
|
| 298 |
+
"optimizer_config": null,
|
| 299 |
+
"grad_clip_method": "value",
|
| 300 |
+
"optimizer_offload_gradients": false,
|
| 301 |
+
"fuse_optimizer": false,
|
| 302 |
+
"optimizer_release_gradients": false,
|
| 303 |
+
"push_to_hub": false,
|
| 304 |
+
"publishing_config": null,
|
| 305 |
+
"post_upload_script": null,
|
| 306 |
+
"push_checkpoints_to_hub": false,
|
| 307 |
+
"push_to_hub_background": false,
|
| 308 |
+
"hub_model_id": "quzo/fl2",
|
| 309 |
+
"model_card_private": false,
|
| 310 |
+
"model_card_safe_for_work": false,
|
| 311 |
+
"model_card_note": null,
|
| 312 |
+
"report_to": "none",
|
| 313 |
+
"custom_tracker": null,
|
| 314 |
+
"checkpoint_step_interval": 200,
|
| 315 |
+
"checkpoint_epoch_interval": null,
|
| 316 |
+
"checkpointing_rolling_steps": 0,
|
| 317 |
+
"checkpointing_use_tempdir": false,
|
| 318 |
+
"checkpoints_rolling_total_limit": 1,
|
| 319 |
+
"tracker_run_name": "simpletuner-testing",
|
| 320 |
+
"tracker_project_name": "simpletuner",
|
| 321 |
+
"tracker_image_layout": "gallery",
|
| 322 |
+
"enable_watermark": false,
|
| 323 |
+
"framerate": null,
|
| 324 |
+
"seed_for_each_device": true,
|
| 325 |
+
"snr_weight": 1.0,
|
| 326 |
+
"webhook_config": [
|
| 327 |
+
{
|
| 328 |
+
"webhook_type": "raw",
|
| 329 |
+
"callback_url": "http://0.0.0.0:8001/callback",
|
| 330 |
+
"log_level": "info",
|
| 331 |
+
"ssl_no_verify": false
|
| 332 |
+
}
|
| 333 |
+
],
|
| 334 |
+
"webhook_reporting_interval": null,
|
| 335 |
+
"distillation_method": null,
|
| 336 |
+
"distillation_config": null,
|
| 337 |
+
"ema_validation": "comparison",
|
| 338 |
+
"local_rank": -1,
|
| 339 |
+
"ltx_train_mode": "i2v",
|
| 340 |
+
"ltx_i2v_prob": 0.1,
|
| 341 |
+
"ltx_partial_noise_fraction": 0.05,
|
| 342 |
+
"ltx_protect_first_frame": false,
|
| 343 |
+
"offload_param_path": null,
|
| 344 |
+
"offset_noise": false,
|
| 345 |
+
"quantize_activations": false,
|
| 346 |
+
"refiner_training": false,
|
| 347 |
+
"refiner_training_invert_schedule": false,
|
| 348 |
+
"refiner_training_strength": 0.2,
|
| 349 |
+
"sdxl_refiner_uses_full_range": false,
|
| 350 |
+
"sana_complex_human_instruction": "Given a user prompt, generate an 'Enhanced prompt' that provides detailed visual descriptions suitable for image generation. Evaluate the level of detail in the user prompt:\n- If the prompt is simple, focus on adding specifics about colors, shapes, sizes, textures, and spatial relationships to create vivid and concrete scenes.\n- Do not add detail that is not in the original prompt! Do not mention colors if the User Prompt does not!\n- If the prompt is already detailed, refine and enhance the existing details slightly without overcomplicating.\n- If the prompt contains explicit material, do not remove this information from the output, which would damage the performance of downstream classifiers.\nHere are examples of how to transform or refine prompts:\n- User Prompt: A cat sleeping -> Enhanced: A cat sleeping peacefully, showcasing the joy of pet ownership. Cute floof kitty cat gatto.\n- User Prompt: A busy city street -> Enhanced: A bustling city street scene featuring a crowd of people.\nPlease generate only the enhanced description for the prompt below and avoid including any additional commentary or evaluations:\nUser Prompt: ",
|
| 351 |
+
"vae_path": "black-forest-labs/FLUX.2-dev",
|
| 352 |
+
"accelerator_project_config": {
|
| 353 |
+
"project_dir": "/home/Ubuntu/.simpletuner/output",
|
| 354 |
+
"logging_dir": "/home/Ubuntu/.simpletuner/output/logs",
|
| 355 |
+
"automatic_checkpoint_naming": false,
|
| 356 |
+
"total_limit": null,
|
| 357 |
+
"iteration": 20,
|
| 358 |
+
"save_on_each_node": false
|
| 359 |
+
},
|
| 360 |
+
"process_group_kwargs": {
|
| 361 |
+
"backend": "nccl",
|
| 362 |
+
"init_method": null,
|
| 363 |
+
"timeout": "1:30:00"
|
| 364 |
+
},
|
| 365 |
+
"is_quantized": false,
|
| 366 |
+
"weight_dtype": "torch.bfloat16",
|
| 367 |
+
"disable_accelerator": false,
|
| 368 |
+
"lora_initialisation_style": true,
|
| 369 |
+
"checkpointing_steps": 200,
|
| 370 |
+
"use_fsdp": false,
|
| 371 |
+
"model_type_label": "Flux.2",
|
| 372 |
+
"use_deepspeed_optimizer": false,
|
| 373 |
+
"use_deepspeed_scheduler": false,
|
| 374 |
+
"base_weight_dtype": "torch.bfloat16",
|
| 375 |
+
"is_quanto": false,
|
| 376 |
+
"is_torchao": false,
|
| 377 |
+
"is_bnb": false,
|
| 378 |
+
"flow_matching": true,
|
| 379 |
+
"overrode_max_train_steps": false,
|
| 380 |
+
"total_num_batches": 6,
|
| 381 |
+
"num_update_steps_per_epoch": 6,
|
| 382 |
+
"total_batch_size": 2,
|
| 383 |
+
"enable_adamw_bf16": true,
|
| 384 |
+
"is_schedulefree": false,
|
| 385 |
+
"is_lr_scheduler_disabled": false,
|
| 386 |
+
"total_steps_remaining_at_start": 5000
|
| 387 |
+
}
|
checkpoint-4000/training_state-training-images.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"aspect_ratio_bucket_indices": {"1.0": ["/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im008.png", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im021.png"], "0.74": ["/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0014.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9743.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9827.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9935.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0003.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9796.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9890.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9865.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9260.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0090.jpg"]}, "buckets": ["0.74"], "exhausted_buckets": ["1.0"], "batch_size": 2, "current_bucket": 0, "seen_images": {"/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im008.png": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im021.png": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9796.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0003.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9935.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0014.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9827.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9743.jpg": true}, "current_epoch": 667}
|
checkpoint-4000/training_state.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"global_step": 4000, "epoch_step": 4666, "epoch": 667, "exhausted_backends": [], "repeats": {"training-images": 0}}
|