quzo commited on
Commit
42a38ce
·
verified ·
1 Parent(s): 1680d6e

Upload 32 files

Browse files
checkpoint-3200/README.md ADDED
@@ -0,0 +1,129 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: "black-forest-labs/FLUX.2-dev"
4
+ tags:
5
+ - flux2
6
+ - flux2-diffusers
7
+ - text-to-image
8
+ - image-to-image
9
+ - diffusers
10
+ - simpletuner
11
+ - not-for-all-audiences
12
+ - lora
13
+
14
+ - template:sd-lora
15
+ - standard
16
+ pipeline_tag: text-to-image
17
+ inference: true
18
+
19
+ ---
20
+
21
+ # quzo/fl2
22
+
23
+ This is a PEFT LoRA derived from [black-forest-labs/FLUX.2-dev](https://huggingface.co/black-forest-labs/FLUX.2-dev).
24
+
25
+ The main validation prompt used during training was:
26
+ ```
27
+ bm82 man
28
+ ```
29
+
30
+
31
+ ## Validation settings
32
+ - CFG: `7.5`
33
+ - CFG Rescale: `0.0`
34
+ - Steps: `20`
35
+ - Sampler: `FlowMatchEulerDiscreteScheduler`
36
+ - Seed: `None`
37
+ - Resolution: `1024x1024`
38
+
39
+
40
+ Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
41
+
42
+
43
+
44
+
45
+ <Gallery />
46
+
47
+ The text encoder **was not** trained.
48
+ You may reuse the base model text encoder for inference.
49
+
50
+
51
+ ## Training settings
52
+
53
+ - Training epochs: 533
54
+ - Training steps: 3200
55
+ - Learning rate: 0.0001
56
+ - Learning rate schedule: constant_with_warmup
57
+ - Warmup steps: 0
58
+ - Max grad value: 2.0
59
+ - Effective batch size: 2
60
+ - Micro-batch size: 2
61
+ - Gradient accumulation steps: 1
62
+ - Number of GPUs: 1
63
+ - Gradient checkpointing: True
64
+ - Prediction type: flow_matching[]
65
+ - Optimizer: adamw_bf16
66
+ - Trainable parameter precision: Pure BF16
67
+ - Base model precision: `no_change`
68
+ - Caption dropout probability: 0.1%
69
+
70
+
71
+
72
+ - LoRA Rank: 16
73
+ - LoRA Alpha: 16.0
74
+ - LoRA Dropout: 0.1
75
+ - LoRA initialisation style: default
76
+ - LoRA mode: Standard
77
+
78
+
79
+ ## Datasets
80
+
81
+ ### training-images
82
+ - Repeats: 0
83
+ - Total number of images: 12
84
+ - Total number of aspect buckets: 2
85
+ - Resolution: 1.048576 megapixels
86
+ - Cropped: False
87
+ - Crop style: None
88
+ - Crop aspect: None
89
+ - Used for regularisation data: No
90
+
91
+
92
+ ## Inference
93
+
94
+
95
+ ```python
96
+ import torch
97
+ from diffusers import DiffusionPipeline
98
+
99
+ model_id = 'black-forest-labs/FLUX.2-dev'
100
+ adapter_id = 'quzo/fl2'
101
+ pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
102
+ pipeline.load_lora_weights(adapter_id)
103
+
104
+ prompt = "bm82 man"
105
+ negative_prompt = 'blurry, cropped, ugly'
106
+
107
+ ## Optional: quantise the model to save on vram.
108
+ ## Note: The model was not quantised during training, so it is not necessary to quantise it during inference time.
109
+ #from optimum.quanto import quantize, freeze, qint8
110
+ #quantize(pipeline.transformer, weights=qint8)
111
+ #freeze(pipeline.transformer)
112
+
113
+ pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
114
+ model_output = pipeline(
115
+ prompt=prompt,
116
+ negative_prompt=negative_prompt,
117
+ num_inference_steps=20,
118
+ generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
119
+ width=1024,
120
+ height=1024,
121
+ guidance_scale=7.5,
122
+ ).images[0]
123
+
124
+ model_output.save("output.png", format="PNG")
125
+
126
+ ```
127
+
128
+
129
+
checkpoint-3200/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcd3a3a7c184a0031b7eef26313fd7020b0b6fcca1949725feca4d9993b0f896
3
+ size 321008843
checkpoint-3200/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88ade5a67008a9aba49b0b6aabf1dd60c57d072029937bf5fb092c49a212413f
3
+ size 106979136
checkpoint-3200/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05c0092c4c618520653c8345b12f78775995d65375428b89ab85aaa646f7482d
3
+ size 14757
checkpoint-3200/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce701db56c2af726940feb1b070907060c446d2e48225a0937aba9bfefcf32e6
3
+ size 1401
checkpoint-3200/simpletuner_config.json ADDED
@@ -0,0 +1,387 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_family": "flux2",
3
+ "model_flavour": "dev",
4
+ "controlnet": false,
5
+ "pretrained_model_name_or_path": "black-forest-labs/FLUX.2-dev",
6
+ "output_dir": "/home/Ubuntu/.simpletuner/output",
7
+ "logging_dir": "/home/Ubuntu/.simpletuner/output/logs",
8
+ "model_type": "lora",
9
+ "seed": null,
10
+ "resolution": 1024,
11
+ "resume_from_checkpoint": null,
12
+ "prediction_type": null,
13
+ "pretrained_vae_model_name_or_path": "black-forest-labs/FLUX.2-dev",
14
+ "vae_dtype": "bf16",
15
+ "vae_cache_ondemand": false,
16
+ "vae_cache_disable": false,
17
+ "accelerator_cache_clear_interval": null,
18
+ "aspect_bucket_rounding": 2,
19
+ "base_model_precision": "no_change",
20
+ "text_encoder_1_precision": "no_change",
21
+ "text_encoder_2_precision": "no_change",
22
+ "text_encoder_3_precision": "no_change",
23
+ "text_encoder_4_precision": "no_change",
24
+ "gradient_checkpointing_interval": null,
25
+ "offload_during_startup": false,
26
+ "quantize_via": "accelerator",
27
+ "wan_force_2_1_time_embedding": false,
28
+ "fuse_qkv_projections": false,
29
+ "rescale_betas_zero_snr": false,
30
+ "control": false,
31
+ "controlnet_custom_config": null,
32
+ "controlnet_model_name_or_path": null,
33
+ "tread_config": null,
34
+ "pretrained_transformer_model_name_or_path": null,
35
+ "pretrained_transformer_subfolder": "transformer",
36
+ "pretrained_unet_model_name_or_path": null,
37
+ "pretrained_unet_subfolder": "unet",
38
+ "pretrained_t5_model_name_or_path": null,
39
+ "revision": null,
40
+ "variant": null,
41
+ "base_model_default_dtype": "bf16",
42
+ "unet_attention_slice": false,
43
+ "num_train_epochs": 834,
44
+ "max_train_steps": 5000,
45
+ "train_batch_size": 2,
46
+ "learning_rate": 0.0001,
47
+ "optimizer": "adamw_bf16",
48
+ "lr_scheduler": "constant_with_warmup",
49
+ "gradient_accumulation_steps": 1,
50
+ "lr_warmup_steps": 0,
51
+ "checkpoints_total_limit": 5,
52
+ "gradient_checkpointing": true,
53
+ "enable_group_offload": false,
54
+ "ramtorch": false,
55
+ "ramtorch_target_modules": null,
56
+ "ramtorch_text_encoder": false,
57
+ "ramtorch_vae": false,
58
+ "ramtorch_controlnet": false,
59
+ "group_offload_type": "block_level",
60
+ "group_offload_blocks_per_group": 1,
61
+ "group_offload_use_stream": false,
62
+ "group_offload_to_disk_path": "",
63
+ "group_offload_text_encoder": false,
64
+ "group_offload_vae": false,
65
+ "offload_during_save": false,
66
+ "enable_chunked_feed_forward": false,
67
+ "feed_forward_chunk_size": null,
68
+ "train_text_encoder": false,
69
+ "text_encoder_lr": null,
70
+ "lyrics_embedder_train": false,
71
+ "lyrics_embedder_optimizer": null,
72
+ "lyrics_embedder_lr": null,
73
+ "lyrics_embedder_lr_scheduler": null,
74
+ "lr_num_cycles": 1,
75
+ "lr_power": 0.8,
76
+ "use_soft_min_snr": false,
77
+ "use_ema": false,
78
+ "ema_device": "cpu",
79
+ "ema_cpu_only": false,
80
+ "ema_update_interval": 1,
81
+ "ema_foreach_disable": false,
82
+ "ema_decay": 0.995,
83
+ "lora_rank": 16,
84
+ "lora_alpha": 16.0,
85
+ "lora_type": "standard",
86
+ "lora_dropout": 0.1,
87
+ "lora_init_type": "default",
88
+ "peft_lora_mode": "standard",
89
+ "singlora_ramp_up_steps": 0,
90
+ "init_lora": null,
91
+ "lycoris_config": "/home/Ubuntu/.simpletuner/configs/config/lycoris_config.json",
92
+ "init_lokr_norm": null,
93
+ "flux_lora_target": "all",
94
+ "acestep_lora_target": "attn_qkv+linear_qkv",
95
+ "use_dora": false,
96
+ "resolution_type": "pixel_area",
97
+ "data_backend_config": "/home/Ubuntu/.simpletuner/configs/111/multidatabackend-bartek_mizak.json",
98
+ "caption_strategy": "filename",
99
+ "conditioning_multidataset_sampling": "random",
100
+ "instance_prompt": null,
101
+ "parquet_caption_column": null,
102
+ "parquet_filename_column": null,
103
+ "ignore_missing_files": false,
104
+ "vae_cache_scan_behaviour": "recreate",
105
+ "vae_enable_slicing": false,
106
+ "vae_enable_tiling": false,
107
+ "vae_batch_size": 4,
108
+ "max_upscale_threshold": null,
109
+ "caption_dropout_probability": 0.1,
110
+ "tokenizer_max_length": 512,
111
+ "audio_max_duration_seconds": null,
112
+ "audio_min_duration_seconds": null,
113
+ "audio_channels": 1,
114
+ "audio_duration_interval": 3.0,
115
+ "audio_truncation_mode": "beginning",
116
+ "validation_step_interval": 100,
117
+ "validation_epoch_interval": null,
118
+ "disable_benchmark": false,
119
+ "validation_preview": false,
120
+ "validation_preview_steps": 1,
121
+ "validation_prompt": "bm82 man",
122
+ "validation_lyrics": null,
123
+ "validation_audio_duration": 30.0,
124
+ "num_validation_images": 1,
125
+ "num_eval_images": 4,
126
+ "eval_steps_interval": null,
127
+ "eval_timesteps": 28,
128
+ "eval_dataset_pooling": false,
129
+ "evaluation_type": "none",
130
+ "pretrained_evaluation_model_name_or_path": "openai/clip-vit-large-patch14-336",
131
+ "validation_guidance": 7.5,
132
+ "validation_num_inference_steps": 20,
133
+ "validation_on_startup": false,
134
+ "validation_method": "simpletuner-local",
135
+ "validation_external_script": null,
136
+ "validation_external_background": false,
137
+ "validation_using_datasets": false,
138
+ "validation_torch_compile": false,
139
+ "validation_guidance_real": 1.0,
140
+ "validation_no_cfg_until_timestep": 2,
141
+ "validation_negative_prompt": "blurry, cropped, ugly",
142
+ "validation_randomize": false,
143
+ "validation_seed": null,
144
+ "validation_multigpu": "batch-parallel",
145
+ "validation_disable": false,
146
+ "validation_prompt_library": false,
147
+ "user_prompt_library": null,
148
+ "eval_dataset_id": null,
149
+ "validation_stitch_input_location": "left",
150
+ "validation_guidance_rescale": 0.0,
151
+ "validation_disable_unconditional": false,
152
+ "validation_guidance_skip_layers": null,
153
+ "validation_guidance_skip_layers_start": 0.01,
154
+ "validation_guidance_skip_layers_stop": 0.2,
155
+ "validation_guidance_skip_scale": 2.8,
156
+ "validation_lycoris_strength": 1.0,
157
+ "validation_noise_scheduler": null,
158
+ "validation_num_video_frames": null,
159
+ "validation_resolution": "1024x1024",
160
+ "validation_seed_source": "cpu",
161
+ "validation_torch_compile_mode": "max-autotune",
162
+ "validation_adapter_path": null,
163
+ "validation_adapter_name": null,
164
+ "validation_adapter_strength": 1.0,
165
+ "validation_adapter_mode": "adapter_only",
166
+ "validation_adapter_config": null,
167
+ "i_know_what_i_am_doing": false,
168
+ "flow_sigmoid_scale": 1.0,
169
+ "flux_fast_schedule": false,
170
+ "flow_use_uniform_schedule": false,
171
+ "flow_use_beta_schedule": false,
172
+ "flow_beta_schedule_alpha": 2.0,
173
+ "flow_beta_schedule_beta": 2.0,
174
+ "flow_schedule_shift": 3.0,
175
+ "flow_schedule_auto_shift": false,
176
+ "flux_guidance_mode": "constant",
177
+ "flux_attention_masked_training": false,
178
+ "flux_guidance_value": 1.0,
179
+ "flux_guidance_min": 0.0,
180
+ "flux_guidance_max": 4.0,
181
+ "t5_padding": "unmodified",
182
+ "sd3_clip_uncond_behaviour": "empty_string",
183
+ "sd3_t5_uncond_behaviour": null,
184
+ "soft_min_snr_sigma_data": null,
185
+ "mixed_precision": "bf16",
186
+ "attention_mechanism": "diffusers",
187
+ "sla_config": null,
188
+ "sageattention_usage": {
189
+ "_value_": "inference",
190
+ "_name_": "INFERENCE",
191
+ "__objclass__": "{'_generate_next_value_': <staticmethod(<function Enum._generate_next_value_ at 0x7649ca9cdf80>)>, '__module__': 'simpletuner.helpers.training.attention_backend', 'allows_training': <property object at 0x764872f93880>, 'allows_inference': <property object at 0x764872f93790>, 'from_raw': <classmethod(<function AttentionBackendMode.from_raw at 0x764872fe85e0>)>, '_new_member_': <built-in method __new__ of type object at 0xa234e0>, '_use_args_': True, '_member_names_': ['TRAINING', 'INFERENCE', 'TRAINING_AND_INFERENCE'], '_member_map_': {'TRAINING': <AttentionBackendMode.TRAINING: 'training'>, 'INFERENCE': <AttentionBackendMode.INFERENCE: 'inference'>, 'TRAINING_AND_INFERENCE': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>}, '_value2member_map_': {'training': <AttentionBackendMode.TRAINING: 'training'>, 'inference': <AttentionBackendMode.INFERENCE: 'inference'>, 'training+inference': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>}, '_unhashable_values_': [], '_member_type_': <class 'str'>, '_value_repr_': <slot wrapper '__repr__' of 'str' objects>, '__dict__': <attribute '__dict__' of 'AttentionBackendMode' objects>, '__weakref__': <attribute '__weakref__' of 'AttentionBackendMode' objects>, '__doc__': None, 'TRAINING': <AttentionBackendMode.TRAINING: 'training'>, 'INFERENCE': <AttentionBackendMode.INFERENCE: 'inference'>, 'TRAINING_AND_INFERENCE': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>, '__repr__': <function Enum.__repr__ at 0x7649ca9ce0c0>, '__str__': <function Enum.__str__ at 0x7649ca9ce160>, '__format__': <function Enum.__format__ at 0x7649ca9ce2a0>, '__new__': <function Enum.__new__ at 0x7649ca9cde40>}",
192
+ "_sort_order_": 1
193
+ },
194
+ "disable_tf32": false,
195
+ "set_grads_to_none": false,
196
+ "noise_offset": 0.1,
197
+ "noise_offset_probability": 0.25,
198
+ "input_perturbation": 0.0,
199
+ "input_perturbation_steps": 0,
200
+ "lr_end": "4e-7",
201
+ "lr_scale": false,
202
+ "lr_scale_sqrt": false,
203
+ "ignore_final_epochs": false,
204
+ "freeze_encoder_before": 12,
205
+ "freeze_encoder_after": 17,
206
+ "freeze_encoder_strategy": "after",
207
+ "layer_freeze_strategy": null,
208
+ "fully_unload_text_encoder": false,
209
+ "save_text_encoder": false,
210
+ "text_encoder_limit": 100,
211
+ "prepend_instance_prompt": false,
212
+ "only_instance_prompt": false,
213
+ "data_aesthetic_score": 7.0,
214
+ "delete_unwanted_images": false,
215
+ "delete_problematic_images": false,
216
+ "disable_bucket_pruning": false,
217
+ "allow_dataset_oversubscription": false,
218
+ "disable_segmented_timestep_sampling": false,
219
+ "preserve_data_backend_cache": false,
220
+ "override_dataset_config": false,
221
+ "cache_dir": "/home/Ubuntu/.simpletuner/output/cache",
222
+ "cache_dir_text": "cache",
223
+ "cache_dir_vae": "",
224
+ "compress_disk_cache": true,
225
+ "aspect_bucket_disable_rebuild": false,
226
+ "keep_vae_loaded": false,
227
+ "skip_file_discovery": "",
228
+ "data_backend_sampling": "auto-weighting",
229
+ "image_processing_batch_size": 32,
230
+ "write_batch_size": 128,
231
+ "read_batch_size": 25,
232
+ "enable_multiprocessing": false,
233
+ "accelerate_config": null,
234
+ "deepspeed_config": null,
235
+ "fsdp_enable": false,
236
+ "fsdp_version": 2,
237
+ "fsdp_reshard_after_forward": false,
238
+ "fsdp_state_dict_type": "SHARDED_STATE_DICT",
239
+ "fsdp_cpu_ram_efficient_loading": false,
240
+ "fsdp_auto_wrap_policy": "TRANSFORMER_BASED_WRAP",
241
+ "fsdp_limit_all_gathers": false,
242
+ "fsdp_cpu_offload": false,
243
+ "fsdp_activation_checkpointing": false,
244
+ "fsdp_transformer_layer_cls_to_wrap": null,
245
+ "context_parallel_size": 1,
246
+ "context_parallel_comm_strategy": "allgather",
247
+ "num_processes": 1,
248
+ "num_machines": 1,
249
+ "accelerate_extra_args": null,
250
+ "main_process_ip": "127.0.0.1",
251
+ "main_process_port": 29500,
252
+ "machine_rank": 0,
253
+ "same_network": true,
254
+ "dynamo_backend": "no",
255
+ "dynamo_mode": "",
256
+ "dynamo_fullgraph": false,
257
+ "dynamo_dynamic": false,
258
+ "dynamo_use_regional_compilation": false,
259
+ "max_workers": 32,
260
+ "aws_max_pool_connections": 128,
261
+ "torch_num_threads": 8,
262
+ "dataloader_prefetch": false,
263
+ "dataloader_prefetch_qlen": 10,
264
+ "aspect_bucket_worker_count": 12,
265
+ "aspect_bucket_alignment": 16,
266
+ "minimum_image_size": null,
267
+ "maximum_image_size": null,
268
+ "target_downsample_size": null,
269
+ "metadata_update_interval": 3600,
270
+ "debug_aspect_buckets": false,
271
+ "debug_dataset_loader": false,
272
+ "print_filenames": false,
273
+ "print_sampler_statistics": false,
274
+ "timestep_bias_strategy": null,
275
+ "timestep_bias_begin": 0,
276
+ "timestep_bias_end": 1000,
277
+ "timestep_bias_multiplier": 1.0,
278
+ "timestep_bias_portion": 0.25,
279
+ "training_scheduler_timestep_spacing": "trailing",
280
+ "inference_scheduler_timestep_spacing": "trailing",
281
+ "loss_type": "l2",
282
+ "huber_schedule": "snr",
283
+ "huber_c": 0.1,
284
+ "snr_gamma": null,
285
+ "masked_loss_probability": 1.0,
286
+ "hidream_use_load_balancing_loss": false,
287
+ "hidream_load_balancing_loss_weight": null,
288
+ "adam_beta1": 0.9,
289
+ "adam_beta2": 0.999,
290
+ "optimizer_beta1": null,
291
+ "optimizer_beta2": null,
292
+ "optimizer_cpu_offload_method": null,
293
+ "gradient_precision": null,
294
+ "adam_weight_decay": 0.01,
295
+ "adam_epsilon": 1e-08,
296
+ "prodigy_steps": null,
297
+ "max_grad_norm": 2.0,
298
+ "optimizer_config": null,
299
+ "grad_clip_method": "value",
300
+ "optimizer_offload_gradients": false,
301
+ "fuse_optimizer": false,
302
+ "optimizer_release_gradients": false,
303
+ "push_to_hub": false,
304
+ "publishing_config": null,
305
+ "post_upload_script": null,
306
+ "push_checkpoints_to_hub": false,
307
+ "push_to_hub_background": false,
308
+ "hub_model_id": "quzo/fl2",
309
+ "model_card_private": false,
310
+ "model_card_safe_for_work": false,
311
+ "model_card_note": null,
312
+ "report_to": "none",
313
+ "custom_tracker": null,
314
+ "checkpoint_step_interval": 200,
315
+ "checkpoint_epoch_interval": null,
316
+ "checkpointing_rolling_steps": 0,
317
+ "checkpointing_use_tempdir": false,
318
+ "checkpoints_rolling_total_limit": 1,
319
+ "tracker_run_name": "simpletuner-testing",
320
+ "tracker_project_name": "simpletuner",
321
+ "tracker_image_layout": "gallery",
322
+ "enable_watermark": false,
323
+ "framerate": null,
324
+ "seed_for_each_device": true,
325
+ "snr_weight": 1.0,
326
+ "webhook_config": [
327
+ {
328
+ "webhook_type": "raw",
329
+ "callback_url": "http://0.0.0.0:8001/callback",
330
+ "log_level": "info",
331
+ "ssl_no_verify": false
332
+ }
333
+ ],
334
+ "webhook_reporting_interval": null,
335
+ "distillation_method": null,
336
+ "distillation_config": null,
337
+ "ema_validation": "comparison",
338
+ "local_rank": -1,
339
+ "ltx_train_mode": "i2v",
340
+ "ltx_i2v_prob": 0.1,
341
+ "ltx_partial_noise_fraction": 0.05,
342
+ "ltx_protect_first_frame": false,
343
+ "offload_param_path": null,
344
+ "offset_noise": false,
345
+ "quantize_activations": false,
346
+ "refiner_training": false,
347
+ "refiner_training_invert_schedule": false,
348
+ "refiner_training_strength": 0.2,
349
+ "sdxl_refiner_uses_full_range": false,
350
+ "sana_complex_human_instruction": "Given a user prompt, generate an 'Enhanced prompt' that provides detailed visual descriptions suitable for image generation. Evaluate the level of detail in the user prompt:\n- If the prompt is simple, focus on adding specifics about colors, shapes, sizes, textures, and spatial relationships to create vivid and concrete scenes.\n- Do not add detail that is not in the original prompt! Do not mention colors if the User Prompt does not!\n- If the prompt is already detailed, refine and enhance the existing details slightly without overcomplicating.\n- If the prompt contains explicit material, do not remove this information from the output, which would damage the performance of downstream classifiers.\nHere are examples of how to transform or refine prompts:\n- User Prompt: A cat sleeping -> Enhanced: A cat sleeping peacefully, showcasing the joy of pet ownership. Cute floof kitty cat gatto.\n- User Prompt: A busy city street -> Enhanced: A bustling city street scene featuring a crowd of people.\nPlease generate only the enhanced description for the prompt below and avoid including any additional commentary or evaluations:\nUser Prompt: ",
351
+ "vae_path": "black-forest-labs/FLUX.2-dev",
352
+ "accelerator_project_config": {
353
+ "project_dir": "/home/Ubuntu/.simpletuner/output",
354
+ "logging_dir": "/home/Ubuntu/.simpletuner/output/logs",
355
+ "automatic_checkpoint_naming": false,
356
+ "total_limit": null,
357
+ "iteration": 16,
358
+ "save_on_each_node": false
359
+ },
360
+ "process_group_kwargs": {
361
+ "backend": "nccl",
362
+ "init_method": null,
363
+ "timeout": "1:30:00"
364
+ },
365
+ "is_quantized": false,
366
+ "weight_dtype": "torch.bfloat16",
367
+ "disable_accelerator": false,
368
+ "lora_initialisation_style": true,
369
+ "checkpointing_steps": 200,
370
+ "use_fsdp": false,
371
+ "model_type_label": "Flux.2",
372
+ "use_deepspeed_optimizer": false,
373
+ "use_deepspeed_scheduler": false,
374
+ "base_weight_dtype": "torch.bfloat16",
375
+ "is_quanto": false,
376
+ "is_torchao": false,
377
+ "is_bnb": false,
378
+ "flow_matching": true,
379
+ "overrode_max_train_steps": false,
380
+ "total_num_batches": 6,
381
+ "num_update_steps_per_epoch": 6,
382
+ "total_batch_size": 2,
383
+ "enable_adamw_bf16": true,
384
+ "is_schedulefree": false,
385
+ "is_lr_scheduler_disabled": false,
386
+ "total_steps_remaining_at_start": 5000
387
+ }
checkpoint-3200/training_state-training-images.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"aspect_ratio_bucket_indices": {"1.0": ["/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im008.png", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im021.png"], "0.74": ["/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0014.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9743.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9827.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9935.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0003.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9796.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9890.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9865.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9260.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0090.jpg"]}, "buckets": ["1.0", "0.74"], "exhausted_buckets": [], "batch_size": 2, "current_bucket": 1, "seen_images": {"/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im008.png": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im021.png": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9743.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9796.jpg": true}, "current_epoch": 534}
checkpoint-3200/training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"global_step": 3200, "epoch_step": 3733, "epoch": 534, "exhausted_backends": [], "repeats": {"training-images": 0}}
checkpoint-3400/README.md ADDED
@@ -0,0 +1,129 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: "black-forest-labs/FLUX.2-dev"
4
+ tags:
5
+ - flux2
6
+ - flux2-diffusers
7
+ - text-to-image
8
+ - image-to-image
9
+ - diffusers
10
+ - simpletuner
11
+ - not-for-all-audiences
12
+ - lora
13
+
14
+ - template:sd-lora
15
+ - standard
16
+ pipeline_tag: text-to-image
17
+ inference: true
18
+
19
+ ---
20
+
21
+ # quzo/fl2
22
+
23
+ This is a PEFT LoRA derived from [black-forest-labs/FLUX.2-dev](https://huggingface.co/black-forest-labs/FLUX.2-dev).
24
+
25
+ The main validation prompt used during training was:
26
+ ```
27
+ bm82 man
28
+ ```
29
+
30
+
31
+ ## Validation settings
32
+ - CFG: `7.5`
33
+ - CFG Rescale: `0.0`
34
+ - Steps: `20`
35
+ - Sampler: `FlowMatchEulerDiscreteScheduler`
36
+ - Seed: `None`
37
+ - Resolution: `1024x1024`
38
+
39
+
40
+ Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
41
+
42
+
43
+
44
+
45
+ <Gallery />
46
+
47
+ The text encoder **was not** trained.
48
+ You may reuse the base model text encoder for inference.
49
+
50
+
51
+ ## Training settings
52
+
53
+ - Training epochs: 566
54
+ - Training steps: 3400
55
+ - Learning rate: 0.0001
56
+ - Learning rate schedule: constant_with_warmup
57
+ - Warmup steps: 0
58
+ - Max grad value: 2.0
59
+ - Effective batch size: 2
60
+ - Micro-batch size: 2
61
+ - Gradient accumulation steps: 1
62
+ - Number of GPUs: 1
63
+ - Gradient checkpointing: True
64
+ - Prediction type: flow_matching[]
65
+ - Optimizer: adamw_bf16
66
+ - Trainable parameter precision: Pure BF16
67
+ - Base model precision: `no_change`
68
+ - Caption dropout probability: 0.1%
69
+
70
+
71
+
72
+ - LoRA Rank: 16
73
+ - LoRA Alpha: 16.0
74
+ - LoRA Dropout: 0.1
75
+ - LoRA initialisation style: default
76
+ - LoRA mode: Standard
77
+
78
+
79
+ ## Datasets
80
+
81
+ ### training-images
82
+ - Repeats: 0
83
+ - Total number of images: 12
84
+ - Total number of aspect buckets: 1
85
+ - Resolution: 1.048576 megapixels
86
+ - Cropped: False
87
+ - Crop style: None
88
+ - Crop aspect: None
89
+ - Used for regularisation data: No
90
+
91
+
92
+ ## Inference
93
+
94
+
95
+ ```python
96
+ import torch
97
+ from diffusers import DiffusionPipeline
98
+
99
+ model_id = 'black-forest-labs/FLUX.2-dev'
100
+ adapter_id = 'quzo/fl2'
101
+ pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
102
+ pipeline.load_lora_weights(adapter_id)
103
+
104
+ prompt = "bm82 man"
105
+ negative_prompt = 'blurry, cropped, ugly'
106
+
107
+ ## Optional: quantise the model to save on vram.
108
+ ## Note: The model was not quantised during training, so it is not necessary to quantise it during inference time.
109
+ #from optimum.quanto import quantize, freeze, qint8
110
+ #quantize(pipeline.transformer, weights=qint8)
111
+ #freeze(pipeline.transformer)
112
+
113
+ pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
114
+ model_output = pipeline(
115
+ prompt=prompt,
116
+ negative_prompt=negative_prompt,
117
+ num_inference_steps=20,
118
+ generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
119
+ width=1024,
120
+ height=1024,
121
+ guidance_scale=7.5,
122
+ ).images[0]
123
+
124
+ model_output.save("output.png", format="PNG")
125
+
126
+ ```
127
+
128
+
129
+
checkpoint-3400/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a1cf4adf651ea45be3c0204a409ab22a3913476ee69840d1ffce3c8b45697c1
3
+ size 321008843
checkpoint-3400/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:414a5aff8b7277e413b9f104c7edd4bbcd12ab0d261beb22649b90a62d304fc4
3
+ size 106979136
checkpoint-3400/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01b906f76c1ba1117bc03af59e66750f2205f8a2f01c5e6240e20dc436024aae
3
+ size 14885
checkpoint-3400/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f02eb14703c391ff6b0683ceeea335d568ec18788de76c6f5aa899ec9eb225ae
3
+ size 1401
checkpoint-3400/simpletuner_config.json ADDED
@@ -0,0 +1,387 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_family": "flux2",
3
+ "model_flavour": "dev",
4
+ "controlnet": false,
5
+ "pretrained_model_name_or_path": "black-forest-labs/FLUX.2-dev",
6
+ "output_dir": "/home/Ubuntu/.simpletuner/output",
7
+ "logging_dir": "/home/Ubuntu/.simpletuner/output/logs",
8
+ "model_type": "lora",
9
+ "seed": null,
10
+ "resolution": 1024,
11
+ "resume_from_checkpoint": null,
12
+ "prediction_type": null,
13
+ "pretrained_vae_model_name_or_path": "black-forest-labs/FLUX.2-dev",
14
+ "vae_dtype": "bf16",
15
+ "vae_cache_ondemand": false,
16
+ "vae_cache_disable": false,
17
+ "accelerator_cache_clear_interval": null,
18
+ "aspect_bucket_rounding": 2,
19
+ "base_model_precision": "no_change",
20
+ "text_encoder_1_precision": "no_change",
21
+ "text_encoder_2_precision": "no_change",
22
+ "text_encoder_3_precision": "no_change",
23
+ "text_encoder_4_precision": "no_change",
24
+ "gradient_checkpointing_interval": null,
25
+ "offload_during_startup": false,
26
+ "quantize_via": "accelerator",
27
+ "wan_force_2_1_time_embedding": false,
28
+ "fuse_qkv_projections": false,
29
+ "rescale_betas_zero_snr": false,
30
+ "control": false,
31
+ "controlnet_custom_config": null,
32
+ "controlnet_model_name_or_path": null,
33
+ "tread_config": null,
34
+ "pretrained_transformer_model_name_or_path": null,
35
+ "pretrained_transformer_subfolder": "transformer",
36
+ "pretrained_unet_model_name_or_path": null,
37
+ "pretrained_unet_subfolder": "unet",
38
+ "pretrained_t5_model_name_or_path": null,
39
+ "revision": null,
40
+ "variant": null,
41
+ "base_model_default_dtype": "bf16",
42
+ "unet_attention_slice": false,
43
+ "num_train_epochs": 834,
44
+ "max_train_steps": 5000,
45
+ "train_batch_size": 2,
46
+ "learning_rate": 0.0001,
47
+ "optimizer": "adamw_bf16",
48
+ "lr_scheduler": "constant_with_warmup",
49
+ "gradient_accumulation_steps": 1,
50
+ "lr_warmup_steps": 0,
51
+ "checkpoints_total_limit": 5,
52
+ "gradient_checkpointing": true,
53
+ "enable_group_offload": false,
54
+ "ramtorch": false,
55
+ "ramtorch_target_modules": null,
56
+ "ramtorch_text_encoder": false,
57
+ "ramtorch_vae": false,
58
+ "ramtorch_controlnet": false,
59
+ "group_offload_type": "block_level",
60
+ "group_offload_blocks_per_group": 1,
61
+ "group_offload_use_stream": false,
62
+ "group_offload_to_disk_path": "",
63
+ "group_offload_text_encoder": false,
64
+ "group_offload_vae": false,
65
+ "offload_during_save": false,
66
+ "enable_chunked_feed_forward": false,
67
+ "feed_forward_chunk_size": null,
68
+ "train_text_encoder": false,
69
+ "text_encoder_lr": null,
70
+ "lyrics_embedder_train": false,
71
+ "lyrics_embedder_optimizer": null,
72
+ "lyrics_embedder_lr": null,
73
+ "lyrics_embedder_lr_scheduler": null,
74
+ "lr_num_cycles": 1,
75
+ "lr_power": 0.8,
76
+ "use_soft_min_snr": false,
77
+ "use_ema": false,
78
+ "ema_device": "cpu",
79
+ "ema_cpu_only": false,
80
+ "ema_update_interval": 1,
81
+ "ema_foreach_disable": false,
82
+ "ema_decay": 0.995,
83
+ "lora_rank": 16,
84
+ "lora_alpha": 16.0,
85
+ "lora_type": "standard",
86
+ "lora_dropout": 0.1,
87
+ "lora_init_type": "default",
88
+ "peft_lora_mode": "standard",
89
+ "singlora_ramp_up_steps": 0,
90
+ "init_lora": null,
91
+ "lycoris_config": "/home/Ubuntu/.simpletuner/configs/config/lycoris_config.json",
92
+ "init_lokr_norm": null,
93
+ "flux_lora_target": "all",
94
+ "acestep_lora_target": "attn_qkv+linear_qkv",
95
+ "use_dora": false,
96
+ "resolution_type": "pixel_area",
97
+ "data_backend_config": "/home/Ubuntu/.simpletuner/configs/111/multidatabackend-bartek_mizak.json",
98
+ "caption_strategy": "filename",
99
+ "conditioning_multidataset_sampling": "random",
100
+ "instance_prompt": null,
101
+ "parquet_caption_column": null,
102
+ "parquet_filename_column": null,
103
+ "ignore_missing_files": false,
104
+ "vae_cache_scan_behaviour": "recreate",
105
+ "vae_enable_slicing": false,
106
+ "vae_enable_tiling": false,
107
+ "vae_batch_size": 4,
108
+ "max_upscale_threshold": null,
109
+ "caption_dropout_probability": 0.1,
110
+ "tokenizer_max_length": 512,
111
+ "audio_max_duration_seconds": null,
112
+ "audio_min_duration_seconds": null,
113
+ "audio_channels": 1,
114
+ "audio_duration_interval": 3.0,
115
+ "audio_truncation_mode": "beginning",
116
+ "validation_step_interval": 100,
117
+ "validation_epoch_interval": null,
118
+ "disable_benchmark": false,
119
+ "validation_preview": false,
120
+ "validation_preview_steps": 1,
121
+ "validation_prompt": "bm82 man",
122
+ "validation_lyrics": null,
123
+ "validation_audio_duration": 30.0,
124
+ "num_validation_images": 1,
125
+ "num_eval_images": 4,
126
+ "eval_steps_interval": null,
127
+ "eval_timesteps": 28,
128
+ "eval_dataset_pooling": false,
129
+ "evaluation_type": "none",
130
+ "pretrained_evaluation_model_name_or_path": "openai/clip-vit-large-patch14-336",
131
+ "validation_guidance": 7.5,
132
+ "validation_num_inference_steps": 20,
133
+ "validation_on_startup": false,
134
+ "validation_method": "simpletuner-local",
135
+ "validation_external_script": null,
136
+ "validation_external_background": false,
137
+ "validation_using_datasets": false,
138
+ "validation_torch_compile": false,
139
+ "validation_guidance_real": 1.0,
140
+ "validation_no_cfg_until_timestep": 2,
141
+ "validation_negative_prompt": "blurry, cropped, ugly",
142
+ "validation_randomize": false,
143
+ "validation_seed": null,
144
+ "validation_multigpu": "batch-parallel",
145
+ "validation_disable": false,
146
+ "validation_prompt_library": false,
147
+ "user_prompt_library": null,
148
+ "eval_dataset_id": null,
149
+ "validation_stitch_input_location": "left",
150
+ "validation_guidance_rescale": 0.0,
151
+ "validation_disable_unconditional": false,
152
+ "validation_guidance_skip_layers": null,
153
+ "validation_guidance_skip_layers_start": 0.01,
154
+ "validation_guidance_skip_layers_stop": 0.2,
155
+ "validation_guidance_skip_scale": 2.8,
156
+ "validation_lycoris_strength": 1.0,
157
+ "validation_noise_scheduler": null,
158
+ "validation_num_video_frames": null,
159
+ "validation_resolution": "1024x1024",
160
+ "validation_seed_source": "cpu",
161
+ "validation_torch_compile_mode": "max-autotune",
162
+ "validation_adapter_path": null,
163
+ "validation_adapter_name": null,
164
+ "validation_adapter_strength": 1.0,
165
+ "validation_adapter_mode": "adapter_only",
166
+ "validation_adapter_config": null,
167
+ "i_know_what_i_am_doing": false,
168
+ "flow_sigmoid_scale": 1.0,
169
+ "flux_fast_schedule": false,
170
+ "flow_use_uniform_schedule": false,
171
+ "flow_use_beta_schedule": false,
172
+ "flow_beta_schedule_alpha": 2.0,
173
+ "flow_beta_schedule_beta": 2.0,
174
+ "flow_schedule_shift": 3.0,
175
+ "flow_schedule_auto_shift": false,
176
+ "flux_guidance_mode": "constant",
177
+ "flux_attention_masked_training": false,
178
+ "flux_guidance_value": 1.0,
179
+ "flux_guidance_min": 0.0,
180
+ "flux_guidance_max": 4.0,
181
+ "t5_padding": "unmodified",
182
+ "sd3_clip_uncond_behaviour": "empty_string",
183
+ "sd3_t5_uncond_behaviour": null,
184
+ "soft_min_snr_sigma_data": null,
185
+ "mixed_precision": "bf16",
186
+ "attention_mechanism": "diffusers",
187
+ "sla_config": null,
188
+ "sageattention_usage": {
189
+ "_value_": "inference",
190
+ "_name_": "INFERENCE",
191
+ "__objclass__": "{'_generate_next_value_': <staticmethod(<function Enum._generate_next_value_ at 0x7649ca9cdf80>)>, '__module__': 'simpletuner.helpers.training.attention_backend', 'allows_training': <property object at 0x764872f93880>, 'allows_inference': <property object at 0x764872f93790>, 'from_raw': <classmethod(<function AttentionBackendMode.from_raw at 0x764872fe85e0>)>, '_new_member_': <built-in method __new__ of type object at 0xa234e0>, '_use_args_': True, '_member_names_': ['TRAINING', 'INFERENCE', 'TRAINING_AND_INFERENCE'], '_member_map_': {'TRAINING': <AttentionBackendMode.TRAINING: 'training'>, 'INFERENCE': <AttentionBackendMode.INFERENCE: 'inference'>, 'TRAINING_AND_INFERENCE': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>}, '_value2member_map_': {'training': <AttentionBackendMode.TRAINING: 'training'>, 'inference': <AttentionBackendMode.INFERENCE: 'inference'>, 'training+inference': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>}, '_unhashable_values_': [], '_member_type_': <class 'str'>, '_value_repr_': <slot wrapper '__repr__' of 'str' objects>, '__dict__': <attribute '__dict__' of 'AttentionBackendMode' objects>, '__weakref__': <attribute '__weakref__' of 'AttentionBackendMode' objects>, '__doc__': None, 'TRAINING': <AttentionBackendMode.TRAINING: 'training'>, 'INFERENCE': <AttentionBackendMode.INFERENCE: 'inference'>, 'TRAINING_AND_INFERENCE': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>, '__repr__': <function Enum.__repr__ at 0x7649ca9ce0c0>, '__str__': <function Enum.__str__ at 0x7649ca9ce160>, '__format__': <function Enum.__format__ at 0x7649ca9ce2a0>, '__new__': <function Enum.__new__ at 0x7649ca9cde40>}",
192
+ "_sort_order_": 1
193
+ },
194
+ "disable_tf32": false,
195
+ "set_grads_to_none": false,
196
+ "noise_offset": 0.1,
197
+ "noise_offset_probability": 0.25,
198
+ "input_perturbation": 0.0,
199
+ "input_perturbation_steps": 0,
200
+ "lr_end": "4e-7",
201
+ "lr_scale": false,
202
+ "lr_scale_sqrt": false,
203
+ "ignore_final_epochs": false,
204
+ "freeze_encoder_before": 12,
205
+ "freeze_encoder_after": 17,
206
+ "freeze_encoder_strategy": "after",
207
+ "layer_freeze_strategy": null,
208
+ "fully_unload_text_encoder": false,
209
+ "save_text_encoder": false,
210
+ "text_encoder_limit": 100,
211
+ "prepend_instance_prompt": false,
212
+ "only_instance_prompt": false,
213
+ "data_aesthetic_score": 7.0,
214
+ "delete_unwanted_images": false,
215
+ "delete_problematic_images": false,
216
+ "disable_bucket_pruning": false,
217
+ "allow_dataset_oversubscription": false,
218
+ "disable_segmented_timestep_sampling": false,
219
+ "preserve_data_backend_cache": false,
220
+ "override_dataset_config": false,
221
+ "cache_dir": "/home/Ubuntu/.simpletuner/output/cache",
222
+ "cache_dir_text": "cache",
223
+ "cache_dir_vae": "",
224
+ "compress_disk_cache": true,
225
+ "aspect_bucket_disable_rebuild": false,
226
+ "keep_vae_loaded": false,
227
+ "skip_file_discovery": "",
228
+ "data_backend_sampling": "auto-weighting",
229
+ "image_processing_batch_size": 32,
230
+ "write_batch_size": 128,
231
+ "read_batch_size": 25,
232
+ "enable_multiprocessing": false,
233
+ "accelerate_config": null,
234
+ "deepspeed_config": null,
235
+ "fsdp_enable": false,
236
+ "fsdp_version": 2,
237
+ "fsdp_reshard_after_forward": false,
238
+ "fsdp_state_dict_type": "SHARDED_STATE_DICT",
239
+ "fsdp_cpu_ram_efficient_loading": false,
240
+ "fsdp_auto_wrap_policy": "TRANSFORMER_BASED_WRAP",
241
+ "fsdp_limit_all_gathers": false,
242
+ "fsdp_cpu_offload": false,
243
+ "fsdp_activation_checkpointing": false,
244
+ "fsdp_transformer_layer_cls_to_wrap": null,
245
+ "context_parallel_size": 1,
246
+ "context_parallel_comm_strategy": "allgather",
247
+ "num_processes": 1,
248
+ "num_machines": 1,
249
+ "accelerate_extra_args": null,
250
+ "main_process_ip": "127.0.0.1",
251
+ "main_process_port": 29500,
252
+ "machine_rank": 0,
253
+ "same_network": true,
254
+ "dynamo_backend": "no",
255
+ "dynamo_mode": "",
256
+ "dynamo_fullgraph": false,
257
+ "dynamo_dynamic": false,
258
+ "dynamo_use_regional_compilation": false,
259
+ "max_workers": 32,
260
+ "aws_max_pool_connections": 128,
261
+ "torch_num_threads": 8,
262
+ "dataloader_prefetch": false,
263
+ "dataloader_prefetch_qlen": 10,
264
+ "aspect_bucket_worker_count": 12,
265
+ "aspect_bucket_alignment": 16,
266
+ "minimum_image_size": null,
267
+ "maximum_image_size": null,
268
+ "target_downsample_size": null,
269
+ "metadata_update_interval": 3600,
270
+ "debug_aspect_buckets": false,
271
+ "debug_dataset_loader": false,
272
+ "print_filenames": false,
273
+ "print_sampler_statistics": false,
274
+ "timestep_bias_strategy": null,
275
+ "timestep_bias_begin": 0,
276
+ "timestep_bias_end": 1000,
277
+ "timestep_bias_multiplier": 1.0,
278
+ "timestep_bias_portion": 0.25,
279
+ "training_scheduler_timestep_spacing": "trailing",
280
+ "inference_scheduler_timestep_spacing": "trailing",
281
+ "loss_type": "l2",
282
+ "huber_schedule": "snr",
283
+ "huber_c": 0.1,
284
+ "snr_gamma": null,
285
+ "masked_loss_probability": 1.0,
286
+ "hidream_use_load_balancing_loss": false,
287
+ "hidream_load_balancing_loss_weight": null,
288
+ "adam_beta1": 0.9,
289
+ "adam_beta2": 0.999,
290
+ "optimizer_beta1": null,
291
+ "optimizer_beta2": null,
292
+ "optimizer_cpu_offload_method": null,
293
+ "gradient_precision": null,
294
+ "adam_weight_decay": 0.01,
295
+ "adam_epsilon": 1e-08,
296
+ "prodigy_steps": null,
297
+ "max_grad_norm": 2.0,
298
+ "optimizer_config": null,
299
+ "grad_clip_method": "value",
300
+ "optimizer_offload_gradients": false,
301
+ "fuse_optimizer": false,
302
+ "optimizer_release_gradients": false,
303
+ "push_to_hub": false,
304
+ "publishing_config": null,
305
+ "post_upload_script": null,
306
+ "push_checkpoints_to_hub": false,
307
+ "push_to_hub_background": false,
308
+ "hub_model_id": "quzo/fl2",
309
+ "model_card_private": false,
310
+ "model_card_safe_for_work": false,
311
+ "model_card_note": null,
312
+ "report_to": "none",
313
+ "custom_tracker": null,
314
+ "checkpoint_step_interval": 200,
315
+ "checkpoint_epoch_interval": null,
316
+ "checkpointing_rolling_steps": 0,
317
+ "checkpointing_use_tempdir": false,
318
+ "checkpoints_rolling_total_limit": 1,
319
+ "tracker_run_name": "simpletuner-testing",
320
+ "tracker_project_name": "simpletuner",
321
+ "tracker_image_layout": "gallery",
322
+ "enable_watermark": false,
323
+ "framerate": null,
324
+ "seed_for_each_device": true,
325
+ "snr_weight": 1.0,
326
+ "webhook_config": [
327
+ {
328
+ "webhook_type": "raw",
329
+ "callback_url": "http://0.0.0.0:8001/callback",
330
+ "log_level": "info",
331
+ "ssl_no_verify": false
332
+ }
333
+ ],
334
+ "webhook_reporting_interval": null,
335
+ "distillation_method": null,
336
+ "distillation_config": null,
337
+ "ema_validation": "comparison",
338
+ "local_rank": -1,
339
+ "ltx_train_mode": "i2v",
340
+ "ltx_i2v_prob": 0.1,
341
+ "ltx_partial_noise_fraction": 0.05,
342
+ "ltx_protect_first_frame": false,
343
+ "offload_param_path": null,
344
+ "offset_noise": false,
345
+ "quantize_activations": false,
346
+ "refiner_training": false,
347
+ "refiner_training_invert_schedule": false,
348
+ "refiner_training_strength": 0.2,
349
+ "sdxl_refiner_uses_full_range": false,
350
+ "sana_complex_human_instruction": "Given a user prompt, generate an 'Enhanced prompt' that provides detailed visual descriptions suitable for image generation. Evaluate the level of detail in the user prompt:\n- If the prompt is simple, focus on adding specifics about colors, shapes, sizes, textures, and spatial relationships to create vivid and concrete scenes.\n- Do not add detail that is not in the original prompt! Do not mention colors if the User Prompt does not!\n- If the prompt is already detailed, refine and enhance the existing details slightly without overcomplicating.\n- If the prompt contains explicit material, do not remove this information from the output, which would damage the performance of downstream classifiers.\nHere are examples of how to transform or refine prompts:\n- User Prompt: A cat sleeping -> Enhanced: A cat sleeping peacefully, showcasing the joy of pet ownership. Cute floof kitty cat gatto.\n- User Prompt: A busy city street -> Enhanced: A bustling city street scene featuring a crowd of people.\nPlease generate only the enhanced description for the prompt below and avoid including any additional commentary or evaluations:\nUser Prompt: ",
351
+ "vae_path": "black-forest-labs/FLUX.2-dev",
352
+ "accelerator_project_config": {
353
+ "project_dir": "/home/Ubuntu/.simpletuner/output",
354
+ "logging_dir": "/home/Ubuntu/.simpletuner/output/logs",
355
+ "automatic_checkpoint_naming": false,
356
+ "total_limit": null,
357
+ "iteration": 17,
358
+ "save_on_each_node": false
359
+ },
360
+ "process_group_kwargs": {
361
+ "backend": "nccl",
362
+ "init_method": null,
363
+ "timeout": "1:30:00"
364
+ },
365
+ "is_quantized": false,
366
+ "weight_dtype": "torch.bfloat16",
367
+ "disable_accelerator": false,
368
+ "lora_initialisation_style": true,
369
+ "checkpointing_steps": 200,
370
+ "use_fsdp": false,
371
+ "model_type_label": "Flux.2",
372
+ "use_deepspeed_optimizer": false,
373
+ "use_deepspeed_scheduler": false,
374
+ "base_weight_dtype": "torch.bfloat16",
375
+ "is_quanto": false,
376
+ "is_torchao": false,
377
+ "is_bnb": false,
378
+ "flow_matching": true,
379
+ "overrode_max_train_steps": false,
380
+ "total_num_batches": 6,
381
+ "num_update_steps_per_epoch": 6,
382
+ "total_batch_size": 2,
383
+ "enable_adamw_bf16": true,
384
+ "is_schedulefree": false,
385
+ "is_lr_scheduler_disabled": false,
386
+ "total_steps_remaining_at_start": 5000
387
+ }
checkpoint-3400/training_state-training-images.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"aspect_ratio_bucket_indices": {"1.0": ["/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im008.png", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im021.png"], "0.74": ["/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0014.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9743.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9827.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9935.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0003.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9796.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9890.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9865.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9260.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0090.jpg"]}, "buckets": ["0.74"], "exhausted_buckets": ["1.0"], "batch_size": 2, "current_bucket": 0, "seen_images": {"/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im021.png": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im008.png": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9935.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9890.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0003.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9743.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9827.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9260.jpg": true}, "current_epoch": 567}
checkpoint-3400/training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"global_step": 3400, "epoch_step": 3966, "epoch": 567, "exhausted_backends": [], "repeats": {"training-images": 0}}
checkpoint-3600/README.md ADDED
@@ -0,0 +1,129 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: "black-forest-labs/FLUX.2-dev"
4
+ tags:
5
+ - flux2
6
+ - flux2-diffusers
7
+ - text-to-image
8
+ - image-to-image
9
+ - diffusers
10
+ - simpletuner
11
+ - not-for-all-audiences
12
+ - lora
13
+
14
+ - template:sd-lora
15
+ - standard
16
+ pipeline_tag: text-to-image
17
+ inference: true
18
+
19
+ ---
20
+
21
+ # quzo/fl2
22
+
23
+ This is a PEFT LoRA derived from [black-forest-labs/FLUX.2-dev](https://huggingface.co/black-forest-labs/FLUX.2-dev).
24
+
25
+ The main validation prompt used during training was:
26
+ ```
27
+ bm82 man
28
+ ```
29
+
30
+
31
+ ## Validation settings
32
+ - CFG: `7.5`
33
+ - CFG Rescale: `0.0`
34
+ - Steps: `20`
35
+ - Sampler: `FlowMatchEulerDiscreteScheduler`
36
+ - Seed: `None`
37
+ - Resolution: `1024x1024`
38
+
39
+
40
+ Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
41
+
42
+
43
+
44
+
45
+ <Gallery />
46
+
47
+ The text encoder **was not** trained.
48
+ You may reuse the base model text encoder for inference.
49
+
50
+
51
+ ## Training settings
52
+
53
+ - Training epochs: 599
54
+ - Training steps: 3600
55
+ - Learning rate: 0.0001
56
+ - Learning rate schedule: constant_with_warmup
57
+ - Warmup steps: 0
58
+ - Max grad value: 2.0
59
+ - Effective batch size: 2
60
+ - Micro-batch size: 2
61
+ - Gradient accumulation steps: 1
62
+ - Number of GPUs: 1
63
+ - Gradient checkpointing: True
64
+ - Prediction type: flow_matching[]
65
+ - Optimizer: adamw_bf16
66
+ - Trainable parameter precision: Pure BF16
67
+ - Base model precision: `no_change`
68
+ - Caption dropout probability: 0.1%
69
+
70
+
71
+
72
+ - LoRA Rank: 16
73
+ - LoRA Alpha: 16.0
74
+ - LoRA Dropout: 0.1
75
+ - LoRA initialisation style: default
76
+ - LoRA mode: Standard
77
+
78
+
79
+ ## Datasets
80
+
81
+ ### training-images
82
+ - Repeats: 0
83
+ - Total number of images: 12
84
+ - Total number of aspect buckets: 1
85
+ - Resolution: 1.048576 megapixels
86
+ - Cropped: False
87
+ - Crop style: None
88
+ - Crop aspect: None
89
+ - Used for regularisation data: No
90
+
91
+
92
+ ## Inference
93
+
94
+
95
+ ```python
96
+ import torch
97
+ from diffusers import DiffusionPipeline
98
+
99
+ model_id = 'black-forest-labs/FLUX.2-dev'
100
+ adapter_id = 'quzo/fl2'
101
+ pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
102
+ pipeline.load_lora_weights(adapter_id)
103
+
104
+ prompt = "bm82 man"
105
+ negative_prompt = 'blurry, cropped, ugly'
106
+
107
+ ## Optional: quantise the model to save on vram.
108
+ ## Note: The model was not quantised during training, so it is not necessary to quantise it during inference time.
109
+ #from optimum.quanto import quantize, freeze, qint8
110
+ #quantize(pipeline.transformer, weights=qint8)
111
+ #freeze(pipeline.transformer)
112
+
113
+ pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
114
+ model_output = pipeline(
115
+ prompt=prompt,
116
+ negative_prompt=negative_prompt,
117
+ num_inference_steps=20,
118
+ generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
119
+ width=1024,
120
+ height=1024,
121
+ guidance_scale=7.5,
122
+ ).images[0]
123
+
124
+ model_output.save("output.png", format="PNG")
125
+
126
+ ```
127
+
128
+
129
+
checkpoint-3600/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f020ed312a541740949018b668e757324f91d58a2c4e8cf7be9b56b4b7ce1cb9
3
+ size 321008843
checkpoint-3600/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5daaa3075a66bf8aad8942dfd83df8683047f13a9fbdd8cbd35e3090e8789f7
3
+ size 106979136
checkpoint-3600/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2890a0caf51b59c66e1dac8a4761f5003fbe092a9bd8d548920ac59abfae98c5
3
+ size 14693
checkpoint-3600/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6322143ad2efc56c6c09c3066b83311a4b8987351a4fe633c0f1c92f289e9d4f
3
+ size 1401
checkpoint-3600/simpletuner_config.json ADDED
@@ -0,0 +1,387 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_family": "flux2",
3
+ "model_flavour": "dev",
4
+ "controlnet": false,
5
+ "pretrained_model_name_or_path": "black-forest-labs/FLUX.2-dev",
6
+ "output_dir": "/home/Ubuntu/.simpletuner/output",
7
+ "logging_dir": "/home/Ubuntu/.simpletuner/output/logs",
8
+ "model_type": "lora",
9
+ "seed": null,
10
+ "resolution": 1024,
11
+ "resume_from_checkpoint": null,
12
+ "prediction_type": null,
13
+ "pretrained_vae_model_name_or_path": "black-forest-labs/FLUX.2-dev",
14
+ "vae_dtype": "bf16",
15
+ "vae_cache_ondemand": false,
16
+ "vae_cache_disable": false,
17
+ "accelerator_cache_clear_interval": null,
18
+ "aspect_bucket_rounding": 2,
19
+ "base_model_precision": "no_change",
20
+ "text_encoder_1_precision": "no_change",
21
+ "text_encoder_2_precision": "no_change",
22
+ "text_encoder_3_precision": "no_change",
23
+ "text_encoder_4_precision": "no_change",
24
+ "gradient_checkpointing_interval": null,
25
+ "offload_during_startup": false,
26
+ "quantize_via": "accelerator",
27
+ "wan_force_2_1_time_embedding": false,
28
+ "fuse_qkv_projections": false,
29
+ "rescale_betas_zero_snr": false,
30
+ "control": false,
31
+ "controlnet_custom_config": null,
32
+ "controlnet_model_name_or_path": null,
33
+ "tread_config": null,
34
+ "pretrained_transformer_model_name_or_path": null,
35
+ "pretrained_transformer_subfolder": "transformer",
36
+ "pretrained_unet_model_name_or_path": null,
37
+ "pretrained_unet_subfolder": "unet",
38
+ "pretrained_t5_model_name_or_path": null,
39
+ "revision": null,
40
+ "variant": null,
41
+ "base_model_default_dtype": "bf16",
42
+ "unet_attention_slice": false,
43
+ "num_train_epochs": 834,
44
+ "max_train_steps": 5000,
45
+ "train_batch_size": 2,
46
+ "learning_rate": 0.0001,
47
+ "optimizer": "adamw_bf16",
48
+ "lr_scheduler": "constant_with_warmup",
49
+ "gradient_accumulation_steps": 1,
50
+ "lr_warmup_steps": 0,
51
+ "checkpoints_total_limit": 5,
52
+ "gradient_checkpointing": true,
53
+ "enable_group_offload": false,
54
+ "ramtorch": false,
55
+ "ramtorch_target_modules": null,
56
+ "ramtorch_text_encoder": false,
57
+ "ramtorch_vae": false,
58
+ "ramtorch_controlnet": false,
59
+ "group_offload_type": "block_level",
60
+ "group_offload_blocks_per_group": 1,
61
+ "group_offload_use_stream": false,
62
+ "group_offload_to_disk_path": "",
63
+ "group_offload_text_encoder": false,
64
+ "group_offload_vae": false,
65
+ "offload_during_save": false,
66
+ "enable_chunked_feed_forward": false,
67
+ "feed_forward_chunk_size": null,
68
+ "train_text_encoder": false,
69
+ "text_encoder_lr": null,
70
+ "lyrics_embedder_train": false,
71
+ "lyrics_embedder_optimizer": null,
72
+ "lyrics_embedder_lr": null,
73
+ "lyrics_embedder_lr_scheduler": null,
74
+ "lr_num_cycles": 1,
75
+ "lr_power": 0.8,
76
+ "use_soft_min_snr": false,
77
+ "use_ema": false,
78
+ "ema_device": "cpu",
79
+ "ema_cpu_only": false,
80
+ "ema_update_interval": 1,
81
+ "ema_foreach_disable": false,
82
+ "ema_decay": 0.995,
83
+ "lora_rank": 16,
84
+ "lora_alpha": 16.0,
85
+ "lora_type": "standard",
86
+ "lora_dropout": 0.1,
87
+ "lora_init_type": "default",
88
+ "peft_lora_mode": "standard",
89
+ "singlora_ramp_up_steps": 0,
90
+ "init_lora": null,
91
+ "lycoris_config": "/home/Ubuntu/.simpletuner/configs/config/lycoris_config.json",
92
+ "init_lokr_norm": null,
93
+ "flux_lora_target": "all",
94
+ "acestep_lora_target": "attn_qkv+linear_qkv",
95
+ "use_dora": false,
96
+ "resolution_type": "pixel_area",
97
+ "data_backend_config": "/home/Ubuntu/.simpletuner/configs/111/multidatabackend-bartek_mizak.json",
98
+ "caption_strategy": "filename",
99
+ "conditioning_multidataset_sampling": "random",
100
+ "instance_prompt": null,
101
+ "parquet_caption_column": null,
102
+ "parquet_filename_column": null,
103
+ "ignore_missing_files": false,
104
+ "vae_cache_scan_behaviour": "recreate",
105
+ "vae_enable_slicing": false,
106
+ "vae_enable_tiling": false,
107
+ "vae_batch_size": 4,
108
+ "max_upscale_threshold": null,
109
+ "caption_dropout_probability": 0.1,
110
+ "tokenizer_max_length": 512,
111
+ "audio_max_duration_seconds": null,
112
+ "audio_min_duration_seconds": null,
113
+ "audio_channels": 1,
114
+ "audio_duration_interval": 3.0,
115
+ "audio_truncation_mode": "beginning",
116
+ "validation_step_interval": 100,
117
+ "validation_epoch_interval": null,
118
+ "disable_benchmark": false,
119
+ "validation_preview": false,
120
+ "validation_preview_steps": 1,
121
+ "validation_prompt": "bm82 man",
122
+ "validation_lyrics": null,
123
+ "validation_audio_duration": 30.0,
124
+ "num_validation_images": 1,
125
+ "num_eval_images": 4,
126
+ "eval_steps_interval": null,
127
+ "eval_timesteps": 28,
128
+ "eval_dataset_pooling": false,
129
+ "evaluation_type": "none",
130
+ "pretrained_evaluation_model_name_or_path": "openai/clip-vit-large-patch14-336",
131
+ "validation_guidance": 7.5,
132
+ "validation_num_inference_steps": 20,
133
+ "validation_on_startup": false,
134
+ "validation_method": "simpletuner-local",
135
+ "validation_external_script": null,
136
+ "validation_external_background": false,
137
+ "validation_using_datasets": false,
138
+ "validation_torch_compile": false,
139
+ "validation_guidance_real": 1.0,
140
+ "validation_no_cfg_until_timestep": 2,
141
+ "validation_negative_prompt": "blurry, cropped, ugly",
142
+ "validation_randomize": false,
143
+ "validation_seed": null,
144
+ "validation_multigpu": "batch-parallel",
145
+ "validation_disable": false,
146
+ "validation_prompt_library": false,
147
+ "user_prompt_library": null,
148
+ "eval_dataset_id": null,
149
+ "validation_stitch_input_location": "left",
150
+ "validation_guidance_rescale": 0.0,
151
+ "validation_disable_unconditional": false,
152
+ "validation_guidance_skip_layers": null,
153
+ "validation_guidance_skip_layers_start": 0.01,
154
+ "validation_guidance_skip_layers_stop": 0.2,
155
+ "validation_guidance_skip_scale": 2.8,
156
+ "validation_lycoris_strength": 1.0,
157
+ "validation_noise_scheduler": null,
158
+ "validation_num_video_frames": null,
159
+ "validation_resolution": "1024x1024",
160
+ "validation_seed_source": "cpu",
161
+ "validation_torch_compile_mode": "max-autotune",
162
+ "validation_adapter_path": null,
163
+ "validation_adapter_name": null,
164
+ "validation_adapter_strength": 1.0,
165
+ "validation_adapter_mode": "adapter_only",
166
+ "validation_adapter_config": null,
167
+ "i_know_what_i_am_doing": false,
168
+ "flow_sigmoid_scale": 1.0,
169
+ "flux_fast_schedule": false,
170
+ "flow_use_uniform_schedule": false,
171
+ "flow_use_beta_schedule": false,
172
+ "flow_beta_schedule_alpha": 2.0,
173
+ "flow_beta_schedule_beta": 2.0,
174
+ "flow_schedule_shift": 3.0,
175
+ "flow_schedule_auto_shift": false,
176
+ "flux_guidance_mode": "constant",
177
+ "flux_attention_masked_training": false,
178
+ "flux_guidance_value": 1.0,
179
+ "flux_guidance_min": 0.0,
180
+ "flux_guidance_max": 4.0,
181
+ "t5_padding": "unmodified",
182
+ "sd3_clip_uncond_behaviour": "empty_string",
183
+ "sd3_t5_uncond_behaviour": null,
184
+ "soft_min_snr_sigma_data": null,
185
+ "mixed_precision": "bf16",
186
+ "attention_mechanism": "diffusers",
187
+ "sla_config": null,
188
+ "sageattention_usage": {
189
+ "_value_": "inference",
190
+ "_name_": "INFERENCE",
191
+ "__objclass__": "{'_generate_next_value_': <staticmethod(<function Enum._generate_next_value_ at 0x7649ca9cdf80>)>, '__module__': 'simpletuner.helpers.training.attention_backend', 'allows_training': <property object at 0x764872f93880>, 'allows_inference': <property object at 0x764872f93790>, 'from_raw': <classmethod(<function AttentionBackendMode.from_raw at 0x764872fe85e0>)>, '_new_member_': <built-in method __new__ of type object at 0xa234e0>, '_use_args_': True, '_member_names_': ['TRAINING', 'INFERENCE', 'TRAINING_AND_INFERENCE'], '_member_map_': {'TRAINING': <AttentionBackendMode.TRAINING: 'training'>, 'INFERENCE': <AttentionBackendMode.INFERENCE: 'inference'>, 'TRAINING_AND_INFERENCE': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>}, '_value2member_map_': {'training': <AttentionBackendMode.TRAINING: 'training'>, 'inference': <AttentionBackendMode.INFERENCE: 'inference'>, 'training+inference': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>}, '_unhashable_values_': [], '_member_type_': <class 'str'>, '_value_repr_': <slot wrapper '__repr__' of 'str' objects>, '__dict__': <attribute '__dict__' of 'AttentionBackendMode' objects>, '__weakref__': <attribute '__weakref__' of 'AttentionBackendMode' objects>, '__doc__': None, 'TRAINING': <AttentionBackendMode.TRAINING: 'training'>, 'INFERENCE': <AttentionBackendMode.INFERENCE: 'inference'>, 'TRAINING_AND_INFERENCE': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>, '__repr__': <function Enum.__repr__ at 0x7649ca9ce0c0>, '__str__': <function Enum.__str__ at 0x7649ca9ce160>, '__format__': <function Enum.__format__ at 0x7649ca9ce2a0>, '__new__': <function Enum.__new__ at 0x7649ca9cde40>}",
192
+ "_sort_order_": 1
193
+ },
194
+ "disable_tf32": false,
195
+ "set_grads_to_none": false,
196
+ "noise_offset": 0.1,
197
+ "noise_offset_probability": 0.25,
198
+ "input_perturbation": 0.0,
199
+ "input_perturbation_steps": 0,
200
+ "lr_end": "4e-7",
201
+ "lr_scale": false,
202
+ "lr_scale_sqrt": false,
203
+ "ignore_final_epochs": false,
204
+ "freeze_encoder_before": 12,
205
+ "freeze_encoder_after": 17,
206
+ "freeze_encoder_strategy": "after",
207
+ "layer_freeze_strategy": null,
208
+ "fully_unload_text_encoder": false,
209
+ "save_text_encoder": false,
210
+ "text_encoder_limit": 100,
211
+ "prepend_instance_prompt": false,
212
+ "only_instance_prompt": false,
213
+ "data_aesthetic_score": 7.0,
214
+ "delete_unwanted_images": false,
215
+ "delete_problematic_images": false,
216
+ "disable_bucket_pruning": false,
217
+ "allow_dataset_oversubscription": false,
218
+ "disable_segmented_timestep_sampling": false,
219
+ "preserve_data_backend_cache": false,
220
+ "override_dataset_config": false,
221
+ "cache_dir": "/home/Ubuntu/.simpletuner/output/cache",
222
+ "cache_dir_text": "cache",
223
+ "cache_dir_vae": "",
224
+ "compress_disk_cache": true,
225
+ "aspect_bucket_disable_rebuild": false,
226
+ "keep_vae_loaded": false,
227
+ "skip_file_discovery": "",
228
+ "data_backend_sampling": "auto-weighting",
229
+ "image_processing_batch_size": 32,
230
+ "write_batch_size": 128,
231
+ "read_batch_size": 25,
232
+ "enable_multiprocessing": false,
233
+ "accelerate_config": null,
234
+ "deepspeed_config": null,
235
+ "fsdp_enable": false,
236
+ "fsdp_version": 2,
237
+ "fsdp_reshard_after_forward": false,
238
+ "fsdp_state_dict_type": "SHARDED_STATE_DICT",
239
+ "fsdp_cpu_ram_efficient_loading": false,
240
+ "fsdp_auto_wrap_policy": "TRANSFORMER_BASED_WRAP",
241
+ "fsdp_limit_all_gathers": false,
242
+ "fsdp_cpu_offload": false,
243
+ "fsdp_activation_checkpointing": false,
244
+ "fsdp_transformer_layer_cls_to_wrap": null,
245
+ "context_parallel_size": 1,
246
+ "context_parallel_comm_strategy": "allgather",
247
+ "num_processes": 1,
248
+ "num_machines": 1,
249
+ "accelerate_extra_args": null,
250
+ "main_process_ip": "127.0.0.1",
251
+ "main_process_port": 29500,
252
+ "machine_rank": 0,
253
+ "same_network": true,
254
+ "dynamo_backend": "no",
255
+ "dynamo_mode": "",
256
+ "dynamo_fullgraph": false,
257
+ "dynamo_dynamic": false,
258
+ "dynamo_use_regional_compilation": false,
259
+ "max_workers": 32,
260
+ "aws_max_pool_connections": 128,
261
+ "torch_num_threads": 8,
262
+ "dataloader_prefetch": false,
263
+ "dataloader_prefetch_qlen": 10,
264
+ "aspect_bucket_worker_count": 12,
265
+ "aspect_bucket_alignment": 16,
266
+ "minimum_image_size": null,
267
+ "maximum_image_size": null,
268
+ "target_downsample_size": null,
269
+ "metadata_update_interval": 3600,
270
+ "debug_aspect_buckets": false,
271
+ "debug_dataset_loader": false,
272
+ "print_filenames": false,
273
+ "print_sampler_statistics": false,
274
+ "timestep_bias_strategy": null,
275
+ "timestep_bias_begin": 0,
276
+ "timestep_bias_end": 1000,
277
+ "timestep_bias_multiplier": 1.0,
278
+ "timestep_bias_portion": 0.25,
279
+ "training_scheduler_timestep_spacing": "trailing",
280
+ "inference_scheduler_timestep_spacing": "trailing",
281
+ "loss_type": "l2",
282
+ "huber_schedule": "snr",
283
+ "huber_c": 0.1,
284
+ "snr_gamma": null,
285
+ "masked_loss_probability": 1.0,
286
+ "hidream_use_load_balancing_loss": false,
287
+ "hidream_load_balancing_loss_weight": null,
288
+ "adam_beta1": 0.9,
289
+ "adam_beta2": 0.999,
290
+ "optimizer_beta1": null,
291
+ "optimizer_beta2": null,
292
+ "optimizer_cpu_offload_method": null,
293
+ "gradient_precision": null,
294
+ "adam_weight_decay": 0.01,
295
+ "adam_epsilon": 1e-08,
296
+ "prodigy_steps": null,
297
+ "max_grad_norm": 2.0,
298
+ "optimizer_config": null,
299
+ "grad_clip_method": "value",
300
+ "optimizer_offload_gradients": false,
301
+ "fuse_optimizer": false,
302
+ "optimizer_release_gradients": false,
303
+ "push_to_hub": false,
304
+ "publishing_config": null,
305
+ "post_upload_script": null,
306
+ "push_checkpoints_to_hub": false,
307
+ "push_to_hub_background": false,
308
+ "hub_model_id": "quzo/fl2",
309
+ "model_card_private": false,
310
+ "model_card_safe_for_work": false,
311
+ "model_card_note": null,
312
+ "report_to": "none",
313
+ "custom_tracker": null,
314
+ "checkpoint_step_interval": 200,
315
+ "checkpoint_epoch_interval": null,
316
+ "checkpointing_rolling_steps": 0,
317
+ "checkpointing_use_tempdir": false,
318
+ "checkpoints_rolling_total_limit": 1,
319
+ "tracker_run_name": "simpletuner-testing",
320
+ "tracker_project_name": "simpletuner",
321
+ "tracker_image_layout": "gallery",
322
+ "enable_watermark": false,
323
+ "framerate": null,
324
+ "seed_for_each_device": true,
325
+ "snr_weight": 1.0,
326
+ "webhook_config": [
327
+ {
328
+ "webhook_type": "raw",
329
+ "callback_url": "http://0.0.0.0:8001/callback",
330
+ "log_level": "info",
331
+ "ssl_no_verify": false
332
+ }
333
+ ],
334
+ "webhook_reporting_interval": null,
335
+ "distillation_method": null,
336
+ "distillation_config": null,
337
+ "ema_validation": "comparison",
338
+ "local_rank": -1,
339
+ "ltx_train_mode": "i2v",
340
+ "ltx_i2v_prob": 0.1,
341
+ "ltx_partial_noise_fraction": 0.05,
342
+ "ltx_protect_first_frame": false,
343
+ "offload_param_path": null,
344
+ "offset_noise": false,
345
+ "quantize_activations": false,
346
+ "refiner_training": false,
347
+ "refiner_training_invert_schedule": false,
348
+ "refiner_training_strength": 0.2,
349
+ "sdxl_refiner_uses_full_range": false,
350
+ "sana_complex_human_instruction": "Given a user prompt, generate an 'Enhanced prompt' that provides detailed visual descriptions suitable for image generation. Evaluate the level of detail in the user prompt:\n- If the prompt is simple, focus on adding specifics about colors, shapes, sizes, textures, and spatial relationships to create vivid and concrete scenes.\n- Do not add detail that is not in the original prompt! Do not mention colors if the User Prompt does not!\n- If the prompt is already detailed, refine and enhance the existing details slightly without overcomplicating.\n- If the prompt contains explicit material, do not remove this information from the output, which would damage the performance of downstream classifiers.\nHere are examples of how to transform or refine prompts:\n- User Prompt: A cat sleeping -> Enhanced: A cat sleeping peacefully, showcasing the joy of pet ownership. Cute floof kitty cat gatto.\n- User Prompt: A busy city street -> Enhanced: A bustling city street scene featuring a crowd of people.\nPlease generate only the enhanced description for the prompt below and avoid including any additional commentary or evaluations:\nUser Prompt: ",
351
+ "vae_path": "black-forest-labs/FLUX.2-dev",
352
+ "accelerator_project_config": {
353
+ "project_dir": "/home/Ubuntu/.simpletuner/output",
354
+ "logging_dir": "/home/Ubuntu/.simpletuner/output/logs",
355
+ "automatic_checkpoint_naming": false,
356
+ "total_limit": null,
357
+ "iteration": 18,
358
+ "save_on_each_node": false
359
+ },
360
+ "process_group_kwargs": {
361
+ "backend": "nccl",
362
+ "init_method": null,
363
+ "timeout": "1:30:00"
364
+ },
365
+ "is_quantized": false,
366
+ "weight_dtype": "torch.bfloat16",
367
+ "disable_accelerator": false,
368
+ "lora_initialisation_style": true,
369
+ "checkpointing_steps": 200,
370
+ "use_fsdp": false,
371
+ "model_type_label": "Flux.2",
372
+ "use_deepspeed_optimizer": false,
373
+ "use_deepspeed_scheduler": false,
374
+ "base_weight_dtype": "torch.bfloat16",
375
+ "is_quanto": false,
376
+ "is_torchao": false,
377
+ "is_bnb": false,
378
+ "flow_matching": true,
379
+ "overrode_max_train_steps": false,
380
+ "total_num_batches": 6,
381
+ "num_update_steps_per_epoch": 6,
382
+ "total_batch_size": 2,
383
+ "enable_adamw_bf16": true,
384
+ "is_schedulefree": false,
385
+ "is_lr_scheduler_disabled": false,
386
+ "total_steps_remaining_at_start": 5000
387
+ }
checkpoint-3600/training_state-training-images.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"aspect_ratio_bucket_indices": {"1.0": ["/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im008.png", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im021.png"], "0.74": ["/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0014.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9743.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9827.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9935.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0003.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9796.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9890.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9865.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9260.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0090.jpg"]}, "buckets": ["0.74"], "exhausted_buckets": ["1.0"], "batch_size": 2, "current_bucket": 0, "seen_images": {"/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im008.png": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im021.png": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0014.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0003.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9827.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9260.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0090.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9890.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9743.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9865.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9796.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9935.jpg": true}, "current_epoch": 600}
checkpoint-3600/training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"global_step": 3600, "epoch_step": 4199, "epoch": 600, "exhausted_backends": [], "repeats": {"training-images": 0}}
checkpoint-4000/README.md ADDED
@@ -0,0 +1,129 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: "black-forest-labs/FLUX.2-dev"
4
+ tags:
5
+ - flux2
6
+ - flux2-diffusers
7
+ - text-to-image
8
+ - image-to-image
9
+ - diffusers
10
+ - simpletuner
11
+ - not-for-all-audiences
12
+ - lora
13
+
14
+ - template:sd-lora
15
+ - standard
16
+ pipeline_tag: text-to-image
17
+ inference: true
18
+
19
+ ---
20
+
21
+ # quzo/fl2
22
+
23
+ This is a PEFT LoRA derived from [black-forest-labs/FLUX.2-dev](https://huggingface.co/black-forest-labs/FLUX.2-dev).
24
+
25
+ The main validation prompt used during training was:
26
+ ```
27
+ bm82 man
28
+ ```
29
+
30
+
31
+ ## Validation settings
32
+ - CFG: `7.5`
33
+ - CFG Rescale: `0.0`
34
+ - Steps: `20`
35
+ - Sampler: `FlowMatchEulerDiscreteScheduler`
36
+ - Seed: `None`
37
+ - Resolution: `1024x1024`
38
+
39
+
40
+ Note: The validation settings are not necessarily the same as the [training settings](#training-settings).
41
+
42
+
43
+
44
+
45
+ <Gallery />
46
+
47
+ The text encoder **was not** trained.
48
+ You may reuse the base model text encoder for inference.
49
+
50
+
51
+ ## Training settings
52
+
53
+ - Training epochs: 666
54
+ - Training steps: 4000
55
+ - Learning rate: 0.0001
56
+ - Learning rate schedule: constant_with_warmup
57
+ - Warmup steps: 0
58
+ - Max grad value: 2.0
59
+ - Effective batch size: 2
60
+ - Micro-batch size: 2
61
+ - Gradient accumulation steps: 1
62
+ - Number of GPUs: 1
63
+ - Gradient checkpointing: True
64
+ - Prediction type: flow_matching[]
65
+ - Optimizer: adamw_bf16
66
+ - Trainable parameter precision: Pure BF16
67
+ - Base model precision: `no_change`
68
+ - Caption dropout probability: 0.1%
69
+
70
+
71
+
72
+ - LoRA Rank: 16
73
+ - LoRA Alpha: 16.0
74
+ - LoRA Dropout: 0.1
75
+ - LoRA initialisation style: default
76
+ - LoRA mode: Standard
77
+
78
+
79
+ ## Datasets
80
+
81
+ ### training-images
82
+ - Repeats: 0
83
+ - Total number of images: 12
84
+ - Total number of aspect buckets: 1
85
+ - Resolution: 1.048576 megapixels
86
+ - Cropped: False
87
+ - Crop style: None
88
+ - Crop aspect: None
89
+ - Used for regularisation data: No
90
+
91
+
92
+ ## Inference
93
+
94
+
95
+ ```python
96
+ import torch
97
+ from diffusers import DiffusionPipeline
98
+
99
+ model_id = 'black-forest-labs/FLUX.2-dev'
100
+ adapter_id = 'quzo/fl2'
101
+ pipeline = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16) # loading directly in bf16
102
+ pipeline.load_lora_weights(adapter_id)
103
+
104
+ prompt = "bm82 man"
105
+ negative_prompt = 'blurry, cropped, ugly'
106
+
107
+ ## Optional: quantise the model to save on vram.
108
+ ## Note: The model was not quantised during training, so it is not necessary to quantise it during inference time.
109
+ #from optimum.quanto import quantize, freeze, qint8
110
+ #quantize(pipeline.transformer, weights=qint8)
111
+ #freeze(pipeline.transformer)
112
+
113
+ pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') # the pipeline is already in its target precision level
114
+ model_output = pipeline(
115
+ prompt=prompt,
116
+ negative_prompt=negative_prompt,
117
+ num_inference_steps=20,
118
+ generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(42),
119
+ width=1024,
120
+ height=1024,
121
+ guidance_scale=7.5,
122
+ ).images[0]
123
+
124
+ model_output.save("output.png", format="PNG")
125
+
126
+ ```
127
+
128
+
129
+
checkpoint-4000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb5e08c3dd2f66e213afe2a25b1326f19804647e293077e94c7081620e01658e
3
+ size 321008843
checkpoint-4000/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:263d549f7e01a872a1f228806b722556cb672a1e30876f93eb1601f703299642
3
+ size 106979136
checkpoint-4000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:042666dd1082a430bd12fc8536141c84d36683e4390d2ea7c2f635ce006f9880
3
+ size 14757
checkpoint-4000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ceee27b48e7dcd059bd1aa420fb9860574384cf52a7228f1287f54609342ffe
3
+ size 1401
checkpoint-4000/simpletuner_config.json ADDED
@@ -0,0 +1,387 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_family": "flux2",
3
+ "model_flavour": "dev",
4
+ "controlnet": false,
5
+ "pretrained_model_name_or_path": "black-forest-labs/FLUX.2-dev",
6
+ "output_dir": "/home/Ubuntu/.simpletuner/output",
7
+ "logging_dir": "/home/Ubuntu/.simpletuner/output/logs",
8
+ "model_type": "lora",
9
+ "seed": null,
10
+ "resolution": 1024,
11
+ "resume_from_checkpoint": null,
12
+ "prediction_type": null,
13
+ "pretrained_vae_model_name_or_path": "black-forest-labs/FLUX.2-dev",
14
+ "vae_dtype": "bf16",
15
+ "vae_cache_ondemand": false,
16
+ "vae_cache_disable": false,
17
+ "accelerator_cache_clear_interval": null,
18
+ "aspect_bucket_rounding": 2,
19
+ "base_model_precision": "no_change",
20
+ "text_encoder_1_precision": "no_change",
21
+ "text_encoder_2_precision": "no_change",
22
+ "text_encoder_3_precision": "no_change",
23
+ "text_encoder_4_precision": "no_change",
24
+ "gradient_checkpointing_interval": null,
25
+ "offload_during_startup": false,
26
+ "quantize_via": "accelerator",
27
+ "wan_force_2_1_time_embedding": false,
28
+ "fuse_qkv_projections": false,
29
+ "rescale_betas_zero_snr": false,
30
+ "control": false,
31
+ "controlnet_custom_config": null,
32
+ "controlnet_model_name_or_path": null,
33
+ "tread_config": null,
34
+ "pretrained_transformer_model_name_or_path": null,
35
+ "pretrained_transformer_subfolder": "transformer",
36
+ "pretrained_unet_model_name_or_path": null,
37
+ "pretrained_unet_subfolder": "unet",
38
+ "pretrained_t5_model_name_or_path": null,
39
+ "revision": null,
40
+ "variant": null,
41
+ "base_model_default_dtype": "bf16",
42
+ "unet_attention_slice": false,
43
+ "num_train_epochs": 834,
44
+ "max_train_steps": 5000,
45
+ "train_batch_size": 2,
46
+ "learning_rate": 0.0001,
47
+ "optimizer": "adamw_bf16",
48
+ "lr_scheduler": "constant_with_warmup",
49
+ "gradient_accumulation_steps": 1,
50
+ "lr_warmup_steps": 0,
51
+ "checkpoints_total_limit": 5,
52
+ "gradient_checkpointing": true,
53
+ "enable_group_offload": false,
54
+ "ramtorch": false,
55
+ "ramtorch_target_modules": null,
56
+ "ramtorch_text_encoder": false,
57
+ "ramtorch_vae": false,
58
+ "ramtorch_controlnet": false,
59
+ "group_offload_type": "block_level",
60
+ "group_offload_blocks_per_group": 1,
61
+ "group_offload_use_stream": false,
62
+ "group_offload_to_disk_path": "",
63
+ "group_offload_text_encoder": false,
64
+ "group_offload_vae": false,
65
+ "offload_during_save": false,
66
+ "enable_chunked_feed_forward": false,
67
+ "feed_forward_chunk_size": null,
68
+ "train_text_encoder": false,
69
+ "text_encoder_lr": null,
70
+ "lyrics_embedder_train": false,
71
+ "lyrics_embedder_optimizer": null,
72
+ "lyrics_embedder_lr": null,
73
+ "lyrics_embedder_lr_scheduler": null,
74
+ "lr_num_cycles": 1,
75
+ "lr_power": 0.8,
76
+ "use_soft_min_snr": false,
77
+ "use_ema": false,
78
+ "ema_device": "cpu",
79
+ "ema_cpu_only": false,
80
+ "ema_update_interval": 1,
81
+ "ema_foreach_disable": false,
82
+ "ema_decay": 0.995,
83
+ "lora_rank": 16,
84
+ "lora_alpha": 16.0,
85
+ "lora_type": "standard",
86
+ "lora_dropout": 0.1,
87
+ "lora_init_type": "default",
88
+ "peft_lora_mode": "standard",
89
+ "singlora_ramp_up_steps": 0,
90
+ "init_lora": null,
91
+ "lycoris_config": "/home/Ubuntu/.simpletuner/configs/config/lycoris_config.json",
92
+ "init_lokr_norm": null,
93
+ "flux_lora_target": "all",
94
+ "acestep_lora_target": "attn_qkv+linear_qkv",
95
+ "use_dora": false,
96
+ "resolution_type": "pixel_area",
97
+ "data_backend_config": "/home/Ubuntu/.simpletuner/configs/111/multidatabackend-bartek_mizak.json",
98
+ "caption_strategy": "filename",
99
+ "conditioning_multidataset_sampling": "random",
100
+ "instance_prompt": null,
101
+ "parquet_caption_column": null,
102
+ "parquet_filename_column": null,
103
+ "ignore_missing_files": false,
104
+ "vae_cache_scan_behaviour": "recreate",
105
+ "vae_enable_slicing": false,
106
+ "vae_enable_tiling": false,
107
+ "vae_batch_size": 4,
108
+ "max_upscale_threshold": null,
109
+ "caption_dropout_probability": 0.1,
110
+ "tokenizer_max_length": 512,
111
+ "audio_max_duration_seconds": null,
112
+ "audio_min_duration_seconds": null,
113
+ "audio_channels": 1,
114
+ "audio_duration_interval": 3.0,
115
+ "audio_truncation_mode": "beginning",
116
+ "validation_step_interval": 100,
117
+ "validation_epoch_interval": null,
118
+ "disable_benchmark": false,
119
+ "validation_preview": false,
120
+ "validation_preview_steps": 1,
121
+ "validation_prompt": "bm82 man",
122
+ "validation_lyrics": null,
123
+ "validation_audio_duration": 30.0,
124
+ "num_validation_images": 1,
125
+ "num_eval_images": 4,
126
+ "eval_steps_interval": null,
127
+ "eval_timesteps": 28,
128
+ "eval_dataset_pooling": false,
129
+ "evaluation_type": "none",
130
+ "pretrained_evaluation_model_name_or_path": "openai/clip-vit-large-patch14-336",
131
+ "validation_guidance": 7.5,
132
+ "validation_num_inference_steps": 20,
133
+ "validation_on_startup": false,
134
+ "validation_method": "simpletuner-local",
135
+ "validation_external_script": null,
136
+ "validation_external_background": false,
137
+ "validation_using_datasets": false,
138
+ "validation_torch_compile": false,
139
+ "validation_guidance_real": 1.0,
140
+ "validation_no_cfg_until_timestep": 2,
141
+ "validation_negative_prompt": "blurry, cropped, ugly",
142
+ "validation_randomize": false,
143
+ "validation_seed": null,
144
+ "validation_multigpu": "batch-parallel",
145
+ "validation_disable": false,
146
+ "validation_prompt_library": false,
147
+ "user_prompt_library": null,
148
+ "eval_dataset_id": null,
149
+ "validation_stitch_input_location": "left",
150
+ "validation_guidance_rescale": 0.0,
151
+ "validation_disable_unconditional": false,
152
+ "validation_guidance_skip_layers": null,
153
+ "validation_guidance_skip_layers_start": 0.01,
154
+ "validation_guidance_skip_layers_stop": 0.2,
155
+ "validation_guidance_skip_scale": 2.8,
156
+ "validation_lycoris_strength": 1.0,
157
+ "validation_noise_scheduler": null,
158
+ "validation_num_video_frames": null,
159
+ "validation_resolution": "1024x1024",
160
+ "validation_seed_source": "cpu",
161
+ "validation_torch_compile_mode": "max-autotune",
162
+ "validation_adapter_path": null,
163
+ "validation_adapter_name": null,
164
+ "validation_adapter_strength": 1.0,
165
+ "validation_adapter_mode": "adapter_only",
166
+ "validation_adapter_config": null,
167
+ "i_know_what_i_am_doing": false,
168
+ "flow_sigmoid_scale": 1.0,
169
+ "flux_fast_schedule": false,
170
+ "flow_use_uniform_schedule": false,
171
+ "flow_use_beta_schedule": false,
172
+ "flow_beta_schedule_alpha": 2.0,
173
+ "flow_beta_schedule_beta": 2.0,
174
+ "flow_schedule_shift": 3.0,
175
+ "flow_schedule_auto_shift": false,
176
+ "flux_guidance_mode": "constant",
177
+ "flux_attention_masked_training": false,
178
+ "flux_guidance_value": 1.0,
179
+ "flux_guidance_min": 0.0,
180
+ "flux_guidance_max": 4.0,
181
+ "t5_padding": "unmodified",
182
+ "sd3_clip_uncond_behaviour": "empty_string",
183
+ "sd3_t5_uncond_behaviour": null,
184
+ "soft_min_snr_sigma_data": null,
185
+ "mixed_precision": "bf16",
186
+ "attention_mechanism": "diffusers",
187
+ "sla_config": null,
188
+ "sageattention_usage": {
189
+ "_value_": "inference",
190
+ "_name_": "INFERENCE",
191
+ "__objclass__": "{'_generate_next_value_': <staticmethod(<function Enum._generate_next_value_ at 0x7649ca9cdf80>)>, '__module__': 'simpletuner.helpers.training.attention_backend', 'allows_training': <property object at 0x764872f93880>, 'allows_inference': <property object at 0x764872f93790>, 'from_raw': <classmethod(<function AttentionBackendMode.from_raw at 0x764872fe85e0>)>, '_new_member_': <built-in method __new__ of type object at 0xa234e0>, '_use_args_': True, '_member_names_': ['TRAINING', 'INFERENCE', 'TRAINING_AND_INFERENCE'], '_member_map_': {'TRAINING': <AttentionBackendMode.TRAINING: 'training'>, 'INFERENCE': <AttentionBackendMode.INFERENCE: 'inference'>, 'TRAINING_AND_INFERENCE': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>}, '_value2member_map_': {'training': <AttentionBackendMode.TRAINING: 'training'>, 'inference': <AttentionBackendMode.INFERENCE: 'inference'>, 'training+inference': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>}, '_unhashable_values_': [], '_member_type_': <class 'str'>, '_value_repr_': <slot wrapper '__repr__' of 'str' objects>, '__dict__': <attribute '__dict__' of 'AttentionBackendMode' objects>, '__weakref__': <attribute '__weakref__' of 'AttentionBackendMode' objects>, '__doc__': None, 'TRAINING': <AttentionBackendMode.TRAINING: 'training'>, 'INFERENCE': <AttentionBackendMode.INFERENCE: 'inference'>, 'TRAINING_AND_INFERENCE': <AttentionBackendMode.TRAINING_AND_INFERENCE: 'training+inference'>, '__repr__': <function Enum.__repr__ at 0x7649ca9ce0c0>, '__str__': <function Enum.__str__ at 0x7649ca9ce160>, '__format__': <function Enum.__format__ at 0x7649ca9ce2a0>, '__new__': <function Enum.__new__ at 0x7649ca9cde40>}",
192
+ "_sort_order_": 1
193
+ },
194
+ "disable_tf32": false,
195
+ "set_grads_to_none": false,
196
+ "noise_offset": 0.1,
197
+ "noise_offset_probability": 0.25,
198
+ "input_perturbation": 0.0,
199
+ "input_perturbation_steps": 0,
200
+ "lr_end": "4e-7",
201
+ "lr_scale": false,
202
+ "lr_scale_sqrt": false,
203
+ "ignore_final_epochs": false,
204
+ "freeze_encoder_before": 12,
205
+ "freeze_encoder_after": 17,
206
+ "freeze_encoder_strategy": "after",
207
+ "layer_freeze_strategy": null,
208
+ "fully_unload_text_encoder": false,
209
+ "save_text_encoder": false,
210
+ "text_encoder_limit": 100,
211
+ "prepend_instance_prompt": false,
212
+ "only_instance_prompt": false,
213
+ "data_aesthetic_score": 7.0,
214
+ "delete_unwanted_images": false,
215
+ "delete_problematic_images": false,
216
+ "disable_bucket_pruning": false,
217
+ "allow_dataset_oversubscription": false,
218
+ "disable_segmented_timestep_sampling": false,
219
+ "preserve_data_backend_cache": false,
220
+ "override_dataset_config": false,
221
+ "cache_dir": "/home/Ubuntu/.simpletuner/output/cache",
222
+ "cache_dir_text": "cache",
223
+ "cache_dir_vae": "",
224
+ "compress_disk_cache": true,
225
+ "aspect_bucket_disable_rebuild": false,
226
+ "keep_vae_loaded": false,
227
+ "skip_file_discovery": "",
228
+ "data_backend_sampling": "auto-weighting",
229
+ "image_processing_batch_size": 32,
230
+ "write_batch_size": 128,
231
+ "read_batch_size": 25,
232
+ "enable_multiprocessing": false,
233
+ "accelerate_config": null,
234
+ "deepspeed_config": null,
235
+ "fsdp_enable": false,
236
+ "fsdp_version": 2,
237
+ "fsdp_reshard_after_forward": false,
238
+ "fsdp_state_dict_type": "SHARDED_STATE_DICT",
239
+ "fsdp_cpu_ram_efficient_loading": false,
240
+ "fsdp_auto_wrap_policy": "TRANSFORMER_BASED_WRAP",
241
+ "fsdp_limit_all_gathers": false,
242
+ "fsdp_cpu_offload": false,
243
+ "fsdp_activation_checkpointing": false,
244
+ "fsdp_transformer_layer_cls_to_wrap": null,
245
+ "context_parallel_size": 1,
246
+ "context_parallel_comm_strategy": "allgather",
247
+ "num_processes": 1,
248
+ "num_machines": 1,
249
+ "accelerate_extra_args": null,
250
+ "main_process_ip": "127.0.0.1",
251
+ "main_process_port": 29500,
252
+ "machine_rank": 0,
253
+ "same_network": true,
254
+ "dynamo_backend": "no",
255
+ "dynamo_mode": "",
256
+ "dynamo_fullgraph": false,
257
+ "dynamo_dynamic": false,
258
+ "dynamo_use_regional_compilation": false,
259
+ "max_workers": 32,
260
+ "aws_max_pool_connections": 128,
261
+ "torch_num_threads": 8,
262
+ "dataloader_prefetch": false,
263
+ "dataloader_prefetch_qlen": 10,
264
+ "aspect_bucket_worker_count": 12,
265
+ "aspect_bucket_alignment": 16,
266
+ "minimum_image_size": null,
267
+ "maximum_image_size": null,
268
+ "target_downsample_size": null,
269
+ "metadata_update_interval": 3600,
270
+ "debug_aspect_buckets": false,
271
+ "debug_dataset_loader": false,
272
+ "print_filenames": false,
273
+ "print_sampler_statistics": false,
274
+ "timestep_bias_strategy": null,
275
+ "timestep_bias_begin": 0,
276
+ "timestep_bias_end": 1000,
277
+ "timestep_bias_multiplier": 1.0,
278
+ "timestep_bias_portion": 0.25,
279
+ "training_scheduler_timestep_spacing": "trailing",
280
+ "inference_scheduler_timestep_spacing": "trailing",
281
+ "loss_type": "l2",
282
+ "huber_schedule": "snr",
283
+ "huber_c": 0.1,
284
+ "snr_gamma": null,
285
+ "masked_loss_probability": 1.0,
286
+ "hidream_use_load_balancing_loss": false,
287
+ "hidream_load_balancing_loss_weight": null,
288
+ "adam_beta1": 0.9,
289
+ "adam_beta2": 0.999,
290
+ "optimizer_beta1": null,
291
+ "optimizer_beta2": null,
292
+ "optimizer_cpu_offload_method": null,
293
+ "gradient_precision": null,
294
+ "adam_weight_decay": 0.01,
295
+ "adam_epsilon": 1e-08,
296
+ "prodigy_steps": null,
297
+ "max_grad_norm": 2.0,
298
+ "optimizer_config": null,
299
+ "grad_clip_method": "value",
300
+ "optimizer_offload_gradients": false,
301
+ "fuse_optimizer": false,
302
+ "optimizer_release_gradients": false,
303
+ "push_to_hub": false,
304
+ "publishing_config": null,
305
+ "post_upload_script": null,
306
+ "push_checkpoints_to_hub": false,
307
+ "push_to_hub_background": false,
308
+ "hub_model_id": "quzo/fl2",
309
+ "model_card_private": false,
310
+ "model_card_safe_for_work": false,
311
+ "model_card_note": null,
312
+ "report_to": "none",
313
+ "custom_tracker": null,
314
+ "checkpoint_step_interval": 200,
315
+ "checkpoint_epoch_interval": null,
316
+ "checkpointing_rolling_steps": 0,
317
+ "checkpointing_use_tempdir": false,
318
+ "checkpoints_rolling_total_limit": 1,
319
+ "tracker_run_name": "simpletuner-testing",
320
+ "tracker_project_name": "simpletuner",
321
+ "tracker_image_layout": "gallery",
322
+ "enable_watermark": false,
323
+ "framerate": null,
324
+ "seed_for_each_device": true,
325
+ "snr_weight": 1.0,
326
+ "webhook_config": [
327
+ {
328
+ "webhook_type": "raw",
329
+ "callback_url": "http://0.0.0.0:8001/callback",
330
+ "log_level": "info",
331
+ "ssl_no_verify": false
332
+ }
333
+ ],
334
+ "webhook_reporting_interval": null,
335
+ "distillation_method": null,
336
+ "distillation_config": null,
337
+ "ema_validation": "comparison",
338
+ "local_rank": -1,
339
+ "ltx_train_mode": "i2v",
340
+ "ltx_i2v_prob": 0.1,
341
+ "ltx_partial_noise_fraction": 0.05,
342
+ "ltx_protect_first_frame": false,
343
+ "offload_param_path": null,
344
+ "offset_noise": false,
345
+ "quantize_activations": false,
346
+ "refiner_training": false,
347
+ "refiner_training_invert_schedule": false,
348
+ "refiner_training_strength": 0.2,
349
+ "sdxl_refiner_uses_full_range": false,
350
+ "sana_complex_human_instruction": "Given a user prompt, generate an 'Enhanced prompt' that provides detailed visual descriptions suitable for image generation. Evaluate the level of detail in the user prompt:\n- If the prompt is simple, focus on adding specifics about colors, shapes, sizes, textures, and spatial relationships to create vivid and concrete scenes.\n- Do not add detail that is not in the original prompt! Do not mention colors if the User Prompt does not!\n- If the prompt is already detailed, refine and enhance the existing details slightly without overcomplicating.\n- If the prompt contains explicit material, do not remove this information from the output, which would damage the performance of downstream classifiers.\nHere are examples of how to transform or refine prompts:\n- User Prompt: A cat sleeping -> Enhanced: A cat sleeping peacefully, showcasing the joy of pet ownership. Cute floof kitty cat gatto.\n- User Prompt: A busy city street -> Enhanced: A bustling city street scene featuring a crowd of people.\nPlease generate only the enhanced description for the prompt below and avoid including any additional commentary or evaluations:\nUser Prompt: ",
351
+ "vae_path": "black-forest-labs/FLUX.2-dev",
352
+ "accelerator_project_config": {
353
+ "project_dir": "/home/Ubuntu/.simpletuner/output",
354
+ "logging_dir": "/home/Ubuntu/.simpletuner/output/logs",
355
+ "automatic_checkpoint_naming": false,
356
+ "total_limit": null,
357
+ "iteration": 20,
358
+ "save_on_each_node": false
359
+ },
360
+ "process_group_kwargs": {
361
+ "backend": "nccl",
362
+ "init_method": null,
363
+ "timeout": "1:30:00"
364
+ },
365
+ "is_quantized": false,
366
+ "weight_dtype": "torch.bfloat16",
367
+ "disable_accelerator": false,
368
+ "lora_initialisation_style": true,
369
+ "checkpointing_steps": 200,
370
+ "use_fsdp": false,
371
+ "model_type_label": "Flux.2",
372
+ "use_deepspeed_optimizer": false,
373
+ "use_deepspeed_scheduler": false,
374
+ "base_weight_dtype": "torch.bfloat16",
375
+ "is_quanto": false,
376
+ "is_torchao": false,
377
+ "is_bnb": false,
378
+ "flow_matching": true,
379
+ "overrode_max_train_steps": false,
380
+ "total_num_batches": 6,
381
+ "num_update_steps_per_epoch": 6,
382
+ "total_batch_size": 2,
383
+ "enable_adamw_bf16": true,
384
+ "is_schedulefree": false,
385
+ "is_lr_scheduler_disabled": false,
386
+ "total_steps_remaining_at_start": 5000
387
+ }
checkpoint-4000/training_state-training-images.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"aspect_ratio_bucket_indices": {"1.0": ["/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im008.png", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im021.png"], "0.74": ["/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0014.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9743.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9827.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9935.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0003.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9796.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9890.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9865.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9260.jpg", "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0090.jpg"]}, "buckets": ["0.74"], "exhausted_buckets": ["1.0"], "batch_size": 2, "current_bucket": 0, "seen_images": {"/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im008.png": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/im021.png": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9796.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0003.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9935.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_0014.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9827.jpg": true, "/home/Ubuntu/.simpletuner/datasets/bartek_mizak/img_9743.jpg": true}, "current_epoch": 667}
checkpoint-4000/training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"global_step": 4000, "epoch_step": 4666, "epoch": 667, "exhausted_backends": [], "repeats": {"training-images": 0}}