job: extension config: name: trained_lora process: - type: sd_trainer training_folder: /home/bhara/ai-toolkit/output device: cuda:0 trigger_word: UNST network: type: lora linear: 16 linear_alpha: 16 network_kwargs: only_if_contains: - transformer.single_transformer_blocks.'7.proj_out - transformer.single_transformer_blocks.12.proj_out - transformer.single_transformer_blocks.16.proj_out - transformer.single_transformer_blocks.20'.proj_out save: dtype: float16 save_every: 100 max_step_saves_to_keep: 4 datasets: - folder_path: /home/bhara/ai-toolkit/dataset/images_and_captions caption_ext: txt caption_dropout_rate: 0.05 cache_latents_to_disk: false resolution: - 512 - 768 - 1024 train: batch_size: 1 steps: 10 lr: 0.0004 gradient_checkpointing: true optimizer: adamw8bit disable_sampling: true dtype: bf16 model: name_or_path: black-forest-labs/FLUX.1-dev is_flux: true quantize: false sample: sample_every: 10000 width: 1024 height: 1024 prompts: [] seed: 42 sample_steps: 28