rayli1107 commited on
Commit
083219d
·
verified ·
1 Parent(s): a33c54e

End of training

Browse files
README.md CHANGED
@@ -12,6 +12,13 @@ tags:
12
  - flux
13
  - flux-diffusers
14
  - template:sd-lora
 
 
 
 
 
 
 
15
  ---
16
 
17
  <!-- This model card has been generated automatically according to the information the training script had access to. You
 
12
  - flux
13
  - flux-diffusers
14
  - template:sd-lora
15
+ - text-to-image
16
+ - diffusers-training
17
+ - diffusers
18
+ - lora
19
+ - flux
20
+ - flux-diffusers
21
+ - template:sd-lora
22
  ---
23
 
24
  <!-- This model card has been generated automatically according to the information the training script had access to. You
checkpoint-500/optimizer.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cbef20aef3972e4b8be80dba7d9fdc8da28af856ac401a9dbe081e5ff886b235
3
  size 23562884
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b1ca5ade99ee2310c9da4115d6f3cf80ef7c5d3dd84b39fd8b5915f5b9c78d1
3
  size 23562884
checkpoint-500/pytorch_lora_weights.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dba7f3a6fdfabbf0cac7bff081a2a85aa57171f76a6a0aaf61e5b0272a286531
3
  size 22504080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83d1875eeeb8ffd92d3411c89cccdaf83441406f6f489d571446861bc81b52ed
3
  size 22504080
checkpoint-500/random_states_0.pkl CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1ff63348c8bf2e90a9e4c4ad8dfe68081f757743f2f4f0a743e52799f944480c
3
  size 14344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3daeac355e657259fe6900290c6ba537320ff4a08e7c40e326bca77105fd476
3
  size 14344
logs/dreambooth-flux-dev-lora/1743331491.0675344/events.out.tfevents.1743331491.8ee1a30f46eb.4491.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21711bce5c0924ae10bef4d5506fa8ca92562e446f85f15b2fa11baf6629458f
3
+ size 3223
logs/dreambooth-flux-dev-lora/1743331491.0693886/hparams.yml ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.0001
5
+ adam_weight_decay_text_encoder: 0.001
6
+ allow_tf32: false
7
+ cache_dir: null
8
+ cache_latents: true
9
+ caption_column: null
10
+ center_crop: false
11
+ checkpointing_steps: 500
12
+ checkpoints_total_limit: null
13
+ class_data_dir: null
14
+ class_prompt: null
15
+ dataloader_num_workers: 0
16
+ dataset_config_name: null
17
+ dataset_name: null
18
+ gradient_accumulation_steps: 4
19
+ gradient_checkpointing: false
20
+ guidance_scale: 1.0
21
+ hub_model_id: null
22
+ hub_token: null
23
+ image_column: image
24
+ instance_data_dir: input_florawhite
25
+ instance_prompt: florawhite
26
+ learning_rate: 0.0001
27
+ local_rank: -1
28
+ logging_dir: logs
29
+ logit_mean: 0.0
30
+ logit_std: 1.0
31
+ lora_layers: null
32
+ lr_num_cycles: 1
33
+ lr_power: 1.0
34
+ lr_scheduler: constant
35
+ lr_warmup_steps: 0
36
+ max_grad_norm: 1.0
37
+ max_sequence_length: 512
38
+ max_train_steps: 500
39
+ mixed_precision: bf16
40
+ mode_scale: 1.29
41
+ num_class_images: 100
42
+ num_train_epochs: 100
43
+ num_validation_images: 4
44
+ optimizer: AdamW
45
+ output_dir: trained-flux
46
+ pretrained_model_name_or_path: black-forest-labs/FLUX.1-dev
47
+ prior_generation_precision: null
48
+ prior_loss_weight: 1.0
49
+ prodigy_beta3: null
50
+ prodigy_decouple: true
51
+ prodigy_safeguard_warmup: true
52
+ prodigy_use_bias_correction: true
53
+ push_to_hub: true
54
+ random_flip: false
55
+ rank: 4
56
+ repeats: 1
57
+ report_to: tensorboard
58
+ resolution: 512
59
+ resume_from_checkpoint: null
60
+ revision: null
61
+ sample_batch_size: 4
62
+ scale_lr: false
63
+ seed: 0
64
+ text_encoder_lr: 5.0e-06
65
+ train_batch_size: 1
66
+ train_text_encoder: false
67
+ upcast_before_saving: false
68
+ use_8bit_adam: true
69
+ validation_epochs: 50
70
+ validation_prompt: null
71
+ variant: null
72
+ weighting_scheme: none
73
+ with_prior_preservation: false
logs/dreambooth-flux-dev-lora/1743331914.605265/events.out.tfevents.1743331914.8ee1a30f46eb.975.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae83a2eea76e50baa4b3d3ed60f3936a6def354816250856ca097b77a733e50d
3
+ size 3223
logs/dreambooth-flux-dev-lora/1743331914.6073563/hparams.yml ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.0001
5
+ adam_weight_decay_text_encoder: 0.001
6
+ allow_tf32: false
7
+ cache_dir: null
8
+ cache_latents: true
9
+ caption_column: null
10
+ center_crop: false
11
+ checkpointing_steps: 500
12
+ checkpoints_total_limit: null
13
+ class_data_dir: null
14
+ class_prompt: null
15
+ dataloader_num_workers: 0
16
+ dataset_config_name: null
17
+ dataset_name: null
18
+ gradient_accumulation_steps: 4
19
+ gradient_checkpointing: false
20
+ guidance_scale: 1.0
21
+ hub_model_id: null
22
+ hub_token: null
23
+ image_column: image
24
+ instance_data_dir: input_florawhite
25
+ instance_prompt: florawhite
26
+ learning_rate: 0.0001
27
+ local_rank: -1
28
+ logging_dir: logs
29
+ logit_mean: 0.0
30
+ logit_std: 1.0
31
+ lora_layers: null
32
+ lr_num_cycles: 1
33
+ lr_power: 1.0
34
+ lr_scheduler: constant
35
+ lr_warmup_steps: 0
36
+ max_grad_norm: 1.0
37
+ max_sequence_length: 512
38
+ max_train_steps: 500
39
+ mixed_precision: bf16
40
+ mode_scale: 1.29
41
+ num_class_images: 100
42
+ num_train_epochs: 100
43
+ num_validation_images: 4
44
+ optimizer: AdamW
45
+ output_dir: trained-flux
46
+ pretrained_model_name_or_path: black-forest-labs/FLUX.1-dev
47
+ prior_generation_precision: null
48
+ prior_loss_weight: 1.0
49
+ prodigy_beta3: null
50
+ prodigy_decouple: true
51
+ prodigy_safeguard_warmup: true
52
+ prodigy_use_bias_correction: true
53
+ push_to_hub: true
54
+ random_flip: false
55
+ rank: 4
56
+ repeats: 1
57
+ report_to: tensorboard
58
+ resolution: 512
59
+ resume_from_checkpoint: null
60
+ revision: null
61
+ sample_batch_size: 4
62
+ scale_lr: false
63
+ seed: 0
64
+ text_encoder_lr: 5.0e-06
65
+ train_batch_size: 1
66
+ train_text_encoder: false
67
+ upcast_before_saving: false
68
+ use_8bit_adam: true
69
+ validation_epochs: 50
70
+ validation_prompt: null
71
+ variant: null
72
+ weighting_scheme: none
73
+ with_prior_preservation: false
logs/dreambooth-flux-dev-lora/events.out.tfevents.1743331491.8ee1a30f46eb.4491.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:029e46a796df06c5c6b275c7e2b83ccdcfea9c0b81f23b7c9db7d2ad80229715
3
+ size 88
logs/dreambooth-flux-dev-lora/events.out.tfevents.1743331914.8ee1a30f46eb.975.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9d175833eca3b62e51a4f36b9e469f9ef00f6b348f77fff4216eb0d3051c308
3
+ size 158704
pytorch_lora_weights.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dba7f3a6fdfabbf0cac7bff081a2a85aa57171f76a6a0aaf61e5b0272a286531
3
  size 22504080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83d1875eeeb8ffd92d3411c89cccdaf83441406f6f489d571446861bc81b52ed
3
  size 22504080