SidXXD commited on
Commit
6ad280b
·
verified ·
1 Parent(s): 4a78b02

End of training

Browse files
Files changed (18) hide show
  1. logs/custom-diffusion/1739958609.2958472/events.out.tfevents.1739958609.cudahpc06.2182443.1 +3 -0
  2. logs/custom-diffusion/1739958609.2995698/hparams.yml +58 -0
  3. logs/custom-diffusion/events.out.tfevents.1739958609.cudahpc06.2182443.0 +3 -0
  4. svd_plots/down_block_0_attentions_0_transformer_blocks_0_attn2_processor_to_k_custom_diffusion.png +0 -0
  5. svd_plots/down_block_0_attentions_0_transformer_blocks_0_attn2_processor_to_v_custom_diffusion.png +0 -0
  6. svd_plots/down_block_1_attentions_0_transformer_blocks_0_attn2_processor_to_k_custom_diffusion.png +0 -0
  7. svd_plots/down_block_1_attentions_0_transformer_blocks_0_attn2_processor_to_v_custom_diffusion.png +0 -0
  8. svd_plots/down_block_2_attentions_0_transformer_blocks_0_attn2_processor_to_k_custom_diffusion.png +0 -0
  9. svd_plots/down_block_2_attentions_0_transformer_blocks_0_attn2_processor_to_v_custom_diffusion.png +0 -0
  10. svd_plots/mid_block_attentions_0_transformer_blocks_0_attn2_processor_to_k_custom_diffusion.png +0 -0
  11. svd_plots/mid_block_attentions_0_transformer_blocks_0_attn2_processor_to_v_custom_diffusion.png +0 -0
  12. svd_plots/up_block_1_attentions_0_transformer_blocks_0_attn2_processor_to_k_custom_diffusion.png +0 -0
  13. svd_plots/up_block_1_attentions_0_transformer_blocks_0_attn2_processor_to_v_custom_diffusion.png +0 -0
  14. svd_plots/up_block_2_attentions_0_transformer_blocks_0_attn2_processor_to_k_custom_diffusion.png +0 -0
  15. svd_plots/up_block_2_attentions_0_transformer_blocks_0_attn2_processor_to_v_custom_diffusion.png +0 -0
  16. svd_plots/up_block_3_attentions_0_transformer_blocks_0_attn2_processor_to_k_custom_diffusion.png +0 -0
  17. svd_plots/up_block_3_attentions_0_transformer_blocks_0_attn2_processor_to_v_custom_diffusion.png +0 -0
  18. svd_results.json +0 -0
logs/custom-diffusion/1739958609.2958472/events.out.tfevents.1739958609.cudahpc06.2182443.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e22c33803da442ec7334d6bc9b29c5fef2c0a56701aa1bdbb4567b373661321c
3
+ size 2797
logs/custom-diffusion/1739958609.2995698/hparams.yml ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ attn_maps_path_name: null
7
+ center_crop: false
8
+ checkpointing_steps: 250
9
+ checkpoints_total_limit: null
10
+ class_data_dir: null
11
+ class_prompt: person
12
+ concepts_list: null
13
+ dataloader_num_workers: 2
14
+ enable_xformers_memory_efficient_attention: false
15
+ freeze_model: crossattn_kv
16
+ gradient_accumulation_steps: 1
17
+ gradient_checkpointing: false
18
+ hflip: true
19
+ hub_model_id: null
20
+ hub_token: null
21
+ initializer_token: ktn+pll+ucd
22
+ instance_data_dir: /home/xide/miniconda3/code/data/Untarget_training_img_20/Group_for_caat/f-26/p1
23
+ instance_prompt: photo of a <new1> person
24
+ learning_rate: 1.0e-05
25
+ local_rank: -1
26
+ logging_dir: logs
27
+ lr_scheduler: constant
28
+ lr_warmup_steps: 0
29
+ max_grad_norm: 1.0
30
+ max_train_steps: 2
31
+ mixed_precision: null
32
+ modifier_token: <new1>
33
+ no_safe_serialization: true
34
+ noaug: false
35
+ num_class_images: 200
36
+ num_train_epochs: 1
37
+ num_validation_images: 2
38
+ output_dir: /home/xide/miniconda3/code/diffusers/examples/custom_diffusion/model/debug
39
+ pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
40
+ prior_generation_precision: null
41
+ prior_loss_weight: 0.0
42
+ push_to_hub: true
43
+ real_prior: false
44
+ report_to: tensorboard
45
+ resolution: 512
46
+ resume_from_checkpoint: null
47
+ revision: null
48
+ sample_batch_size: 4
49
+ scale_lr: true
50
+ seed: 42
51
+ set_grads_to_none: false
52
+ tokenizer_name: null
53
+ train_batch_size: 2
54
+ use_8bit_adam: false
55
+ validation_prompt: null
56
+ validation_steps: 50
57
+ variant: null
58
+ with_prior_preservation: false
logs/custom-diffusion/events.out.tfevents.1739958609.cudahpc06.2182443.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9f052ea96d2f993464876da2e51439c7e8e39c56fcbd3f8325eac6bfab7fc65
3
+ size 252
svd_plots/down_block_0_attentions_0_transformer_blocks_0_attn2_processor_to_k_custom_diffusion.png ADDED
svd_plots/down_block_0_attentions_0_transformer_blocks_0_attn2_processor_to_v_custom_diffusion.png ADDED
svd_plots/down_block_1_attentions_0_transformer_blocks_0_attn2_processor_to_k_custom_diffusion.png ADDED
svd_plots/down_block_1_attentions_0_transformer_blocks_0_attn2_processor_to_v_custom_diffusion.png ADDED
svd_plots/down_block_2_attentions_0_transformer_blocks_0_attn2_processor_to_k_custom_diffusion.png ADDED
svd_plots/down_block_2_attentions_0_transformer_blocks_0_attn2_processor_to_v_custom_diffusion.png ADDED
svd_plots/mid_block_attentions_0_transformer_blocks_0_attn2_processor_to_k_custom_diffusion.png ADDED
svd_plots/mid_block_attentions_0_transformer_blocks_0_attn2_processor_to_v_custom_diffusion.png ADDED
svd_plots/up_block_1_attentions_0_transformer_blocks_0_attn2_processor_to_k_custom_diffusion.png ADDED
svd_plots/up_block_1_attentions_0_transformer_blocks_0_attn2_processor_to_v_custom_diffusion.png ADDED
svd_plots/up_block_2_attentions_0_transformer_blocks_0_attn2_processor_to_k_custom_diffusion.png ADDED
svd_plots/up_block_2_attentions_0_transformer_blocks_0_attn2_processor_to_v_custom_diffusion.png ADDED
svd_plots/up_block_3_attentions_0_transformer_blocks_0_attn2_processor_to_k_custom_diffusion.png ADDED
svd_plots/up_block_3_attentions_0_transformer_blocks_0_attn2_processor_to_v_custom_diffusion.png ADDED
svd_results.json ADDED
The diff for this file is too large to render. See raw diff