SidXXD commited on
Commit
9107827
·
verified ·
1 Parent(s): de35342

End of training

Browse files
logs/custom-diffusion/1739957676.0361822/events.out.tfevents.1739957676.cudahpc06.2139148.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7140763337331deae508a64724894f65c77316c8ff6228adbda6a68849df884
3
+ size 2797
logs/custom-diffusion/1739957676.039834/hparams.yml ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ attn_maps_path_name: null
7
+ center_crop: false
8
+ checkpointing_steps: 250
9
+ checkpoints_total_limit: null
10
+ class_data_dir: null
11
+ class_prompt: person
12
+ concepts_list: null
13
+ dataloader_num_workers: 2
14
+ enable_xformers_memory_efficient_attention: false
15
+ freeze_model: crossattn_kv
16
+ gradient_accumulation_steps: 1
17
+ gradient_checkpointing: false
18
+ hflip: true
19
+ hub_model_id: null
20
+ hub_token: null
21
+ initializer_token: ktn+pll+ucd
22
+ instance_data_dir: /home/xide/miniconda3/code/data/Untarget_training_img_20/Group_for_caat/f-26/p1
23
+ instance_prompt: photo of a <new1> person
24
+ learning_rate: 1.0e-05
25
+ local_rank: -1
26
+ logging_dir: logs
27
+ lr_scheduler: constant
28
+ lr_warmup_steps: 0
29
+ max_grad_norm: 1.0
30
+ max_train_steps: 2
31
+ mixed_precision: null
32
+ modifier_token: <new1>
33
+ no_safe_serialization: true
34
+ noaug: false
35
+ num_class_images: 200
36
+ num_train_epochs: 1
37
+ num_validation_images: 2
38
+ output_dir: /home/xide/miniconda3/code/diffusers/examples/custom_diffusion/model/debug
39
+ pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
40
+ prior_generation_precision: null
41
+ prior_loss_weight: 0.0
42
+ push_to_hub: true
43
+ real_prior: false
44
+ report_to: tensorboard
45
+ resolution: 512
46
+ resume_from_checkpoint: null
47
+ revision: null
48
+ sample_batch_size: 4
49
+ scale_lr: true
50
+ seed: 42
51
+ set_grads_to_none: false
52
+ tokenizer_name: null
53
+ train_batch_size: 2
54
+ use_8bit_adam: false
55
+ validation_prompt: null
56
+ validation_steps: 50
57
+ variant: null
58
+ with_prior_preservation: false
logs/custom-diffusion/events.out.tfevents.1739957676.cudahpc06.2139148.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed2ccc14b604a79fde12c8e0b961b51f5d5b203f8a95dcc7f950c6ffb914e06e
3
+ size 252
svd_plots/mid_block_svd_attentions_0_proj_in.png ADDED
svd_plots/mid_block_svd_attentions_0_transformer_blocks_0_attn1_to_k.png ADDED
svd_plots/mid_block_svd_attentions_0_transformer_blocks_0_attn1_to_out_0.png ADDED
svd_plots/mid_block_svd_attentions_0_transformer_blocks_0_attn1_to_q.png ADDED
svd_plots/mid_block_svd_attentions_0_transformer_blocks_0_attn1_to_v.png ADDED
svd_plots/mid_block_svd_attentions_0_transformer_blocks_0_attn2_to_k.png ADDED
svd_plots/mid_block_svd_attentions_0_transformer_blocks_0_attn2_to_out_0.png ADDED
svd_plots/mid_block_svd_attentions_0_transformer_blocks_0_attn2_to_q.png ADDED
svd_plots/mid_block_svd_attentions_0_transformer_blocks_0_attn2_to_v.png ADDED