zli12321 commited on
Commit
083256c
·
verified ·
1 Parent(s): 62e23b4

Upload folder using huggingface_hub

Browse files
Files changed (24) hide show
  1. .gitattributes +6 -0
  2. 10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/checkpoint-600.safetensors +3 -0
  3. 10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/events.out.tfevents.1761118962.edb6fde9143d.2647.0 +3 -0
  4. 10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/text2image-fine-tune/1761119615.7413323/events.out.tfevents.1761119615.edb6fde9143d.2647.2 +3 -0
  5. 10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/text2image-fine-tune/1761119615.8940296/hparams.yml +89 -0
  6. 10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/text2image-fine-tune/events.out.tfevents.1761119615.edb6fde9143d.2647.1 +3 -0
  7. 10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/network_structure.txt +1606 -0
  8. 10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/ad23r2-the.gif +3 -0
  9. 10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/clip_ad23r2-the.png +3 -0
  10. 10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/mask_ad23r2-the.gif +3 -0
  11. 10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/trainable_params.txt +802 -0
  12. 10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/transformer3d_structure.txt +49 -0
  13. 10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/checkpoint-600.safetensors +3 -0
  14. 10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/events.out.tfevents.1761137386.edb6fde9143d.95968.0 +3 -0
  15. 10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/text2image-fine-tune/1761138020.9281862/events.out.tfevents.1761138020.edb6fde9143d.95968.2 +3 -0
  16. 10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/text2image-fine-tune/1761138020.940236/.ipynb_checkpoints/hparams-checkpoint.yml +89 -0
  17. 10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/text2image-fine-tune/1761138020.940236/hparams.yml +89 -0
  18. 10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/text2image-fine-tune/events.out.tfevents.1761138020.edb6fde9143d.95968.1 +3 -0
  19. 10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/network_structure.txt +1606 -0
  20. 10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/ad23r2-the.gif +3 -0
  21. 10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/clip_ad23r2-the.png +3 -0
  22. 10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/mask_ad23r2-the.gif +3 -0
  23. 10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/trainable_params.txt +802 -0
  24. 10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/transformer3d_structure.txt +49 -0
.gitattributes CHANGED
@@ -33,3 +33,9 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ 10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/ad23r2-the.gif filter=lfs diff=lfs merge=lfs -text
37
+ 10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/clip_ad23r2-the.png filter=lfs diff=lfs merge=lfs -text
38
+ 10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/mask_ad23r2-the.gif filter=lfs diff=lfs merge=lfs -text
39
+ 10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/ad23r2-the.gif filter=lfs diff=lfs merge=lfs -text
40
+ 10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/clip_ad23r2-the.png filter=lfs diff=lfs merge=lfs -text
41
+ 10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/mask_ad23r2-the.gif filter=lfs diff=lfs merge=lfs -text
10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/checkpoint-600.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f29e4994032d800673ee193be76af295dee36f9c9e8b2a4ef69b4617369fbdb
3
+ size 1236150882
10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/events.out.tfevents.1761118962.edb6fde9143d.2647.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:822e3b24bc28694a0ffc689fc6b738b21614d5d1904c0081f4d6780fb3086ece
3
+ size 88
10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/text2image-fine-tune/1761119615.7413323/events.out.tfevents.1761119615.edb6fde9143d.2647.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b1d139e52f04150220d9bc54a19cb2752013a750270e46646399abc12a6c3a5
3
+ size 4517
10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/text2image-fine-tune/1761119615.8940296/hparams.yml ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-10
4
+ adam_weight_decay: 0.03
5
+ allow_tf32: false
6
+ auto_tile_batch_size: false
7
+ boundary_type: high
8
+ cache_dir: null
9
+ checkpointing_steps: 100
10
+ checkpoints_total_limit: null
11
+ config_path: /workspace/Project/VideoRAG/VideoX-Fun/config/wan2.2/wan_civitai_i2v.yaml
12
+ dataloader_num_workers: 4
13
+ enable_bucket: true
14
+ enable_text_encoder_in_dataloader: false
15
+ gradient_accumulation_steps: 1
16
+ gradient_checkpointing: true
17
+ height: 720
18
+ hub_model_id: null
19
+ hub_token: null
20
+ image_sample_size: 1024
21
+ input_perturbation: 0
22
+ keep_all_node_same_token_length: false
23
+ learning_rate: 0.0001
24
+ local_rank: 0
25
+ logging_dir: logs
26
+ logit_mean: 0.0
27
+ logit_std: 1.0
28
+ lora_skip_name: null
29
+ low_vram: false
30
+ lr_scheduler: constant
31
+ lr_warmup_steps: 500
32
+ max_grad_norm: 0.05
33
+ max_train_samples: null
34
+ max_train_steps: 60000
35
+ mixed_precision: bf16
36
+ mode_scale: 1.29
37
+ motion_sub_loss: false
38
+ motion_sub_loss_ratio: 0.25
39
+ multi_stream: false
40
+ network_alpha: 128
41
+ noise_offset: 0
42
+ noise_share_in_frames: false
43
+ noise_share_in_frames_ratio: 0.5
44
+ non_ema_revision: null
45
+ num_train_epochs: 5000
46
+ output_dir: /workspace/Models/10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4
47
+ prediction_type: null
48
+ pretrained_model_name_or_path: /workspace/hfhome/hub/models--Wan-AI--Wan2.2-I2V-A14B/snapshots/206a9ee1b7bfaaf8f7e4d81335650533490646a3
49
+ push_to_hub: false
50
+ random_flip: false
51
+ random_frame_crop: false
52
+ random_hw_adapt: true
53
+ random_ratio_crop: false
54
+ rank: 128
55
+ report_to: tensorboard
56
+ resume_from_checkpoint: null
57
+ revision: null
58
+ save_state: false
59
+ scale_lr: false
60
+ seed: 42
61
+ snr_loss: false
62
+ token_sample_size: 512
63
+ tokenizer_max_length: 512
64
+ tracker_project_name: text2image-fine-tune
65
+ train_batch_size: 2
66
+ train_data_dir: /workspace/Data/05_mixed_data_Large
67
+ train_data_meta: /workspace/Data/05_mixed_data_Large/train_data_mixed_Large.json
68
+ train_mode: i2v
69
+ train_sampling_steps: 1000
70
+ train_text_encoder: false
71
+ training_with_video_token_length: true
72
+ transformer_path: null
73
+ uniform_sampling: true
74
+ use_8bit_adam: false
75
+ use_came: false
76
+ use_deepspeed: false
77
+ use_ema: false
78
+ use_fsdp: false
79
+ vae_mini_batch: 1
80
+ vae_path: null
81
+ validation_epochs: 5
82
+ validation_steps: 200000000
83
+ variant: null
84
+ video_repeat: 1
85
+ video_sample_n_frames: 81
86
+ video_sample_size: 1024
87
+ video_sample_stride: 1
88
+ weighting_scheme: none
89
+ width: 1280
10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/text2image-fine-tune/events.out.tfevents.1761119615.edb6fde9143d.2647.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab910c83d3cd3bab1fcada437e1c898728def99153a2e57a210c820f0cc0fce7
3
+ size 29508
10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/network_structure.txt ADDED
@@ -0,0 +1,1606 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ LoRANetwork(
2
+ (lora_unet__time_projection_1): LoRAModule(
3
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
4
+ (lora_up): Linear(in_features=128, out_features=30720, bias=False)
5
+ )
6
+ (lora_unet__blocks_0_self_attn_q): LoRAModule(
7
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
8
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
9
+ )
10
+ (lora_unet__blocks_0_self_attn_k): LoRAModule(
11
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
12
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
13
+ )
14
+ (lora_unet__blocks_0_self_attn_v): LoRAModule(
15
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
16
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
17
+ )
18
+ (lora_unet__blocks_0_self_attn_o): LoRAModule(
19
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
20
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
21
+ )
22
+ (lora_unet__blocks_0_cross_attn_q): LoRAModule(
23
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
24
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
25
+ )
26
+ (lora_unet__blocks_0_cross_attn_k): LoRAModule(
27
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
28
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
29
+ )
30
+ (lora_unet__blocks_0_cross_attn_v): LoRAModule(
31
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
32
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
33
+ )
34
+ (lora_unet__blocks_0_cross_attn_o): LoRAModule(
35
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
36
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
37
+ )
38
+ (lora_unet__blocks_0_ffn_0): LoRAModule(
39
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
40
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
41
+ )
42
+ (lora_unet__blocks_0_ffn_2): LoRAModule(
43
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
44
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
45
+ )
46
+ (lora_unet__blocks_1_self_attn_q): LoRAModule(
47
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
48
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
49
+ )
50
+ (lora_unet__blocks_1_self_attn_k): LoRAModule(
51
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
52
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
53
+ )
54
+ (lora_unet__blocks_1_self_attn_v): LoRAModule(
55
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
56
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
57
+ )
58
+ (lora_unet__blocks_1_self_attn_o): LoRAModule(
59
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
60
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
61
+ )
62
+ (lora_unet__blocks_1_cross_attn_q): LoRAModule(
63
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
64
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
65
+ )
66
+ (lora_unet__blocks_1_cross_attn_k): LoRAModule(
67
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
68
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
69
+ )
70
+ (lora_unet__blocks_1_cross_attn_v): LoRAModule(
71
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
72
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
73
+ )
74
+ (lora_unet__blocks_1_cross_attn_o): LoRAModule(
75
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
76
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
77
+ )
78
+ (lora_unet__blocks_1_ffn_0): LoRAModule(
79
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
80
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
81
+ )
82
+ (lora_unet__blocks_1_ffn_2): LoRAModule(
83
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
84
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
85
+ )
86
+ (lora_unet__blocks_2_self_attn_q): LoRAModule(
87
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
88
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
89
+ )
90
+ (lora_unet__blocks_2_self_attn_k): LoRAModule(
91
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
92
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
93
+ )
94
+ (lora_unet__blocks_2_self_attn_v): LoRAModule(
95
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
96
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
97
+ )
98
+ (lora_unet__blocks_2_self_attn_o): LoRAModule(
99
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
100
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
101
+ )
102
+ (lora_unet__blocks_2_cross_attn_q): LoRAModule(
103
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
104
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
105
+ )
106
+ (lora_unet__blocks_2_cross_attn_k): LoRAModule(
107
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
108
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
109
+ )
110
+ (lora_unet__blocks_2_cross_attn_v): LoRAModule(
111
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
112
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
113
+ )
114
+ (lora_unet__blocks_2_cross_attn_o): LoRAModule(
115
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
116
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
117
+ )
118
+ (lora_unet__blocks_2_ffn_0): LoRAModule(
119
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
120
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
121
+ )
122
+ (lora_unet__blocks_2_ffn_2): LoRAModule(
123
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
124
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
125
+ )
126
+ (lora_unet__blocks_3_self_attn_q): LoRAModule(
127
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
128
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
129
+ )
130
+ (lora_unet__blocks_3_self_attn_k): LoRAModule(
131
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
132
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
133
+ )
134
+ (lora_unet__blocks_3_self_attn_v): LoRAModule(
135
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
136
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
137
+ )
138
+ (lora_unet__blocks_3_self_attn_o): LoRAModule(
139
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
140
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
141
+ )
142
+ (lora_unet__blocks_3_cross_attn_q): LoRAModule(
143
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
144
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
145
+ )
146
+ (lora_unet__blocks_3_cross_attn_k): LoRAModule(
147
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
148
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
149
+ )
150
+ (lora_unet__blocks_3_cross_attn_v): LoRAModule(
151
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
152
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
153
+ )
154
+ (lora_unet__blocks_3_cross_attn_o): LoRAModule(
155
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
156
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
157
+ )
158
+ (lora_unet__blocks_3_ffn_0): LoRAModule(
159
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
160
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
161
+ )
162
+ (lora_unet__blocks_3_ffn_2): LoRAModule(
163
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
164
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
165
+ )
166
+ (lora_unet__blocks_4_self_attn_q): LoRAModule(
167
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
168
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
169
+ )
170
+ (lora_unet__blocks_4_self_attn_k): LoRAModule(
171
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
172
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
173
+ )
174
+ (lora_unet__blocks_4_self_attn_v): LoRAModule(
175
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
176
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
177
+ )
178
+ (lora_unet__blocks_4_self_attn_o): LoRAModule(
179
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
180
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
181
+ )
182
+ (lora_unet__blocks_4_cross_attn_q): LoRAModule(
183
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
184
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
185
+ )
186
+ (lora_unet__blocks_4_cross_attn_k): LoRAModule(
187
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
188
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
189
+ )
190
+ (lora_unet__blocks_4_cross_attn_v): LoRAModule(
191
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
192
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
193
+ )
194
+ (lora_unet__blocks_4_cross_attn_o): LoRAModule(
195
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
196
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
197
+ )
198
+ (lora_unet__blocks_4_ffn_0): LoRAModule(
199
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
200
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
201
+ )
202
+ (lora_unet__blocks_4_ffn_2): LoRAModule(
203
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
204
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
205
+ )
206
+ (lora_unet__blocks_5_self_attn_q): LoRAModule(
207
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
208
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
209
+ )
210
+ (lora_unet__blocks_5_self_attn_k): LoRAModule(
211
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
212
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
213
+ )
214
+ (lora_unet__blocks_5_self_attn_v): LoRAModule(
215
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
216
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
217
+ )
218
+ (lora_unet__blocks_5_self_attn_o): LoRAModule(
219
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
220
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
221
+ )
222
+ (lora_unet__blocks_5_cross_attn_q): LoRAModule(
223
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
224
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
225
+ )
226
+ (lora_unet__blocks_5_cross_attn_k): LoRAModule(
227
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
228
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
229
+ )
230
+ (lora_unet__blocks_5_cross_attn_v): LoRAModule(
231
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
232
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
233
+ )
234
+ (lora_unet__blocks_5_cross_attn_o): LoRAModule(
235
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
236
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
237
+ )
238
+ (lora_unet__blocks_5_ffn_0): LoRAModule(
239
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
240
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
241
+ )
242
+ (lora_unet__blocks_5_ffn_2): LoRAModule(
243
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
244
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
245
+ )
246
+ (lora_unet__blocks_6_self_attn_q): LoRAModule(
247
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
248
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
249
+ )
250
+ (lora_unet__blocks_6_self_attn_k): LoRAModule(
251
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
252
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
253
+ )
254
+ (lora_unet__blocks_6_self_attn_v): LoRAModule(
255
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
256
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
257
+ )
258
+ (lora_unet__blocks_6_self_attn_o): LoRAModule(
259
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
260
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
261
+ )
262
+ (lora_unet__blocks_6_cross_attn_q): LoRAModule(
263
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
264
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
265
+ )
266
+ (lora_unet__blocks_6_cross_attn_k): LoRAModule(
267
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
268
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
269
+ )
270
+ (lora_unet__blocks_6_cross_attn_v): LoRAModule(
271
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
272
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
273
+ )
274
+ (lora_unet__blocks_6_cross_attn_o): LoRAModule(
275
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
276
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
277
+ )
278
+ (lora_unet__blocks_6_ffn_0): LoRAModule(
279
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
280
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
281
+ )
282
+ (lora_unet__blocks_6_ffn_2): LoRAModule(
283
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
284
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
285
+ )
286
+ (lora_unet__blocks_7_self_attn_q): LoRAModule(
287
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
288
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
289
+ )
290
+ (lora_unet__blocks_7_self_attn_k): LoRAModule(
291
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
292
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
293
+ )
294
+ (lora_unet__blocks_7_self_attn_v): LoRAModule(
295
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
296
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
297
+ )
298
+ (lora_unet__blocks_7_self_attn_o): LoRAModule(
299
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
300
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
301
+ )
302
+ (lora_unet__blocks_7_cross_attn_q): LoRAModule(
303
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
304
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
305
+ )
306
+ (lora_unet__blocks_7_cross_attn_k): LoRAModule(
307
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
308
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
309
+ )
310
+ (lora_unet__blocks_7_cross_attn_v): LoRAModule(
311
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
312
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
313
+ )
314
+ (lora_unet__blocks_7_cross_attn_o): LoRAModule(
315
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
316
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
317
+ )
318
+ (lora_unet__blocks_7_ffn_0): LoRAModule(
319
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
320
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
321
+ )
322
+ (lora_unet__blocks_7_ffn_2): LoRAModule(
323
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
324
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
325
+ )
326
+ (lora_unet__blocks_8_self_attn_q): LoRAModule(
327
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
328
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
329
+ )
330
+ (lora_unet__blocks_8_self_attn_k): LoRAModule(
331
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
332
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
333
+ )
334
+ (lora_unet__blocks_8_self_attn_v): LoRAModule(
335
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
336
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
337
+ )
338
+ (lora_unet__blocks_8_self_attn_o): LoRAModule(
339
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
340
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
341
+ )
342
+ (lora_unet__blocks_8_cross_attn_q): LoRAModule(
343
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
344
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
345
+ )
346
+ (lora_unet__blocks_8_cross_attn_k): LoRAModule(
347
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
348
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
349
+ )
350
+ (lora_unet__blocks_8_cross_attn_v): LoRAModule(
351
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
352
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
353
+ )
354
+ (lora_unet__blocks_8_cross_attn_o): LoRAModule(
355
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
356
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
357
+ )
358
+ (lora_unet__blocks_8_ffn_0): LoRAModule(
359
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
360
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
361
+ )
362
+ (lora_unet__blocks_8_ffn_2): LoRAModule(
363
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
364
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
365
+ )
366
+ (lora_unet__blocks_9_self_attn_q): LoRAModule(
367
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
368
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
369
+ )
370
+ (lora_unet__blocks_9_self_attn_k): LoRAModule(
371
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
372
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
373
+ )
374
+ (lora_unet__blocks_9_self_attn_v): LoRAModule(
375
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
376
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
377
+ )
378
+ (lora_unet__blocks_9_self_attn_o): LoRAModule(
379
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
380
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
381
+ )
382
+ (lora_unet__blocks_9_cross_attn_q): LoRAModule(
383
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
384
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
385
+ )
386
+ (lora_unet__blocks_9_cross_attn_k): LoRAModule(
387
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
388
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
389
+ )
390
+ (lora_unet__blocks_9_cross_attn_v): LoRAModule(
391
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
392
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
393
+ )
394
+ (lora_unet__blocks_9_cross_attn_o): LoRAModule(
395
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
396
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
397
+ )
398
+ (lora_unet__blocks_9_ffn_0): LoRAModule(
399
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
400
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
401
+ )
402
+ (lora_unet__blocks_9_ffn_2): LoRAModule(
403
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
404
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
405
+ )
406
+ (lora_unet__blocks_10_self_attn_q): LoRAModule(
407
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
408
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
409
+ )
410
+ (lora_unet__blocks_10_self_attn_k): LoRAModule(
411
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
412
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
413
+ )
414
+ (lora_unet__blocks_10_self_attn_v): LoRAModule(
415
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
416
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
417
+ )
418
+ (lora_unet__blocks_10_self_attn_o): LoRAModule(
419
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
420
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
421
+ )
422
+ (lora_unet__blocks_10_cross_attn_q): LoRAModule(
423
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
424
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
425
+ )
426
+ (lora_unet__blocks_10_cross_attn_k): LoRAModule(
427
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
428
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
429
+ )
430
+ (lora_unet__blocks_10_cross_attn_v): LoRAModule(
431
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
432
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
433
+ )
434
+ (lora_unet__blocks_10_cross_attn_o): LoRAModule(
435
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
436
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
437
+ )
438
+ (lora_unet__blocks_10_ffn_0): LoRAModule(
439
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
440
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
441
+ )
442
+ (lora_unet__blocks_10_ffn_2): LoRAModule(
443
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
444
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
445
+ )
446
+ (lora_unet__blocks_11_self_attn_q): LoRAModule(
447
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
448
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
449
+ )
450
+ (lora_unet__blocks_11_self_attn_k): LoRAModule(
451
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
452
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
453
+ )
454
+ (lora_unet__blocks_11_self_attn_v): LoRAModule(
455
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
456
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
457
+ )
458
+ (lora_unet__blocks_11_self_attn_o): LoRAModule(
459
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
460
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
461
+ )
462
+ (lora_unet__blocks_11_cross_attn_q): LoRAModule(
463
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
464
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
465
+ )
466
+ (lora_unet__blocks_11_cross_attn_k): LoRAModule(
467
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
468
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
469
+ )
470
+ (lora_unet__blocks_11_cross_attn_v): LoRAModule(
471
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
472
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
473
+ )
474
+ (lora_unet__blocks_11_cross_attn_o): LoRAModule(
475
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
476
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
477
+ )
478
+ (lora_unet__blocks_11_ffn_0): LoRAModule(
479
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
480
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
481
+ )
482
+ (lora_unet__blocks_11_ffn_2): LoRAModule(
483
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
484
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
485
+ )
486
+ (lora_unet__blocks_12_self_attn_q): LoRAModule(
487
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
488
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
489
+ )
490
+ (lora_unet__blocks_12_self_attn_k): LoRAModule(
491
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
492
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
493
+ )
494
+ (lora_unet__blocks_12_self_attn_v): LoRAModule(
495
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
496
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
497
+ )
498
+ (lora_unet__blocks_12_self_attn_o): LoRAModule(
499
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
500
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
501
+ )
502
+ (lora_unet__blocks_12_cross_attn_q): LoRAModule(
503
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
504
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
505
+ )
506
+ (lora_unet__blocks_12_cross_attn_k): LoRAModule(
507
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
508
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
509
+ )
510
+ (lora_unet__blocks_12_cross_attn_v): LoRAModule(
511
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
512
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
513
+ )
514
+ (lora_unet__blocks_12_cross_attn_o): LoRAModule(
515
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
516
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
517
+ )
518
+ (lora_unet__blocks_12_ffn_0): LoRAModule(
519
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
520
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
521
+ )
522
+ (lora_unet__blocks_12_ffn_2): LoRAModule(
523
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
524
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
525
+ )
526
+ (lora_unet__blocks_13_self_attn_q): LoRAModule(
527
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
528
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
529
+ )
530
+ (lora_unet__blocks_13_self_attn_k): LoRAModule(
531
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
532
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
533
+ )
534
+ (lora_unet__blocks_13_self_attn_v): LoRAModule(
535
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
536
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
537
+ )
538
+ (lora_unet__blocks_13_self_attn_o): LoRAModule(
539
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
540
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
541
+ )
542
+ (lora_unet__blocks_13_cross_attn_q): LoRAModule(
543
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
544
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
545
+ )
546
+ (lora_unet__blocks_13_cross_attn_k): LoRAModule(
547
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
548
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
549
+ )
550
+ (lora_unet__blocks_13_cross_attn_v): LoRAModule(
551
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
552
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
553
+ )
554
+ (lora_unet__blocks_13_cross_attn_o): LoRAModule(
555
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
556
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
557
+ )
558
+ (lora_unet__blocks_13_ffn_0): LoRAModule(
559
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
560
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
561
+ )
562
+ (lora_unet__blocks_13_ffn_2): LoRAModule(
563
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
564
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
565
+ )
566
+ (lora_unet__blocks_14_self_attn_q): LoRAModule(
567
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
568
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
569
+ )
570
+ (lora_unet__blocks_14_self_attn_k): LoRAModule(
571
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
572
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
573
+ )
574
+ (lora_unet__blocks_14_self_attn_v): LoRAModule(
575
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
576
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
577
+ )
578
+ (lora_unet__blocks_14_self_attn_o): LoRAModule(
579
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
580
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
581
+ )
582
+ (lora_unet__blocks_14_cross_attn_q): LoRAModule(
583
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
584
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
585
+ )
586
+ (lora_unet__blocks_14_cross_attn_k): LoRAModule(
587
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
588
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
589
+ )
590
+ (lora_unet__blocks_14_cross_attn_v): LoRAModule(
591
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
592
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
593
+ )
594
+ (lora_unet__blocks_14_cross_attn_o): LoRAModule(
595
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
596
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
597
+ )
598
+ (lora_unet__blocks_14_ffn_0): LoRAModule(
599
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
600
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
601
+ )
602
+ (lora_unet__blocks_14_ffn_2): LoRAModule(
603
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
604
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
605
+ )
606
+ (lora_unet__blocks_15_self_attn_q): LoRAModule(
607
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
608
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
609
+ )
610
+ (lora_unet__blocks_15_self_attn_k): LoRAModule(
611
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
612
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
613
+ )
614
+ (lora_unet__blocks_15_self_attn_v): LoRAModule(
615
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
616
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
617
+ )
618
+ (lora_unet__blocks_15_self_attn_o): LoRAModule(
619
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
620
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
621
+ )
622
+ (lora_unet__blocks_15_cross_attn_q): LoRAModule(
623
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
624
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
625
+ )
626
+ (lora_unet__blocks_15_cross_attn_k): LoRAModule(
627
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
628
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
629
+ )
630
+ (lora_unet__blocks_15_cross_attn_v): LoRAModule(
631
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
632
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
633
+ )
634
+ (lora_unet__blocks_15_cross_attn_o): LoRAModule(
635
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
636
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
637
+ )
638
+ (lora_unet__blocks_15_ffn_0): LoRAModule(
639
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
640
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
641
+ )
642
+ (lora_unet__blocks_15_ffn_2): LoRAModule(
643
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
644
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
645
+ )
646
+ (lora_unet__blocks_16_self_attn_q): LoRAModule(
647
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
648
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
649
+ )
650
+ (lora_unet__blocks_16_self_attn_k): LoRAModule(
651
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
652
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
653
+ )
654
+ (lora_unet__blocks_16_self_attn_v): LoRAModule(
655
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
656
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
657
+ )
658
+ (lora_unet__blocks_16_self_attn_o): LoRAModule(
659
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
660
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
661
+ )
662
+ (lora_unet__blocks_16_cross_attn_q): LoRAModule(
663
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
664
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
665
+ )
666
+ (lora_unet__blocks_16_cross_attn_k): LoRAModule(
667
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
668
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
669
+ )
670
+ (lora_unet__blocks_16_cross_attn_v): LoRAModule(
671
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
672
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
673
+ )
674
+ (lora_unet__blocks_16_cross_attn_o): LoRAModule(
675
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
676
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
677
+ )
678
+ (lora_unet__blocks_16_ffn_0): LoRAModule(
679
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
680
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
681
+ )
682
+ (lora_unet__blocks_16_ffn_2): LoRAModule(
683
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
684
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
685
+ )
686
+ (lora_unet__blocks_17_self_attn_q): LoRAModule(
687
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
688
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
689
+ )
690
+ (lora_unet__blocks_17_self_attn_k): LoRAModule(
691
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
692
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
693
+ )
694
+ (lora_unet__blocks_17_self_attn_v): LoRAModule(
695
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
696
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
697
+ )
698
+ (lora_unet__blocks_17_self_attn_o): LoRAModule(
699
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
700
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
701
+ )
702
+ (lora_unet__blocks_17_cross_attn_q): LoRAModule(
703
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
704
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
705
+ )
706
+ (lora_unet__blocks_17_cross_attn_k): LoRAModule(
707
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
708
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
709
+ )
710
+ (lora_unet__blocks_17_cross_attn_v): LoRAModule(
711
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
712
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
713
+ )
714
+ (lora_unet__blocks_17_cross_attn_o): LoRAModule(
715
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
716
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
717
+ )
718
+ (lora_unet__blocks_17_ffn_0): LoRAModule(
719
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
720
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
721
+ )
722
+ (lora_unet__blocks_17_ffn_2): LoRAModule(
723
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
724
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
725
+ )
726
+ (lora_unet__blocks_18_self_attn_q): LoRAModule(
727
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
728
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
729
+ )
730
+ (lora_unet__blocks_18_self_attn_k): LoRAModule(
731
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
732
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
733
+ )
734
+ (lora_unet__blocks_18_self_attn_v): LoRAModule(
735
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
736
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
737
+ )
738
+ (lora_unet__blocks_18_self_attn_o): LoRAModule(
739
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
740
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
741
+ )
742
+ (lora_unet__blocks_18_cross_attn_q): LoRAModule(
743
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
744
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
745
+ )
746
+ (lora_unet__blocks_18_cross_attn_k): LoRAModule(
747
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
748
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
749
+ )
750
+ (lora_unet__blocks_18_cross_attn_v): LoRAModule(
751
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
752
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
753
+ )
754
+ (lora_unet__blocks_18_cross_attn_o): LoRAModule(
755
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
756
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
757
+ )
758
+ (lora_unet__blocks_18_ffn_0): LoRAModule(
759
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
760
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
761
+ )
762
+ (lora_unet__blocks_18_ffn_2): LoRAModule(
763
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
764
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
765
+ )
766
+ (lora_unet__blocks_19_self_attn_q): LoRAModule(
767
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
768
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
769
+ )
770
+ (lora_unet__blocks_19_self_attn_k): LoRAModule(
771
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
772
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
773
+ )
774
+ (lora_unet__blocks_19_self_attn_v): LoRAModule(
775
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
776
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
777
+ )
778
+ (lora_unet__blocks_19_self_attn_o): LoRAModule(
779
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
780
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
781
+ )
782
+ (lora_unet__blocks_19_cross_attn_q): LoRAModule(
783
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
784
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
785
+ )
786
+ (lora_unet__blocks_19_cross_attn_k): LoRAModule(
787
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
788
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
789
+ )
790
+ (lora_unet__blocks_19_cross_attn_v): LoRAModule(
791
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
792
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
793
+ )
794
+ (lora_unet__blocks_19_cross_attn_o): LoRAModule(
795
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
796
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
797
+ )
798
+ (lora_unet__blocks_19_ffn_0): LoRAModule(
799
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
800
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
801
+ )
802
+ (lora_unet__blocks_19_ffn_2): LoRAModule(
803
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
804
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
805
+ )
806
+ (lora_unet__blocks_20_self_attn_q): LoRAModule(
807
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
808
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
809
+ )
810
+ (lora_unet__blocks_20_self_attn_k): LoRAModule(
811
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
812
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
813
+ )
814
+ (lora_unet__blocks_20_self_attn_v): LoRAModule(
815
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
816
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
817
+ )
818
+ (lora_unet__blocks_20_self_attn_o): LoRAModule(
819
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
820
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
821
+ )
822
+ (lora_unet__blocks_20_cross_attn_q): LoRAModule(
823
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
824
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
825
+ )
826
+ (lora_unet__blocks_20_cross_attn_k): LoRAModule(
827
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
828
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
829
+ )
830
+ (lora_unet__blocks_20_cross_attn_v): LoRAModule(
831
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
832
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
833
+ )
834
+ (lora_unet__blocks_20_cross_attn_o): LoRAModule(
835
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
836
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
837
+ )
838
+ (lora_unet__blocks_20_ffn_0): LoRAModule(
839
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
840
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
841
+ )
842
+ (lora_unet__blocks_20_ffn_2): LoRAModule(
843
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
844
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
845
+ )
846
+ (lora_unet__blocks_21_self_attn_q): LoRAModule(
847
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
848
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
849
+ )
850
+ (lora_unet__blocks_21_self_attn_k): LoRAModule(
851
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
852
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
853
+ )
854
+ (lora_unet__blocks_21_self_attn_v): LoRAModule(
855
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
856
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
857
+ )
858
+ (lora_unet__blocks_21_self_attn_o): LoRAModule(
859
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
860
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
861
+ )
862
+ (lora_unet__blocks_21_cross_attn_q): LoRAModule(
863
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
864
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
865
+ )
866
+ (lora_unet__blocks_21_cross_attn_k): LoRAModule(
867
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
868
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
869
+ )
870
+ (lora_unet__blocks_21_cross_attn_v): LoRAModule(
871
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
872
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
873
+ )
874
+ (lora_unet__blocks_21_cross_attn_o): LoRAModule(
875
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
876
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
877
+ )
878
+ (lora_unet__blocks_21_ffn_0): LoRAModule(
879
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
880
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
881
+ )
882
+ (lora_unet__blocks_21_ffn_2): LoRAModule(
883
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
884
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
885
+ )
886
+ (lora_unet__blocks_22_self_attn_q): LoRAModule(
887
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
888
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
889
+ )
890
+ (lora_unet__blocks_22_self_attn_k): LoRAModule(
891
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
892
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
893
+ )
894
+ (lora_unet__blocks_22_self_attn_v): LoRAModule(
895
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
896
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
897
+ )
898
+ (lora_unet__blocks_22_self_attn_o): LoRAModule(
899
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
900
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
901
+ )
902
+ (lora_unet__blocks_22_cross_attn_q): LoRAModule(
903
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
904
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
905
+ )
906
+ (lora_unet__blocks_22_cross_attn_k): LoRAModule(
907
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
908
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
909
+ )
910
+ (lora_unet__blocks_22_cross_attn_v): LoRAModule(
911
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
912
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
913
+ )
914
+ (lora_unet__blocks_22_cross_attn_o): LoRAModule(
915
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
916
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
917
+ )
918
+ (lora_unet__blocks_22_ffn_0): LoRAModule(
919
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
920
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
921
+ )
922
+ (lora_unet__blocks_22_ffn_2): LoRAModule(
923
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
924
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
925
+ )
926
+ (lora_unet__blocks_23_self_attn_q): LoRAModule(
927
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
928
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
929
+ )
930
+ (lora_unet__blocks_23_self_attn_k): LoRAModule(
931
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
932
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
933
+ )
934
+ (lora_unet__blocks_23_self_attn_v): LoRAModule(
935
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
936
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
937
+ )
938
+ (lora_unet__blocks_23_self_attn_o): LoRAModule(
939
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
940
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
941
+ )
942
+ (lora_unet__blocks_23_cross_attn_q): LoRAModule(
943
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
944
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
945
+ )
946
+ (lora_unet__blocks_23_cross_attn_k): LoRAModule(
947
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
948
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
949
+ )
950
+ (lora_unet__blocks_23_cross_attn_v): LoRAModule(
951
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
952
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
953
+ )
954
+ (lora_unet__blocks_23_cross_attn_o): LoRAModule(
955
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
956
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
957
+ )
958
+ (lora_unet__blocks_23_ffn_0): LoRAModule(
959
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
960
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
961
+ )
962
+ (lora_unet__blocks_23_ffn_2): LoRAModule(
963
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
964
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
965
+ )
966
+ (lora_unet__blocks_24_self_attn_q): LoRAModule(
967
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
968
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
969
+ )
970
+ (lora_unet__blocks_24_self_attn_k): LoRAModule(
971
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
972
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
973
+ )
974
+ (lora_unet__blocks_24_self_attn_v): LoRAModule(
975
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
976
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
977
+ )
978
+ (lora_unet__blocks_24_self_attn_o): LoRAModule(
979
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
980
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
981
+ )
982
+ (lora_unet__blocks_24_cross_attn_q): LoRAModule(
983
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
984
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
985
+ )
986
+ (lora_unet__blocks_24_cross_attn_k): LoRAModule(
987
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
988
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
989
+ )
990
+ (lora_unet__blocks_24_cross_attn_v): LoRAModule(
991
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
992
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
993
+ )
994
+ (lora_unet__blocks_24_cross_attn_o): LoRAModule(
995
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
996
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
997
+ )
998
+ (lora_unet__blocks_24_ffn_0): LoRAModule(
999
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1000
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1001
+ )
1002
+ (lora_unet__blocks_24_ffn_2): LoRAModule(
1003
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1004
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1005
+ )
1006
+ (lora_unet__blocks_25_self_attn_q): LoRAModule(
1007
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1008
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1009
+ )
1010
+ (lora_unet__blocks_25_self_attn_k): LoRAModule(
1011
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1012
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1013
+ )
1014
+ (lora_unet__blocks_25_self_attn_v): LoRAModule(
1015
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1016
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1017
+ )
1018
+ (lora_unet__blocks_25_self_attn_o): LoRAModule(
1019
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1020
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1021
+ )
1022
+ (lora_unet__blocks_25_cross_attn_q): LoRAModule(
1023
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1024
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1025
+ )
1026
+ (lora_unet__blocks_25_cross_attn_k): LoRAModule(
1027
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1028
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1029
+ )
1030
+ (lora_unet__blocks_25_cross_attn_v): LoRAModule(
1031
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1032
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1033
+ )
1034
+ (lora_unet__blocks_25_cross_attn_o): LoRAModule(
1035
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1036
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1037
+ )
1038
+ (lora_unet__blocks_25_ffn_0): LoRAModule(
1039
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1040
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1041
+ )
1042
+ (lora_unet__blocks_25_ffn_2): LoRAModule(
1043
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1044
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1045
+ )
1046
+ (lora_unet__blocks_26_self_attn_q): LoRAModule(
1047
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1048
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1049
+ )
1050
+ (lora_unet__blocks_26_self_attn_k): LoRAModule(
1051
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1052
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1053
+ )
1054
+ (lora_unet__blocks_26_self_attn_v): LoRAModule(
1055
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1056
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1057
+ )
1058
+ (lora_unet__blocks_26_self_attn_o): LoRAModule(
1059
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1060
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1061
+ )
1062
+ (lora_unet__blocks_26_cross_attn_q): LoRAModule(
1063
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1064
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1065
+ )
1066
+ (lora_unet__blocks_26_cross_attn_k): LoRAModule(
1067
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1068
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1069
+ )
1070
+ (lora_unet__blocks_26_cross_attn_v): LoRAModule(
1071
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1072
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1073
+ )
1074
+ (lora_unet__blocks_26_cross_attn_o): LoRAModule(
1075
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1076
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1077
+ )
1078
+ (lora_unet__blocks_26_ffn_0): LoRAModule(
1079
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1080
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1081
+ )
1082
+ (lora_unet__blocks_26_ffn_2): LoRAModule(
1083
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1084
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1085
+ )
1086
+ (lora_unet__blocks_27_self_attn_q): LoRAModule(
1087
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1088
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1089
+ )
1090
+ (lora_unet__blocks_27_self_attn_k): LoRAModule(
1091
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1092
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1093
+ )
1094
+ (lora_unet__blocks_27_self_attn_v): LoRAModule(
1095
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1096
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1097
+ )
1098
+ (lora_unet__blocks_27_self_attn_o): LoRAModule(
1099
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1100
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1101
+ )
1102
+ (lora_unet__blocks_27_cross_attn_q): LoRAModule(
1103
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1104
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1105
+ )
1106
+ (lora_unet__blocks_27_cross_attn_k): LoRAModule(
1107
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1108
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1109
+ )
1110
+ (lora_unet__blocks_27_cross_attn_v): LoRAModule(
1111
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1112
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1113
+ )
1114
+ (lora_unet__blocks_27_cross_attn_o): LoRAModule(
1115
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1116
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1117
+ )
1118
+ (lora_unet__blocks_27_ffn_0): LoRAModule(
1119
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1120
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1121
+ )
1122
+ (lora_unet__blocks_27_ffn_2): LoRAModule(
1123
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1124
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1125
+ )
1126
+ (lora_unet__blocks_28_self_attn_q): LoRAModule(
1127
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1128
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1129
+ )
1130
+ (lora_unet__blocks_28_self_attn_k): LoRAModule(
1131
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1132
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1133
+ )
1134
+ (lora_unet__blocks_28_self_attn_v): LoRAModule(
1135
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1136
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1137
+ )
1138
+ (lora_unet__blocks_28_self_attn_o): LoRAModule(
1139
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1140
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1141
+ )
1142
+ (lora_unet__blocks_28_cross_attn_q): LoRAModule(
1143
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1144
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1145
+ )
1146
+ (lora_unet__blocks_28_cross_attn_k): LoRAModule(
1147
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1148
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1149
+ )
1150
+ (lora_unet__blocks_28_cross_attn_v): LoRAModule(
1151
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1152
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1153
+ )
1154
+ (lora_unet__blocks_28_cross_attn_o): LoRAModule(
1155
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1156
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1157
+ )
1158
+ (lora_unet__blocks_28_ffn_0): LoRAModule(
1159
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1160
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1161
+ )
1162
+ (lora_unet__blocks_28_ffn_2): LoRAModule(
1163
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1164
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1165
+ )
1166
+ (lora_unet__blocks_29_self_attn_q): LoRAModule(
1167
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1168
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1169
+ )
1170
+ (lora_unet__blocks_29_self_attn_k): LoRAModule(
1171
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1172
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1173
+ )
1174
+ (lora_unet__blocks_29_self_attn_v): LoRAModule(
1175
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1176
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1177
+ )
1178
+ (lora_unet__blocks_29_self_attn_o): LoRAModule(
1179
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1180
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1181
+ )
1182
+ (lora_unet__blocks_29_cross_attn_q): LoRAModule(
1183
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1184
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1185
+ )
1186
+ (lora_unet__blocks_29_cross_attn_k): LoRAModule(
1187
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1188
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1189
+ )
1190
+ (lora_unet__blocks_29_cross_attn_v): LoRAModule(
1191
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1192
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1193
+ )
1194
+ (lora_unet__blocks_29_cross_attn_o): LoRAModule(
1195
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1196
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1197
+ )
1198
+ (lora_unet__blocks_29_ffn_0): LoRAModule(
1199
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1200
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1201
+ )
1202
+ (lora_unet__blocks_29_ffn_2): LoRAModule(
1203
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1204
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1205
+ )
1206
+ (lora_unet__blocks_30_self_attn_q): LoRAModule(
1207
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1208
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1209
+ )
1210
+ (lora_unet__blocks_30_self_attn_k): LoRAModule(
1211
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1212
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1213
+ )
1214
+ (lora_unet__blocks_30_self_attn_v): LoRAModule(
1215
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1216
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1217
+ )
1218
+ (lora_unet__blocks_30_self_attn_o): LoRAModule(
1219
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1220
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1221
+ )
1222
+ (lora_unet__blocks_30_cross_attn_q): LoRAModule(
1223
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1224
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1225
+ )
1226
+ (lora_unet__blocks_30_cross_attn_k): LoRAModule(
1227
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1228
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1229
+ )
1230
+ (lora_unet__blocks_30_cross_attn_v): LoRAModule(
1231
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1232
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1233
+ )
1234
+ (lora_unet__blocks_30_cross_attn_o): LoRAModule(
1235
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1236
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1237
+ )
1238
+ (lora_unet__blocks_30_ffn_0): LoRAModule(
1239
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1240
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1241
+ )
1242
+ (lora_unet__blocks_30_ffn_2): LoRAModule(
1243
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1244
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1245
+ )
1246
+ (lora_unet__blocks_31_self_attn_q): LoRAModule(
1247
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1248
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1249
+ )
1250
+ (lora_unet__blocks_31_self_attn_k): LoRAModule(
1251
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1252
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1253
+ )
1254
+ (lora_unet__blocks_31_self_attn_v): LoRAModule(
1255
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1256
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1257
+ )
1258
+ (lora_unet__blocks_31_self_attn_o): LoRAModule(
1259
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1260
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1261
+ )
1262
+ (lora_unet__blocks_31_cross_attn_q): LoRAModule(
1263
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1264
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1265
+ )
1266
+ (lora_unet__blocks_31_cross_attn_k): LoRAModule(
1267
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1268
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1269
+ )
1270
+ (lora_unet__blocks_31_cross_attn_v): LoRAModule(
1271
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1272
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1273
+ )
1274
+ (lora_unet__blocks_31_cross_attn_o): LoRAModule(
1275
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1276
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1277
+ )
1278
+ (lora_unet__blocks_31_ffn_0): LoRAModule(
1279
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1280
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1281
+ )
1282
+ (lora_unet__blocks_31_ffn_2): LoRAModule(
1283
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1284
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1285
+ )
1286
+ (lora_unet__blocks_32_self_attn_q): LoRAModule(
1287
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1288
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1289
+ )
1290
+ (lora_unet__blocks_32_self_attn_k): LoRAModule(
1291
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1292
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1293
+ )
1294
+ (lora_unet__blocks_32_self_attn_v): LoRAModule(
1295
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1296
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1297
+ )
1298
+ (lora_unet__blocks_32_self_attn_o): LoRAModule(
1299
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1300
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1301
+ )
1302
+ (lora_unet__blocks_32_cross_attn_q): LoRAModule(
1303
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1304
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1305
+ )
1306
+ (lora_unet__blocks_32_cross_attn_k): LoRAModule(
1307
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1308
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1309
+ )
1310
+ (lora_unet__blocks_32_cross_attn_v): LoRAModule(
1311
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1312
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1313
+ )
1314
+ (lora_unet__blocks_32_cross_attn_o): LoRAModule(
1315
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1316
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1317
+ )
1318
+ (lora_unet__blocks_32_ffn_0): LoRAModule(
1319
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1320
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1321
+ )
1322
+ (lora_unet__blocks_32_ffn_2): LoRAModule(
1323
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1324
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1325
+ )
1326
+ (lora_unet__blocks_33_self_attn_q): LoRAModule(
1327
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1328
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1329
+ )
1330
+ (lora_unet__blocks_33_self_attn_k): LoRAModule(
1331
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1332
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1333
+ )
1334
+ (lora_unet__blocks_33_self_attn_v): LoRAModule(
1335
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1336
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1337
+ )
1338
+ (lora_unet__blocks_33_self_attn_o): LoRAModule(
1339
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1340
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1341
+ )
1342
+ (lora_unet__blocks_33_cross_attn_q): LoRAModule(
1343
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1344
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1345
+ )
1346
+ (lora_unet__blocks_33_cross_attn_k): LoRAModule(
1347
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1348
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1349
+ )
1350
+ (lora_unet__blocks_33_cross_attn_v): LoRAModule(
1351
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1352
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1353
+ )
1354
+ (lora_unet__blocks_33_cross_attn_o): LoRAModule(
1355
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1356
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1357
+ )
1358
+ (lora_unet__blocks_33_ffn_0): LoRAModule(
1359
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1360
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1361
+ )
1362
+ (lora_unet__blocks_33_ffn_2): LoRAModule(
1363
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1364
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1365
+ )
1366
+ (lora_unet__blocks_34_self_attn_q): LoRAModule(
1367
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1368
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1369
+ )
1370
+ (lora_unet__blocks_34_self_attn_k): LoRAModule(
1371
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1372
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1373
+ )
1374
+ (lora_unet__blocks_34_self_attn_v): LoRAModule(
1375
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1376
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1377
+ )
1378
+ (lora_unet__blocks_34_self_attn_o): LoRAModule(
1379
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1380
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1381
+ )
1382
+ (lora_unet__blocks_34_cross_attn_q): LoRAModule(
1383
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1384
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1385
+ )
1386
+ (lora_unet__blocks_34_cross_attn_k): LoRAModule(
1387
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1388
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1389
+ )
1390
+ (lora_unet__blocks_34_cross_attn_v): LoRAModule(
1391
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1392
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1393
+ )
1394
+ (lora_unet__blocks_34_cross_attn_o): LoRAModule(
1395
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1396
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1397
+ )
1398
+ (lora_unet__blocks_34_ffn_0): LoRAModule(
1399
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1400
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1401
+ )
1402
+ (lora_unet__blocks_34_ffn_2): LoRAModule(
1403
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1404
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1405
+ )
1406
+ (lora_unet__blocks_35_self_attn_q): LoRAModule(
1407
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1408
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1409
+ )
1410
+ (lora_unet__blocks_35_self_attn_k): LoRAModule(
1411
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1412
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1413
+ )
1414
+ (lora_unet__blocks_35_self_attn_v): LoRAModule(
1415
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1416
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1417
+ )
1418
+ (lora_unet__blocks_35_self_attn_o): LoRAModule(
1419
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1420
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1421
+ )
1422
+ (lora_unet__blocks_35_cross_attn_q): LoRAModule(
1423
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1424
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1425
+ )
1426
+ (lora_unet__blocks_35_cross_attn_k): LoRAModule(
1427
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1428
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1429
+ )
1430
+ (lora_unet__blocks_35_cross_attn_v): LoRAModule(
1431
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1432
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1433
+ )
1434
+ (lora_unet__blocks_35_cross_attn_o): LoRAModule(
1435
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1436
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1437
+ )
1438
+ (lora_unet__blocks_35_ffn_0): LoRAModule(
1439
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1440
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1441
+ )
1442
+ (lora_unet__blocks_35_ffn_2): LoRAModule(
1443
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1444
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1445
+ )
1446
+ (lora_unet__blocks_36_self_attn_q): LoRAModule(
1447
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1448
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1449
+ )
1450
+ (lora_unet__blocks_36_self_attn_k): LoRAModule(
1451
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1452
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1453
+ )
1454
+ (lora_unet__blocks_36_self_attn_v): LoRAModule(
1455
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1456
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1457
+ )
1458
+ (lora_unet__blocks_36_self_attn_o): LoRAModule(
1459
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1460
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1461
+ )
1462
+ (lora_unet__blocks_36_cross_attn_q): LoRAModule(
1463
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1464
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1465
+ )
1466
+ (lora_unet__blocks_36_cross_attn_k): LoRAModule(
1467
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1468
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1469
+ )
1470
+ (lora_unet__blocks_36_cross_attn_v): LoRAModule(
1471
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1472
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1473
+ )
1474
+ (lora_unet__blocks_36_cross_attn_o): LoRAModule(
1475
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1476
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1477
+ )
1478
+ (lora_unet__blocks_36_ffn_0): LoRAModule(
1479
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1480
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1481
+ )
1482
+ (lora_unet__blocks_36_ffn_2): LoRAModule(
1483
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1484
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1485
+ )
1486
+ (lora_unet__blocks_37_self_attn_q): LoRAModule(
1487
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1488
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1489
+ )
1490
+ (lora_unet__blocks_37_self_attn_k): LoRAModule(
1491
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1492
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1493
+ )
1494
+ (lora_unet__blocks_37_self_attn_v): LoRAModule(
1495
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1496
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1497
+ )
1498
+ (lora_unet__blocks_37_self_attn_o): LoRAModule(
1499
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1500
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1501
+ )
1502
+ (lora_unet__blocks_37_cross_attn_q): LoRAModule(
1503
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1504
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1505
+ )
1506
+ (lora_unet__blocks_37_cross_attn_k): LoRAModule(
1507
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1508
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1509
+ )
1510
+ (lora_unet__blocks_37_cross_attn_v): LoRAModule(
1511
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1512
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1513
+ )
1514
+ (lora_unet__blocks_37_cross_attn_o): LoRAModule(
1515
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1516
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1517
+ )
1518
+ (lora_unet__blocks_37_ffn_0): LoRAModule(
1519
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1520
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1521
+ )
1522
+ (lora_unet__blocks_37_ffn_2): LoRAModule(
1523
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1524
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1525
+ )
1526
+ (lora_unet__blocks_38_self_attn_q): LoRAModule(
1527
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1528
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1529
+ )
1530
+ (lora_unet__blocks_38_self_attn_k): LoRAModule(
1531
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1532
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1533
+ )
1534
+ (lora_unet__blocks_38_self_attn_v): LoRAModule(
1535
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1536
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1537
+ )
1538
+ (lora_unet__blocks_38_self_attn_o): LoRAModule(
1539
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1540
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1541
+ )
1542
+ (lora_unet__blocks_38_cross_attn_q): LoRAModule(
1543
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1544
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1545
+ )
1546
+ (lora_unet__blocks_38_cross_attn_k): LoRAModule(
1547
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1548
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1549
+ )
1550
+ (lora_unet__blocks_38_cross_attn_v): LoRAModule(
1551
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1552
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1553
+ )
1554
+ (lora_unet__blocks_38_cross_attn_o): LoRAModule(
1555
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1556
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1557
+ )
1558
+ (lora_unet__blocks_38_ffn_0): LoRAModule(
1559
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1560
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1561
+ )
1562
+ (lora_unet__blocks_38_ffn_2): LoRAModule(
1563
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1564
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1565
+ )
1566
+ (lora_unet__blocks_39_self_attn_q): LoRAModule(
1567
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1568
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1569
+ )
1570
+ (lora_unet__blocks_39_self_attn_k): LoRAModule(
1571
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1572
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1573
+ )
1574
+ (lora_unet__blocks_39_self_attn_v): LoRAModule(
1575
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1576
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1577
+ )
1578
+ (lora_unet__blocks_39_self_attn_o): LoRAModule(
1579
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1580
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1581
+ )
1582
+ (lora_unet__blocks_39_cross_attn_q): LoRAModule(
1583
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1584
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1585
+ )
1586
+ (lora_unet__blocks_39_cross_attn_k): LoRAModule(
1587
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1588
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1589
+ )
1590
+ (lora_unet__blocks_39_cross_attn_v): LoRAModule(
1591
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1592
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1593
+ )
1594
+ (lora_unet__blocks_39_cross_attn_o): LoRAModule(
1595
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1596
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1597
+ )
1598
+ (lora_unet__blocks_39_ffn_0): LoRAModule(
1599
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1600
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1601
+ )
1602
+ (lora_unet__blocks_39_ffn_2): LoRAModule(
1603
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1604
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1605
+ )
1606
+ )
10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/ad23r2-the.gif ADDED

Git LFS Details

  • SHA256: 7bc46a6d13ba867ed57ddfca71e60830e382ef15a638b68b5f0fe535e69dab98
  • Pointer size: 132 Bytes
  • Size of remote file: 8.56 MB
10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/clip_ad23r2-the.png ADDED

Git LFS Details

  • SHA256: 73255781a48d8eb6307bca0e8e73a79ec50550b7d7d9d20362007b5f6fc71023
  • Pointer size: 131 Bytes
  • Size of remote file: 481 kB
10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/mask_ad23r2-the.gif ADDED

Git LFS Details

  • SHA256: a8d268dfc9cde6fd60bb72df8477e39360a8a10a86fd2e7e848cc3a20dfce46e
  • Pointer size: 131 Bytes
  • Size of remote file: 259 kB
10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/trainable_params.txt ADDED
@@ -0,0 +1,802 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ lora_unet__time_projection_1.lora_down.weight
2
+ lora_unet__time_projection_1.lora_up.weight
3
+ lora_unet__blocks_0_self_attn_q.lora_down.weight
4
+ lora_unet__blocks_0_self_attn_q.lora_up.weight
5
+ lora_unet__blocks_0_self_attn_k.lora_down.weight
6
+ lora_unet__blocks_0_self_attn_k.lora_up.weight
7
+ lora_unet__blocks_0_self_attn_v.lora_down.weight
8
+ lora_unet__blocks_0_self_attn_v.lora_up.weight
9
+ lora_unet__blocks_0_self_attn_o.lora_down.weight
10
+ lora_unet__blocks_0_self_attn_o.lora_up.weight
11
+ lora_unet__blocks_0_cross_attn_q.lora_down.weight
12
+ lora_unet__blocks_0_cross_attn_q.lora_up.weight
13
+ lora_unet__blocks_0_cross_attn_k.lora_down.weight
14
+ lora_unet__blocks_0_cross_attn_k.lora_up.weight
15
+ lora_unet__blocks_0_cross_attn_v.lora_down.weight
16
+ lora_unet__blocks_0_cross_attn_v.lora_up.weight
17
+ lora_unet__blocks_0_cross_attn_o.lora_down.weight
18
+ lora_unet__blocks_0_cross_attn_o.lora_up.weight
19
+ lora_unet__blocks_0_ffn_0.lora_down.weight
20
+ lora_unet__blocks_0_ffn_0.lora_up.weight
21
+ lora_unet__blocks_0_ffn_2.lora_down.weight
22
+ lora_unet__blocks_0_ffn_2.lora_up.weight
23
+ lora_unet__blocks_1_self_attn_q.lora_down.weight
24
+ lora_unet__blocks_1_self_attn_q.lora_up.weight
25
+ lora_unet__blocks_1_self_attn_k.lora_down.weight
26
+ lora_unet__blocks_1_self_attn_k.lora_up.weight
27
+ lora_unet__blocks_1_self_attn_v.lora_down.weight
28
+ lora_unet__blocks_1_self_attn_v.lora_up.weight
29
+ lora_unet__blocks_1_self_attn_o.lora_down.weight
30
+ lora_unet__blocks_1_self_attn_o.lora_up.weight
31
+ lora_unet__blocks_1_cross_attn_q.lora_down.weight
32
+ lora_unet__blocks_1_cross_attn_q.lora_up.weight
33
+ lora_unet__blocks_1_cross_attn_k.lora_down.weight
34
+ lora_unet__blocks_1_cross_attn_k.lora_up.weight
35
+ lora_unet__blocks_1_cross_attn_v.lora_down.weight
36
+ lora_unet__blocks_1_cross_attn_v.lora_up.weight
37
+ lora_unet__blocks_1_cross_attn_o.lora_down.weight
38
+ lora_unet__blocks_1_cross_attn_o.lora_up.weight
39
+ lora_unet__blocks_1_ffn_0.lora_down.weight
40
+ lora_unet__blocks_1_ffn_0.lora_up.weight
41
+ lora_unet__blocks_1_ffn_2.lora_down.weight
42
+ lora_unet__blocks_1_ffn_2.lora_up.weight
43
+ lora_unet__blocks_2_self_attn_q.lora_down.weight
44
+ lora_unet__blocks_2_self_attn_q.lora_up.weight
45
+ lora_unet__blocks_2_self_attn_k.lora_down.weight
46
+ lora_unet__blocks_2_self_attn_k.lora_up.weight
47
+ lora_unet__blocks_2_self_attn_v.lora_down.weight
48
+ lora_unet__blocks_2_self_attn_v.lora_up.weight
49
+ lora_unet__blocks_2_self_attn_o.lora_down.weight
50
+ lora_unet__blocks_2_self_attn_o.lora_up.weight
51
+ lora_unet__blocks_2_cross_attn_q.lora_down.weight
52
+ lora_unet__blocks_2_cross_attn_q.lora_up.weight
53
+ lora_unet__blocks_2_cross_attn_k.lora_down.weight
54
+ lora_unet__blocks_2_cross_attn_k.lora_up.weight
55
+ lora_unet__blocks_2_cross_attn_v.lora_down.weight
56
+ lora_unet__blocks_2_cross_attn_v.lora_up.weight
57
+ lora_unet__blocks_2_cross_attn_o.lora_down.weight
58
+ lora_unet__blocks_2_cross_attn_o.lora_up.weight
59
+ lora_unet__blocks_2_ffn_0.lora_down.weight
60
+ lora_unet__blocks_2_ffn_0.lora_up.weight
61
+ lora_unet__blocks_2_ffn_2.lora_down.weight
62
+ lora_unet__blocks_2_ffn_2.lora_up.weight
63
+ lora_unet__blocks_3_self_attn_q.lora_down.weight
64
+ lora_unet__blocks_3_self_attn_q.lora_up.weight
65
+ lora_unet__blocks_3_self_attn_k.lora_down.weight
66
+ lora_unet__blocks_3_self_attn_k.lora_up.weight
67
+ lora_unet__blocks_3_self_attn_v.lora_down.weight
68
+ lora_unet__blocks_3_self_attn_v.lora_up.weight
69
+ lora_unet__blocks_3_self_attn_o.lora_down.weight
70
+ lora_unet__blocks_3_self_attn_o.lora_up.weight
71
+ lora_unet__blocks_3_cross_attn_q.lora_down.weight
72
+ lora_unet__blocks_3_cross_attn_q.lora_up.weight
73
+ lora_unet__blocks_3_cross_attn_k.lora_down.weight
74
+ lora_unet__blocks_3_cross_attn_k.lora_up.weight
75
+ lora_unet__blocks_3_cross_attn_v.lora_down.weight
76
+ lora_unet__blocks_3_cross_attn_v.lora_up.weight
77
+ lora_unet__blocks_3_cross_attn_o.lora_down.weight
78
+ lora_unet__blocks_3_cross_attn_o.lora_up.weight
79
+ lora_unet__blocks_3_ffn_0.lora_down.weight
80
+ lora_unet__blocks_3_ffn_0.lora_up.weight
81
+ lora_unet__blocks_3_ffn_2.lora_down.weight
82
+ lora_unet__blocks_3_ffn_2.lora_up.weight
83
+ lora_unet__blocks_4_self_attn_q.lora_down.weight
84
+ lora_unet__blocks_4_self_attn_q.lora_up.weight
85
+ lora_unet__blocks_4_self_attn_k.lora_down.weight
86
+ lora_unet__blocks_4_self_attn_k.lora_up.weight
87
+ lora_unet__blocks_4_self_attn_v.lora_down.weight
88
+ lora_unet__blocks_4_self_attn_v.lora_up.weight
89
+ lora_unet__blocks_4_self_attn_o.lora_down.weight
90
+ lora_unet__blocks_4_self_attn_o.lora_up.weight
91
+ lora_unet__blocks_4_cross_attn_q.lora_down.weight
92
+ lora_unet__blocks_4_cross_attn_q.lora_up.weight
93
+ lora_unet__blocks_4_cross_attn_k.lora_down.weight
94
+ lora_unet__blocks_4_cross_attn_k.lora_up.weight
95
+ lora_unet__blocks_4_cross_attn_v.lora_down.weight
96
+ lora_unet__blocks_4_cross_attn_v.lora_up.weight
97
+ lora_unet__blocks_4_cross_attn_o.lora_down.weight
98
+ lora_unet__blocks_4_cross_attn_o.lora_up.weight
99
+ lora_unet__blocks_4_ffn_0.lora_down.weight
100
+ lora_unet__blocks_4_ffn_0.lora_up.weight
101
+ lora_unet__blocks_4_ffn_2.lora_down.weight
102
+ lora_unet__blocks_4_ffn_2.lora_up.weight
103
+ lora_unet__blocks_5_self_attn_q.lora_down.weight
104
+ lora_unet__blocks_5_self_attn_q.lora_up.weight
105
+ lora_unet__blocks_5_self_attn_k.lora_down.weight
106
+ lora_unet__blocks_5_self_attn_k.lora_up.weight
107
+ lora_unet__blocks_5_self_attn_v.lora_down.weight
108
+ lora_unet__blocks_5_self_attn_v.lora_up.weight
109
+ lora_unet__blocks_5_self_attn_o.lora_down.weight
110
+ lora_unet__blocks_5_self_attn_o.lora_up.weight
111
+ lora_unet__blocks_5_cross_attn_q.lora_down.weight
112
+ lora_unet__blocks_5_cross_attn_q.lora_up.weight
113
+ lora_unet__blocks_5_cross_attn_k.lora_down.weight
114
+ lora_unet__blocks_5_cross_attn_k.lora_up.weight
115
+ lora_unet__blocks_5_cross_attn_v.lora_down.weight
116
+ lora_unet__blocks_5_cross_attn_v.lora_up.weight
117
+ lora_unet__blocks_5_cross_attn_o.lora_down.weight
118
+ lora_unet__blocks_5_cross_attn_o.lora_up.weight
119
+ lora_unet__blocks_5_ffn_0.lora_down.weight
120
+ lora_unet__blocks_5_ffn_0.lora_up.weight
121
+ lora_unet__blocks_5_ffn_2.lora_down.weight
122
+ lora_unet__blocks_5_ffn_2.lora_up.weight
123
+ lora_unet__blocks_6_self_attn_q.lora_down.weight
124
+ lora_unet__blocks_6_self_attn_q.lora_up.weight
125
+ lora_unet__blocks_6_self_attn_k.lora_down.weight
126
+ lora_unet__blocks_6_self_attn_k.lora_up.weight
127
+ lora_unet__blocks_6_self_attn_v.lora_down.weight
128
+ lora_unet__blocks_6_self_attn_v.lora_up.weight
129
+ lora_unet__blocks_6_self_attn_o.lora_down.weight
130
+ lora_unet__blocks_6_self_attn_o.lora_up.weight
131
+ lora_unet__blocks_6_cross_attn_q.lora_down.weight
132
+ lora_unet__blocks_6_cross_attn_q.lora_up.weight
133
+ lora_unet__blocks_6_cross_attn_k.lora_down.weight
134
+ lora_unet__blocks_6_cross_attn_k.lora_up.weight
135
+ lora_unet__blocks_6_cross_attn_v.lora_down.weight
136
+ lora_unet__blocks_6_cross_attn_v.lora_up.weight
137
+ lora_unet__blocks_6_cross_attn_o.lora_down.weight
138
+ lora_unet__blocks_6_cross_attn_o.lora_up.weight
139
+ lora_unet__blocks_6_ffn_0.lora_down.weight
140
+ lora_unet__blocks_6_ffn_0.lora_up.weight
141
+ lora_unet__blocks_6_ffn_2.lora_down.weight
142
+ lora_unet__blocks_6_ffn_2.lora_up.weight
143
+ lora_unet__blocks_7_self_attn_q.lora_down.weight
144
+ lora_unet__blocks_7_self_attn_q.lora_up.weight
145
+ lora_unet__blocks_7_self_attn_k.lora_down.weight
146
+ lora_unet__blocks_7_self_attn_k.lora_up.weight
147
+ lora_unet__blocks_7_self_attn_v.lora_down.weight
148
+ lora_unet__blocks_7_self_attn_v.lora_up.weight
149
+ lora_unet__blocks_7_self_attn_o.lora_down.weight
150
+ lora_unet__blocks_7_self_attn_o.lora_up.weight
151
+ lora_unet__blocks_7_cross_attn_q.lora_down.weight
152
+ lora_unet__blocks_7_cross_attn_q.lora_up.weight
153
+ lora_unet__blocks_7_cross_attn_k.lora_down.weight
154
+ lora_unet__blocks_7_cross_attn_k.lora_up.weight
155
+ lora_unet__blocks_7_cross_attn_v.lora_down.weight
156
+ lora_unet__blocks_7_cross_attn_v.lora_up.weight
157
+ lora_unet__blocks_7_cross_attn_o.lora_down.weight
158
+ lora_unet__blocks_7_cross_attn_o.lora_up.weight
159
+ lora_unet__blocks_7_ffn_0.lora_down.weight
160
+ lora_unet__blocks_7_ffn_0.lora_up.weight
161
+ lora_unet__blocks_7_ffn_2.lora_down.weight
162
+ lora_unet__blocks_7_ffn_2.lora_up.weight
163
+ lora_unet__blocks_8_self_attn_q.lora_down.weight
164
+ lora_unet__blocks_8_self_attn_q.lora_up.weight
165
+ lora_unet__blocks_8_self_attn_k.lora_down.weight
166
+ lora_unet__blocks_8_self_attn_k.lora_up.weight
167
+ lora_unet__blocks_8_self_attn_v.lora_down.weight
168
+ lora_unet__blocks_8_self_attn_v.lora_up.weight
169
+ lora_unet__blocks_8_self_attn_o.lora_down.weight
170
+ lora_unet__blocks_8_self_attn_o.lora_up.weight
171
+ lora_unet__blocks_8_cross_attn_q.lora_down.weight
172
+ lora_unet__blocks_8_cross_attn_q.lora_up.weight
173
+ lora_unet__blocks_8_cross_attn_k.lora_down.weight
174
+ lora_unet__blocks_8_cross_attn_k.lora_up.weight
175
+ lora_unet__blocks_8_cross_attn_v.lora_down.weight
176
+ lora_unet__blocks_8_cross_attn_v.lora_up.weight
177
+ lora_unet__blocks_8_cross_attn_o.lora_down.weight
178
+ lora_unet__blocks_8_cross_attn_o.lora_up.weight
179
+ lora_unet__blocks_8_ffn_0.lora_down.weight
180
+ lora_unet__blocks_8_ffn_0.lora_up.weight
181
+ lora_unet__blocks_8_ffn_2.lora_down.weight
182
+ lora_unet__blocks_8_ffn_2.lora_up.weight
183
+ lora_unet__blocks_9_self_attn_q.lora_down.weight
184
+ lora_unet__blocks_9_self_attn_q.lora_up.weight
185
+ lora_unet__blocks_9_self_attn_k.lora_down.weight
186
+ lora_unet__blocks_9_self_attn_k.lora_up.weight
187
+ lora_unet__blocks_9_self_attn_v.lora_down.weight
188
+ lora_unet__blocks_9_self_attn_v.lora_up.weight
189
+ lora_unet__blocks_9_self_attn_o.lora_down.weight
190
+ lora_unet__blocks_9_self_attn_o.lora_up.weight
191
+ lora_unet__blocks_9_cross_attn_q.lora_down.weight
192
+ lora_unet__blocks_9_cross_attn_q.lora_up.weight
193
+ lora_unet__blocks_9_cross_attn_k.lora_down.weight
194
+ lora_unet__blocks_9_cross_attn_k.lora_up.weight
195
+ lora_unet__blocks_9_cross_attn_v.lora_down.weight
196
+ lora_unet__blocks_9_cross_attn_v.lora_up.weight
197
+ lora_unet__blocks_9_cross_attn_o.lora_down.weight
198
+ lora_unet__blocks_9_cross_attn_o.lora_up.weight
199
+ lora_unet__blocks_9_ffn_0.lora_down.weight
200
+ lora_unet__blocks_9_ffn_0.lora_up.weight
201
+ lora_unet__blocks_9_ffn_2.lora_down.weight
202
+ lora_unet__blocks_9_ffn_2.lora_up.weight
203
+ lora_unet__blocks_10_self_attn_q.lora_down.weight
204
+ lora_unet__blocks_10_self_attn_q.lora_up.weight
205
+ lora_unet__blocks_10_self_attn_k.lora_down.weight
206
+ lora_unet__blocks_10_self_attn_k.lora_up.weight
207
+ lora_unet__blocks_10_self_attn_v.lora_down.weight
208
+ lora_unet__blocks_10_self_attn_v.lora_up.weight
209
+ lora_unet__blocks_10_self_attn_o.lora_down.weight
210
+ lora_unet__blocks_10_self_attn_o.lora_up.weight
211
+ lora_unet__blocks_10_cross_attn_q.lora_down.weight
212
+ lora_unet__blocks_10_cross_attn_q.lora_up.weight
213
+ lora_unet__blocks_10_cross_attn_k.lora_down.weight
214
+ lora_unet__blocks_10_cross_attn_k.lora_up.weight
215
+ lora_unet__blocks_10_cross_attn_v.lora_down.weight
216
+ lora_unet__blocks_10_cross_attn_v.lora_up.weight
217
+ lora_unet__blocks_10_cross_attn_o.lora_down.weight
218
+ lora_unet__blocks_10_cross_attn_o.lora_up.weight
219
+ lora_unet__blocks_10_ffn_0.lora_down.weight
220
+ lora_unet__blocks_10_ffn_0.lora_up.weight
221
+ lora_unet__blocks_10_ffn_2.lora_down.weight
222
+ lora_unet__blocks_10_ffn_2.lora_up.weight
223
+ lora_unet__blocks_11_self_attn_q.lora_down.weight
224
+ lora_unet__blocks_11_self_attn_q.lora_up.weight
225
+ lora_unet__blocks_11_self_attn_k.lora_down.weight
226
+ lora_unet__blocks_11_self_attn_k.lora_up.weight
227
+ lora_unet__blocks_11_self_attn_v.lora_down.weight
228
+ lora_unet__blocks_11_self_attn_v.lora_up.weight
229
+ lora_unet__blocks_11_self_attn_o.lora_down.weight
230
+ lora_unet__blocks_11_self_attn_o.lora_up.weight
231
+ lora_unet__blocks_11_cross_attn_q.lora_down.weight
232
+ lora_unet__blocks_11_cross_attn_q.lora_up.weight
233
+ lora_unet__blocks_11_cross_attn_k.lora_down.weight
234
+ lora_unet__blocks_11_cross_attn_k.lora_up.weight
235
+ lora_unet__blocks_11_cross_attn_v.lora_down.weight
236
+ lora_unet__blocks_11_cross_attn_v.lora_up.weight
237
+ lora_unet__blocks_11_cross_attn_o.lora_down.weight
238
+ lora_unet__blocks_11_cross_attn_o.lora_up.weight
239
+ lora_unet__blocks_11_ffn_0.lora_down.weight
240
+ lora_unet__blocks_11_ffn_0.lora_up.weight
241
+ lora_unet__blocks_11_ffn_2.lora_down.weight
242
+ lora_unet__blocks_11_ffn_2.lora_up.weight
243
+ lora_unet__blocks_12_self_attn_q.lora_down.weight
244
+ lora_unet__blocks_12_self_attn_q.lora_up.weight
245
+ lora_unet__blocks_12_self_attn_k.lora_down.weight
246
+ lora_unet__blocks_12_self_attn_k.lora_up.weight
247
+ lora_unet__blocks_12_self_attn_v.lora_down.weight
248
+ lora_unet__blocks_12_self_attn_v.lora_up.weight
249
+ lora_unet__blocks_12_self_attn_o.lora_down.weight
250
+ lora_unet__blocks_12_self_attn_o.lora_up.weight
251
+ lora_unet__blocks_12_cross_attn_q.lora_down.weight
252
+ lora_unet__blocks_12_cross_attn_q.lora_up.weight
253
+ lora_unet__blocks_12_cross_attn_k.lora_down.weight
254
+ lora_unet__blocks_12_cross_attn_k.lora_up.weight
255
+ lora_unet__blocks_12_cross_attn_v.lora_down.weight
256
+ lora_unet__blocks_12_cross_attn_v.lora_up.weight
257
+ lora_unet__blocks_12_cross_attn_o.lora_down.weight
258
+ lora_unet__blocks_12_cross_attn_o.lora_up.weight
259
+ lora_unet__blocks_12_ffn_0.lora_down.weight
260
+ lora_unet__blocks_12_ffn_0.lora_up.weight
261
+ lora_unet__blocks_12_ffn_2.lora_down.weight
262
+ lora_unet__blocks_12_ffn_2.lora_up.weight
263
+ lora_unet__blocks_13_self_attn_q.lora_down.weight
264
+ lora_unet__blocks_13_self_attn_q.lora_up.weight
265
+ lora_unet__blocks_13_self_attn_k.lora_down.weight
266
+ lora_unet__blocks_13_self_attn_k.lora_up.weight
267
+ lora_unet__blocks_13_self_attn_v.lora_down.weight
268
+ lora_unet__blocks_13_self_attn_v.lora_up.weight
269
+ lora_unet__blocks_13_self_attn_o.lora_down.weight
270
+ lora_unet__blocks_13_self_attn_o.lora_up.weight
271
+ lora_unet__blocks_13_cross_attn_q.lora_down.weight
272
+ lora_unet__blocks_13_cross_attn_q.lora_up.weight
273
+ lora_unet__blocks_13_cross_attn_k.lora_down.weight
274
+ lora_unet__blocks_13_cross_attn_k.lora_up.weight
275
+ lora_unet__blocks_13_cross_attn_v.lora_down.weight
276
+ lora_unet__blocks_13_cross_attn_v.lora_up.weight
277
+ lora_unet__blocks_13_cross_attn_o.lora_down.weight
278
+ lora_unet__blocks_13_cross_attn_o.lora_up.weight
279
+ lora_unet__blocks_13_ffn_0.lora_down.weight
280
+ lora_unet__blocks_13_ffn_0.lora_up.weight
281
+ lora_unet__blocks_13_ffn_2.lora_down.weight
282
+ lora_unet__blocks_13_ffn_2.lora_up.weight
283
+ lora_unet__blocks_14_self_attn_q.lora_down.weight
284
+ lora_unet__blocks_14_self_attn_q.lora_up.weight
285
+ lora_unet__blocks_14_self_attn_k.lora_down.weight
286
+ lora_unet__blocks_14_self_attn_k.lora_up.weight
287
+ lora_unet__blocks_14_self_attn_v.lora_down.weight
288
+ lora_unet__blocks_14_self_attn_v.lora_up.weight
289
+ lora_unet__blocks_14_self_attn_o.lora_down.weight
290
+ lora_unet__blocks_14_self_attn_o.lora_up.weight
291
+ lora_unet__blocks_14_cross_attn_q.lora_down.weight
292
+ lora_unet__blocks_14_cross_attn_q.lora_up.weight
293
+ lora_unet__blocks_14_cross_attn_k.lora_down.weight
294
+ lora_unet__blocks_14_cross_attn_k.lora_up.weight
295
+ lora_unet__blocks_14_cross_attn_v.lora_down.weight
296
+ lora_unet__blocks_14_cross_attn_v.lora_up.weight
297
+ lora_unet__blocks_14_cross_attn_o.lora_down.weight
298
+ lora_unet__blocks_14_cross_attn_o.lora_up.weight
299
+ lora_unet__blocks_14_ffn_0.lora_down.weight
300
+ lora_unet__blocks_14_ffn_0.lora_up.weight
301
+ lora_unet__blocks_14_ffn_2.lora_down.weight
302
+ lora_unet__blocks_14_ffn_2.lora_up.weight
303
+ lora_unet__blocks_15_self_attn_q.lora_down.weight
304
+ lora_unet__blocks_15_self_attn_q.lora_up.weight
305
+ lora_unet__blocks_15_self_attn_k.lora_down.weight
306
+ lora_unet__blocks_15_self_attn_k.lora_up.weight
307
+ lora_unet__blocks_15_self_attn_v.lora_down.weight
308
+ lora_unet__blocks_15_self_attn_v.lora_up.weight
309
+ lora_unet__blocks_15_self_attn_o.lora_down.weight
310
+ lora_unet__blocks_15_self_attn_o.lora_up.weight
311
+ lora_unet__blocks_15_cross_attn_q.lora_down.weight
312
+ lora_unet__blocks_15_cross_attn_q.lora_up.weight
313
+ lora_unet__blocks_15_cross_attn_k.lora_down.weight
314
+ lora_unet__blocks_15_cross_attn_k.lora_up.weight
315
+ lora_unet__blocks_15_cross_attn_v.lora_down.weight
316
+ lora_unet__blocks_15_cross_attn_v.lora_up.weight
317
+ lora_unet__blocks_15_cross_attn_o.lora_down.weight
318
+ lora_unet__blocks_15_cross_attn_o.lora_up.weight
319
+ lora_unet__blocks_15_ffn_0.lora_down.weight
320
+ lora_unet__blocks_15_ffn_0.lora_up.weight
321
+ lora_unet__blocks_15_ffn_2.lora_down.weight
322
+ lora_unet__blocks_15_ffn_2.lora_up.weight
323
+ lora_unet__blocks_16_self_attn_q.lora_down.weight
324
+ lora_unet__blocks_16_self_attn_q.lora_up.weight
325
+ lora_unet__blocks_16_self_attn_k.lora_down.weight
326
+ lora_unet__blocks_16_self_attn_k.lora_up.weight
327
+ lora_unet__blocks_16_self_attn_v.lora_down.weight
328
+ lora_unet__blocks_16_self_attn_v.lora_up.weight
329
+ lora_unet__blocks_16_self_attn_o.lora_down.weight
330
+ lora_unet__blocks_16_self_attn_o.lora_up.weight
331
+ lora_unet__blocks_16_cross_attn_q.lora_down.weight
332
+ lora_unet__blocks_16_cross_attn_q.lora_up.weight
333
+ lora_unet__blocks_16_cross_attn_k.lora_down.weight
334
+ lora_unet__blocks_16_cross_attn_k.lora_up.weight
335
+ lora_unet__blocks_16_cross_attn_v.lora_down.weight
336
+ lora_unet__blocks_16_cross_attn_v.lora_up.weight
337
+ lora_unet__blocks_16_cross_attn_o.lora_down.weight
338
+ lora_unet__blocks_16_cross_attn_o.lora_up.weight
339
+ lora_unet__blocks_16_ffn_0.lora_down.weight
340
+ lora_unet__blocks_16_ffn_0.lora_up.weight
341
+ lora_unet__blocks_16_ffn_2.lora_down.weight
342
+ lora_unet__blocks_16_ffn_2.lora_up.weight
343
+ lora_unet__blocks_17_self_attn_q.lora_down.weight
344
+ lora_unet__blocks_17_self_attn_q.lora_up.weight
345
+ lora_unet__blocks_17_self_attn_k.lora_down.weight
346
+ lora_unet__blocks_17_self_attn_k.lora_up.weight
347
+ lora_unet__blocks_17_self_attn_v.lora_down.weight
348
+ lora_unet__blocks_17_self_attn_v.lora_up.weight
349
+ lora_unet__blocks_17_self_attn_o.lora_down.weight
350
+ lora_unet__blocks_17_self_attn_o.lora_up.weight
351
+ lora_unet__blocks_17_cross_attn_q.lora_down.weight
352
+ lora_unet__blocks_17_cross_attn_q.lora_up.weight
353
+ lora_unet__blocks_17_cross_attn_k.lora_down.weight
354
+ lora_unet__blocks_17_cross_attn_k.lora_up.weight
355
+ lora_unet__blocks_17_cross_attn_v.lora_down.weight
356
+ lora_unet__blocks_17_cross_attn_v.lora_up.weight
357
+ lora_unet__blocks_17_cross_attn_o.lora_down.weight
358
+ lora_unet__blocks_17_cross_attn_o.lora_up.weight
359
+ lora_unet__blocks_17_ffn_0.lora_down.weight
360
+ lora_unet__blocks_17_ffn_0.lora_up.weight
361
+ lora_unet__blocks_17_ffn_2.lora_down.weight
362
+ lora_unet__blocks_17_ffn_2.lora_up.weight
363
+ lora_unet__blocks_18_self_attn_q.lora_down.weight
364
+ lora_unet__blocks_18_self_attn_q.lora_up.weight
365
+ lora_unet__blocks_18_self_attn_k.lora_down.weight
366
+ lora_unet__blocks_18_self_attn_k.lora_up.weight
367
+ lora_unet__blocks_18_self_attn_v.lora_down.weight
368
+ lora_unet__blocks_18_self_attn_v.lora_up.weight
369
+ lora_unet__blocks_18_self_attn_o.lora_down.weight
370
+ lora_unet__blocks_18_self_attn_o.lora_up.weight
371
+ lora_unet__blocks_18_cross_attn_q.lora_down.weight
372
+ lora_unet__blocks_18_cross_attn_q.lora_up.weight
373
+ lora_unet__blocks_18_cross_attn_k.lora_down.weight
374
+ lora_unet__blocks_18_cross_attn_k.lora_up.weight
375
+ lora_unet__blocks_18_cross_attn_v.lora_down.weight
376
+ lora_unet__blocks_18_cross_attn_v.lora_up.weight
377
+ lora_unet__blocks_18_cross_attn_o.lora_down.weight
378
+ lora_unet__blocks_18_cross_attn_o.lora_up.weight
379
+ lora_unet__blocks_18_ffn_0.lora_down.weight
380
+ lora_unet__blocks_18_ffn_0.lora_up.weight
381
+ lora_unet__blocks_18_ffn_2.lora_down.weight
382
+ lora_unet__blocks_18_ffn_2.lora_up.weight
383
+ lora_unet__blocks_19_self_attn_q.lora_down.weight
384
+ lora_unet__blocks_19_self_attn_q.lora_up.weight
385
+ lora_unet__blocks_19_self_attn_k.lora_down.weight
386
+ lora_unet__blocks_19_self_attn_k.lora_up.weight
387
+ lora_unet__blocks_19_self_attn_v.lora_down.weight
388
+ lora_unet__blocks_19_self_attn_v.lora_up.weight
389
+ lora_unet__blocks_19_self_attn_o.lora_down.weight
390
+ lora_unet__blocks_19_self_attn_o.lora_up.weight
391
+ lora_unet__blocks_19_cross_attn_q.lora_down.weight
392
+ lora_unet__blocks_19_cross_attn_q.lora_up.weight
393
+ lora_unet__blocks_19_cross_attn_k.lora_down.weight
394
+ lora_unet__blocks_19_cross_attn_k.lora_up.weight
395
+ lora_unet__blocks_19_cross_attn_v.lora_down.weight
396
+ lora_unet__blocks_19_cross_attn_v.lora_up.weight
397
+ lora_unet__blocks_19_cross_attn_o.lora_down.weight
398
+ lora_unet__blocks_19_cross_attn_o.lora_up.weight
399
+ lora_unet__blocks_19_ffn_0.lora_down.weight
400
+ lora_unet__blocks_19_ffn_0.lora_up.weight
401
+ lora_unet__blocks_19_ffn_2.lora_down.weight
402
+ lora_unet__blocks_19_ffn_2.lora_up.weight
403
+ lora_unet__blocks_20_self_attn_q.lora_down.weight
404
+ lora_unet__blocks_20_self_attn_q.lora_up.weight
405
+ lora_unet__blocks_20_self_attn_k.lora_down.weight
406
+ lora_unet__blocks_20_self_attn_k.lora_up.weight
407
+ lora_unet__blocks_20_self_attn_v.lora_down.weight
408
+ lora_unet__blocks_20_self_attn_v.lora_up.weight
409
+ lora_unet__blocks_20_self_attn_o.lora_down.weight
410
+ lora_unet__blocks_20_self_attn_o.lora_up.weight
411
+ lora_unet__blocks_20_cross_attn_q.lora_down.weight
412
+ lora_unet__blocks_20_cross_attn_q.lora_up.weight
413
+ lora_unet__blocks_20_cross_attn_k.lora_down.weight
414
+ lora_unet__blocks_20_cross_attn_k.lora_up.weight
415
+ lora_unet__blocks_20_cross_attn_v.lora_down.weight
416
+ lora_unet__blocks_20_cross_attn_v.lora_up.weight
417
+ lora_unet__blocks_20_cross_attn_o.lora_down.weight
418
+ lora_unet__blocks_20_cross_attn_o.lora_up.weight
419
+ lora_unet__blocks_20_ffn_0.lora_down.weight
420
+ lora_unet__blocks_20_ffn_0.lora_up.weight
421
+ lora_unet__blocks_20_ffn_2.lora_down.weight
422
+ lora_unet__blocks_20_ffn_2.lora_up.weight
423
+ lora_unet__blocks_21_self_attn_q.lora_down.weight
424
+ lora_unet__blocks_21_self_attn_q.lora_up.weight
425
+ lora_unet__blocks_21_self_attn_k.lora_down.weight
426
+ lora_unet__blocks_21_self_attn_k.lora_up.weight
427
+ lora_unet__blocks_21_self_attn_v.lora_down.weight
428
+ lora_unet__blocks_21_self_attn_v.lora_up.weight
429
+ lora_unet__blocks_21_self_attn_o.lora_down.weight
430
+ lora_unet__blocks_21_self_attn_o.lora_up.weight
431
+ lora_unet__blocks_21_cross_attn_q.lora_down.weight
432
+ lora_unet__blocks_21_cross_attn_q.lora_up.weight
433
+ lora_unet__blocks_21_cross_attn_k.lora_down.weight
434
+ lora_unet__blocks_21_cross_attn_k.lora_up.weight
435
+ lora_unet__blocks_21_cross_attn_v.lora_down.weight
436
+ lora_unet__blocks_21_cross_attn_v.lora_up.weight
437
+ lora_unet__blocks_21_cross_attn_o.lora_down.weight
438
+ lora_unet__blocks_21_cross_attn_o.lora_up.weight
439
+ lora_unet__blocks_21_ffn_0.lora_down.weight
440
+ lora_unet__blocks_21_ffn_0.lora_up.weight
441
+ lora_unet__blocks_21_ffn_2.lora_down.weight
442
+ lora_unet__blocks_21_ffn_2.lora_up.weight
443
+ lora_unet__blocks_22_self_attn_q.lora_down.weight
444
+ lora_unet__blocks_22_self_attn_q.lora_up.weight
445
+ lora_unet__blocks_22_self_attn_k.lora_down.weight
446
+ lora_unet__blocks_22_self_attn_k.lora_up.weight
447
+ lora_unet__blocks_22_self_attn_v.lora_down.weight
448
+ lora_unet__blocks_22_self_attn_v.lora_up.weight
449
+ lora_unet__blocks_22_self_attn_o.lora_down.weight
450
+ lora_unet__blocks_22_self_attn_o.lora_up.weight
451
+ lora_unet__blocks_22_cross_attn_q.lora_down.weight
452
+ lora_unet__blocks_22_cross_attn_q.lora_up.weight
453
+ lora_unet__blocks_22_cross_attn_k.lora_down.weight
454
+ lora_unet__blocks_22_cross_attn_k.lora_up.weight
455
+ lora_unet__blocks_22_cross_attn_v.lora_down.weight
456
+ lora_unet__blocks_22_cross_attn_v.lora_up.weight
457
+ lora_unet__blocks_22_cross_attn_o.lora_down.weight
458
+ lora_unet__blocks_22_cross_attn_o.lora_up.weight
459
+ lora_unet__blocks_22_ffn_0.lora_down.weight
460
+ lora_unet__blocks_22_ffn_0.lora_up.weight
461
+ lora_unet__blocks_22_ffn_2.lora_down.weight
462
+ lora_unet__blocks_22_ffn_2.lora_up.weight
463
+ lora_unet__blocks_23_self_attn_q.lora_down.weight
464
+ lora_unet__blocks_23_self_attn_q.lora_up.weight
465
+ lora_unet__blocks_23_self_attn_k.lora_down.weight
466
+ lora_unet__blocks_23_self_attn_k.lora_up.weight
467
+ lora_unet__blocks_23_self_attn_v.lora_down.weight
468
+ lora_unet__blocks_23_self_attn_v.lora_up.weight
469
+ lora_unet__blocks_23_self_attn_o.lora_down.weight
470
+ lora_unet__blocks_23_self_attn_o.lora_up.weight
471
+ lora_unet__blocks_23_cross_attn_q.lora_down.weight
472
+ lora_unet__blocks_23_cross_attn_q.lora_up.weight
473
+ lora_unet__blocks_23_cross_attn_k.lora_down.weight
474
+ lora_unet__blocks_23_cross_attn_k.lora_up.weight
475
+ lora_unet__blocks_23_cross_attn_v.lora_down.weight
476
+ lora_unet__blocks_23_cross_attn_v.lora_up.weight
477
+ lora_unet__blocks_23_cross_attn_o.lora_down.weight
478
+ lora_unet__blocks_23_cross_attn_o.lora_up.weight
479
+ lora_unet__blocks_23_ffn_0.lora_down.weight
480
+ lora_unet__blocks_23_ffn_0.lora_up.weight
481
+ lora_unet__blocks_23_ffn_2.lora_down.weight
482
+ lora_unet__blocks_23_ffn_2.lora_up.weight
483
+ lora_unet__blocks_24_self_attn_q.lora_down.weight
484
+ lora_unet__blocks_24_self_attn_q.lora_up.weight
485
+ lora_unet__blocks_24_self_attn_k.lora_down.weight
486
+ lora_unet__blocks_24_self_attn_k.lora_up.weight
487
+ lora_unet__blocks_24_self_attn_v.lora_down.weight
488
+ lora_unet__blocks_24_self_attn_v.lora_up.weight
489
+ lora_unet__blocks_24_self_attn_o.lora_down.weight
490
+ lora_unet__blocks_24_self_attn_o.lora_up.weight
491
+ lora_unet__blocks_24_cross_attn_q.lora_down.weight
492
+ lora_unet__blocks_24_cross_attn_q.lora_up.weight
493
+ lora_unet__blocks_24_cross_attn_k.lora_down.weight
494
+ lora_unet__blocks_24_cross_attn_k.lora_up.weight
495
+ lora_unet__blocks_24_cross_attn_v.lora_down.weight
496
+ lora_unet__blocks_24_cross_attn_v.lora_up.weight
497
+ lora_unet__blocks_24_cross_attn_o.lora_down.weight
498
+ lora_unet__blocks_24_cross_attn_o.lora_up.weight
499
+ lora_unet__blocks_24_ffn_0.lora_down.weight
500
+ lora_unet__blocks_24_ffn_0.lora_up.weight
501
+ lora_unet__blocks_24_ffn_2.lora_down.weight
502
+ lora_unet__blocks_24_ffn_2.lora_up.weight
503
+ lora_unet__blocks_25_self_attn_q.lora_down.weight
504
+ lora_unet__blocks_25_self_attn_q.lora_up.weight
505
+ lora_unet__blocks_25_self_attn_k.lora_down.weight
506
+ lora_unet__blocks_25_self_attn_k.lora_up.weight
507
+ lora_unet__blocks_25_self_attn_v.lora_down.weight
508
+ lora_unet__blocks_25_self_attn_v.lora_up.weight
509
+ lora_unet__blocks_25_self_attn_o.lora_down.weight
510
+ lora_unet__blocks_25_self_attn_o.lora_up.weight
511
+ lora_unet__blocks_25_cross_attn_q.lora_down.weight
512
+ lora_unet__blocks_25_cross_attn_q.lora_up.weight
513
+ lora_unet__blocks_25_cross_attn_k.lora_down.weight
514
+ lora_unet__blocks_25_cross_attn_k.lora_up.weight
515
+ lora_unet__blocks_25_cross_attn_v.lora_down.weight
516
+ lora_unet__blocks_25_cross_attn_v.lora_up.weight
517
+ lora_unet__blocks_25_cross_attn_o.lora_down.weight
518
+ lora_unet__blocks_25_cross_attn_o.lora_up.weight
519
+ lora_unet__blocks_25_ffn_0.lora_down.weight
520
+ lora_unet__blocks_25_ffn_0.lora_up.weight
521
+ lora_unet__blocks_25_ffn_2.lora_down.weight
522
+ lora_unet__blocks_25_ffn_2.lora_up.weight
523
+ lora_unet__blocks_26_self_attn_q.lora_down.weight
524
+ lora_unet__blocks_26_self_attn_q.lora_up.weight
525
+ lora_unet__blocks_26_self_attn_k.lora_down.weight
526
+ lora_unet__blocks_26_self_attn_k.lora_up.weight
527
+ lora_unet__blocks_26_self_attn_v.lora_down.weight
528
+ lora_unet__blocks_26_self_attn_v.lora_up.weight
529
+ lora_unet__blocks_26_self_attn_o.lora_down.weight
530
+ lora_unet__blocks_26_self_attn_o.lora_up.weight
531
+ lora_unet__blocks_26_cross_attn_q.lora_down.weight
532
+ lora_unet__blocks_26_cross_attn_q.lora_up.weight
533
+ lora_unet__blocks_26_cross_attn_k.lora_down.weight
534
+ lora_unet__blocks_26_cross_attn_k.lora_up.weight
535
+ lora_unet__blocks_26_cross_attn_v.lora_down.weight
536
+ lora_unet__blocks_26_cross_attn_v.lora_up.weight
537
+ lora_unet__blocks_26_cross_attn_o.lora_down.weight
538
+ lora_unet__blocks_26_cross_attn_o.lora_up.weight
539
+ lora_unet__blocks_26_ffn_0.lora_down.weight
540
+ lora_unet__blocks_26_ffn_0.lora_up.weight
541
+ lora_unet__blocks_26_ffn_2.lora_down.weight
542
+ lora_unet__blocks_26_ffn_2.lora_up.weight
543
+ lora_unet__blocks_27_self_attn_q.lora_down.weight
544
+ lora_unet__blocks_27_self_attn_q.lora_up.weight
545
+ lora_unet__blocks_27_self_attn_k.lora_down.weight
546
+ lora_unet__blocks_27_self_attn_k.lora_up.weight
547
+ lora_unet__blocks_27_self_attn_v.lora_down.weight
548
+ lora_unet__blocks_27_self_attn_v.lora_up.weight
549
+ lora_unet__blocks_27_self_attn_o.lora_down.weight
550
+ lora_unet__blocks_27_self_attn_o.lora_up.weight
551
+ lora_unet__blocks_27_cross_attn_q.lora_down.weight
552
+ lora_unet__blocks_27_cross_attn_q.lora_up.weight
553
+ lora_unet__blocks_27_cross_attn_k.lora_down.weight
554
+ lora_unet__blocks_27_cross_attn_k.lora_up.weight
555
+ lora_unet__blocks_27_cross_attn_v.lora_down.weight
556
+ lora_unet__blocks_27_cross_attn_v.lora_up.weight
557
+ lora_unet__blocks_27_cross_attn_o.lora_down.weight
558
+ lora_unet__blocks_27_cross_attn_o.lora_up.weight
559
+ lora_unet__blocks_27_ffn_0.lora_down.weight
560
+ lora_unet__blocks_27_ffn_0.lora_up.weight
561
+ lora_unet__blocks_27_ffn_2.lora_down.weight
562
+ lora_unet__blocks_27_ffn_2.lora_up.weight
563
+ lora_unet__blocks_28_self_attn_q.lora_down.weight
564
+ lora_unet__blocks_28_self_attn_q.lora_up.weight
565
+ lora_unet__blocks_28_self_attn_k.lora_down.weight
566
+ lora_unet__blocks_28_self_attn_k.lora_up.weight
567
+ lora_unet__blocks_28_self_attn_v.lora_down.weight
568
+ lora_unet__blocks_28_self_attn_v.lora_up.weight
569
+ lora_unet__blocks_28_self_attn_o.lora_down.weight
570
+ lora_unet__blocks_28_self_attn_o.lora_up.weight
571
+ lora_unet__blocks_28_cross_attn_q.lora_down.weight
572
+ lora_unet__blocks_28_cross_attn_q.lora_up.weight
573
+ lora_unet__blocks_28_cross_attn_k.lora_down.weight
574
+ lora_unet__blocks_28_cross_attn_k.lora_up.weight
575
+ lora_unet__blocks_28_cross_attn_v.lora_down.weight
576
+ lora_unet__blocks_28_cross_attn_v.lora_up.weight
577
+ lora_unet__blocks_28_cross_attn_o.lora_down.weight
578
+ lora_unet__blocks_28_cross_attn_o.lora_up.weight
579
+ lora_unet__blocks_28_ffn_0.lora_down.weight
580
+ lora_unet__blocks_28_ffn_0.lora_up.weight
581
+ lora_unet__blocks_28_ffn_2.lora_down.weight
582
+ lora_unet__blocks_28_ffn_2.lora_up.weight
583
+ lora_unet__blocks_29_self_attn_q.lora_down.weight
584
+ lora_unet__blocks_29_self_attn_q.lora_up.weight
585
+ lora_unet__blocks_29_self_attn_k.lora_down.weight
586
+ lora_unet__blocks_29_self_attn_k.lora_up.weight
587
+ lora_unet__blocks_29_self_attn_v.lora_down.weight
588
+ lora_unet__blocks_29_self_attn_v.lora_up.weight
589
+ lora_unet__blocks_29_self_attn_o.lora_down.weight
590
+ lora_unet__blocks_29_self_attn_o.lora_up.weight
591
+ lora_unet__blocks_29_cross_attn_q.lora_down.weight
592
+ lora_unet__blocks_29_cross_attn_q.lora_up.weight
593
+ lora_unet__blocks_29_cross_attn_k.lora_down.weight
594
+ lora_unet__blocks_29_cross_attn_k.lora_up.weight
595
+ lora_unet__blocks_29_cross_attn_v.lora_down.weight
596
+ lora_unet__blocks_29_cross_attn_v.lora_up.weight
597
+ lora_unet__blocks_29_cross_attn_o.lora_down.weight
598
+ lora_unet__blocks_29_cross_attn_o.lora_up.weight
599
+ lora_unet__blocks_29_ffn_0.lora_down.weight
600
+ lora_unet__blocks_29_ffn_0.lora_up.weight
601
+ lora_unet__blocks_29_ffn_2.lora_down.weight
602
+ lora_unet__blocks_29_ffn_2.lora_up.weight
603
+ lora_unet__blocks_30_self_attn_q.lora_down.weight
604
+ lora_unet__blocks_30_self_attn_q.lora_up.weight
605
+ lora_unet__blocks_30_self_attn_k.lora_down.weight
606
+ lora_unet__blocks_30_self_attn_k.lora_up.weight
607
+ lora_unet__blocks_30_self_attn_v.lora_down.weight
608
+ lora_unet__blocks_30_self_attn_v.lora_up.weight
609
+ lora_unet__blocks_30_self_attn_o.lora_down.weight
610
+ lora_unet__blocks_30_self_attn_o.lora_up.weight
611
+ lora_unet__blocks_30_cross_attn_q.lora_down.weight
612
+ lora_unet__blocks_30_cross_attn_q.lora_up.weight
613
+ lora_unet__blocks_30_cross_attn_k.lora_down.weight
614
+ lora_unet__blocks_30_cross_attn_k.lora_up.weight
615
+ lora_unet__blocks_30_cross_attn_v.lora_down.weight
616
+ lora_unet__blocks_30_cross_attn_v.lora_up.weight
617
+ lora_unet__blocks_30_cross_attn_o.lora_down.weight
618
+ lora_unet__blocks_30_cross_attn_o.lora_up.weight
619
+ lora_unet__blocks_30_ffn_0.lora_down.weight
620
+ lora_unet__blocks_30_ffn_0.lora_up.weight
621
+ lora_unet__blocks_30_ffn_2.lora_down.weight
622
+ lora_unet__blocks_30_ffn_2.lora_up.weight
623
+ lora_unet__blocks_31_self_attn_q.lora_down.weight
624
+ lora_unet__blocks_31_self_attn_q.lora_up.weight
625
+ lora_unet__blocks_31_self_attn_k.lora_down.weight
626
+ lora_unet__blocks_31_self_attn_k.lora_up.weight
627
+ lora_unet__blocks_31_self_attn_v.lora_down.weight
628
+ lora_unet__blocks_31_self_attn_v.lora_up.weight
629
+ lora_unet__blocks_31_self_attn_o.lora_down.weight
630
+ lora_unet__blocks_31_self_attn_o.lora_up.weight
631
+ lora_unet__blocks_31_cross_attn_q.lora_down.weight
632
+ lora_unet__blocks_31_cross_attn_q.lora_up.weight
633
+ lora_unet__blocks_31_cross_attn_k.lora_down.weight
634
+ lora_unet__blocks_31_cross_attn_k.lora_up.weight
635
+ lora_unet__blocks_31_cross_attn_v.lora_down.weight
636
+ lora_unet__blocks_31_cross_attn_v.lora_up.weight
637
+ lora_unet__blocks_31_cross_attn_o.lora_down.weight
638
+ lora_unet__blocks_31_cross_attn_o.lora_up.weight
639
+ lora_unet__blocks_31_ffn_0.lora_down.weight
640
+ lora_unet__blocks_31_ffn_0.lora_up.weight
641
+ lora_unet__blocks_31_ffn_2.lora_down.weight
642
+ lora_unet__blocks_31_ffn_2.lora_up.weight
643
+ lora_unet__blocks_32_self_attn_q.lora_down.weight
644
+ lora_unet__blocks_32_self_attn_q.lora_up.weight
645
+ lora_unet__blocks_32_self_attn_k.lora_down.weight
646
+ lora_unet__blocks_32_self_attn_k.lora_up.weight
647
+ lora_unet__blocks_32_self_attn_v.lora_down.weight
648
+ lora_unet__blocks_32_self_attn_v.lora_up.weight
649
+ lora_unet__blocks_32_self_attn_o.lora_down.weight
650
+ lora_unet__blocks_32_self_attn_o.lora_up.weight
651
+ lora_unet__blocks_32_cross_attn_q.lora_down.weight
652
+ lora_unet__blocks_32_cross_attn_q.lora_up.weight
653
+ lora_unet__blocks_32_cross_attn_k.lora_down.weight
654
+ lora_unet__blocks_32_cross_attn_k.lora_up.weight
655
+ lora_unet__blocks_32_cross_attn_v.lora_down.weight
656
+ lora_unet__blocks_32_cross_attn_v.lora_up.weight
657
+ lora_unet__blocks_32_cross_attn_o.lora_down.weight
658
+ lora_unet__blocks_32_cross_attn_o.lora_up.weight
659
+ lora_unet__blocks_32_ffn_0.lora_down.weight
660
+ lora_unet__blocks_32_ffn_0.lora_up.weight
661
+ lora_unet__blocks_32_ffn_2.lora_down.weight
662
+ lora_unet__blocks_32_ffn_2.lora_up.weight
663
+ lora_unet__blocks_33_self_attn_q.lora_down.weight
664
+ lora_unet__blocks_33_self_attn_q.lora_up.weight
665
+ lora_unet__blocks_33_self_attn_k.lora_down.weight
666
+ lora_unet__blocks_33_self_attn_k.lora_up.weight
667
+ lora_unet__blocks_33_self_attn_v.lora_down.weight
668
+ lora_unet__blocks_33_self_attn_v.lora_up.weight
669
+ lora_unet__blocks_33_self_attn_o.lora_down.weight
670
+ lora_unet__blocks_33_self_attn_o.lora_up.weight
671
+ lora_unet__blocks_33_cross_attn_q.lora_down.weight
672
+ lora_unet__blocks_33_cross_attn_q.lora_up.weight
673
+ lora_unet__blocks_33_cross_attn_k.lora_down.weight
674
+ lora_unet__blocks_33_cross_attn_k.lora_up.weight
675
+ lora_unet__blocks_33_cross_attn_v.lora_down.weight
676
+ lora_unet__blocks_33_cross_attn_v.lora_up.weight
677
+ lora_unet__blocks_33_cross_attn_o.lora_down.weight
678
+ lora_unet__blocks_33_cross_attn_o.lora_up.weight
679
+ lora_unet__blocks_33_ffn_0.lora_down.weight
680
+ lora_unet__blocks_33_ffn_0.lora_up.weight
681
+ lora_unet__blocks_33_ffn_2.lora_down.weight
682
+ lora_unet__blocks_33_ffn_2.lora_up.weight
683
+ lora_unet__blocks_34_self_attn_q.lora_down.weight
684
+ lora_unet__blocks_34_self_attn_q.lora_up.weight
685
+ lora_unet__blocks_34_self_attn_k.lora_down.weight
686
+ lora_unet__blocks_34_self_attn_k.lora_up.weight
687
+ lora_unet__blocks_34_self_attn_v.lora_down.weight
688
+ lora_unet__blocks_34_self_attn_v.lora_up.weight
689
+ lora_unet__blocks_34_self_attn_o.lora_down.weight
690
+ lora_unet__blocks_34_self_attn_o.lora_up.weight
691
+ lora_unet__blocks_34_cross_attn_q.lora_down.weight
692
+ lora_unet__blocks_34_cross_attn_q.lora_up.weight
693
+ lora_unet__blocks_34_cross_attn_k.lora_down.weight
694
+ lora_unet__blocks_34_cross_attn_k.lora_up.weight
695
+ lora_unet__blocks_34_cross_attn_v.lora_down.weight
696
+ lora_unet__blocks_34_cross_attn_v.lora_up.weight
697
+ lora_unet__blocks_34_cross_attn_o.lora_down.weight
698
+ lora_unet__blocks_34_cross_attn_o.lora_up.weight
699
+ lora_unet__blocks_34_ffn_0.lora_down.weight
700
+ lora_unet__blocks_34_ffn_0.lora_up.weight
701
+ lora_unet__blocks_34_ffn_2.lora_down.weight
702
+ lora_unet__blocks_34_ffn_2.lora_up.weight
703
+ lora_unet__blocks_35_self_attn_q.lora_down.weight
704
+ lora_unet__blocks_35_self_attn_q.lora_up.weight
705
+ lora_unet__blocks_35_self_attn_k.lora_down.weight
706
+ lora_unet__blocks_35_self_attn_k.lora_up.weight
707
+ lora_unet__blocks_35_self_attn_v.lora_down.weight
708
+ lora_unet__blocks_35_self_attn_v.lora_up.weight
709
+ lora_unet__blocks_35_self_attn_o.lora_down.weight
710
+ lora_unet__blocks_35_self_attn_o.lora_up.weight
711
+ lora_unet__blocks_35_cross_attn_q.lora_down.weight
712
+ lora_unet__blocks_35_cross_attn_q.lora_up.weight
713
+ lora_unet__blocks_35_cross_attn_k.lora_down.weight
714
+ lora_unet__blocks_35_cross_attn_k.lora_up.weight
715
+ lora_unet__blocks_35_cross_attn_v.lora_down.weight
716
+ lora_unet__blocks_35_cross_attn_v.lora_up.weight
717
+ lora_unet__blocks_35_cross_attn_o.lora_down.weight
718
+ lora_unet__blocks_35_cross_attn_o.lora_up.weight
719
+ lora_unet__blocks_35_ffn_0.lora_down.weight
720
+ lora_unet__blocks_35_ffn_0.lora_up.weight
721
+ lora_unet__blocks_35_ffn_2.lora_down.weight
722
+ lora_unet__blocks_35_ffn_2.lora_up.weight
723
+ lora_unet__blocks_36_self_attn_q.lora_down.weight
724
+ lora_unet__blocks_36_self_attn_q.lora_up.weight
725
+ lora_unet__blocks_36_self_attn_k.lora_down.weight
726
+ lora_unet__blocks_36_self_attn_k.lora_up.weight
727
+ lora_unet__blocks_36_self_attn_v.lora_down.weight
728
+ lora_unet__blocks_36_self_attn_v.lora_up.weight
729
+ lora_unet__blocks_36_self_attn_o.lora_down.weight
730
+ lora_unet__blocks_36_self_attn_o.lora_up.weight
731
+ lora_unet__blocks_36_cross_attn_q.lora_down.weight
732
+ lora_unet__blocks_36_cross_attn_q.lora_up.weight
733
+ lora_unet__blocks_36_cross_attn_k.lora_down.weight
734
+ lora_unet__blocks_36_cross_attn_k.lora_up.weight
735
+ lora_unet__blocks_36_cross_attn_v.lora_down.weight
736
+ lora_unet__blocks_36_cross_attn_v.lora_up.weight
737
+ lora_unet__blocks_36_cross_attn_o.lora_down.weight
738
+ lora_unet__blocks_36_cross_attn_o.lora_up.weight
739
+ lora_unet__blocks_36_ffn_0.lora_down.weight
740
+ lora_unet__blocks_36_ffn_0.lora_up.weight
741
+ lora_unet__blocks_36_ffn_2.lora_down.weight
742
+ lora_unet__blocks_36_ffn_2.lora_up.weight
743
+ lora_unet__blocks_37_self_attn_q.lora_down.weight
744
+ lora_unet__blocks_37_self_attn_q.lora_up.weight
745
+ lora_unet__blocks_37_self_attn_k.lora_down.weight
746
+ lora_unet__blocks_37_self_attn_k.lora_up.weight
747
+ lora_unet__blocks_37_self_attn_v.lora_down.weight
748
+ lora_unet__blocks_37_self_attn_v.lora_up.weight
749
+ lora_unet__blocks_37_self_attn_o.lora_down.weight
750
+ lora_unet__blocks_37_self_attn_o.lora_up.weight
751
+ lora_unet__blocks_37_cross_attn_q.lora_down.weight
752
+ lora_unet__blocks_37_cross_attn_q.lora_up.weight
753
+ lora_unet__blocks_37_cross_attn_k.lora_down.weight
754
+ lora_unet__blocks_37_cross_attn_k.lora_up.weight
755
+ lora_unet__blocks_37_cross_attn_v.lora_down.weight
756
+ lora_unet__blocks_37_cross_attn_v.lora_up.weight
757
+ lora_unet__blocks_37_cross_attn_o.lora_down.weight
758
+ lora_unet__blocks_37_cross_attn_o.lora_up.weight
759
+ lora_unet__blocks_37_ffn_0.lora_down.weight
760
+ lora_unet__blocks_37_ffn_0.lora_up.weight
761
+ lora_unet__blocks_37_ffn_2.lora_down.weight
762
+ lora_unet__blocks_37_ffn_2.lora_up.weight
763
+ lora_unet__blocks_38_self_attn_q.lora_down.weight
764
+ lora_unet__blocks_38_self_attn_q.lora_up.weight
765
+ lora_unet__blocks_38_self_attn_k.lora_down.weight
766
+ lora_unet__blocks_38_self_attn_k.lora_up.weight
767
+ lora_unet__blocks_38_self_attn_v.lora_down.weight
768
+ lora_unet__blocks_38_self_attn_v.lora_up.weight
769
+ lora_unet__blocks_38_self_attn_o.lora_down.weight
770
+ lora_unet__blocks_38_self_attn_o.lora_up.weight
771
+ lora_unet__blocks_38_cross_attn_q.lora_down.weight
772
+ lora_unet__blocks_38_cross_attn_q.lora_up.weight
773
+ lora_unet__blocks_38_cross_attn_k.lora_down.weight
774
+ lora_unet__blocks_38_cross_attn_k.lora_up.weight
775
+ lora_unet__blocks_38_cross_attn_v.lora_down.weight
776
+ lora_unet__blocks_38_cross_attn_v.lora_up.weight
777
+ lora_unet__blocks_38_cross_attn_o.lora_down.weight
778
+ lora_unet__blocks_38_cross_attn_o.lora_up.weight
779
+ lora_unet__blocks_38_ffn_0.lora_down.weight
780
+ lora_unet__blocks_38_ffn_0.lora_up.weight
781
+ lora_unet__blocks_38_ffn_2.lora_down.weight
782
+ lora_unet__blocks_38_ffn_2.lora_up.weight
783
+ lora_unet__blocks_39_self_attn_q.lora_down.weight
784
+ lora_unet__blocks_39_self_attn_q.lora_up.weight
785
+ lora_unet__blocks_39_self_attn_k.lora_down.weight
786
+ lora_unet__blocks_39_self_attn_k.lora_up.weight
787
+ lora_unet__blocks_39_self_attn_v.lora_down.weight
788
+ lora_unet__blocks_39_self_attn_v.lora_up.weight
789
+ lora_unet__blocks_39_self_attn_o.lora_down.weight
790
+ lora_unet__blocks_39_self_attn_o.lora_up.weight
791
+ lora_unet__blocks_39_cross_attn_q.lora_down.weight
792
+ lora_unet__blocks_39_cross_attn_q.lora_up.weight
793
+ lora_unet__blocks_39_cross_attn_k.lora_down.weight
794
+ lora_unet__blocks_39_cross_attn_k.lora_up.weight
795
+ lora_unet__blocks_39_cross_attn_v.lora_down.weight
796
+ lora_unet__blocks_39_cross_attn_v.lora_up.weight
797
+ lora_unet__blocks_39_cross_attn_o.lora_down.weight
798
+ lora_unet__blocks_39_cross_attn_o.lora_up.weight
799
+ lora_unet__blocks_39_ffn_0.lora_down.weight
800
+ lora_unet__blocks_39_ffn_0.lora_up.weight
801
+ lora_unet__blocks_39_ffn_2.lora_down.weight
802
+ lora_unet__blocks_39_ffn_2.lora_up.weight
10_LargeMixedDatset_wan_14bHigh_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/transformer3d_structure.txt ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Wan2_2Transformer3DModel(
2
+ (patch_embedding): Conv3d(36, 5120, kernel_size=(1, 2, 2), stride=(1, 2, 2))
3
+ (text_embedding): Sequential(
4
+ (0): Linear(in_features=4096, out_features=5120, bias=True)
5
+ (1): GELU(approximate='tanh')
6
+ (2): Linear(in_features=5120, out_features=5120, bias=True)
7
+ )
8
+ (time_embedding): Sequential(
9
+ (0): Linear(in_features=256, out_features=5120, bias=True)
10
+ (1): SiLU()
11
+ (2): Linear(in_features=5120, out_features=5120, bias=True)
12
+ )
13
+ (time_projection): Sequential(
14
+ (0): SiLU()
15
+ (1): Linear(in_features=5120, out_features=30720, bias=True)
16
+ )
17
+ (blocks): ModuleList(
18
+ (0-39): 40 x WanAttentionBlock(
19
+ (norm1): WanLayerNorm((5120,), eps=1e-06, elementwise_affine=False)
20
+ (self_attn): WanSelfAttention(
21
+ (q): Linear(in_features=5120, out_features=5120, bias=True)
22
+ (k): Linear(in_features=5120, out_features=5120, bias=True)
23
+ (v): Linear(in_features=5120, out_features=5120, bias=True)
24
+ (o): Linear(in_features=5120, out_features=5120, bias=True)
25
+ (norm_q): WanRMSNorm()
26
+ (norm_k): WanRMSNorm()
27
+ )
28
+ (norm3): WanLayerNorm((5120,), eps=1e-06, elementwise_affine=True)
29
+ (cross_attn): WanCrossAttention(
30
+ (q): Linear(in_features=5120, out_features=5120, bias=True)
31
+ (k): Linear(in_features=5120, out_features=5120, bias=True)
32
+ (v): Linear(in_features=5120, out_features=5120, bias=True)
33
+ (o): Linear(in_features=5120, out_features=5120, bias=True)
34
+ (norm_q): WanRMSNorm()
35
+ (norm_k): WanRMSNorm()
36
+ )
37
+ (norm2): WanLayerNorm((5120,), eps=1e-06, elementwise_affine=False)
38
+ (ffn): Sequential(
39
+ (0): Linear(in_features=5120, out_features=13824, bias=True)
40
+ (1): GELU(approximate='tanh')
41
+ (2): Linear(in_features=13824, out_features=5120, bias=True)
42
+ )
43
+ )
44
+ )
45
+ (head): Head(
46
+ (norm): WanLayerNorm((5120,), eps=1e-06, elementwise_affine=False)
47
+ (head): Linear(in_features=5120, out_features=64, bias=True)
48
+ )
49
+ )
10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/checkpoint-600.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3469743e51c6c7730a9d0ced98e7f5b77c019b8b36e8b252ab454ce5bc16c310
3
+ size 1236150882
10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/events.out.tfevents.1761137386.edb6fde9143d.95968.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a3e88e3a1e3c9d036590191fd705f025de13fa1b42be2a17eeb7be5937dea58
3
+ size 88
10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/text2image-fine-tune/1761138020.9281862/events.out.tfevents.1761138020.edb6fde9143d.95968.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfbc26dbdfe7947f8f94dbd3b32783ffc4226297edef92b66da511981bbad805
3
+ size 4515
10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/text2image-fine-tune/1761138020.940236/.ipynb_checkpoints/hparams-checkpoint.yml ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-10
4
+ adam_weight_decay: 0.03
5
+ allow_tf32: false
6
+ auto_tile_batch_size: false
7
+ boundary_type: low
8
+ cache_dir: null
9
+ checkpointing_steps: 100
10
+ checkpoints_total_limit: null
11
+ config_path: /workspace/Project/VideoRAG/VideoX-Fun/config/wan2.2/wan_civitai_i2v.yaml
12
+ dataloader_num_workers: 4
13
+ enable_bucket: true
14
+ enable_text_encoder_in_dataloader: false
15
+ gradient_accumulation_steps: 1
16
+ gradient_checkpointing: true
17
+ height: 720
18
+ hub_model_id: null
19
+ hub_token: null
20
+ image_sample_size: 1024
21
+ input_perturbation: 0
22
+ keep_all_node_same_token_length: false
23
+ learning_rate: 0.0001
24
+ local_rank: 0
25
+ logging_dir: logs
26
+ logit_mean: 0.0
27
+ logit_std: 1.0
28
+ lora_skip_name: null
29
+ low_vram: false
30
+ lr_scheduler: constant
31
+ lr_warmup_steps: 500
32
+ max_grad_norm: 0.05
33
+ max_train_samples: null
34
+ max_train_steps: 60000
35
+ mixed_precision: bf16
36
+ mode_scale: 1.29
37
+ motion_sub_loss: false
38
+ motion_sub_loss_ratio: 0.25
39
+ multi_stream: false
40
+ network_alpha: 128
41
+ noise_offset: 0
42
+ noise_share_in_frames: false
43
+ noise_share_in_frames_ratio: 0.5
44
+ non_ema_revision: null
45
+ num_train_epochs: 5000
46
+ output_dir: /workspace/Models/10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4
47
+ prediction_type: null
48
+ pretrained_model_name_or_path: /workspace/hfhome/hub/models--Wan-AI--Wan2.2-I2V-A14B/snapshots/206a9ee1b7bfaaf8f7e4d81335650533490646a3
49
+ push_to_hub: false
50
+ random_flip: false
51
+ random_frame_crop: false
52
+ random_hw_adapt: true
53
+ random_ratio_crop: false
54
+ rank: 128
55
+ report_to: tensorboard
56
+ resume_from_checkpoint: null
57
+ revision: null
58
+ save_state: false
59
+ scale_lr: false
60
+ seed: 42
61
+ snr_loss: false
62
+ token_sample_size: 512
63
+ tokenizer_max_length: 512
64
+ tracker_project_name: text2image-fine-tune
65
+ train_batch_size: 2
66
+ train_data_dir: /workspace/Data/05_mixed_data_Large
67
+ train_data_meta: /workspace/Data/05_mixed_data_Large/train_data_mixed_Large.json
68
+ train_mode: i2v
69
+ train_sampling_steps: 1000
70
+ train_text_encoder: false
71
+ training_with_video_token_length: true
72
+ transformer_path: null
73
+ uniform_sampling: true
74
+ use_8bit_adam: false
75
+ use_came: false
76
+ use_deepspeed: false
77
+ use_ema: false
78
+ use_fsdp: false
79
+ vae_mini_batch: 1
80
+ vae_path: null
81
+ validation_epochs: 5
82
+ validation_steps: 200000000
83
+ variant: null
84
+ video_repeat: 1
85
+ video_sample_n_frames: 81
86
+ video_sample_size: 1024
87
+ video_sample_stride: 1
88
+ weighting_scheme: none
89
+ width: 1280
10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/text2image-fine-tune/1761138020.940236/hparams.yml ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-10
4
+ adam_weight_decay: 0.03
5
+ allow_tf32: false
6
+ auto_tile_batch_size: false
7
+ boundary_type: low
8
+ cache_dir: null
9
+ checkpointing_steps: 100
10
+ checkpoints_total_limit: null
11
+ config_path: /workspace/Project/VideoRAG/VideoX-Fun/config/wan2.2/wan_civitai_i2v.yaml
12
+ dataloader_num_workers: 4
13
+ enable_bucket: true
14
+ enable_text_encoder_in_dataloader: false
15
+ gradient_accumulation_steps: 1
16
+ gradient_checkpointing: true
17
+ height: 720
18
+ hub_model_id: null
19
+ hub_token: null
20
+ image_sample_size: 1024
21
+ input_perturbation: 0
22
+ keep_all_node_same_token_length: false
23
+ learning_rate: 0.0001
24
+ local_rank: 0
25
+ logging_dir: logs
26
+ logit_mean: 0.0
27
+ logit_std: 1.0
28
+ lora_skip_name: null
29
+ low_vram: false
30
+ lr_scheduler: constant
31
+ lr_warmup_steps: 500
32
+ max_grad_norm: 0.05
33
+ max_train_samples: null
34
+ max_train_steps: 60000
35
+ mixed_precision: bf16
36
+ mode_scale: 1.29
37
+ motion_sub_loss: false
38
+ motion_sub_loss_ratio: 0.25
39
+ multi_stream: false
40
+ network_alpha: 128
41
+ noise_offset: 0
42
+ noise_share_in_frames: false
43
+ noise_share_in_frames_ratio: 0.5
44
+ non_ema_revision: null
45
+ num_train_epochs: 5000
46
+ output_dir: /workspace/Models/10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4
47
+ prediction_type: null
48
+ pretrained_model_name_or_path: /workspace/hfhome/hub/models--Wan-AI--Wan2.2-I2V-A14B/snapshots/206a9ee1b7bfaaf8f7e4d81335650533490646a3
49
+ push_to_hub: false
50
+ random_flip: false
51
+ random_frame_crop: false
52
+ random_hw_adapt: true
53
+ random_ratio_crop: false
54
+ rank: 128
55
+ report_to: tensorboard
56
+ resume_from_checkpoint: null
57
+ revision: null
58
+ save_state: false
59
+ scale_lr: false
60
+ seed: 42
61
+ snr_loss: false
62
+ token_sample_size: 512
63
+ tokenizer_max_length: 512
64
+ tracker_project_name: text2image-fine-tune
65
+ train_batch_size: 2
66
+ train_data_dir: /workspace/Data/05_mixed_data_Large
67
+ train_data_meta: /workspace/Data/05_mixed_data_Large/train_data_mixed_Large.json
68
+ train_mode: i2v
69
+ train_sampling_steps: 1000
70
+ train_text_encoder: false
71
+ training_with_video_token_length: true
72
+ transformer_path: null
73
+ uniform_sampling: true
74
+ use_8bit_adam: false
75
+ use_came: false
76
+ use_deepspeed: false
77
+ use_ema: false
78
+ use_fsdp: false
79
+ vae_mini_batch: 1
80
+ vae_path: null
81
+ validation_epochs: 5
82
+ validation_steps: 200000000
83
+ variant: null
84
+ video_repeat: 1
85
+ video_sample_n_frames: 81
86
+ video_sample_size: 1024
87
+ video_sample_stride: 1
88
+ weighting_scheme: none
89
+ width: 1280
10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/logs/text2image-fine-tune/events.out.tfevents.1761138020.edb6fde9143d.95968.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75d8507a7e0418c0ce0f3f07b4b1f15a3841919bf6e978289ae65328263bec2a
3
+ size 31664
10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/network_structure.txt ADDED
@@ -0,0 +1,1606 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ LoRANetwork(
2
+ (lora_unet__time_projection_1): LoRAModule(
3
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
4
+ (lora_up): Linear(in_features=128, out_features=30720, bias=False)
5
+ )
6
+ (lora_unet__blocks_0_self_attn_q): LoRAModule(
7
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
8
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
9
+ )
10
+ (lora_unet__blocks_0_self_attn_k): LoRAModule(
11
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
12
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
13
+ )
14
+ (lora_unet__blocks_0_self_attn_v): LoRAModule(
15
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
16
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
17
+ )
18
+ (lora_unet__blocks_0_self_attn_o): LoRAModule(
19
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
20
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
21
+ )
22
+ (lora_unet__blocks_0_cross_attn_q): LoRAModule(
23
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
24
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
25
+ )
26
+ (lora_unet__blocks_0_cross_attn_k): LoRAModule(
27
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
28
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
29
+ )
30
+ (lora_unet__blocks_0_cross_attn_v): LoRAModule(
31
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
32
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
33
+ )
34
+ (lora_unet__blocks_0_cross_attn_o): LoRAModule(
35
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
36
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
37
+ )
38
+ (lora_unet__blocks_0_ffn_0): LoRAModule(
39
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
40
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
41
+ )
42
+ (lora_unet__blocks_0_ffn_2): LoRAModule(
43
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
44
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
45
+ )
46
+ (lora_unet__blocks_1_self_attn_q): LoRAModule(
47
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
48
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
49
+ )
50
+ (lora_unet__blocks_1_self_attn_k): LoRAModule(
51
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
52
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
53
+ )
54
+ (lora_unet__blocks_1_self_attn_v): LoRAModule(
55
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
56
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
57
+ )
58
+ (lora_unet__blocks_1_self_attn_o): LoRAModule(
59
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
60
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
61
+ )
62
+ (lora_unet__blocks_1_cross_attn_q): LoRAModule(
63
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
64
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
65
+ )
66
+ (lora_unet__blocks_1_cross_attn_k): LoRAModule(
67
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
68
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
69
+ )
70
+ (lora_unet__blocks_1_cross_attn_v): LoRAModule(
71
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
72
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
73
+ )
74
+ (lora_unet__blocks_1_cross_attn_o): LoRAModule(
75
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
76
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
77
+ )
78
+ (lora_unet__blocks_1_ffn_0): LoRAModule(
79
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
80
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
81
+ )
82
+ (lora_unet__blocks_1_ffn_2): LoRAModule(
83
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
84
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
85
+ )
86
+ (lora_unet__blocks_2_self_attn_q): LoRAModule(
87
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
88
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
89
+ )
90
+ (lora_unet__blocks_2_self_attn_k): LoRAModule(
91
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
92
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
93
+ )
94
+ (lora_unet__blocks_2_self_attn_v): LoRAModule(
95
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
96
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
97
+ )
98
+ (lora_unet__blocks_2_self_attn_o): LoRAModule(
99
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
100
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
101
+ )
102
+ (lora_unet__blocks_2_cross_attn_q): LoRAModule(
103
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
104
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
105
+ )
106
+ (lora_unet__blocks_2_cross_attn_k): LoRAModule(
107
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
108
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
109
+ )
110
+ (lora_unet__blocks_2_cross_attn_v): LoRAModule(
111
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
112
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
113
+ )
114
+ (lora_unet__blocks_2_cross_attn_o): LoRAModule(
115
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
116
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
117
+ )
118
+ (lora_unet__blocks_2_ffn_0): LoRAModule(
119
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
120
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
121
+ )
122
+ (lora_unet__blocks_2_ffn_2): LoRAModule(
123
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
124
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
125
+ )
126
+ (lora_unet__blocks_3_self_attn_q): LoRAModule(
127
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
128
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
129
+ )
130
+ (lora_unet__blocks_3_self_attn_k): LoRAModule(
131
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
132
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
133
+ )
134
+ (lora_unet__blocks_3_self_attn_v): LoRAModule(
135
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
136
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
137
+ )
138
+ (lora_unet__blocks_3_self_attn_o): LoRAModule(
139
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
140
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
141
+ )
142
+ (lora_unet__blocks_3_cross_attn_q): LoRAModule(
143
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
144
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
145
+ )
146
+ (lora_unet__blocks_3_cross_attn_k): LoRAModule(
147
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
148
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
149
+ )
150
+ (lora_unet__blocks_3_cross_attn_v): LoRAModule(
151
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
152
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
153
+ )
154
+ (lora_unet__blocks_3_cross_attn_o): LoRAModule(
155
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
156
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
157
+ )
158
+ (lora_unet__blocks_3_ffn_0): LoRAModule(
159
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
160
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
161
+ )
162
+ (lora_unet__blocks_3_ffn_2): LoRAModule(
163
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
164
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
165
+ )
166
+ (lora_unet__blocks_4_self_attn_q): LoRAModule(
167
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
168
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
169
+ )
170
+ (lora_unet__blocks_4_self_attn_k): LoRAModule(
171
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
172
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
173
+ )
174
+ (lora_unet__blocks_4_self_attn_v): LoRAModule(
175
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
176
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
177
+ )
178
+ (lora_unet__blocks_4_self_attn_o): LoRAModule(
179
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
180
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
181
+ )
182
+ (lora_unet__blocks_4_cross_attn_q): LoRAModule(
183
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
184
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
185
+ )
186
+ (lora_unet__blocks_4_cross_attn_k): LoRAModule(
187
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
188
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
189
+ )
190
+ (lora_unet__blocks_4_cross_attn_v): LoRAModule(
191
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
192
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
193
+ )
194
+ (lora_unet__blocks_4_cross_attn_o): LoRAModule(
195
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
196
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
197
+ )
198
+ (lora_unet__blocks_4_ffn_0): LoRAModule(
199
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
200
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
201
+ )
202
+ (lora_unet__blocks_4_ffn_2): LoRAModule(
203
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
204
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
205
+ )
206
+ (lora_unet__blocks_5_self_attn_q): LoRAModule(
207
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
208
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
209
+ )
210
+ (lora_unet__blocks_5_self_attn_k): LoRAModule(
211
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
212
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
213
+ )
214
+ (lora_unet__blocks_5_self_attn_v): LoRAModule(
215
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
216
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
217
+ )
218
+ (lora_unet__blocks_5_self_attn_o): LoRAModule(
219
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
220
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
221
+ )
222
+ (lora_unet__blocks_5_cross_attn_q): LoRAModule(
223
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
224
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
225
+ )
226
+ (lora_unet__blocks_5_cross_attn_k): LoRAModule(
227
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
228
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
229
+ )
230
+ (lora_unet__blocks_5_cross_attn_v): LoRAModule(
231
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
232
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
233
+ )
234
+ (lora_unet__blocks_5_cross_attn_o): LoRAModule(
235
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
236
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
237
+ )
238
+ (lora_unet__blocks_5_ffn_0): LoRAModule(
239
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
240
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
241
+ )
242
+ (lora_unet__blocks_5_ffn_2): LoRAModule(
243
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
244
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
245
+ )
246
+ (lora_unet__blocks_6_self_attn_q): LoRAModule(
247
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
248
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
249
+ )
250
+ (lora_unet__blocks_6_self_attn_k): LoRAModule(
251
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
252
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
253
+ )
254
+ (lora_unet__blocks_6_self_attn_v): LoRAModule(
255
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
256
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
257
+ )
258
+ (lora_unet__blocks_6_self_attn_o): LoRAModule(
259
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
260
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
261
+ )
262
+ (lora_unet__blocks_6_cross_attn_q): LoRAModule(
263
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
264
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
265
+ )
266
+ (lora_unet__blocks_6_cross_attn_k): LoRAModule(
267
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
268
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
269
+ )
270
+ (lora_unet__blocks_6_cross_attn_v): LoRAModule(
271
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
272
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
273
+ )
274
+ (lora_unet__blocks_6_cross_attn_o): LoRAModule(
275
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
276
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
277
+ )
278
+ (lora_unet__blocks_6_ffn_0): LoRAModule(
279
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
280
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
281
+ )
282
+ (lora_unet__blocks_6_ffn_2): LoRAModule(
283
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
284
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
285
+ )
286
+ (lora_unet__blocks_7_self_attn_q): LoRAModule(
287
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
288
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
289
+ )
290
+ (lora_unet__blocks_7_self_attn_k): LoRAModule(
291
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
292
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
293
+ )
294
+ (lora_unet__blocks_7_self_attn_v): LoRAModule(
295
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
296
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
297
+ )
298
+ (lora_unet__blocks_7_self_attn_o): LoRAModule(
299
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
300
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
301
+ )
302
+ (lora_unet__blocks_7_cross_attn_q): LoRAModule(
303
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
304
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
305
+ )
306
+ (lora_unet__blocks_7_cross_attn_k): LoRAModule(
307
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
308
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
309
+ )
310
+ (lora_unet__blocks_7_cross_attn_v): LoRAModule(
311
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
312
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
313
+ )
314
+ (lora_unet__blocks_7_cross_attn_o): LoRAModule(
315
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
316
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
317
+ )
318
+ (lora_unet__blocks_7_ffn_0): LoRAModule(
319
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
320
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
321
+ )
322
+ (lora_unet__blocks_7_ffn_2): LoRAModule(
323
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
324
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
325
+ )
326
+ (lora_unet__blocks_8_self_attn_q): LoRAModule(
327
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
328
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
329
+ )
330
+ (lora_unet__blocks_8_self_attn_k): LoRAModule(
331
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
332
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
333
+ )
334
+ (lora_unet__blocks_8_self_attn_v): LoRAModule(
335
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
336
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
337
+ )
338
+ (lora_unet__blocks_8_self_attn_o): LoRAModule(
339
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
340
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
341
+ )
342
+ (lora_unet__blocks_8_cross_attn_q): LoRAModule(
343
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
344
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
345
+ )
346
+ (lora_unet__blocks_8_cross_attn_k): LoRAModule(
347
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
348
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
349
+ )
350
+ (lora_unet__blocks_8_cross_attn_v): LoRAModule(
351
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
352
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
353
+ )
354
+ (lora_unet__blocks_8_cross_attn_o): LoRAModule(
355
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
356
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
357
+ )
358
+ (lora_unet__blocks_8_ffn_0): LoRAModule(
359
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
360
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
361
+ )
362
+ (lora_unet__blocks_8_ffn_2): LoRAModule(
363
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
364
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
365
+ )
366
+ (lora_unet__blocks_9_self_attn_q): LoRAModule(
367
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
368
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
369
+ )
370
+ (lora_unet__blocks_9_self_attn_k): LoRAModule(
371
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
372
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
373
+ )
374
+ (lora_unet__blocks_9_self_attn_v): LoRAModule(
375
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
376
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
377
+ )
378
+ (lora_unet__blocks_9_self_attn_o): LoRAModule(
379
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
380
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
381
+ )
382
+ (lora_unet__blocks_9_cross_attn_q): LoRAModule(
383
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
384
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
385
+ )
386
+ (lora_unet__blocks_9_cross_attn_k): LoRAModule(
387
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
388
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
389
+ )
390
+ (lora_unet__blocks_9_cross_attn_v): LoRAModule(
391
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
392
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
393
+ )
394
+ (lora_unet__blocks_9_cross_attn_o): LoRAModule(
395
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
396
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
397
+ )
398
+ (lora_unet__blocks_9_ffn_0): LoRAModule(
399
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
400
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
401
+ )
402
+ (lora_unet__blocks_9_ffn_2): LoRAModule(
403
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
404
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
405
+ )
406
+ (lora_unet__blocks_10_self_attn_q): LoRAModule(
407
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
408
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
409
+ )
410
+ (lora_unet__blocks_10_self_attn_k): LoRAModule(
411
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
412
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
413
+ )
414
+ (lora_unet__blocks_10_self_attn_v): LoRAModule(
415
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
416
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
417
+ )
418
+ (lora_unet__blocks_10_self_attn_o): LoRAModule(
419
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
420
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
421
+ )
422
+ (lora_unet__blocks_10_cross_attn_q): LoRAModule(
423
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
424
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
425
+ )
426
+ (lora_unet__blocks_10_cross_attn_k): LoRAModule(
427
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
428
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
429
+ )
430
+ (lora_unet__blocks_10_cross_attn_v): LoRAModule(
431
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
432
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
433
+ )
434
+ (lora_unet__blocks_10_cross_attn_o): LoRAModule(
435
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
436
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
437
+ )
438
+ (lora_unet__blocks_10_ffn_0): LoRAModule(
439
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
440
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
441
+ )
442
+ (lora_unet__blocks_10_ffn_2): LoRAModule(
443
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
444
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
445
+ )
446
+ (lora_unet__blocks_11_self_attn_q): LoRAModule(
447
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
448
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
449
+ )
450
+ (lora_unet__blocks_11_self_attn_k): LoRAModule(
451
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
452
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
453
+ )
454
+ (lora_unet__blocks_11_self_attn_v): LoRAModule(
455
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
456
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
457
+ )
458
+ (lora_unet__blocks_11_self_attn_o): LoRAModule(
459
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
460
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
461
+ )
462
+ (lora_unet__blocks_11_cross_attn_q): LoRAModule(
463
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
464
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
465
+ )
466
+ (lora_unet__blocks_11_cross_attn_k): LoRAModule(
467
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
468
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
469
+ )
470
+ (lora_unet__blocks_11_cross_attn_v): LoRAModule(
471
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
472
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
473
+ )
474
+ (lora_unet__blocks_11_cross_attn_o): LoRAModule(
475
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
476
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
477
+ )
478
+ (lora_unet__blocks_11_ffn_0): LoRAModule(
479
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
480
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
481
+ )
482
+ (lora_unet__blocks_11_ffn_2): LoRAModule(
483
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
484
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
485
+ )
486
+ (lora_unet__blocks_12_self_attn_q): LoRAModule(
487
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
488
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
489
+ )
490
+ (lora_unet__blocks_12_self_attn_k): LoRAModule(
491
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
492
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
493
+ )
494
+ (lora_unet__blocks_12_self_attn_v): LoRAModule(
495
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
496
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
497
+ )
498
+ (lora_unet__blocks_12_self_attn_o): LoRAModule(
499
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
500
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
501
+ )
502
+ (lora_unet__blocks_12_cross_attn_q): LoRAModule(
503
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
504
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
505
+ )
506
+ (lora_unet__blocks_12_cross_attn_k): LoRAModule(
507
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
508
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
509
+ )
510
+ (lora_unet__blocks_12_cross_attn_v): LoRAModule(
511
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
512
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
513
+ )
514
+ (lora_unet__blocks_12_cross_attn_o): LoRAModule(
515
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
516
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
517
+ )
518
+ (lora_unet__blocks_12_ffn_0): LoRAModule(
519
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
520
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
521
+ )
522
+ (lora_unet__blocks_12_ffn_2): LoRAModule(
523
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
524
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
525
+ )
526
+ (lora_unet__blocks_13_self_attn_q): LoRAModule(
527
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
528
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
529
+ )
530
+ (lora_unet__blocks_13_self_attn_k): LoRAModule(
531
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
532
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
533
+ )
534
+ (lora_unet__blocks_13_self_attn_v): LoRAModule(
535
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
536
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
537
+ )
538
+ (lora_unet__blocks_13_self_attn_o): LoRAModule(
539
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
540
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
541
+ )
542
+ (lora_unet__blocks_13_cross_attn_q): LoRAModule(
543
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
544
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
545
+ )
546
+ (lora_unet__blocks_13_cross_attn_k): LoRAModule(
547
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
548
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
549
+ )
550
+ (lora_unet__blocks_13_cross_attn_v): LoRAModule(
551
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
552
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
553
+ )
554
+ (lora_unet__blocks_13_cross_attn_o): LoRAModule(
555
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
556
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
557
+ )
558
+ (lora_unet__blocks_13_ffn_0): LoRAModule(
559
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
560
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
561
+ )
562
+ (lora_unet__blocks_13_ffn_2): LoRAModule(
563
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
564
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
565
+ )
566
+ (lora_unet__blocks_14_self_attn_q): LoRAModule(
567
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
568
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
569
+ )
570
+ (lora_unet__blocks_14_self_attn_k): LoRAModule(
571
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
572
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
573
+ )
574
+ (lora_unet__blocks_14_self_attn_v): LoRAModule(
575
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
576
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
577
+ )
578
+ (lora_unet__blocks_14_self_attn_o): LoRAModule(
579
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
580
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
581
+ )
582
+ (lora_unet__blocks_14_cross_attn_q): LoRAModule(
583
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
584
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
585
+ )
586
+ (lora_unet__blocks_14_cross_attn_k): LoRAModule(
587
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
588
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
589
+ )
590
+ (lora_unet__blocks_14_cross_attn_v): LoRAModule(
591
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
592
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
593
+ )
594
+ (lora_unet__blocks_14_cross_attn_o): LoRAModule(
595
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
596
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
597
+ )
598
+ (lora_unet__blocks_14_ffn_0): LoRAModule(
599
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
600
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
601
+ )
602
+ (lora_unet__blocks_14_ffn_2): LoRAModule(
603
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
604
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
605
+ )
606
+ (lora_unet__blocks_15_self_attn_q): LoRAModule(
607
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
608
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
609
+ )
610
+ (lora_unet__blocks_15_self_attn_k): LoRAModule(
611
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
612
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
613
+ )
614
+ (lora_unet__blocks_15_self_attn_v): LoRAModule(
615
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
616
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
617
+ )
618
+ (lora_unet__blocks_15_self_attn_o): LoRAModule(
619
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
620
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
621
+ )
622
+ (lora_unet__blocks_15_cross_attn_q): LoRAModule(
623
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
624
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
625
+ )
626
+ (lora_unet__blocks_15_cross_attn_k): LoRAModule(
627
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
628
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
629
+ )
630
+ (lora_unet__blocks_15_cross_attn_v): LoRAModule(
631
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
632
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
633
+ )
634
+ (lora_unet__blocks_15_cross_attn_o): LoRAModule(
635
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
636
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
637
+ )
638
+ (lora_unet__blocks_15_ffn_0): LoRAModule(
639
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
640
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
641
+ )
642
+ (lora_unet__blocks_15_ffn_2): LoRAModule(
643
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
644
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
645
+ )
646
+ (lora_unet__blocks_16_self_attn_q): LoRAModule(
647
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
648
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
649
+ )
650
+ (lora_unet__blocks_16_self_attn_k): LoRAModule(
651
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
652
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
653
+ )
654
+ (lora_unet__blocks_16_self_attn_v): LoRAModule(
655
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
656
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
657
+ )
658
+ (lora_unet__blocks_16_self_attn_o): LoRAModule(
659
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
660
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
661
+ )
662
+ (lora_unet__blocks_16_cross_attn_q): LoRAModule(
663
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
664
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
665
+ )
666
+ (lora_unet__blocks_16_cross_attn_k): LoRAModule(
667
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
668
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
669
+ )
670
+ (lora_unet__blocks_16_cross_attn_v): LoRAModule(
671
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
672
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
673
+ )
674
+ (lora_unet__blocks_16_cross_attn_o): LoRAModule(
675
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
676
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
677
+ )
678
+ (lora_unet__blocks_16_ffn_0): LoRAModule(
679
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
680
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
681
+ )
682
+ (lora_unet__blocks_16_ffn_2): LoRAModule(
683
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
684
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
685
+ )
686
+ (lora_unet__blocks_17_self_attn_q): LoRAModule(
687
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
688
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
689
+ )
690
+ (lora_unet__blocks_17_self_attn_k): LoRAModule(
691
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
692
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
693
+ )
694
+ (lora_unet__blocks_17_self_attn_v): LoRAModule(
695
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
696
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
697
+ )
698
+ (lora_unet__blocks_17_self_attn_o): LoRAModule(
699
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
700
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
701
+ )
702
+ (lora_unet__blocks_17_cross_attn_q): LoRAModule(
703
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
704
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
705
+ )
706
+ (lora_unet__blocks_17_cross_attn_k): LoRAModule(
707
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
708
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
709
+ )
710
+ (lora_unet__blocks_17_cross_attn_v): LoRAModule(
711
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
712
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
713
+ )
714
+ (lora_unet__blocks_17_cross_attn_o): LoRAModule(
715
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
716
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
717
+ )
718
+ (lora_unet__blocks_17_ffn_0): LoRAModule(
719
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
720
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
721
+ )
722
+ (lora_unet__blocks_17_ffn_2): LoRAModule(
723
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
724
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
725
+ )
726
+ (lora_unet__blocks_18_self_attn_q): LoRAModule(
727
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
728
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
729
+ )
730
+ (lora_unet__blocks_18_self_attn_k): LoRAModule(
731
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
732
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
733
+ )
734
+ (lora_unet__blocks_18_self_attn_v): LoRAModule(
735
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
736
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
737
+ )
738
+ (lora_unet__blocks_18_self_attn_o): LoRAModule(
739
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
740
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
741
+ )
742
+ (lora_unet__blocks_18_cross_attn_q): LoRAModule(
743
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
744
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
745
+ )
746
+ (lora_unet__blocks_18_cross_attn_k): LoRAModule(
747
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
748
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
749
+ )
750
+ (lora_unet__blocks_18_cross_attn_v): LoRAModule(
751
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
752
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
753
+ )
754
+ (lora_unet__blocks_18_cross_attn_o): LoRAModule(
755
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
756
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
757
+ )
758
+ (lora_unet__blocks_18_ffn_0): LoRAModule(
759
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
760
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
761
+ )
762
+ (lora_unet__blocks_18_ffn_2): LoRAModule(
763
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
764
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
765
+ )
766
+ (lora_unet__blocks_19_self_attn_q): LoRAModule(
767
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
768
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
769
+ )
770
+ (lora_unet__blocks_19_self_attn_k): LoRAModule(
771
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
772
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
773
+ )
774
+ (lora_unet__blocks_19_self_attn_v): LoRAModule(
775
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
776
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
777
+ )
778
+ (lora_unet__blocks_19_self_attn_o): LoRAModule(
779
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
780
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
781
+ )
782
+ (lora_unet__blocks_19_cross_attn_q): LoRAModule(
783
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
784
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
785
+ )
786
+ (lora_unet__blocks_19_cross_attn_k): LoRAModule(
787
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
788
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
789
+ )
790
+ (lora_unet__blocks_19_cross_attn_v): LoRAModule(
791
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
792
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
793
+ )
794
+ (lora_unet__blocks_19_cross_attn_o): LoRAModule(
795
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
796
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
797
+ )
798
+ (lora_unet__blocks_19_ffn_0): LoRAModule(
799
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
800
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
801
+ )
802
+ (lora_unet__blocks_19_ffn_2): LoRAModule(
803
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
804
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
805
+ )
806
+ (lora_unet__blocks_20_self_attn_q): LoRAModule(
807
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
808
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
809
+ )
810
+ (lora_unet__blocks_20_self_attn_k): LoRAModule(
811
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
812
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
813
+ )
814
+ (lora_unet__blocks_20_self_attn_v): LoRAModule(
815
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
816
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
817
+ )
818
+ (lora_unet__blocks_20_self_attn_o): LoRAModule(
819
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
820
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
821
+ )
822
+ (lora_unet__blocks_20_cross_attn_q): LoRAModule(
823
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
824
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
825
+ )
826
+ (lora_unet__blocks_20_cross_attn_k): LoRAModule(
827
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
828
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
829
+ )
830
+ (lora_unet__blocks_20_cross_attn_v): LoRAModule(
831
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
832
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
833
+ )
834
+ (lora_unet__blocks_20_cross_attn_o): LoRAModule(
835
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
836
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
837
+ )
838
+ (lora_unet__blocks_20_ffn_0): LoRAModule(
839
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
840
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
841
+ )
842
+ (lora_unet__blocks_20_ffn_2): LoRAModule(
843
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
844
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
845
+ )
846
+ (lora_unet__blocks_21_self_attn_q): LoRAModule(
847
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
848
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
849
+ )
850
+ (lora_unet__blocks_21_self_attn_k): LoRAModule(
851
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
852
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
853
+ )
854
+ (lora_unet__blocks_21_self_attn_v): LoRAModule(
855
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
856
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
857
+ )
858
+ (lora_unet__blocks_21_self_attn_o): LoRAModule(
859
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
860
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
861
+ )
862
+ (lora_unet__blocks_21_cross_attn_q): LoRAModule(
863
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
864
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
865
+ )
866
+ (lora_unet__blocks_21_cross_attn_k): LoRAModule(
867
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
868
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
869
+ )
870
+ (lora_unet__blocks_21_cross_attn_v): LoRAModule(
871
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
872
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
873
+ )
874
+ (lora_unet__blocks_21_cross_attn_o): LoRAModule(
875
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
876
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
877
+ )
878
+ (lora_unet__blocks_21_ffn_0): LoRAModule(
879
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
880
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
881
+ )
882
+ (lora_unet__blocks_21_ffn_2): LoRAModule(
883
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
884
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
885
+ )
886
+ (lora_unet__blocks_22_self_attn_q): LoRAModule(
887
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
888
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
889
+ )
890
+ (lora_unet__blocks_22_self_attn_k): LoRAModule(
891
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
892
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
893
+ )
894
+ (lora_unet__blocks_22_self_attn_v): LoRAModule(
895
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
896
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
897
+ )
898
+ (lora_unet__blocks_22_self_attn_o): LoRAModule(
899
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
900
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
901
+ )
902
+ (lora_unet__blocks_22_cross_attn_q): LoRAModule(
903
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
904
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
905
+ )
906
+ (lora_unet__blocks_22_cross_attn_k): LoRAModule(
907
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
908
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
909
+ )
910
+ (lora_unet__blocks_22_cross_attn_v): LoRAModule(
911
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
912
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
913
+ )
914
+ (lora_unet__blocks_22_cross_attn_o): LoRAModule(
915
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
916
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
917
+ )
918
+ (lora_unet__blocks_22_ffn_0): LoRAModule(
919
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
920
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
921
+ )
922
+ (lora_unet__blocks_22_ffn_2): LoRAModule(
923
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
924
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
925
+ )
926
+ (lora_unet__blocks_23_self_attn_q): LoRAModule(
927
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
928
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
929
+ )
930
+ (lora_unet__blocks_23_self_attn_k): LoRAModule(
931
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
932
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
933
+ )
934
+ (lora_unet__blocks_23_self_attn_v): LoRAModule(
935
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
936
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
937
+ )
938
+ (lora_unet__blocks_23_self_attn_o): LoRAModule(
939
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
940
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
941
+ )
942
+ (lora_unet__blocks_23_cross_attn_q): LoRAModule(
943
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
944
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
945
+ )
946
+ (lora_unet__blocks_23_cross_attn_k): LoRAModule(
947
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
948
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
949
+ )
950
+ (lora_unet__blocks_23_cross_attn_v): LoRAModule(
951
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
952
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
953
+ )
954
+ (lora_unet__blocks_23_cross_attn_o): LoRAModule(
955
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
956
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
957
+ )
958
+ (lora_unet__blocks_23_ffn_0): LoRAModule(
959
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
960
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
961
+ )
962
+ (lora_unet__blocks_23_ffn_2): LoRAModule(
963
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
964
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
965
+ )
966
+ (lora_unet__blocks_24_self_attn_q): LoRAModule(
967
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
968
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
969
+ )
970
+ (lora_unet__blocks_24_self_attn_k): LoRAModule(
971
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
972
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
973
+ )
974
+ (lora_unet__blocks_24_self_attn_v): LoRAModule(
975
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
976
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
977
+ )
978
+ (lora_unet__blocks_24_self_attn_o): LoRAModule(
979
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
980
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
981
+ )
982
+ (lora_unet__blocks_24_cross_attn_q): LoRAModule(
983
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
984
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
985
+ )
986
+ (lora_unet__blocks_24_cross_attn_k): LoRAModule(
987
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
988
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
989
+ )
990
+ (lora_unet__blocks_24_cross_attn_v): LoRAModule(
991
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
992
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
993
+ )
994
+ (lora_unet__blocks_24_cross_attn_o): LoRAModule(
995
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
996
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
997
+ )
998
+ (lora_unet__blocks_24_ffn_0): LoRAModule(
999
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1000
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1001
+ )
1002
+ (lora_unet__blocks_24_ffn_2): LoRAModule(
1003
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1004
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1005
+ )
1006
+ (lora_unet__blocks_25_self_attn_q): LoRAModule(
1007
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1008
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1009
+ )
1010
+ (lora_unet__blocks_25_self_attn_k): LoRAModule(
1011
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1012
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1013
+ )
1014
+ (lora_unet__blocks_25_self_attn_v): LoRAModule(
1015
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1016
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1017
+ )
1018
+ (lora_unet__blocks_25_self_attn_o): LoRAModule(
1019
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1020
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1021
+ )
1022
+ (lora_unet__blocks_25_cross_attn_q): LoRAModule(
1023
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1024
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1025
+ )
1026
+ (lora_unet__blocks_25_cross_attn_k): LoRAModule(
1027
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1028
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1029
+ )
1030
+ (lora_unet__blocks_25_cross_attn_v): LoRAModule(
1031
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1032
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1033
+ )
1034
+ (lora_unet__blocks_25_cross_attn_o): LoRAModule(
1035
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1036
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1037
+ )
1038
+ (lora_unet__blocks_25_ffn_0): LoRAModule(
1039
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1040
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1041
+ )
1042
+ (lora_unet__blocks_25_ffn_2): LoRAModule(
1043
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1044
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1045
+ )
1046
+ (lora_unet__blocks_26_self_attn_q): LoRAModule(
1047
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1048
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1049
+ )
1050
+ (lora_unet__blocks_26_self_attn_k): LoRAModule(
1051
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1052
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1053
+ )
1054
+ (lora_unet__blocks_26_self_attn_v): LoRAModule(
1055
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1056
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1057
+ )
1058
+ (lora_unet__blocks_26_self_attn_o): LoRAModule(
1059
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1060
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1061
+ )
1062
+ (lora_unet__blocks_26_cross_attn_q): LoRAModule(
1063
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1064
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1065
+ )
1066
+ (lora_unet__blocks_26_cross_attn_k): LoRAModule(
1067
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1068
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1069
+ )
1070
+ (lora_unet__blocks_26_cross_attn_v): LoRAModule(
1071
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1072
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1073
+ )
1074
+ (lora_unet__blocks_26_cross_attn_o): LoRAModule(
1075
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1076
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1077
+ )
1078
+ (lora_unet__blocks_26_ffn_0): LoRAModule(
1079
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1080
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1081
+ )
1082
+ (lora_unet__blocks_26_ffn_2): LoRAModule(
1083
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1084
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1085
+ )
1086
+ (lora_unet__blocks_27_self_attn_q): LoRAModule(
1087
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1088
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1089
+ )
1090
+ (lora_unet__blocks_27_self_attn_k): LoRAModule(
1091
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1092
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1093
+ )
1094
+ (lora_unet__blocks_27_self_attn_v): LoRAModule(
1095
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1096
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1097
+ )
1098
+ (lora_unet__blocks_27_self_attn_o): LoRAModule(
1099
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1100
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1101
+ )
1102
+ (lora_unet__blocks_27_cross_attn_q): LoRAModule(
1103
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1104
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1105
+ )
1106
+ (lora_unet__blocks_27_cross_attn_k): LoRAModule(
1107
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1108
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1109
+ )
1110
+ (lora_unet__blocks_27_cross_attn_v): LoRAModule(
1111
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1112
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1113
+ )
1114
+ (lora_unet__blocks_27_cross_attn_o): LoRAModule(
1115
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1116
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1117
+ )
1118
+ (lora_unet__blocks_27_ffn_0): LoRAModule(
1119
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1120
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1121
+ )
1122
+ (lora_unet__blocks_27_ffn_2): LoRAModule(
1123
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1124
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1125
+ )
1126
+ (lora_unet__blocks_28_self_attn_q): LoRAModule(
1127
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1128
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1129
+ )
1130
+ (lora_unet__blocks_28_self_attn_k): LoRAModule(
1131
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1132
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1133
+ )
1134
+ (lora_unet__blocks_28_self_attn_v): LoRAModule(
1135
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1136
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1137
+ )
1138
+ (lora_unet__blocks_28_self_attn_o): LoRAModule(
1139
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1140
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1141
+ )
1142
+ (lora_unet__blocks_28_cross_attn_q): LoRAModule(
1143
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1144
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1145
+ )
1146
+ (lora_unet__blocks_28_cross_attn_k): LoRAModule(
1147
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1148
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1149
+ )
1150
+ (lora_unet__blocks_28_cross_attn_v): LoRAModule(
1151
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1152
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1153
+ )
1154
+ (lora_unet__blocks_28_cross_attn_o): LoRAModule(
1155
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1156
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1157
+ )
1158
+ (lora_unet__blocks_28_ffn_0): LoRAModule(
1159
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1160
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1161
+ )
1162
+ (lora_unet__blocks_28_ffn_2): LoRAModule(
1163
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1164
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1165
+ )
1166
+ (lora_unet__blocks_29_self_attn_q): LoRAModule(
1167
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1168
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1169
+ )
1170
+ (lora_unet__blocks_29_self_attn_k): LoRAModule(
1171
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1172
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1173
+ )
1174
+ (lora_unet__blocks_29_self_attn_v): LoRAModule(
1175
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1176
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1177
+ )
1178
+ (lora_unet__blocks_29_self_attn_o): LoRAModule(
1179
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1180
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1181
+ )
1182
+ (lora_unet__blocks_29_cross_attn_q): LoRAModule(
1183
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1184
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1185
+ )
1186
+ (lora_unet__blocks_29_cross_attn_k): LoRAModule(
1187
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1188
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1189
+ )
1190
+ (lora_unet__blocks_29_cross_attn_v): LoRAModule(
1191
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1192
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1193
+ )
1194
+ (lora_unet__blocks_29_cross_attn_o): LoRAModule(
1195
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1196
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1197
+ )
1198
+ (lora_unet__blocks_29_ffn_0): LoRAModule(
1199
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1200
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1201
+ )
1202
+ (lora_unet__blocks_29_ffn_2): LoRAModule(
1203
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1204
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1205
+ )
1206
+ (lora_unet__blocks_30_self_attn_q): LoRAModule(
1207
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1208
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1209
+ )
1210
+ (lora_unet__blocks_30_self_attn_k): LoRAModule(
1211
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1212
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1213
+ )
1214
+ (lora_unet__blocks_30_self_attn_v): LoRAModule(
1215
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1216
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1217
+ )
1218
+ (lora_unet__blocks_30_self_attn_o): LoRAModule(
1219
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1220
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1221
+ )
1222
+ (lora_unet__blocks_30_cross_attn_q): LoRAModule(
1223
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1224
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1225
+ )
1226
+ (lora_unet__blocks_30_cross_attn_k): LoRAModule(
1227
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1228
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1229
+ )
1230
+ (lora_unet__blocks_30_cross_attn_v): LoRAModule(
1231
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1232
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1233
+ )
1234
+ (lora_unet__blocks_30_cross_attn_o): LoRAModule(
1235
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1236
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1237
+ )
1238
+ (lora_unet__blocks_30_ffn_0): LoRAModule(
1239
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1240
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1241
+ )
1242
+ (lora_unet__blocks_30_ffn_2): LoRAModule(
1243
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1244
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1245
+ )
1246
+ (lora_unet__blocks_31_self_attn_q): LoRAModule(
1247
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1248
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1249
+ )
1250
+ (lora_unet__blocks_31_self_attn_k): LoRAModule(
1251
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1252
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1253
+ )
1254
+ (lora_unet__blocks_31_self_attn_v): LoRAModule(
1255
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1256
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1257
+ )
1258
+ (lora_unet__blocks_31_self_attn_o): LoRAModule(
1259
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1260
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1261
+ )
1262
+ (lora_unet__blocks_31_cross_attn_q): LoRAModule(
1263
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1264
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1265
+ )
1266
+ (lora_unet__blocks_31_cross_attn_k): LoRAModule(
1267
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1268
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1269
+ )
1270
+ (lora_unet__blocks_31_cross_attn_v): LoRAModule(
1271
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1272
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1273
+ )
1274
+ (lora_unet__blocks_31_cross_attn_o): LoRAModule(
1275
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1276
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1277
+ )
1278
+ (lora_unet__blocks_31_ffn_0): LoRAModule(
1279
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1280
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1281
+ )
1282
+ (lora_unet__blocks_31_ffn_2): LoRAModule(
1283
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1284
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1285
+ )
1286
+ (lora_unet__blocks_32_self_attn_q): LoRAModule(
1287
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1288
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1289
+ )
1290
+ (lora_unet__blocks_32_self_attn_k): LoRAModule(
1291
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1292
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1293
+ )
1294
+ (lora_unet__blocks_32_self_attn_v): LoRAModule(
1295
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1296
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1297
+ )
1298
+ (lora_unet__blocks_32_self_attn_o): LoRAModule(
1299
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1300
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1301
+ )
1302
+ (lora_unet__blocks_32_cross_attn_q): LoRAModule(
1303
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1304
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1305
+ )
1306
+ (lora_unet__blocks_32_cross_attn_k): LoRAModule(
1307
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1308
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1309
+ )
1310
+ (lora_unet__blocks_32_cross_attn_v): LoRAModule(
1311
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1312
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1313
+ )
1314
+ (lora_unet__blocks_32_cross_attn_o): LoRAModule(
1315
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1316
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1317
+ )
1318
+ (lora_unet__blocks_32_ffn_0): LoRAModule(
1319
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1320
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1321
+ )
1322
+ (lora_unet__blocks_32_ffn_2): LoRAModule(
1323
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1324
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1325
+ )
1326
+ (lora_unet__blocks_33_self_attn_q): LoRAModule(
1327
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1328
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1329
+ )
1330
+ (lora_unet__blocks_33_self_attn_k): LoRAModule(
1331
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1332
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1333
+ )
1334
+ (lora_unet__blocks_33_self_attn_v): LoRAModule(
1335
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1336
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1337
+ )
1338
+ (lora_unet__blocks_33_self_attn_o): LoRAModule(
1339
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1340
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1341
+ )
1342
+ (lora_unet__blocks_33_cross_attn_q): LoRAModule(
1343
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1344
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1345
+ )
1346
+ (lora_unet__blocks_33_cross_attn_k): LoRAModule(
1347
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1348
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1349
+ )
1350
+ (lora_unet__blocks_33_cross_attn_v): LoRAModule(
1351
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1352
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1353
+ )
1354
+ (lora_unet__blocks_33_cross_attn_o): LoRAModule(
1355
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1356
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1357
+ )
1358
+ (lora_unet__blocks_33_ffn_0): LoRAModule(
1359
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1360
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1361
+ )
1362
+ (lora_unet__blocks_33_ffn_2): LoRAModule(
1363
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1364
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1365
+ )
1366
+ (lora_unet__blocks_34_self_attn_q): LoRAModule(
1367
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1368
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1369
+ )
1370
+ (lora_unet__blocks_34_self_attn_k): LoRAModule(
1371
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1372
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1373
+ )
1374
+ (lora_unet__blocks_34_self_attn_v): LoRAModule(
1375
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1376
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1377
+ )
1378
+ (lora_unet__blocks_34_self_attn_o): LoRAModule(
1379
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1380
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1381
+ )
1382
+ (lora_unet__blocks_34_cross_attn_q): LoRAModule(
1383
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1384
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1385
+ )
1386
+ (lora_unet__blocks_34_cross_attn_k): LoRAModule(
1387
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1388
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1389
+ )
1390
+ (lora_unet__blocks_34_cross_attn_v): LoRAModule(
1391
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1392
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1393
+ )
1394
+ (lora_unet__blocks_34_cross_attn_o): LoRAModule(
1395
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1396
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1397
+ )
1398
+ (lora_unet__blocks_34_ffn_0): LoRAModule(
1399
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1400
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1401
+ )
1402
+ (lora_unet__blocks_34_ffn_2): LoRAModule(
1403
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1404
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1405
+ )
1406
+ (lora_unet__blocks_35_self_attn_q): LoRAModule(
1407
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1408
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1409
+ )
1410
+ (lora_unet__blocks_35_self_attn_k): LoRAModule(
1411
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1412
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1413
+ )
1414
+ (lora_unet__blocks_35_self_attn_v): LoRAModule(
1415
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1416
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1417
+ )
1418
+ (lora_unet__blocks_35_self_attn_o): LoRAModule(
1419
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1420
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1421
+ )
1422
+ (lora_unet__blocks_35_cross_attn_q): LoRAModule(
1423
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1424
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1425
+ )
1426
+ (lora_unet__blocks_35_cross_attn_k): LoRAModule(
1427
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1428
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1429
+ )
1430
+ (lora_unet__blocks_35_cross_attn_v): LoRAModule(
1431
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1432
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1433
+ )
1434
+ (lora_unet__blocks_35_cross_attn_o): LoRAModule(
1435
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1436
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1437
+ )
1438
+ (lora_unet__blocks_35_ffn_0): LoRAModule(
1439
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1440
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1441
+ )
1442
+ (lora_unet__blocks_35_ffn_2): LoRAModule(
1443
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1444
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1445
+ )
1446
+ (lora_unet__blocks_36_self_attn_q): LoRAModule(
1447
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1448
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1449
+ )
1450
+ (lora_unet__blocks_36_self_attn_k): LoRAModule(
1451
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1452
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1453
+ )
1454
+ (lora_unet__blocks_36_self_attn_v): LoRAModule(
1455
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1456
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1457
+ )
1458
+ (lora_unet__blocks_36_self_attn_o): LoRAModule(
1459
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1460
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1461
+ )
1462
+ (lora_unet__blocks_36_cross_attn_q): LoRAModule(
1463
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1464
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1465
+ )
1466
+ (lora_unet__blocks_36_cross_attn_k): LoRAModule(
1467
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1468
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1469
+ )
1470
+ (lora_unet__blocks_36_cross_attn_v): LoRAModule(
1471
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1472
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1473
+ )
1474
+ (lora_unet__blocks_36_cross_attn_o): LoRAModule(
1475
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1476
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1477
+ )
1478
+ (lora_unet__blocks_36_ffn_0): LoRAModule(
1479
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1480
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1481
+ )
1482
+ (lora_unet__blocks_36_ffn_2): LoRAModule(
1483
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1484
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1485
+ )
1486
+ (lora_unet__blocks_37_self_attn_q): LoRAModule(
1487
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1488
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1489
+ )
1490
+ (lora_unet__blocks_37_self_attn_k): LoRAModule(
1491
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1492
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1493
+ )
1494
+ (lora_unet__blocks_37_self_attn_v): LoRAModule(
1495
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1496
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1497
+ )
1498
+ (lora_unet__blocks_37_self_attn_o): LoRAModule(
1499
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1500
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1501
+ )
1502
+ (lora_unet__blocks_37_cross_attn_q): LoRAModule(
1503
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1504
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1505
+ )
1506
+ (lora_unet__blocks_37_cross_attn_k): LoRAModule(
1507
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1508
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1509
+ )
1510
+ (lora_unet__blocks_37_cross_attn_v): LoRAModule(
1511
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1512
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1513
+ )
1514
+ (lora_unet__blocks_37_cross_attn_o): LoRAModule(
1515
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1516
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1517
+ )
1518
+ (lora_unet__blocks_37_ffn_0): LoRAModule(
1519
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1520
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1521
+ )
1522
+ (lora_unet__blocks_37_ffn_2): LoRAModule(
1523
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1524
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1525
+ )
1526
+ (lora_unet__blocks_38_self_attn_q): LoRAModule(
1527
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1528
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1529
+ )
1530
+ (lora_unet__blocks_38_self_attn_k): LoRAModule(
1531
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1532
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1533
+ )
1534
+ (lora_unet__blocks_38_self_attn_v): LoRAModule(
1535
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1536
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1537
+ )
1538
+ (lora_unet__blocks_38_self_attn_o): LoRAModule(
1539
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1540
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1541
+ )
1542
+ (lora_unet__blocks_38_cross_attn_q): LoRAModule(
1543
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1544
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1545
+ )
1546
+ (lora_unet__blocks_38_cross_attn_k): LoRAModule(
1547
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1548
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1549
+ )
1550
+ (lora_unet__blocks_38_cross_attn_v): LoRAModule(
1551
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1552
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1553
+ )
1554
+ (lora_unet__blocks_38_cross_attn_o): LoRAModule(
1555
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1556
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1557
+ )
1558
+ (lora_unet__blocks_38_ffn_0): LoRAModule(
1559
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1560
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1561
+ )
1562
+ (lora_unet__blocks_38_ffn_2): LoRAModule(
1563
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1564
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1565
+ )
1566
+ (lora_unet__blocks_39_self_attn_q): LoRAModule(
1567
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1568
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1569
+ )
1570
+ (lora_unet__blocks_39_self_attn_k): LoRAModule(
1571
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1572
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1573
+ )
1574
+ (lora_unet__blocks_39_self_attn_v): LoRAModule(
1575
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1576
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1577
+ )
1578
+ (lora_unet__blocks_39_self_attn_o): LoRAModule(
1579
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1580
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1581
+ )
1582
+ (lora_unet__blocks_39_cross_attn_q): LoRAModule(
1583
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1584
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1585
+ )
1586
+ (lora_unet__blocks_39_cross_attn_k): LoRAModule(
1587
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1588
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1589
+ )
1590
+ (lora_unet__blocks_39_cross_attn_v): LoRAModule(
1591
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1592
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1593
+ )
1594
+ (lora_unet__blocks_39_cross_attn_o): LoRAModule(
1595
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1596
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1597
+ )
1598
+ (lora_unet__blocks_39_ffn_0): LoRAModule(
1599
+ (lora_down): Linear(in_features=5120, out_features=128, bias=False)
1600
+ (lora_up): Linear(in_features=128, out_features=13824, bias=False)
1601
+ )
1602
+ (lora_unet__blocks_39_ffn_2): LoRAModule(
1603
+ (lora_down): Linear(in_features=13824, out_features=128, bias=False)
1604
+ (lora_up): Linear(in_features=128, out_features=5120, bias=False)
1605
+ )
1606
+ )
10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/ad23r2-the.gif ADDED

Git LFS Details

  • SHA256: 7bc46a6d13ba867ed57ddfca71e60830e382ef15a638b68b5f0fe535e69dab98
  • Pointer size: 132 Bytes
  • Size of remote file: 8.56 MB
10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/clip_ad23r2-the.png ADDED

Git LFS Details

  • SHA256: 73255781a48d8eb6307bca0e8e73a79ec50550b7d7d9d20362007b5f6fc71023
  • Pointer size: 131 Bytes
  • Size of remote file: 481 kB
10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/sanity_check/mask_ad23r2-the.gif ADDED

Git LFS Details

  • SHA256: a8d268dfc9cde6fd60bb72df8477e39360a8a10a86fd2e7e848cc3a20dfce46e
  • Pointer size: 131 Bytes
  • Size of remote file: 259 kB
10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/trainable_params.txt ADDED
@@ -0,0 +1,802 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ lora_unet__time_projection_1.lora_down.weight
2
+ lora_unet__time_projection_1.lora_up.weight
3
+ lora_unet__blocks_0_self_attn_q.lora_down.weight
4
+ lora_unet__blocks_0_self_attn_q.lora_up.weight
5
+ lora_unet__blocks_0_self_attn_k.lora_down.weight
6
+ lora_unet__blocks_0_self_attn_k.lora_up.weight
7
+ lora_unet__blocks_0_self_attn_v.lora_down.weight
8
+ lora_unet__blocks_0_self_attn_v.lora_up.weight
9
+ lora_unet__blocks_0_self_attn_o.lora_down.weight
10
+ lora_unet__blocks_0_self_attn_o.lora_up.weight
11
+ lora_unet__blocks_0_cross_attn_q.lora_down.weight
12
+ lora_unet__blocks_0_cross_attn_q.lora_up.weight
13
+ lora_unet__blocks_0_cross_attn_k.lora_down.weight
14
+ lora_unet__blocks_0_cross_attn_k.lora_up.weight
15
+ lora_unet__blocks_0_cross_attn_v.lora_down.weight
16
+ lora_unet__blocks_0_cross_attn_v.lora_up.weight
17
+ lora_unet__blocks_0_cross_attn_o.lora_down.weight
18
+ lora_unet__blocks_0_cross_attn_o.lora_up.weight
19
+ lora_unet__blocks_0_ffn_0.lora_down.weight
20
+ lora_unet__blocks_0_ffn_0.lora_up.weight
21
+ lora_unet__blocks_0_ffn_2.lora_down.weight
22
+ lora_unet__blocks_0_ffn_2.lora_up.weight
23
+ lora_unet__blocks_1_self_attn_q.lora_down.weight
24
+ lora_unet__blocks_1_self_attn_q.lora_up.weight
25
+ lora_unet__blocks_1_self_attn_k.lora_down.weight
26
+ lora_unet__blocks_1_self_attn_k.lora_up.weight
27
+ lora_unet__blocks_1_self_attn_v.lora_down.weight
28
+ lora_unet__blocks_1_self_attn_v.lora_up.weight
29
+ lora_unet__blocks_1_self_attn_o.lora_down.weight
30
+ lora_unet__blocks_1_self_attn_o.lora_up.weight
31
+ lora_unet__blocks_1_cross_attn_q.lora_down.weight
32
+ lora_unet__blocks_1_cross_attn_q.lora_up.weight
33
+ lora_unet__blocks_1_cross_attn_k.lora_down.weight
34
+ lora_unet__blocks_1_cross_attn_k.lora_up.weight
35
+ lora_unet__blocks_1_cross_attn_v.lora_down.weight
36
+ lora_unet__blocks_1_cross_attn_v.lora_up.weight
37
+ lora_unet__blocks_1_cross_attn_o.lora_down.weight
38
+ lora_unet__blocks_1_cross_attn_o.lora_up.weight
39
+ lora_unet__blocks_1_ffn_0.lora_down.weight
40
+ lora_unet__blocks_1_ffn_0.lora_up.weight
41
+ lora_unet__blocks_1_ffn_2.lora_down.weight
42
+ lora_unet__blocks_1_ffn_2.lora_up.weight
43
+ lora_unet__blocks_2_self_attn_q.lora_down.weight
44
+ lora_unet__blocks_2_self_attn_q.lora_up.weight
45
+ lora_unet__blocks_2_self_attn_k.lora_down.weight
46
+ lora_unet__blocks_2_self_attn_k.lora_up.weight
47
+ lora_unet__blocks_2_self_attn_v.lora_down.weight
48
+ lora_unet__blocks_2_self_attn_v.lora_up.weight
49
+ lora_unet__blocks_2_self_attn_o.lora_down.weight
50
+ lora_unet__blocks_2_self_attn_o.lora_up.weight
51
+ lora_unet__blocks_2_cross_attn_q.lora_down.weight
52
+ lora_unet__blocks_2_cross_attn_q.lora_up.weight
53
+ lora_unet__blocks_2_cross_attn_k.lora_down.weight
54
+ lora_unet__blocks_2_cross_attn_k.lora_up.weight
55
+ lora_unet__blocks_2_cross_attn_v.lora_down.weight
56
+ lora_unet__blocks_2_cross_attn_v.lora_up.weight
57
+ lora_unet__blocks_2_cross_attn_o.lora_down.weight
58
+ lora_unet__blocks_2_cross_attn_o.lora_up.weight
59
+ lora_unet__blocks_2_ffn_0.lora_down.weight
60
+ lora_unet__blocks_2_ffn_0.lora_up.weight
61
+ lora_unet__blocks_2_ffn_2.lora_down.weight
62
+ lora_unet__blocks_2_ffn_2.lora_up.weight
63
+ lora_unet__blocks_3_self_attn_q.lora_down.weight
64
+ lora_unet__blocks_3_self_attn_q.lora_up.weight
65
+ lora_unet__blocks_3_self_attn_k.lora_down.weight
66
+ lora_unet__blocks_3_self_attn_k.lora_up.weight
67
+ lora_unet__blocks_3_self_attn_v.lora_down.weight
68
+ lora_unet__blocks_3_self_attn_v.lora_up.weight
69
+ lora_unet__blocks_3_self_attn_o.lora_down.weight
70
+ lora_unet__blocks_3_self_attn_o.lora_up.weight
71
+ lora_unet__blocks_3_cross_attn_q.lora_down.weight
72
+ lora_unet__blocks_3_cross_attn_q.lora_up.weight
73
+ lora_unet__blocks_3_cross_attn_k.lora_down.weight
74
+ lora_unet__blocks_3_cross_attn_k.lora_up.weight
75
+ lora_unet__blocks_3_cross_attn_v.lora_down.weight
76
+ lora_unet__blocks_3_cross_attn_v.lora_up.weight
77
+ lora_unet__blocks_3_cross_attn_o.lora_down.weight
78
+ lora_unet__blocks_3_cross_attn_o.lora_up.weight
79
+ lora_unet__blocks_3_ffn_0.lora_down.weight
80
+ lora_unet__blocks_3_ffn_0.lora_up.weight
81
+ lora_unet__blocks_3_ffn_2.lora_down.weight
82
+ lora_unet__blocks_3_ffn_2.lora_up.weight
83
+ lora_unet__blocks_4_self_attn_q.lora_down.weight
84
+ lora_unet__blocks_4_self_attn_q.lora_up.weight
85
+ lora_unet__blocks_4_self_attn_k.lora_down.weight
86
+ lora_unet__blocks_4_self_attn_k.lora_up.weight
87
+ lora_unet__blocks_4_self_attn_v.lora_down.weight
88
+ lora_unet__blocks_4_self_attn_v.lora_up.weight
89
+ lora_unet__blocks_4_self_attn_o.lora_down.weight
90
+ lora_unet__blocks_4_self_attn_o.lora_up.weight
91
+ lora_unet__blocks_4_cross_attn_q.lora_down.weight
92
+ lora_unet__blocks_4_cross_attn_q.lora_up.weight
93
+ lora_unet__blocks_4_cross_attn_k.lora_down.weight
94
+ lora_unet__blocks_4_cross_attn_k.lora_up.weight
95
+ lora_unet__blocks_4_cross_attn_v.lora_down.weight
96
+ lora_unet__blocks_4_cross_attn_v.lora_up.weight
97
+ lora_unet__blocks_4_cross_attn_o.lora_down.weight
98
+ lora_unet__blocks_4_cross_attn_o.lora_up.weight
99
+ lora_unet__blocks_4_ffn_0.lora_down.weight
100
+ lora_unet__blocks_4_ffn_0.lora_up.weight
101
+ lora_unet__blocks_4_ffn_2.lora_down.weight
102
+ lora_unet__blocks_4_ffn_2.lora_up.weight
103
+ lora_unet__blocks_5_self_attn_q.lora_down.weight
104
+ lora_unet__blocks_5_self_attn_q.lora_up.weight
105
+ lora_unet__blocks_5_self_attn_k.lora_down.weight
106
+ lora_unet__blocks_5_self_attn_k.lora_up.weight
107
+ lora_unet__blocks_5_self_attn_v.lora_down.weight
108
+ lora_unet__blocks_5_self_attn_v.lora_up.weight
109
+ lora_unet__blocks_5_self_attn_o.lora_down.weight
110
+ lora_unet__blocks_5_self_attn_o.lora_up.weight
111
+ lora_unet__blocks_5_cross_attn_q.lora_down.weight
112
+ lora_unet__blocks_5_cross_attn_q.lora_up.weight
113
+ lora_unet__blocks_5_cross_attn_k.lora_down.weight
114
+ lora_unet__blocks_5_cross_attn_k.lora_up.weight
115
+ lora_unet__blocks_5_cross_attn_v.lora_down.weight
116
+ lora_unet__blocks_5_cross_attn_v.lora_up.weight
117
+ lora_unet__blocks_5_cross_attn_o.lora_down.weight
118
+ lora_unet__blocks_5_cross_attn_o.lora_up.weight
119
+ lora_unet__blocks_5_ffn_0.lora_down.weight
120
+ lora_unet__blocks_5_ffn_0.lora_up.weight
121
+ lora_unet__blocks_5_ffn_2.lora_down.weight
122
+ lora_unet__blocks_5_ffn_2.lora_up.weight
123
+ lora_unet__blocks_6_self_attn_q.lora_down.weight
124
+ lora_unet__blocks_6_self_attn_q.lora_up.weight
125
+ lora_unet__blocks_6_self_attn_k.lora_down.weight
126
+ lora_unet__blocks_6_self_attn_k.lora_up.weight
127
+ lora_unet__blocks_6_self_attn_v.lora_down.weight
128
+ lora_unet__blocks_6_self_attn_v.lora_up.weight
129
+ lora_unet__blocks_6_self_attn_o.lora_down.weight
130
+ lora_unet__blocks_6_self_attn_o.lora_up.weight
131
+ lora_unet__blocks_6_cross_attn_q.lora_down.weight
132
+ lora_unet__blocks_6_cross_attn_q.lora_up.weight
133
+ lora_unet__blocks_6_cross_attn_k.lora_down.weight
134
+ lora_unet__blocks_6_cross_attn_k.lora_up.weight
135
+ lora_unet__blocks_6_cross_attn_v.lora_down.weight
136
+ lora_unet__blocks_6_cross_attn_v.lora_up.weight
137
+ lora_unet__blocks_6_cross_attn_o.lora_down.weight
138
+ lora_unet__blocks_6_cross_attn_o.lora_up.weight
139
+ lora_unet__blocks_6_ffn_0.lora_down.weight
140
+ lora_unet__blocks_6_ffn_0.lora_up.weight
141
+ lora_unet__blocks_6_ffn_2.lora_down.weight
142
+ lora_unet__blocks_6_ffn_2.lora_up.weight
143
+ lora_unet__blocks_7_self_attn_q.lora_down.weight
144
+ lora_unet__blocks_7_self_attn_q.lora_up.weight
145
+ lora_unet__blocks_7_self_attn_k.lora_down.weight
146
+ lora_unet__blocks_7_self_attn_k.lora_up.weight
147
+ lora_unet__blocks_7_self_attn_v.lora_down.weight
148
+ lora_unet__blocks_7_self_attn_v.lora_up.weight
149
+ lora_unet__blocks_7_self_attn_o.lora_down.weight
150
+ lora_unet__blocks_7_self_attn_o.lora_up.weight
151
+ lora_unet__blocks_7_cross_attn_q.lora_down.weight
152
+ lora_unet__blocks_7_cross_attn_q.lora_up.weight
153
+ lora_unet__blocks_7_cross_attn_k.lora_down.weight
154
+ lora_unet__blocks_7_cross_attn_k.lora_up.weight
155
+ lora_unet__blocks_7_cross_attn_v.lora_down.weight
156
+ lora_unet__blocks_7_cross_attn_v.lora_up.weight
157
+ lora_unet__blocks_7_cross_attn_o.lora_down.weight
158
+ lora_unet__blocks_7_cross_attn_o.lora_up.weight
159
+ lora_unet__blocks_7_ffn_0.lora_down.weight
160
+ lora_unet__blocks_7_ffn_0.lora_up.weight
161
+ lora_unet__blocks_7_ffn_2.lora_down.weight
162
+ lora_unet__blocks_7_ffn_2.lora_up.weight
163
+ lora_unet__blocks_8_self_attn_q.lora_down.weight
164
+ lora_unet__blocks_8_self_attn_q.lora_up.weight
165
+ lora_unet__blocks_8_self_attn_k.lora_down.weight
166
+ lora_unet__blocks_8_self_attn_k.lora_up.weight
167
+ lora_unet__blocks_8_self_attn_v.lora_down.weight
168
+ lora_unet__blocks_8_self_attn_v.lora_up.weight
169
+ lora_unet__blocks_8_self_attn_o.lora_down.weight
170
+ lora_unet__blocks_8_self_attn_o.lora_up.weight
171
+ lora_unet__blocks_8_cross_attn_q.lora_down.weight
172
+ lora_unet__blocks_8_cross_attn_q.lora_up.weight
173
+ lora_unet__blocks_8_cross_attn_k.lora_down.weight
174
+ lora_unet__blocks_8_cross_attn_k.lora_up.weight
175
+ lora_unet__blocks_8_cross_attn_v.lora_down.weight
176
+ lora_unet__blocks_8_cross_attn_v.lora_up.weight
177
+ lora_unet__blocks_8_cross_attn_o.lora_down.weight
178
+ lora_unet__blocks_8_cross_attn_o.lora_up.weight
179
+ lora_unet__blocks_8_ffn_0.lora_down.weight
180
+ lora_unet__blocks_8_ffn_0.lora_up.weight
181
+ lora_unet__blocks_8_ffn_2.lora_down.weight
182
+ lora_unet__blocks_8_ffn_2.lora_up.weight
183
+ lora_unet__blocks_9_self_attn_q.lora_down.weight
184
+ lora_unet__blocks_9_self_attn_q.lora_up.weight
185
+ lora_unet__blocks_9_self_attn_k.lora_down.weight
186
+ lora_unet__blocks_9_self_attn_k.lora_up.weight
187
+ lora_unet__blocks_9_self_attn_v.lora_down.weight
188
+ lora_unet__blocks_9_self_attn_v.lora_up.weight
189
+ lora_unet__blocks_9_self_attn_o.lora_down.weight
190
+ lora_unet__blocks_9_self_attn_o.lora_up.weight
191
+ lora_unet__blocks_9_cross_attn_q.lora_down.weight
192
+ lora_unet__blocks_9_cross_attn_q.lora_up.weight
193
+ lora_unet__blocks_9_cross_attn_k.lora_down.weight
194
+ lora_unet__blocks_9_cross_attn_k.lora_up.weight
195
+ lora_unet__blocks_9_cross_attn_v.lora_down.weight
196
+ lora_unet__blocks_9_cross_attn_v.lora_up.weight
197
+ lora_unet__blocks_9_cross_attn_o.lora_down.weight
198
+ lora_unet__blocks_9_cross_attn_o.lora_up.weight
199
+ lora_unet__blocks_9_ffn_0.lora_down.weight
200
+ lora_unet__blocks_9_ffn_0.lora_up.weight
201
+ lora_unet__blocks_9_ffn_2.lora_down.weight
202
+ lora_unet__blocks_9_ffn_2.lora_up.weight
203
+ lora_unet__blocks_10_self_attn_q.lora_down.weight
204
+ lora_unet__blocks_10_self_attn_q.lora_up.weight
205
+ lora_unet__blocks_10_self_attn_k.lora_down.weight
206
+ lora_unet__blocks_10_self_attn_k.lora_up.weight
207
+ lora_unet__blocks_10_self_attn_v.lora_down.weight
208
+ lora_unet__blocks_10_self_attn_v.lora_up.weight
209
+ lora_unet__blocks_10_self_attn_o.lora_down.weight
210
+ lora_unet__blocks_10_self_attn_o.lora_up.weight
211
+ lora_unet__blocks_10_cross_attn_q.lora_down.weight
212
+ lora_unet__blocks_10_cross_attn_q.lora_up.weight
213
+ lora_unet__blocks_10_cross_attn_k.lora_down.weight
214
+ lora_unet__blocks_10_cross_attn_k.lora_up.weight
215
+ lora_unet__blocks_10_cross_attn_v.lora_down.weight
216
+ lora_unet__blocks_10_cross_attn_v.lora_up.weight
217
+ lora_unet__blocks_10_cross_attn_o.lora_down.weight
218
+ lora_unet__blocks_10_cross_attn_o.lora_up.weight
219
+ lora_unet__blocks_10_ffn_0.lora_down.weight
220
+ lora_unet__blocks_10_ffn_0.lora_up.weight
221
+ lora_unet__blocks_10_ffn_2.lora_down.weight
222
+ lora_unet__blocks_10_ffn_2.lora_up.weight
223
+ lora_unet__blocks_11_self_attn_q.lora_down.weight
224
+ lora_unet__blocks_11_self_attn_q.lora_up.weight
225
+ lora_unet__blocks_11_self_attn_k.lora_down.weight
226
+ lora_unet__blocks_11_self_attn_k.lora_up.weight
227
+ lora_unet__blocks_11_self_attn_v.lora_down.weight
228
+ lora_unet__blocks_11_self_attn_v.lora_up.weight
229
+ lora_unet__blocks_11_self_attn_o.lora_down.weight
230
+ lora_unet__blocks_11_self_attn_o.lora_up.weight
231
+ lora_unet__blocks_11_cross_attn_q.lora_down.weight
232
+ lora_unet__blocks_11_cross_attn_q.lora_up.weight
233
+ lora_unet__blocks_11_cross_attn_k.lora_down.weight
234
+ lora_unet__blocks_11_cross_attn_k.lora_up.weight
235
+ lora_unet__blocks_11_cross_attn_v.lora_down.weight
236
+ lora_unet__blocks_11_cross_attn_v.lora_up.weight
237
+ lora_unet__blocks_11_cross_attn_o.lora_down.weight
238
+ lora_unet__blocks_11_cross_attn_o.lora_up.weight
239
+ lora_unet__blocks_11_ffn_0.lora_down.weight
240
+ lora_unet__blocks_11_ffn_0.lora_up.weight
241
+ lora_unet__blocks_11_ffn_2.lora_down.weight
242
+ lora_unet__blocks_11_ffn_2.lora_up.weight
243
+ lora_unet__blocks_12_self_attn_q.lora_down.weight
244
+ lora_unet__blocks_12_self_attn_q.lora_up.weight
245
+ lora_unet__blocks_12_self_attn_k.lora_down.weight
246
+ lora_unet__blocks_12_self_attn_k.lora_up.weight
247
+ lora_unet__blocks_12_self_attn_v.lora_down.weight
248
+ lora_unet__blocks_12_self_attn_v.lora_up.weight
249
+ lora_unet__blocks_12_self_attn_o.lora_down.weight
250
+ lora_unet__blocks_12_self_attn_o.lora_up.weight
251
+ lora_unet__blocks_12_cross_attn_q.lora_down.weight
252
+ lora_unet__blocks_12_cross_attn_q.lora_up.weight
253
+ lora_unet__blocks_12_cross_attn_k.lora_down.weight
254
+ lora_unet__blocks_12_cross_attn_k.lora_up.weight
255
+ lora_unet__blocks_12_cross_attn_v.lora_down.weight
256
+ lora_unet__blocks_12_cross_attn_v.lora_up.weight
257
+ lora_unet__blocks_12_cross_attn_o.lora_down.weight
258
+ lora_unet__blocks_12_cross_attn_o.lora_up.weight
259
+ lora_unet__blocks_12_ffn_0.lora_down.weight
260
+ lora_unet__blocks_12_ffn_0.lora_up.weight
261
+ lora_unet__blocks_12_ffn_2.lora_down.weight
262
+ lora_unet__blocks_12_ffn_2.lora_up.weight
263
+ lora_unet__blocks_13_self_attn_q.lora_down.weight
264
+ lora_unet__blocks_13_self_attn_q.lora_up.weight
265
+ lora_unet__blocks_13_self_attn_k.lora_down.weight
266
+ lora_unet__blocks_13_self_attn_k.lora_up.weight
267
+ lora_unet__blocks_13_self_attn_v.lora_down.weight
268
+ lora_unet__blocks_13_self_attn_v.lora_up.weight
269
+ lora_unet__blocks_13_self_attn_o.lora_down.weight
270
+ lora_unet__blocks_13_self_attn_o.lora_up.weight
271
+ lora_unet__blocks_13_cross_attn_q.lora_down.weight
272
+ lora_unet__blocks_13_cross_attn_q.lora_up.weight
273
+ lora_unet__blocks_13_cross_attn_k.lora_down.weight
274
+ lora_unet__blocks_13_cross_attn_k.lora_up.weight
275
+ lora_unet__blocks_13_cross_attn_v.lora_down.weight
276
+ lora_unet__blocks_13_cross_attn_v.lora_up.weight
277
+ lora_unet__blocks_13_cross_attn_o.lora_down.weight
278
+ lora_unet__blocks_13_cross_attn_o.lora_up.weight
279
+ lora_unet__blocks_13_ffn_0.lora_down.weight
280
+ lora_unet__blocks_13_ffn_0.lora_up.weight
281
+ lora_unet__blocks_13_ffn_2.lora_down.weight
282
+ lora_unet__blocks_13_ffn_2.lora_up.weight
283
+ lora_unet__blocks_14_self_attn_q.lora_down.weight
284
+ lora_unet__blocks_14_self_attn_q.lora_up.weight
285
+ lora_unet__blocks_14_self_attn_k.lora_down.weight
286
+ lora_unet__blocks_14_self_attn_k.lora_up.weight
287
+ lora_unet__blocks_14_self_attn_v.lora_down.weight
288
+ lora_unet__blocks_14_self_attn_v.lora_up.weight
289
+ lora_unet__blocks_14_self_attn_o.lora_down.weight
290
+ lora_unet__blocks_14_self_attn_o.lora_up.weight
291
+ lora_unet__blocks_14_cross_attn_q.lora_down.weight
292
+ lora_unet__blocks_14_cross_attn_q.lora_up.weight
293
+ lora_unet__blocks_14_cross_attn_k.lora_down.weight
294
+ lora_unet__blocks_14_cross_attn_k.lora_up.weight
295
+ lora_unet__blocks_14_cross_attn_v.lora_down.weight
296
+ lora_unet__blocks_14_cross_attn_v.lora_up.weight
297
+ lora_unet__blocks_14_cross_attn_o.lora_down.weight
298
+ lora_unet__blocks_14_cross_attn_o.lora_up.weight
299
+ lora_unet__blocks_14_ffn_0.lora_down.weight
300
+ lora_unet__blocks_14_ffn_0.lora_up.weight
301
+ lora_unet__blocks_14_ffn_2.lora_down.weight
302
+ lora_unet__blocks_14_ffn_2.lora_up.weight
303
+ lora_unet__blocks_15_self_attn_q.lora_down.weight
304
+ lora_unet__blocks_15_self_attn_q.lora_up.weight
305
+ lora_unet__blocks_15_self_attn_k.lora_down.weight
306
+ lora_unet__blocks_15_self_attn_k.lora_up.weight
307
+ lora_unet__blocks_15_self_attn_v.lora_down.weight
308
+ lora_unet__blocks_15_self_attn_v.lora_up.weight
309
+ lora_unet__blocks_15_self_attn_o.lora_down.weight
310
+ lora_unet__blocks_15_self_attn_o.lora_up.weight
311
+ lora_unet__blocks_15_cross_attn_q.lora_down.weight
312
+ lora_unet__blocks_15_cross_attn_q.lora_up.weight
313
+ lora_unet__blocks_15_cross_attn_k.lora_down.weight
314
+ lora_unet__blocks_15_cross_attn_k.lora_up.weight
315
+ lora_unet__blocks_15_cross_attn_v.lora_down.weight
316
+ lora_unet__blocks_15_cross_attn_v.lora_up.weight
317
+ lora_unet__blocks_15_cross_attn_o.lora_down.weight
318
+ lora_unet__blocks_15_cross_attn_o.lora_up.weight
319
+ lora_unet__blocks_15_ffn_0.lora_down.weight
320
+ lora_unet__blocks_15_ffn_0.lora_up.weight
321
+ lora_unet__blocks_15_ffn_2.lora_down.weight
322
+ lora_unet__blocks_15_ffn_2.lora_up.weight
323
+ lora_unet__blocks_16_self_attn_q.lora_down.weight
324
+ lora_unet__blocks_16_self_attn_q.lora_up.weight
325
+ lora_unet__blocks_16_self_attn_k.lora_down.weight
326
+ lora_unet__blocks_16_self_attn_k.lora_up.weight
327
+ lora_unet__blocks_16_self_attn_v.lora_down.weight
328
+ lora_unet__blocks_16_self_attn_v.lora_up.weight
329
+ lora_unet__blocks_16_self_attn_o.lora_down.weight
330
+ lora_unet__blocks_16_self_attn_o.lora_up.weight
331
+ lora_unet__blocks_16_cross_attn_q.lora_down.weight
332
+ lora_unet__blocks_16_cross_attn_q.lora_up.weight
333
+ lora_unet__blocks_16_cross_attn_k.lora_down.weight
334
+ lora_unet__blocks_16_cross_attn_k.lora_up.weight
335
+ lora_unet__blocks_16_cross_attn_v.lora_down.weight
336
+ lora_unet__blocks_16_cross_attn_v.lora_up.weight
337
+ lora_unet__blocks_16_cross_attn_o.lora_down.weight
338
+ lora_unet__blocks_16_cross_attn_o.lora_up.weight
339
+ lora_unet__blocks_16_ffn_0.lora_down.weight
340
+ lora_unet__blocks_16_ffn_0.lora_up.weight
341
+ lora_unet__blocks_16_ffn_2.lora_down.weight
342
+ lora_unet__blocks_16_ffn_2.lora_up.weight
343
+ lora_unet__blocks_17_self_attn_q.lora_down.weight
344
+ lora_unet__blocks_17_self_attn_q.lora_up.weight
345
+ lora_unet__blocks_17_self_attn_k.lora_down.weight
346
+ lora_unet__blocks_17_self_attn_k.lora_up.weight
347
+ lora_unet__blocks_17_self_attn_v.lora_down.weight
348
+ lora_unet__blocks_17_self_attn_v.lora_up.weight
349
+ lora_unet__blocks_17_self_attn_o.lora_down.weight
350
+ lora_unet__blocks_17_self_attn_o.lora_up.weight
351
+ lora_unet__blocks_17_cross_attn_q.lora_down.weight
352
+ lora_unet__blocks_17_cross_attn_q.lora_up.weight
353
+ lora_unet__blocks_17_cross_attn_k.lora_down.weight
354
+ lora_unet__blocks_17_cross_attn_k.lora_up.weight
355
+ lora_unet__blocks_17_cross_attn_v.lora_down.weight
356
+ lora_unet__blocks_17_cross_attn_v.lora_up.weight
357
+ lora_unet__blocks_17_cross_attn_o.lora_down.weight
358
+ lora_unet__blocks_17_cross_attn_o.lora_up.weight
359
+ lora_unet__blocks_17_ffn_0.lora_down.weight
360
+ lora_unet__blocks_17_ffn_0.lora_up.weight
361
+ lora_unet__blocks_17_ffn_2.lora_down.weight
362
+ lora_unet__blocks_17_ffn_2.lora_up.weight
363
+ lora_unet__blocks_18_self_attn_q.lora_down.weight
364
+ lora_unet__blocks_18_self_attn_q.lora_up.weight
365
+ lora_unet__blocks_18_self_attn_k.lora_down.weight
366
+ lora_unet__blocks_18_self_attn_k.lora_up.weight
367
+ lora_unet__blocks_18_self_attn_v.lora_down.weight
368
+ lora_unet__blocks_18_self_attn_v.lora_up.weight
369
+ lora_unet__blocks_18_self_attn_o.lora_down.weight
370
+ lora_unet__blocks_18_self_attn_o.lora_up.weight
371
+ lora_unet__blocks_18_cross_attn_q.lora_down.weight
372
+ lora_unet__blocks_18_cross_attn_q.lora_up.weight
373
+ lora_unet__blocks_18_cross_attn_k.lora_down.weight
374
+ lora_unet__blocks_18_cross_attn_k.lora_up.weight
375
+ lora_unet__blocks_18_cross_attn_v.lora_down.weight
376
+ lora_unet__blocks_18_cross_attn_v.lora_up.weight
377
+ lora_unet__blocks_18_cross_attn_o.lora_down.weight
378
+ lora_unet__blocks_18_cross_attn_o.lora_up.weight
379
+ lora_unet__blocks_18_ffn_0.lora_down.weight
380
+ lora_unet__blocks_18_ffn_0.lora_up.weight
381
+ lora_unet__blocks_18_ffn_2.lora_down.weight
382
+ lora_unet__blocks_18_ffn_2.lora_up.weight
383
+ lora_unet__blocks_19_self_attn_q.lora_down.weight
384
+ lora_unet__blocks_19_self_attn_q.lora_up.weight
385
+ lora_unet__blocks_19_self_attn_k.lora_down.weight
386
+ lora_unet__blocks_19_self_attn_k.lora_up.weight
387
+ lora_unet__blocks_19_self_attn_v.lora_down.weight
388
+ lora_unet__blocks_19_self_attn_v.lora_up.weight
389
+ lora_unet__blocks_19_self_attn_o.lora_down.weight
390
+ lora_unet__blocks_19_self_attn_o.lora_up.weight
391
+ lora_unet__blocks_19_cross_attn_q.lora_down.weight
392
+ lora_unet__blocks_19_cross_attn_q.lora_up.weight
393
+ lora_unet__blocks_19_cross_attn_k.lora_down.weight
394
+ lora_unet__blocks_19_cross_attn_k.lora_up.weight
395
+ lora_unet__blocks_19_cross_attn_v.lora_down.weight
396
+ lora_unet__blocks_19_cross_attn_v.lora_up.weight
397
+ lora_unet__blocks_19_cross_attn_o.lora_down.weight
398
+ lora_unet__blocks_19_cross_attn_o.lora_up.weight
399
+ lora_unet__blocks_19_ffn_0.lora_down.weight
400
+ lora_unet__blocks_19_ffn_0.lora_up.weight
401
+ lora_unet__blocks_19_ffn_2.lora_down.weight
402
+ lora_unet__blocks_19_ffn_2.lora_up.weight
403
+ lora_unet__blocks_20_self_attn_q.lora_down.weight
404
+ lora_unet__blocks_20_self_attn_q.lora_up.weight
405
+ lora_unet__blocks_20_self_attn_k.lora_down.weight
406
+ lora_unet__blocks_20_self_attn_k.lora_up.weight
407
+ lora_unet__blocks_20_self_attn_v.lora_down.weight
408
+ lora_unet__blocks_20_self_attn_v.lora_up.weight
409
+ lora_unet__blocks_20_self_attn_o.lora_down.weight
410
+ lora_unet__blocks_20_self_attn_o.lora_up.weight
411
+ lora_unet__blocks_20_cross_attn_q.lora_down.weight
412
+ lora_unet__blocks_20_cross_attn_q.lora_up.weight
413
+ lora_unet__blocks_20_cross_attn_k.lora_down.weight
414
+ lora_unet__blocks_20_cross_attn_k.lora_up.weight
415
+ lora_unet__blocks_20_cross_attn_v.lora_down.weight
416
+ lora_unet__blocks_20_cross_attn_v.lora_up.weight
417
+ lora_unet__blocks_20_cross_attn_o.lora_down.weight
418
+ lora_unet__blocks_20_cross_attn_o.lora_up.weight
419
+ lora_unet__blocks_20_ffn_0.lora_down.weight
420
+ lora_unet__blocks_20_ffn_0.lora_up.weight
421
+ lora_unet__blocks_20_ffn_2.lora_down.weight
422
+ lora_unet__blocks_20_ffn_2.lora_up.weight
423
+ lora_unet__blocks_21_self_attn_q.lora_down.weight
424
+ lora_unet__blocks_21_self_attn_q.lora_up.weight
425
+ lora_unet__blocks_21_self_attn_k.lora_down.weight
426
+ lora_unet__blocks_21_self_attn_k.lora_up.weight
427
+ lora_unet__blocks_21_self_attn_v.lora_down.weight
428
+ lora_unet__blocks_21_self_attn_v.lora_up.weight
429
+ lora_unet__blocks_21_self_attn_o.lora_down.weight
430
+ lora_unet__blocks_21_self_attn_o.lora_up.weight
431
+ lora_unet__blocks_21_cross_attn_q.lora_down.weight
432
+ lora_unet__blocks_21_cross_attn_q.lora_up.weight
433
+ lora_unet__blocks_21_cross_attn_k.lora_down.weight
434
+ lora_unet__blocks_21_cross_attn_k.lora_up.weight
435
+ lora_unet__blocks_21_cross_attn_v.lora_down.weight
436
+ lora_unet__blocks_21_cross_attn_v.lora_up.weight
437
+ lora_unet__blocks_21_cross_attn_o.lora_down.weight
438
+ lora_unet__blocks_21_cross_attn_o.lora_up.weight
439
+ lora_unet__blocks_21_ffn_0.lora_down.weight
440
+ lora_unet__blocks_21_ffn_0.lora_up.weight
441
+ lora_unet__blocks_21_ffn_2.lora_down.weight
442
+ lora_unet__blocks_21_ffn_2.lora_up.weight
443
+ lora_unet__blocks_22_self_attn_q.lora_down.weight
444
+ lora_unet__blocks_22_self_attn_q.lora_up.weight
445
+ lora_unet__blocks_22_self_attn_k.lora_down.weight
446
+ lora_unet__blocks_22_self_attn_k.lora_up.weight
447
+ lora_unet__blocks_22_self_attn_v.lora_down.weight
448
+ lora_unet__blocks_22_self_attn_v.lora_up.weight
449
+ lora_unet__blocks_22_self_attn_o.lora_down.weight
450
+ lora_unet__blocks_22_self_attn_o.lora_up.weight
451
+ lora_unet__blocks_22_cross_attn_q.lora_down.weight
452
+ lora_unet__blocks_22_cross_attn_q.lora_up.weight
453
+ lora_unet__blocks_22_cross_attn_k.lora_down.weight
454
+ lora_unet__blocks_22_cross_attn_k.lora_up.weight
455
+ lora_unet__blocks_22_cross_attn_v.lora_down.weight
456
+ lora_unet__blocks_22_cross_attn_v.lora_up.weight
457
+ lora_unet__blocks_22_cross_attn_o.lora_down.weight
458
+ lora_unet__blocks_22_cross_attn_o.lora_up.weight
459
+ lora_unet__blocks_22_ffn_0.lora_down.weight
460
+ lora_unet__blocks_22_ffn_0.lora_up.weight
461
+ lora_unet__blocks_22_ffn_2.lora_down.weight
462
+ lora_unet__blocks_22_ffn_2.lora_up.weight
463
+ lora_unet__blocks_23_self_attn_q.lora_down.weight
464
+ lora_unet__blocks_23_self_attn_q.lora_up.weight
465
+ lora_unet__blocks_23_self_attn_k.lora_down.weight
466
+ lora_unet__blocks_23_self_attn_k.lora_up.weight
467
+ lora_unet__blocks_23_self_attn_v.lora_down.weight
468
+ lora_unet__blocks_23_self_attn_v.lora_up.weight
469
+ lora_unet__blocks_23_self_attn_o.lora_down.weight
470
+ lora_unet__blocks_23_self_attn_o.lora_up.weight
471
+ lora_unet__blocks_23_cross_attn_q.lora_down.weight
472
+ lora_unet__blocks_23_cross_attn_q.lora_up.weight
473
+ lora_unet__blocks_23_cross_attn_k.lora_down.weight
474
+ lora_unet__blocks_23_cross_attn_k.lora_up.weight
475
+ lora_unet__blocks_23_cross_attn_v.lora_down.weight
476
+ lora_unet__blocks_23_cross_attn_v.lora_up.weight
477
+ lora_unet__blocks_23_cross_attn_o.lora_down.weight
478
+ lora_unet__blocks_23_cross_attn_o.lora_up.weight
479
+ lora_unet__blocks_23_ffn_0.lora_down.weight
480
+ lora_unet__blocks_23_ffn_0.lora_up.weight
481
+ lora_unet__blocks_23_ffn_2.lora_down.weight
482
+ lora_unet__blocks_23_ffn_2.lora_up.weight
483
+ lora_unet__blocks_24_self_attn_q.lora_down.weight
484
+ lora_unet__blocks_24_self_attn_q.lora_up.weight
485
+ lora_unet__blocks_24_self_attn_k.lora_down.weight
486
+ lora_unet__blocks_24_self_attn_k.lora_up.weight
487
+ lora_unet__blocks_24_self_attn_v.lora_down.weight
488
+ lora_unet__blocks_24_self_attn_v.lora_up.weight
489
+ lora_unet__blocks_24_self_attn_o.lora_down.weight
490
+ lora_unet__blocks_24_self_attn_o.lora_up.weight
491
+ lora_unet__blocks_24_cross_attn_q.lora_down.weight
492
+ lora_unet__blocks_24_cross_attn_q.lora_up.weight
493
+ lora_unet__blocks_24_cross_attn_k.lora_down.weight
494
+ lora_unet__blocks_24_cross_attn_k.lora_up.weight
495
+ lora_unet__blocks_24_cross_attn_v.lora_down.weight
496
+ lora_unet__blocks_24_cross_attn_v.lora_up.weight
497
+ lora_unet__blocks_24_cross_attn_o.lora_down.weight
498
+ lora_unet__blocks_24_cross_attn_o.lora_up.weight
499
+ lora_unet__blocks_24_ffn_0.lora_down.weight
500
+ lora_unet__blocks_24_ffn_0.lora_up.weight
501
+ lora_unet__blocks_24_ffn_2.lora_down.weight
502
+ lora_unet__blocks_24_ffn_2.lora_up.weight
503
+ lora_unet__blocks_25_self_attn_q.lora_down.weight
504
+ lora_unet__blocks_25_self_attn_q.lora_up.weight
505
+ lora_unet__blocks_25_self_attn_k.lora_down.weight
506
+ lora_unet__blocks_25_self_attn_k.lora_up.weight
507
+ lora_unet__blocks_25_self_attn_v.lora_down.weight
508
+ lora_unet__blocks_25_self_attn_v.lora_up.weight
509
+ lora_unet__blocks_25_self_attn_o.lora_down.weight
510
+ lora_unet__blocks_25_self_attn_o.lora_up.weight
511
+ lora_unet__blocks_25_cross_attn_q.lora_down.weight
512
+ lora_unet__blocks_25_cross_attn_q.lora_up.weight
513
+ lora_unet__blocks_25_cross_attn_k.lora_down.weight
514
+ lora_unet__blocks_25_cross_attn_k.lora_up.weight
515
+ lora_unet__blocks_25_cross_attn_v.lora_down.weight
516
+ lora_unet__blocks_25_cross_attn_v.lora_up.weight
517
+ lora_unet__blocks_25_cross_attn_o.lora_down.weight
518
+ lora_unet__blocks_25_cross_attn_o.lora_up.weight
519
+ lora_unet__blocks_25_ffn_0.lora_down.weight
520
+ lora_unet__blocks_25_ffn_0.lora_up.weight
521
+ lora_unet__blocks_25_ffn_2.lora_down.weight
522
+ lora_unet__blocks_25_ffn_2.lora_up.weight
523
+ lora_unet__blocks_26_self_attn_q.lora_down.weight
524
+ lora_unet__blocks_26_self_attn_q.lora_up.weight
525
+ lora_unet__blocks_26_self_attn_k.lora_down.weight
526
+ lora_unet__blocks_26_self_attn_k.lora_up.weight
527
+ lora_unet__blocks_26_self_attn_v.lora_down.weight
528
+ lora_unet__blocks_26_self_attn_v.lora_up.weight
529
+ lora_unet__blocks_26_self_attn_o.lora_down.weight
530
+ lora_unet__blocks_26_self_attn_o.lora_up.weight
531
+ lora_unet__blocks_26_cross_attn_q.lora_down.weight
532
+ lora_unet__blocks_26_cross_attn_q.lora_up.weight
533
+ lora_unet__blocks_26_cross_attn_k.lora_down.weight
534
+ lora_unet__blocks_26_cross_attn_k.lora_up.weight
535
+ lora_unet__blocks_26_cross_attn_v.lora_down.weight
536
+ lora_unet__blocks_26_cross_attn_v.lora_up.weight
537
+ lora_unet__blocks_26_cross_attn_o.lora_down.weight
538
+ lora_unet__blocks_26_cross_attn_o.lora_up.weight
539
+ lora_unet__blocks_26_ffn_0.lora_down.weight
540
+ lora_unet__blocks_26_ffn_0.lora_up.weight
541
+ lora_unet__blocks_26_ffn_2.lora_down.weight
542
+ lora_unet__blocks_26_ffn_2.lora_up.weight
543
+ lora_unet__blocks_27_self_attn_q.lora_down.weight
544
+ lora_unet__blocks_27_self_attn_q.lora_up.weight
545
+ lora_unet__blocks_27_self_attn_k.lora_down.weight
546
+ lora_unet__blocks_27_self_attn_k.lora_up.weight
547
+ lora_unet__blocks_27_self_attn_v.lora_down.weight
548
+ lora_unet__blocks_27_self_attn_v.lora_up.weight
549
+ lora_unet__blocks_27_self_attn_o.lora_down.weight
550
+ lora_unet__blocks_27_self_attn_o.lora_up.weight
551
+ lora_unet__blocks_27_cross_attn_q.lora_down.weight
552
+ lora_unet__blocks_27_cross_attn_q.lora_up.weight
553
+ lora_unet__blocks_27_cross_attn_k.lora_down.weight
554
+ lora_unet__blocks_27_cross_attn_k.lora_up.weight
555
+ lora_unet__blocks_27_cross_attn_v.lora_down.weight
556
+ lora_unet__blocks_27_cross_attn_v.lora_up.weight
557
+ lora_unet__blocks_27_cross_attn_o.lora_down.weight
558
+ lora_unet__blocks_27_cross_attn_o.lora_up.weight
559
+ lora_unet__blocks_27_ffn_0.lora_down.weight
560
+ lora_unet__blocks_27_ffn_0.lora_up.weight
561
+ lora_unet__blocks_27_ffn_2.lora_down.weight
562
+ lora_unet__blocks_27_ffn_2.lora_up.weight
563
+ lora_unet__blocks_28_self_attn_q.lora_down.weight
564
+ lora_unet__blocks_28_self_attn_q.lora_up.weight
565
+ lora_unet__blocks_28_self_attn_k.lora_down.weight
566
+ lora_unet__blocks_28_self_attn_k.lora_up.weight
567
+ lora_unet__blocks_28_self_attn_v.lora_down.weight
568
+ lora_unet__blocks_28_self_attn_v.lora_up.weight
569
+ lora_unet__blocks_28_self_attn_o.lora_down.weight
570
+ lora_unet__blocks_28_self_attn_o.lora_up.weight
571
+ lora_unet__blocks_28_cross_attn_q.lora_down.weight
572
+ lora_unet__blocks_28_cross_attn_q.lora_up.weight
573
+ lora_unet__blocks_28_cross_attn_k.lora_down.weight
574
+ lora_unet__blocks_28_cross_attn_k.lora_up.weight
575
+ lora_unet__blocks_28_cross_attn_v.lora_down.weight
576
+ lora_unet__blocks_28_cross_attn_v.lora_up.weight
577
+ lora_unet__blocks_28_cross_attn_o.lora_down.weight
578
+ lora_unet__blocks_28_cross_attn_o.lora_up.weight
579
+ lora_unet__blocks_28_ffn_0.lora_down.weight
580
+ lora_unet__blocks_28_ffn_0.lora_up.weight
581
+ lora_unet__blocks_28_ffn_2.lora_down.weight
582
+ lora_unet__blocks_28_ffn_2.lora_up.weight
583
+ lora_unet__blocks_29_self_attn_q.lora_down.weight
584
+ lora_unet__blocks_29_self_attn_q.lora_up.weight
585
+ lora_unet__blocks_29_self_attn_k.lora_down.weight
586
+ lora_unet__blocks_29_self_attn_k.lora_up.weight
587
+ lora_unet__blocks_29_self_attn_v.lora_down.weight
588
+ lora_unet__blocks_29_self_attn_v.lora_up.weight
589
+ lora_unet__blocks_29_self_attn_o.lora_down.weight
590
+ lora_unet__blocks_29_self_attn_o.lora_up.weight
591
+ lora_unet__blocks_29_cross_attn_q.lora_down.weight
592
+ lora_unet__blocks_29_cross_attn_q.lora_up.weight
593
+ lora_unet__blocks_29_cross_attn_k.lora_down.weight
594
+ lora_unet__blocks_29_cross_attn_k.lora_up.weight
595
+ lora_unet__blocks_29_cross_attn_v.lora_down.weight
596
+ lora_unet__blocks_29_cross_attn_v.lora_up.weight
597
+ lora_unet__blocks_29_cross_attn_o.lora_down.weight
598
+ lora_unet__blocks_29_cross_attn_o.lora_up.weight
599
+ lora_unet__blocks_29_ffn_0.lora_down.weight
600
+ lora_unet__blocks_29_ffn_0.lora_up.weight
601
+ lora_unet__blocks_29_ffn_2.lora_down.weight
602
+ lora_unet__blocks_29_ffn_2.lora_up.weight
603
+ lora_unet__blocks_30_self_attn_q.lora_down.weight
604
+ lora_unet__blocks_30_self_attn_q.lora_up.weight
605
+ lora_unet__blocks_30_self_attn_k.lora_down.weight
606
+ lora_unet__blocks_30_self_attn_k.lora_up.weight
607
+ lora_unet__blocks_30_self_attn_v.lora_down.weight
608
+ lora_unet__blocks_30_self_attn_v.lora_up.weight
609
+ lora_unet__blocks_30_self_attn_o.lora_down.weight
610
+ lora_unet__blocks_30_self_attn_o.lora_up.weight
611
+ lora_unet__blocks_30_cross_attn_q.lora_down.weight
612
+ lora_unet__blocks_30_cross_attn_q.lora_up.weight
613
+ lora_unet__blocks_30_cross_attn_k.lora_down.weight
614
+ lora_unet__blocks_30_cross_attn_k.lora_up.weight
615
+ lora_unet__blocks_30_cross_attn_v.lora_down.weight
616
+ lora_unet__blocks_30_cross_attn_v.lora_up.weight
617
+ lora_unet__blocks_30_cross_attn_o.lora_down.weight
618
+ lora_unet__blocks_30_cross_attn_o.lora_up.weight
619
+ lora_unet__blocks_30_ffn_0.lora_down.weight
620
+ lora_unet__blocks_30_ffn_0.lora_up.weight
621
+ lora_unet__blocks_30_ffn_2.lora_down.weight
622
+ lora_unet__blocks_30_ffn_2.lora_up.weight
623
+ lora_unet__blocks_31_self_attn_q.lora_down.weight
624
+ lora_unet__blocks_31_self_attn_q.lora_up.weight
625
+ lora_unet__blocks_31_self_attn_k.lora_down.weight
626
+ lora_unet__blocks_31_self_attn_k.lora_up.weight
627
+ lora_unet__blocks_31_self_attn_v.lora_down.weight
628
+ lora_unet__blocks_31_self_attn_v.lora_up.weight
629
+ lora_unet__blocks_31_self_attn_o.lora_down.weight
630
+ lora_unet__blocks_31_self_attn_o.lora_up.weight
631
+ lora_unet__blocks_31_cross_attn_q.lora_down.weight
632
+ lora_unet__blocks_31_cross_attn_q.lora_up.weight
633
+ lora_unet__blocks_31_cross_attn_k.lora_down.weight
634
+ lora_unet__blocks_31_cross_attn_k.lora_up.weight
635
+ lora_unet__blocks_31_cross_attn_v.lora_down.weight
636
+ lora_unet__blocks_31_cross_attn_v.lora_up.weight
637
+ lora_unet__blocks_31_cross_attn_o.lora_down.weight
638
+ lora_unet__blocks_31_cross_attn_o.lora_up.weight
639
+ lora_unet__blocks_31_ffn_0.lora_down.weight
640
+ lora_unet__blocks_31_ffn_0.lora_up.weight
641
+ lora_unet__blocks_31_ffn_2.lora_down.weight
642
+ lora_unet__blocks_31_ffn_2.lora_up.weight
643
+ lora_unet__blocks_32_self_attn_q.lora_down.weight
644
+ lora_unet__blocks_32_self_attn_q.lora_up.weight
645
+ lora_unet__blocks_32_self_attn_k.lora_down.weight
646
+ lora_unet__blocks_32_self_attn_k.lora_up.weight
647
+ lora_unet__blocks_32_self_attn_v.lora_down.weight
648
+ lora_unet__blocks_32_self_attn_v.lora_up.weight
649
+ lora_unet__blocks_32_self_attn_o.lora_down.weight
650
+ lora_unet__blocks_32_self_attn_o.lora_up.weight
651
+ lora_unet__blocks_32_cross_attn_q.lora_down.weight
652
+ lora_unet__blocks_32_cross_attn_q.lora_up.weight
653
+ lora_unet__blocks_32_cross_attn_k.lora_down.weight
654
+ lora_unet__blocks_32_cross_attn_k.lora_up.weight
655
+ lora_unet__blocks_32_cross_attn_v.lora_down.weight
656
+ lora_unet__blocks_32_cross_attn_v.lora_up.weight
657
+ lora_unet__blocks_32_cross_attn_o.lora_down.weight
658
+ lora_unet__blocks_32_cross_attn_o.lora_up.weight
659
+ lora_unet__blocks_32_ffn_0.lora_down.weight
660
+ lora_unet__blocks_32_ffn_0.lora_up.weight
661
+ lora_unet__blocks_32_ffn_2.lora_down.weight
662
+ lora_unet__blocks_32_ffn_2.lora_up.weight
663
+ lora_unet__blocks_33_self_attn_q.lora_down.weight
664
+ lora_unet__blocks_33_self_attn_q.lora_up.weight
665
+ lora_unet__blocks_33_self_attn_k.lora_down.weight
666
+ lora_unet__blocks_33_self_attn_k.lora_up.weight
667
+ lora_unet__blocks_33_self_attn_v.lora_down.weight
668
+ lora_unet__blocks_33_self_attn_v.lora_up.weight
669
+ lora_unet__blocks_33_self_attn_o.lora_down.weight
670
+ lora_unet__blocks_33_self_attn_o.lora_up.weight
671
+ lora_unet__blocks_33_cross_attn_q.lora_down.weight
672
+ lora_unet__blocks_33_cross_attn_q.lora_up.weight
673
+ lora_unet__blocks_33_cross_attn_k.lora_down.weight
674
+ lora_unet__blocks_33_cross_attn_k.lora_up.weight
675
+ lora_unet__blocks_33_cross_attn_v.lora_down.weight
676
+ lora_unet__blocks_33_cross_attn_v.lora_up.weight
677
+ lora_unet__blocks_33_cross_attn_o.lora_down.weight
678
+ lora_unet__blocks_33_cross_attn_o.lora_up.weight
679
+ lora_unet__blocks_33_ffn_0.lora_down.weight
680
+ lora_unet__blocks_33_ffn_0.lora_up.weight
681
+ lora_unet__blocks_33_ffn_2.lora_down.weight
682
+ lora_unet__blocks_33_ffn_2.lora_up.weight
683
+ lora_unet__blocks_34_self_attn_q.lora_down.weight
684
+ lora_unet__blocks_34_self_attn_q.lora_up.weight
685
+ lora_unet__blocks_34_self_attn_k.lora_down.weight
686
+ lora_unet__blocks_34_self_attn_k.lora_up.weight
687
+ lora_unet__blocks_34_self_attn_v.lora_down.weight
688
+ lora_unet__blocks_34_self_attn_v.lora_up.weight
689
+ lora_unet__blocks_34_self_attn_o.lora_down.weight
690
+ lora_unet__blocks_34_self_attn_o.lora_up.weight
691
+ lora_unet__blocks_34_cross_attn_q.lora_down.weight
692
+ lora_unet__blocks_34_cross_attn_q.lora_up.weight
693
+ lora_unet__blocks_34_cross_attn_k.lora_down.weight
694
+ lora_unet__blocks_34_cross_attn_k.lora_up.weight
695
+ lora_unet__blocks_34_cross_attn_v.lora_down.weight
696
+ lora_unet__blocks_34_cross_attn_v.lora_up.weight
697
+ lora_unet__blocks_34_cross_attn_o.lora_down.weight
698
+ lora_unet__blocks_34_cross_attn_o.lora_up.weight
699
+ lora_unet__blocks_34_ffn_0.lora_down.weight
700
+ lora_unet__blocks_34_ffn_0.lora_up.weight
701
+ lora_unet__blocks_34_ffn_2.lora_down.weight
702
+ lora_unet__blocks_34_ffn_2.lora_up.weight
703
+ lora_unet__blocks_35_self_attn_q.lora_down.weight
704
+ lora_unet__blocks_35_self_attn_q.lora_up.weight
705
+ lora_unet__blocks_35_self_attn_k.lora_down.weight
706
+ lora_unet__blocks_35_self_attn_k.lora_up.weight
707
+ lora_unet__blocks_35_self_attn_v.lora_down.weight
708
+ lora_unet__blocks_35_self_attn_v.lora_up.weight
709
+ lora_unet__blocks_35_self_attn_o.lora_down.weight
710
+ lora_unet__blocks_35_self_attn_o.lora_up.weight
711
+ lora_unet__blocks_35_cross_attn_q.lora_down.weight
712
+ lora_unet__blocks_35_cross_attn_q.lora_up.weight
713
+ lora_unet__blocks_35_cross_attn_k.lora_down.weight
714
+ lora_unet__blocks_35_cross_attn_k.lora_up.weight
715
+ lora_unet__blocks_35_cross_attn_v.lora_down.weight
716
+ lora_unet__blocks_35_cross_attn_v.lora_up.weight
717
+ lora_unet__blocks_35_cross_attn_o.lora_down.weight
718
+ lora_unet__blocks_35_cross_attn_o.lora_up.weight
719
+ lora_unet__blocks_35_ffn_0.lora_down.weight
720
+ lora_unet__blocks_35_ffn_0.lora_up.weight
721
+ lora_unet__blocks_35_ffn_2.lora_down.weight
722
+ lora_unet__blocks_35_ffn_2.lora_up.weight
723
+ lora_unet__blocks_36_self_attn_q.lora_down.weight
724
+ lora_unet__blocks_36_self_attn_q.lora_up.weight
725
+ lora_unet__blocks_36_self_attn_k.lora_down.weight
726
+ lora_unet__blocks_36_self_attn_k.lora_up.weight
727
+ lora_unet__blocks_36_self_attn_v.lora_down.weight
728
+ lora_unet__blocks_36_self_attn_v.lora_up.weight
729
+ lora_unet__blocks_36_self_attn_o.lora_down.weight
730
+ lora_unet__blocks_36_self_attn_o.lora_up.weight
731
+ lora_unet__blocks_36_cross_attn_q.lora_down.weight
732
+ lora_unet__blocks_36_cross_attn_q.lora_up.weight
733
+ lora_unet__blocks_36_cross_attn_k.lora_down.weight
734
+ lora_unet__blocks_36_cross_attn_k.lora_up.weight
735
+ lora_unet__blocks_36_cross_attn_v.lora_down.weight
736
+ lora_unet__blocks_36_cross_attn_v.lora_up.weight
737
+ lora_unet__blocks_36_cross_attn_o.lora_down.weight
738
+ lora_unet__blocks_36_cross_attn_o.lora_up.weight
739
+ lora_unet__blocks_36_ffn_0.lora_down.weight
740
+ lora_unet__blocks_36_ffn_0.lora_up.weight
741
+ lora_unet__blocks_36_ffn_2.lora_down.weight
742
+ lora_unet__blocks_36_ffn_2.lora_up.weight
743
+ lora_unet__blocks_37_self_attn_q.lora_down.weight
744
+ lora_unet__blocks_37_self_attn_q.lora_up.weight
745
+ lora_unet__blocks_37_self_attn_k.lora_down.weight
746
+ lora_unet__blocks_37_self_attn_k.lora_up.weight
747
+ lora_unet__blocks_37_self_attn_v.lora_down.weight
748
+ lora_unet__blocks_37_self_attn_v.lora_up.weight
749
+ lora_unet__blocks_37_self_attn_o.lora_down.weight
750
+ lora_unet__blocks_37_self_attn_o.lora_up.weight
751
+ lora_unet__blocks_37_cross_attn_q.lora_down.weight
752
+ lora_unet__blocks_37_cross_attn_q.lora_up.weight
753
+ lora_unet__blocks_37_cross_attn_k.lora_down.weight
754
+ lora_unet__blocks_37_cross_attn_k.lora_up.weight
755
+ lora_unet__blocks_37_cross_attn_v.lora_down.weight
756
+ lora_unet__blocks_37_cross_attn_v.lora_up.weight
757
+ lora_unet__blocks_37_cross_attn_o.lora_down.weight
758
+ lora_unet__blocks_37_cross_attn_o.lora_up.weight
759
+ lora_unet__blocks_37_ffn_0.lora_down.weight
760
+ lora_unet__blocks_37_ffn_0.lora_up.weight
761
+ lora_unet__blocks_37_ffn_2.lora_down.weight
762
+ lora_unet__blocks_37_ffn_2.lora_up.weight
763
+ lora_unet__blocks_38_self_attn_q.lora_down.weight
764
+ lora_unet__blocks_38_self_attn_q.lora_up.weight
765
+ lora_unet__blocks_38_self_attn_k.lora_down.weight
766
+ lora_unet__blocks_38_self_attn_k.lora_up.weight
767
+ lora_unet__blocks_38_self_attn_v.lora_down.weight
768
+ lora_unet__blocks_38_self_attn_v.lora_up.weight
769
+ lora_unet__blocks_38_self_attn_o.lora_down.weight
770
+ lora_unet__blocks_38_self_attn_o.lora_up.weight
771
+ lora_unet__blocks_38_cross_attn_q.lora_down.weight
772
+ lora_unet__blocks_38_cross_attn_q.lora_up.weight
773
+ lora_unet__blocks_38_cross_attn_k.lora_down.weight
774
+ lora_unet__blocks_38_cross_attn_k.lora_up.weight
775
+ lora_unet__blocks_38_cross_attn_v.lora_down.weight
776
+ lora_unet__blocks_38_cross_attn_v.lora_up.weight
777
+ lora_unet__blocks_38_cross_attn_o.lora_down.weight
778
+ lora_unet__blocks_38_cross_attn_o.lora_up.weight
779
+ lora_unet__blocks_38_ffn_0.lora_down.weight
780
+ lora_unet__blocks_38_ffn_0.lora_up.weight
781
+ lora_unet__blocks_38_ffn_2.lora_down.weight
782
+ lora_unet__blocks_38_ffn_2.lora_up.weight
783
+ lora_unet__blocks_39_self_attn_q.lora_down.weight
784
+ lora_unet__blocks_39_self_attn_q.lora_up.weight
785
+ lora_unet__blocks_39_self_attn_k.lora_down.weight
786
+ lora_unet__blocks_39_self_attn_k.lora_up.weight
787
+ lora_unet__blocks_39_self_attn_v.lora_down.weight
788
+ lora_unet__blocks_39_self_attn_v.lora_up.weight
789
+ lora_unet__blocks_39_self_attn_o.lora_down.weight
790
+ lora_unet__blocks_39_self_attn_o.lora_up.weight
791
+ lora_unet__blocks_39_cross_attn_q.lora_down.weight
792
+ lora_unet__blocks_39_cross_attn_q.lora_up.weight
793
+ lora_unet__blocks_39_cross_attn_k.lora_down.weight
794
+ lora_unet__blocks_39_cross_attn_k.lora_up.weight
795
+ lora_unet__blocks_39_cross_attn_v.lora_down.weight
796
+ lora_unet__blocks_39_cross_attn_v.lora_up.weight
797
+ lora_unet__blocks_39_cross_attn_o.lora_down.weight
798
+ lora_unet__blocks_39_cross_attn_o.lora_up.weight
799
+ lora_unet__blocks_39_ffn_0.lora_down.weight
800
+ lora_unet__blocks_39_ffn_0.lora_up.weight
801
+ lora_unet__blocks_39_ffn_2.lora_down.weight
802
+ lora_unet__blocks_39_ffn_2.lora_up.weight
10_LargeMixedDatset_wan_14bLow_f81_LongCaption_StartMatch_run_r128_a128_3obj_Longrun_B4/transformer3d_structure.txt ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Wan2_2Transformer3DModel(
2
+ (patch_embedding): Conv3d(36, 5120, kernel_size=(1, 2, 2), stride=(1, 2, 2))
3
+ (text_embedding): Sequential(
4
+ (0): Linear(in_features=4096, out_features=5120, bias=True)
5
+ (1): GELU(approximate='tanh')
6
+ (2): Linear(in_features=5120, out_features=5120, bias=True)
7
+ )
8
+ (time_embedding): Sequential(
9
+ (0): Linear(in_features=256, out_features=5120, bias=True)
10
+ (1): SiLU()
11
+ (2): Linear(in_features=5120, out_features=5120, bias=True)
12
+ )
13
+ (time_projection): Sequential(
14
+ (0): SiLU()
15
+ (1): Linear(in_features=5120, out_features=30720, bias=True)
16
+ )
17
+ (blocks): ModuleList(
18
+ (0-39): 40 x WanAttentionBlock(
19
+ (norm1): WanLayerNorm((5120,), eps=1e-06, elementwise_affine=False)
20
+ (self_attn): WanSelfAttention(
21
+ (q): Linear(in_features=5120, out_features=5120, bias=True)
22
+ (k): Linear(in_features=5120, out_features=5120, bias=True)
23
+ (v): Linear(in_features=5120, out_features=5120, bias=True)
24
+ (o): Linear(in_features=5120, out_features=5120, bias=True)
25
+ (norm_q): WanRMSNorm()
26
+ (norm_k): WanRMSNorm()
27
+ )
28
+ (norm3): WanLayerNorm((5120,), eps=1e-06, elementwise_affine=True)
29
+ (cross_attn): WanCrossAttention(
30
+ (q): Linear(in_features=5120, out_features=5120, bias=True)
31
+ (k): Linear(in_features=5120, out_features=5120, bias=True)
32
+ (v): Linear(in_features=5120, out_features=5120, bias=True)
33
+ (o): Linear(in_features=5120, out_features=5120, bias=True)
34
+ (norm_q): WanRMSNorm()
35
+ (norm_k): WanRMSNorm()
36
+ )
37
+ (norm2): WanLayerNorm((5120,), eps=1e-06, elementwise_affine=False)
38
+ (ffn): Sequential(
39
+ (0): Linear(in_features=5120, out_features=13824, bias=True)
40
+ (1): GELU(approximate='tanh')
41
+ (2): Linear(in_features=13824, out_features=5120, bias=True)
42
+ )
43
+ )
44
+ )
45
+ (head): Head(
46
+ (norm): WanLayerNorm((5120,), eps=1e-06, elementwise_affine=False)
47
+ (head): Linear(in_features=5120, out_features=64, bias=True)
48
+ )
49
+ )