tryon_checkpoints / tryon_lora_1_23 /low /network_structure.txt
Jayce-Ping's picture
Upload folder using huggingface_hub
fdaa368 verified
LoRANetwork(
(lora_unet__time_projection_1): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=30720, bias=False)
)
(lora_unet__blocks_0_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_0_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_0_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_0_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_0_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_0_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_0_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_0_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_0_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_0_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_1_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_1_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_1_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_1_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_1_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_1_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_1_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_1_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_1_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_1_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_2_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_2_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_2_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_2_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_2_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_2_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_2_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_2_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_2_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_2_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_3_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_3_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_3_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_3_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_3_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_3_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_3_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_3_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_3_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_3_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_4_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_4_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_4_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_4_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_4_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_4_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_4_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_4_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_4_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_4_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_5_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_5_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_5_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_5_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_5_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_5_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_5_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_5_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_5_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_5_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_6_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_6_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_6_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_6_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_6_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_6_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_6_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_6_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_6_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_6_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_7_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_7_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_7_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_7_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_7_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_7_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_7_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_7_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_7_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_7_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_8_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_8_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_8_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_8_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_8_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_8_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_8_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_8_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_8_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_8_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_9_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_9_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_9_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_9_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_9_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_9_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_9_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_9_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_9_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_9_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_10_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_10_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_10_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_10_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_10_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_10_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_10_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_10_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_10_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_10_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_11_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_11_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_11_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_11_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_11_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_11_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_11_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_11_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_11_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_11_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_12_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_12_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_12_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_12_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_12_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_12_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_12_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_12_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_12_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_12_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_13_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_13_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_13_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_13_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_13_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_13_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_13_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_13_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_13_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_13_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_14_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_14_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_14_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_14_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_14_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_14_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_14_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_14_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_14_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_14_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_15_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_15_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_15_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_15_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_15_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_15_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_15_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_15_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_15_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_15_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_16_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_16_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_16_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_16_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_16_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_16_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_16_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_16_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_16_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_16_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_17_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_17_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_17_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_17_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_17_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_17_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_17_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_17_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_17_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_17_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_18_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_18_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_18_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_18_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_18_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_18_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_18_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_18_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_18_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_18_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_19_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_19_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_19_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_19_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_19_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_19_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_19_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_19_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_19_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_19_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_20_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_20_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_20_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_20_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_20_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_20_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_20_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_20_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_20_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_20_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_21_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_21_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_21_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_21_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_21_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_21_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_21_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_21_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_21_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_21_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_22_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_22_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_22_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_22_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_22_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_22_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_22_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_22_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_22_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_22_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_23_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_23_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_23_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_23_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_23_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_23_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_23_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_23_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_23_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_23_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_24_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_24_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_24_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_24_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_24_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_24_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_24_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_24_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_24_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_24_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_25_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_25_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_25_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_25_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_25_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_25_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_25_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_25_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_25_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_25_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_26_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_26_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_26_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_26_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_26_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_26_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_26_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_26_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_26_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_26_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_27_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_27_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_27_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_27_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_27_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_27_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_27_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_27_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_27_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_27_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_28_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_28_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_28_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_28_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_28_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_28_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_28_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_28_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_28_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_28_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_29_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_29_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_29_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_29_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_29_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_29_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_29_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_29_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_29_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_29_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_30_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_30_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_30_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_30_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_30_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_30_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_30_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_30_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_30_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_30_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_31_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_31_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_31_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_31_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_31_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_31_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_31_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_31_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_31_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_31_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_32_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_32_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_32_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_32_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_32_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_32_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_32_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_32_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_32_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_32_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_33_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_33_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_33_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_33_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_33_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_33_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_33_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_33_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_33_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_33_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_34_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_34_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_34_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_34_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_34_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_34_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_34_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_34_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_34_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_34_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_35_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_35_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_35_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_35_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_35_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_35_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_35_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_35_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_35_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_35_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_36_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_36_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_36_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_36_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_36_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_36_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_36_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_36_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_36_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_36_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_37_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_37_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_37_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_37_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_37_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_37_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_37_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_37_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_37_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_37_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_38_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_38_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_38_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_38_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_38_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_38_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_38_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_38_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_38_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_38_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_39_self_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_39_self_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_39_self_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_39_self_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_39_cross_attn_q): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_39_cross_attn_k): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_39_cross_attn_v): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_39_cross_attn_o): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
(lora_unet__blocks_39_ffn_0): LoRAModule(
(lora_down): Linear(in_features=5120, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=13824, bias=False)
)
(lora_unet__blocks_39_ffn_2): LoRAModule(
(lora_down): Linear(in_features=13824, out_features=128, bias=False)
(lora_up): Linear(in_features=128, out_features=5120, bias=False)
)
)