zjuJish commited on
Commit
36de38f
·
verified ·
1 Parent(s): bdb86aa

Upload alpha_work/exp/ckpts/layer_diffusion_sd_ft_3_vid_v4.2_new_bg_scale_v9_sm_v2_stage2_new_loadself_scale/_trainable_params.txt with huggingface_hub

Browse files
alpha_work/exp/ckpts/layer_diffusion_sd_ft_3_vid_v4.2_new_bg_scale_v9_sm_v2_stage2_new_loadself_scale/_trainable_params.txt ADDED
@@ -0,0 +1,432 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q_lora.0.down.weight torch.Size([256, 1280])
2
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q_lora.0.up.weight torch.Size([1280, 256])
3
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q_lora.1.down.weight torch.Size([256, 1280])
4
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q_lora.1.up.weight torch.Size([1280, 256])
5
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q_lora.2.down.weight torch.Size([256, 1280])
6
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q_lora.2.up.weight torch.Size([1280, 256])
7
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k_lora.0.down.weight torch.Size([256, 1280])
8
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k_lora.0.up.weight torch.Size([1280, 256])
9
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k_lora.1.down.weight torch.Size([256, 1280])
10
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k_lora.1.up.weight torch.Size([1280, 256])
11
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k_lora.2.down.weight torch.Size([256, 1280])
12
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k_lora.2.up.weight torch.Size([1280, 256])
13
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v_lora.0.down.weight torch.Size([256, 1280])
14
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v_lora.0.up.weight torch.Size([1280, 256])
15
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v_lora.1.down.weight torch.Size([256, 1280])
16
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v_lora.1.up.weight torch.Size([1280, 256])
17
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v_lora.2.down.weight torch.Size([256, 1280])
18
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v_lora.2.up.weight torch.Size([1280, 256])
19
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out_lora.0.down.weight torch.Size([256, 1280])
20
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out_lora.0.up.weight torch.Size([1280, 256])
21
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out_lora.1.down.weight torch.Size([256, 1280])
22
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out_lora.1.up.weight torch.Size([1280, 256])
23
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out_lora.2.down.weight torch.Size([256, 1280])
24
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out_lora.2.up.weight torch.Size([1280, 256])
25
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q_lora.0.down.weight torch.Size([256, 1280])
26
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q_lora.0.up.weight torch.Size([1280, 256])
27
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q_lora.1.down.weight torch.Size([256, 1280])
28
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q_lora.1.up.weight torch.Size([1280, 256])
29
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q_lora.2.down.weight torch.Size([256, 1280])
30
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q_lora.2.up.weight torch.Size([1280, 256])
31
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k_lora.0.down.weight torch.Size([256, 768])
32
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k_lora.0.up.weight torch.Size([1280, 256])
33
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k_lora.1.down.weight torch.Size([256, 768])
34
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k_lora.1.up.weight torch.Size([1280, 256])
35
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k_lora.2.down.weight torch.Size([256, 768])
36
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k_lora.2.up.weight torch.Size([1280, 256])
37
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v_lora.0.down.weight torch.Size([256, 768])
38
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v_lora.0.up.weight torch.Size([1280, 256])
39
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v_lora.1.down.weight torch.Size([256, 768])
40
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v_lora.1.up.weight torch.Size([1280, 256])
41
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v_lora.2.down.weight torch.Size([256, 768])
42
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v_lora.2.up.weight torch.Size([1280, 256])
43
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out_lora.0.down.weight torch.Size([256, 1280])
44
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out_lora.0.up.weight torch.Size([1280, 256])
45
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out_lora.1.down.weight torch.Size([256, 1280])
46
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out_lora.1.up.weight torch.Size([1280, 256])
47
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out_lora.2.down.weight torch.Size([256, 1280])
48
+ unet.up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out_lora.2.up.weight torch.Size([1280, 256])
49
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q_lora.0.down.weight torch.Size([256, 1280])
50
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q_lora.0.up.weight torch.Size([1280, 256])
51
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q_lora.1.down.weight torch.Size([256, 1280])
52
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q_lora.1.up.weight torch.Size([1280, 256])
53
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q_lora.2.down.weight torch.Size([256, 1280])
54
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q_lora.2.up.weight torch.Size([1280, 256])
55
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k_lora.0.down.weight torch.Size([256, 1280])
56
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k_lora.0.up.weight torch.Size([1280, 256])
57
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k_lora.1.down.weight torch.Size([256, 1280])
58
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k_lora.1.up.weight torch.Size([1280, 256])
59
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k_lora.2.down.weight torch.Size([256, 1280])
60
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k_lora.2.up.weight torch.Size([1280, 256])
61
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v_lora.0.down.weight torch.Size([256, 1280])
62
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v_lora.0.up.weight torch.Size([1280, 256])
63
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v_lora.1.down.weight torch.Size([256, 1280])
64
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v_lora.1.up.weight torch.Size([1280, 256])
65
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v_lora.2.down.weight torch.Size([256, 1280])
66
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v_lora.2.up.weight torch.Size([1280, 256])
67
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out_lora.0.down.weight torch.Size([256, 1280])
68
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out_lora.0.up.weight torch.Size([1280, 256])
69
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out_lora.1.down.weight torch.Size([256, 1280])
70
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out_lora.1.up.weight torch.Size([1280, 256])
71
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out_lora.2.down.weight torch.Size([256, 1280])
72
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out_lora.2.up.weight torch.Size([1280, 256])
73
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q_lora.0.down.weight torch.Size([256, 1280])
74
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q_lora.0.up.weight torch.Size([1280, 256])
75
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q_lora.1.down.weight torch.Size([256, 1280])
76
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q_lora.1.up.weight torch.Size([1280, 256])
77
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q_lora.2.down.weight torch.Size([256, 1280])
78
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q_lora.2.up.weight torch.Size([1280, 256])
79
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k_lora.0.down.weight torch.Size([256, 768])
80
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k_lora.0.up.weight torch.Size([1280, 256])
81
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k_lora.1.down.weight torch.Size([256, 768])
82
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k_lora.1.up.weight torch.Size([1280, 256])
83
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k_lora.2.down.weight torch.Size([256, 768])
84
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k_lora.2.up.weight torch.Size([1280, 256])
85
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v_lora.0.down.weight torch.Size([256, 768])
86
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v_lora.0.up.weight torch.Size([1280, 256])
87
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v_lora.1.down.weight torch.Size([256, 768])
88
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v_lora.1.up.weight torch.Size([1280, 256])
89
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v_lora.2.down.weight torch.Size([256, 768])
90
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v_lora.2.up.weight torch.Size([1280, 256])
91
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out_lora.0.down.weight torch.Size([256, 1280])
92
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out_lora.0.up.weight torch.Size([1280, 256])
93
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out_lora.1.down.weight torch.Size([256, 1280])
94
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out_lora.1.up.weight torch.Size([1280, 256])
95
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out_lora.2.down.weight torch.Size([256, 1280])
96
+ unet.up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out_lora.2.up.weight torch.Size([1280, 256])
97
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_q_lora.0.down.weight torch.Size([256, 1280])
98
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_q_lora.0.up.weight torch.Size([1280, 256])
99
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_q_lora.1.down.weight torch.Size([256, 1280])
100
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_q_lora.1.up.weight torch.Size([1280, 256])
101
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_q_lora.2.down.weight torch.Size([256, 1280])
102
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_q_lora.2.up.weight torch.Size([1280, 256])
103
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_k_lora.0.down.weight torch.Size([256, 1280])
104
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_k_lora.0.up.weight torch.Size([1280, 256])
105
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_k_lora.1.down.weight torch.Size([256, 1280])
106
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_k_lora.1.up.weight torch.Size([1280, 256])
107
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_k_lora.2.down.weight torch.Size([256, 1280])
108
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_k_lora.2.up.weight torch.Size([1280, 256])
109
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_v_lora.0.down.weight torch.Size([256, 1280])
110
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_v_lora.0.up.weight torch.Size([1280, 256])
111
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_v_lora.1.down.weight torch.Size([256, 1280])
112
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_v_lora.1.up.weight torch.Size([1280, 256])
113
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_v_lora.2.down.weight torch.Size([256, 1280])
114
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_v_lora.2.up.weight torch.Size([1280, 256])
115
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_out_lora.0.down.weight torch.Size([256, 1280])
116
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_out_lora.0.up.weight torch.Size([1280, 256])
117
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_out_lora.1.down.weight torch.Size([256, 1280])
118
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_out_lora.1.up.weight torch.Size([1280, 256])
119
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_out_lora.2.down.weight torch.Size([256, 1280])
120
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_out_lora.2.up.weight torch.Size([1280, 256])
121
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_q_lora.0.down.weight torch.Size([256, 1280])
122
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_q_lora.0.up.weight torch.Size([1280, 256])
123
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_q_lora.1.down.weight torch.Size([256, 1280])
124
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_q_lora.1.up.weight torch.Size([1280, 256])
125
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_q_lora.2.down.weight torch.Size([256, 1280])
126
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_q_lora.2.up.weight torch.Size([1280, 256])
127
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_k_lora.0.down.weight torch.Size([256, 768])
128
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_k_lora.0.up.weight torch.Size([1280, 256])
129
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_k_lora.1.down.weight torch.Size([256, 768])
130
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_k_lora.1.up.weight torch.Size([1280, 256])
131
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_k_lora.2.down.weight torch.Size([256, 768])
132
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_k_lora.2.up.weight torch.Size([1280, 256])
133
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_v_lora.0.down.weight torch.Size([256, 768])
134
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_v_lora.0.up.weight torch.Size([1280, 256])
135
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_v_lora.1.down.weight torch.Size([256, 768])
136
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_v_lora.1.up.weight torch.Size([1280, 256])
137
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_v_lora.2.down.weight torch.Size([256, 768])
138
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_v_lora.2.up.weight torch.Size([1280, 256])
139
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_out_lora.0.down.weight torch.Size([256, 1280])
140
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_out_lora.0.up.weight torch.Size([1280, 256])
141
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_out_lora.1.down.weight torch.Size([256, 1280])
142
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_out_lora.1.up.weight torch.Size([1280, 256])
143
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_out_lora.2.down.weight torch.Size([256, 1280])
144
+ unet.up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_out_lora.2.up.weight torch.Size([1280, 256])
145
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q_lora.0.down.weight torch.Size([256, 640])
146
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q_lora.0.up.weight torch.Size([640, 256])
147
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q_lora.1.down.weight torch.Size([256, 640])
148
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q_lora.1.up.weight torch.Size([640, 256])
149
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q_lora.2.down.weight torch.Size([256, 640])
150
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q_lora.2.up.weight torch.Size([640, 256])
151
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k_lora.0.down.weight torch.Size([256, 640])
152
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k_lora.0.up.weight torch.Size([640, 256])
153
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k_lora.1.down.weight torch.Size([256, 640])
154
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k_lora.1.up.weight torch.Size([640, 256])
155
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k_lora.2.down.weight torch.Size([256, 640])
156
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k_lora.2.up.weight torch.Size([640, 256])
157
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v_lora.0.down.weight torch.Size([256, 640])
158
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v_lora.0.up.weight torch.Size([640, 256])
159
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v_lora.1.down.weight torch.Size([256, 640])
160
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v_lora.1.up.weight torch.Size([640, 256])
161
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v_lora.2.down.weight torch.Size([256, 640])
162
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v_lora.2.up.weight torch.Size([640, 256])
163
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out_lora.0.down.weight torch.Size([256, 640])
164
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out_lora.0.up.weight torch.Size([640, 256])
165
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out_lora.1.down.weight torch.Size([256, 640])
166
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out_lora.1.up.weight torch.Size([640, 256])
167
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out_lora.2.down.weight torch.Size([256, 640])
168
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out_lora.2.up.weight torch.Size([640, 256])
169
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q_lora.0.down.weight torch.Size([256, 640])
170
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q_lora.0.up.weight torch.Size([640, 256])
171
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q_lora.1.down.weight torch.Size([256, 640])
172
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q_lora.1.up.weight torch.Size([640, 256])
173
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q_lora.2.down.weight torch.Size([256, 640])
174
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q_lora.2.up.weight torch.Size([640, 256])
175
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k_lora.0.down.weight torch.Size([256, 768])
176
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k_lora.0.up.weight torch.Size([640, 256])
177
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k_lora.1.down.weight torch.Size([256, 768])
178
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k_lora.1.up.weight torch.Size([640, 256])
179
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k_lora.2.down.weight torch.Size([256, 768])
180
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k_lora.2.up.weight torch.Size([640, 256])
181
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v_lora.0.down.weight torch.Size([256, 768])
182
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v_lora.0.up.weight torch.Size([640, 256])
183
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v_lora.1.down.weight torch.Size([256, 768])
184
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v_lora.1.up.weight torch.Size([640, 256])
185
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v_lora.2.down.weight torch.Size([256, 768])
186
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v_lora.2.up.weight torch.Size([640, 256])
187
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out_lora.0.down.weight torch.Size([256, 640])
188
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out_lora.0.up.weight torch.Size([640, 256])
189
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out_lora.1.down.weight torch.Size([256, 640])
190
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out_lora.1.up.weight torch.Size([640, 256])
191
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out_lora.2.down.weight torch.Size([256, 640])
192
+ unet.up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out_lora.2.up.weight torch.Size([640, 256])
193
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q_lora.0.down.weight torch.Size([256, 640])
194
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q_lora.0.up.weight torch.Size([640, 256])
195
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q_lora.1.down.weight torch.Size([256, 640])
196
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q_lora.1.up.weight torch.Size([640, 256])
197
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q_lora.2.down.weight torch.Size([256, 640])
198
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q_lora.2.up.weight torch.Size([640, 256])
199
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k_lora.0.down.weight torch.Size([256, 640])
200
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k_lora.0.up.weight torch.Size([640, 256])
201
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k_lora.1.down.weight torch.Size([256, 640])
202
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k_lora.1.up.weight torch.Size([640, 256])
203
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k_lora.2.down.weight torch.Size([256, 640])
204
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k_lora.2.up.weight torch.Size([640, 256])
205
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v_lora.0.down.weight torch.Size([256, 640])
206
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v_lora.0.up.weight torch.Size([640, 256])
207
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v_lora.1.down.weight torch.Size([256, 640])
208
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v_lora.1.up.weight torch.Size([640, 256])
209
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v_lora.2.down.weight torch.Size([256, 640])
210
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v_lora.2.up.weight torch.Size([640, 256])
211
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out_lora.0.down.weight torch.Size([256, 640])
212
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out_lora.0.up.weight torch.Size([640, 256])
213
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out_lora.1.down.weight torch.Size([256, 640])
214
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out_lora.1.up.weight torch.Size([640, 256])
215
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out_lora.2.down.weight torch.Size([256, 640])
216
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out_lora.2.up.weight torch.Size([640, 256])
217
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q_lora.0.down.weight torch.Size([256, 640])
218
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q_lora.0.up.weight torch.Size([640, 256])
219
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q_lora.1.down.weight torch.Size([256, 640])
220
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q_lora.1.up.weight torch.Size([640, 256])
221
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q_lora.2.down.weight torch.Size([256, 640])
222
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q_lora.2.up.weight torch.Size([640, 256])
223
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k_lora.0.down.weight torch.Size([256, 768])
224
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k_lora.0.up.weight torch.Size([640, 256])
225
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k_lora.1.down.weight torch.Size([256, 768])
226
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k_lora.1.up.weight torch.Size([640, 256])
227
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k_lora.2.down.weight torch.Size([256, 768])
228
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k_lora.2.up.weight torch.Size([640, 256])
229
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v_lora.0.down.weight torch.Size([256, 768])
230
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v_lora.0.up.weight torch.Size([640, 256])
231
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v_lora.1.down.weight torch.Size([256, 768])
232
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v_lora.1.up.weight torch.Size([640, 256])
233
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v_lora.2.down.weight torch.Size([256, 768])
234
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v_lora.2.up.weight torch.Size([640, 256])
235
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out_lora.0.down.weight torch.Size([256, 640])
236
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out_lora.0.up.weight torch.Size([640, 256])
237
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out_lora.1.down.weight torch.Size([256, 640])
238
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out_lora.1.up.weight torch.Size([640, 256])
239
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out_lora.2.down.weight torch.Size([256, 640])
240
+ unet.up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out_lora.2.up.weight torch.Size([640, 256])
241
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_q_lora.0.down.weight torch.Size([256, 640])
242
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_q_lora.0.up.weight torch.Size([640, 256])
243
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_q_lora.1.down.weight torch.Size([256, 640])
244
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_q_lora.1.up.weight torch.Size([640, 256])
245
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_q_lora.2.down.weight torch.Size([256, 640])
246
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_q_lora.2.up.weight torch.Size([640, 256])
247
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_k_lora.0.down.weight torch.Size([256, 640])
248
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_k_lora.0.up.weight torch.Size([640, 256])
249
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_k_lora.1.down.weight torch.Size([256, 640])
250
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_k_lora.1.up.weight torch.Size([640, 256])
251
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_k_lora.2.down.weight torch.Size([256, 640])
252
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_k_lora.2.up.weight torch.Size([640, 256])
253
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_v_lora.0.down.weight torch.Size([256, 640])
254
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_v_lora.0.up.weight torch.Size([640, 256])
255
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_v_lora.1.down.weight torch.Size([256, 640])
256
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_v_lora.1.up.weight torch.Size([640, 256])
257
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_v_lora.2.down.weight torch.Size([256, 640])
258
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_v_lora.2.up.weight torch.Size([640, 256])
259
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_out_lora.0.down.weight torch.Size([256, 640])
260
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_out_lora.0.up.weight torch.Size([640, 256])
261
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_out_lora.1.down.weight torch.Size([256, 640])
262
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_out_lora.1.up.weight torch.Size([640, 256])
263
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_out_lora.2.down.weight torch.Size([256, 640])
264
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_out_lora.2.up.weight torch.Size([640, 256])
265
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_q_lora.0.down.weight torch.Size([256, 640])
266
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_q_lora.0.up.weight torch.Size([640, 256])
267
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_q_lora.1.down.weight torch.Size([256, 640])
268
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_q_lora.1.up.weight torch.Size([640, 256])
269
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_q_lora.2.down.weight torch.Size([256, 640])
270
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_q_lora.2.up.weight torch.Size([640, 256])
271
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_k_lora.0.down.weight torch.Size([256, 768])
272
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_k_lora.0.up.weight torch.Size([640, 256])
273
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_k_lora.1.down.weight torch.Size([256, 768])
274
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_k_lora.1.up.weight torch.Size([640, 256])
275
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_k_lora.2.down.weight torch.Size([256, 768])
276
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_k_lora.2.up.weight torch.Size([640, 256])
277
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_v_lora.0.down.weight torch.Size([256, 768])
278
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_v_lora.0.up.weight torch.Size([640, 256])
279
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_v_lora.1.down.weight torch.Size([256, 768])
280
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_v_lora.1.up.weight torch.Size([640, 256])
281
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_v_lora.2.down.weight torch.Size([256, 768])
282
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_v_lora.2.up.weight torch.Size([640, 256])
283
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_out_lora.0.down.weight torch.Size([256, 640])
284
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_out_lora.0.up.weight torch.Size([640, 256])
285
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_out_lora.1.down.weight torch.Size([256, 640])
286
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_out_lora.1.up.weight torch.Size([640, 256])
287
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_out_lora.2.down.weight torch.Size([256, 640])
288
+ unet.up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_out_lora.2.up.weight torch.Size([640, 256])
289
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_q_lora.0.down.weight torch.Size([256, 320])
290
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_q_lora.0.up.weight torch.Size([320, 256])
291
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_q_lora.1.down.weight torch.Size([256, 320])
292
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_q_lora.1.up.weight torch.Size([320, 256])
293
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_q_lora.2.down.weight torch.Size([256, 320])
294
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_q_lora.2.up.weight torch.Size([320, 256])
295
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_k_lora.0.down.weight torch.Size([256, 320])
296
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_k_lora.0.up.weight torch.Size([320, 256])
297
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_k_lora.1.down.weight torch.Size([256, 320])
298
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_k_lora.1.up.weight torch.Size([320, 256])
299
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_k_lora.2.down.weight torch.Size([256, 320])
300
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_k_lora.2.up.weight torch.Size([320, 256])
301
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_v_lora.0.down.weight torch.Size([256, 320])
302
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_v_lora.0.up.weight torch.Size([320, 256])
303
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_v_lora.1.down.weight torch.Size([256, 320])
304
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_v_lora.1.up.weight torch.Size([320, 256])
305
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_v_lora.2.down.weight torch.Size([256, 320])
306
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_v_lora.2.up.weight torch.Size([320, 256])
307
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_out_lora.0.down.weight torch.Size([256, 320])
308
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_out_lora.0.up.weight torch.Size([320, 256])
309
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_out_lora.1.down.weight torch.Size([256, 320])
310
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_out_lora.1.up.weight torch.Size([320, 256])
311
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_out_lora.2.down.weight torch.Size([256, 320])
312
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_out_lora.2.up.weight torch.Size([320, 256])
313
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_q_lora.0.down.weight torch.Size([256, 320])
314
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_q_lora.0.up.weight torch.Size([320, 256])
315
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_q_lora.1.down.weight torch.Size([256, 320])
316
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_q_lora.1.up.weight torch.Size([320, 256])
317
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_q_lora.2.down.weight torch.Size([256, 320])
318
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_q_lora.2.up.weight torch.Size([320, 256])
319
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_k_lora.0.down.weight torch.Size([256, 768])
320
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_k_lora.0.up.weight torch.Size([320, 256])
321
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_k_lora.1.down.weight torch.Size([256, 768])
322
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_k_lora.1.up.weight torch.Size([320, 256])
323
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_k_lora.2.down.weight torch.Size([256, 768])
324
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_k_lora.2.up.weight torch.Size([320, 256])
325
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_v_lora.0.down.weight torch.Size([256, 768])
326
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_v_lora.0.up.weight torch.Size([320, 256])
327
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_v_lora.1.down.weight torch.Size([256, 768])
328
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_v_lora.1.up.weight torch.Size([320, 256])
329
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_v_lora.2.down.weight torch.Size([256, 768])
330
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_v_lora.2.up.weight torch.Size([320, 256])
331
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_out_lora.0.down.weight torch.Size([256, 320])
332
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_out_lora.0.up.weight torch.Size([320, 256])
333
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_out_lora.1.down.weight torch.Size([256, 320])
334
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_out_lora.1.up.weight torch.Size([320, 256])
335
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_out_lora.2.down.weight torch.Size([256, 320])
336
+ unet.up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_out_lora.2.up.weight torch.Size([320, 256])
337
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_q_lora.0.down.weight torch.Size([256, 320])
338
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_q_lora.0.up.weight torch.Size([320, 256])
339
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_q_lora.1.down.weight torch.Size([256, 320])
340
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_q_lora.1.up.weight torch.Size([320, 256])
341
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_q_lora.2.down.weight torch.Size([256, 320])
342
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_q_lora.2.up.weight torch.Size([320, 256])
343
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_k_lora.0.down.weight torch.Size([256, 320])
344
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_k_lora.0.up.weight torch.Size([320, 256])
345
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_k_lora.1.down.weight torch.Size([256, 320])
346
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_k_lora.1.up.weight torch.Size([320, 256])
347
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_k_lora.2.down.weight torch.Size([256, 320])
348
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_k_lora.2.up.weight torch.Size([320, 256])
349
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_v_lora.0.down.weight torch.Size([256, 320])
350
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_v_lora.0.up.weight torch.Size([320, 256])
351
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_v_lora.1.down.weight torch.Size([256, 320])
352
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_v_lora.1.up.weight torch.Size([320, 256])
353
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_v_lora.2.down.weight torch.Size([256, 320])
354
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_v_lora.2.up.weight torch.Size([320, 256])
355
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_out_lora.0.down.weight torch.Size([256, 320])
356
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_out_lora.0.up.weight torch.Size([320, 256])
357
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_out_lora.1.down.weight torch.Size([256, 320])
358
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_out_lora.1.up.weight torch.Size([320, 256])
359
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_out_lora.2.down.weight torch.Size([256, 320])
360
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_out_lora.2.up.weight torch.Size([320, 256])
361
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_q_lora.0.down.weight torch.Size([256, 320])
362
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_q_lora.0.up.weight torch.Size([320, 256])
363
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_q_lora.1.down.weight torch.Size([256, 320])
364
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_q_lora.1.up.weight torch.Size([320, 256])
365
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_q_lora.2.down.weight torch.Size([256, 320])
366
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_q_lora.2.up.weight torch.Size([320, 256])
367
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_k_lora.0.down.weight torch.Size([256, 768])
368
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_k_lora.0.up.weight torch.Size([320, 256])
369
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_k_lora.1.down.weight torch.Size([256, 768])
370
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_k_lora.1.up.weight torch.Size([320, 256])
371
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_k_lora.2.down.weight torch.Size([256, 768])
372
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_k_lora.2.up.weight torch.Size([320, 256])
373
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_v_lora.0.down.weight torch.Size([256, 768])
374
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_v_lora.0.up.weight torch.Size([320, 256])
375
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_v_lora.1.down.weight torch.Size([256, 768])
376
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_v_lora.1.up.weight torch.Size([320, 256])
377
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_v_lora.2.down.weight torch.Size([256, 768])
378
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_v_lora.2.up.weight torch.Size([320, 256])
379
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_out_lora.0.down.weight torch.Size([256, 320])
380
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_out_lora.0.up.weight torch.Size([320, 256])
381
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_out_lora.1.down.weight torch.Size([256, 320])
382
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_out_lora.1.up.weight torch.Size([320, 256])
383
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_out_lora.2.down.weight torch.Size([256, 320])
384
+ unet.up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_out_lora.2.up.weight torch.Size([320, 256])
385
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_q_lora.0.down.weight torch.Size([256, 320])
386
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_q_lora.0.up.weight torch.Size([320, 256])
387
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_q_lora.1.down.weight torch.Size([256, 320])
388
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_q_lora.1.up.weight torch.Size([320, 256])
389
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_q_lora.2.down.weight torch.Size([256, 320])
390
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_q_lora.2.up.weight torch.Size([320, 256])
391
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_k_lora.0.down.weight torch.Size([256, 320])
392
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_k_lora.0.up.weight torch.Size([320, 256])
393
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_k_lora.1.down.weight torch.Size([256, 320])
394
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_k_lora.1.up.weight torch.Size([320, 256])
395
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_k_lora.2.down.weight torch.Size([256, 320])
396
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_k_lora.2.up.weight torch.Size([320, 256])
397
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_v_lora.0.down.weight torch.Size([256, 320])
398
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_v_lora.0.up.weight torch.Size([320, 256])
399
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_v_lora.1.down.weight torch.Size([256, 320])
400
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_v_lora.1.up.weight torch.Size([320, 256])
401
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_v_lora.2.down.weight torch.Size([256, 320])
402
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_v_lora.2.up.weight torch.Size([320, 256])
403
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_out_lora.0.down.weight torch.Size([256, 320])
404
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_out_lora.0.up.weight torch.Size([320, 256])
405
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_out_lora.1.down.weight torch.Size([256, 320])
406
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_out_lora.1.up.weight torch.Size([320, 256])
407
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_out_lora.2.down.weight torch.Size([256, 320])
408
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_out_lora.2.up.weight torch.Size([320, 256])
409
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_q_lora.0.down.weight torch.Size([256, 320])
410
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_q_lora.0.up.weight torch.Size([320, 256])
411
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_q_lora.1.down.weight torch.Size([256, 320])
412
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_q_lora.1.up.weight torch.Size([320, 256])
413
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_q_lora.2.down.weight torch.Size([256, 320])
414
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_q_lora.2.up.weight torch.Size([320, 256])
415
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_k_lora.0.down.weight torch.Size([256, 768])
416
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_k_lora.0.up.weight torch.Size([320, 256])
417
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_k_lora.1.down.weight torch.Size([256, 768])
418
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_k_lora.1.up.weight torch.Size([320, 256])
419
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_k_lora.2.down.weight torch.Size([256, 768])
420
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_k_lora.2.up.weight torch.Size([320, 256])
421
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_v_lora.0.down.weight torch.Size([256, 768])
422
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_v_lora.0.up.weight torch.Size([320, 256])
423
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_v_lora.1.down.weight torch.Size([256, 768])
424
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_v_lora.1.up.weight torch.Size([320, 256])
425
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_v_lora.2.down.weight torch.Size([256, 768])
426
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_v_lora.2.up.weight torch.Size([320, 256])
427
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_out_lora.0.down.weight torch.Size([256, 320])
428
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_out_lora.0.up.weight torch.Size([320, 256])
429
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_out_lora.1.down.weight torch.Size([256, 320])
430
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_out_lora.1.up.weight torch.Size([320, 256])
431
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_out_lora.2.down.weight torch.Size([256, 320])
432
+ unet.up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_out_lora.2.up.weight torch.Size([320, 256])