sippycoder commited on
Commit
4c14b06
·
1 Parent(s): 78fc60f

NucleusMoE-Image

Browse files
transformer/diffusion_pytorch_model-00001-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b51d597556c6902367f208d059c75cc05382e6ab62490cb77cda04650bb96501
3
- size 5359982992
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad0a79d47a0ea15e1f746cf64ed9b85b210b577e6769f1196075d80e6d5b46d7
3
+ size 5359982584
transformer/diffusion_pytorch_model-00002-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e53ac9da8543bdf0c04ca5b0d75ac56b056fa26123282a7132f3d731abb55144
3
- size 5368037472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d1c04d5ae3e81c7c699dcbb9e2d9ccd2a086f50bad889d0617ca2bdf11da3fd
3
+ size 5368036944
transformer/diffusion_pytorch_model-00003-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6aa90c1de8edf3c3ae31e479f1e0155d24a4adf118049ba234593f4c62710475
3
- size 5119271672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba2dc218dee2032e2e9bbd89548b7962901f28f2f92d63f23ecd6c138e04d612
3
+ size 5119271264
transformer/diffusion_pytorch_model-00004-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:52ae1707aa19c8fb4e2481edf349355df275a1a54e6c476df810327c71dc6ddf
3
- size 5368037472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d00f85e512a49b881513f0ca0123d70dc9963c3d5531a6986aa2fe929b62d2bf
3
+ size 5368036944
transformer/diffusion_pytorch_model-00005-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9aa214df52813cba492dea5328f1c0e0d618db67175d8874fe25e1c3c29bf379
3
- size 5330012176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ef5a35ddd6f01a8fe26a475dc8e2ffc56585cd8797e8a0525578ddea109b681
3
+ size 4977690112
transformer/diffusion_pytorch_model-00006-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:52dd29236d4553417b2298451506891c284e3b925234065f13e6c74473020cd6
3
- size 5042983264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ceb86f9bb8b934dae6ff9bc562482f6a493c97df50a1b8897a3b9e7d64cedf8d
3
+ size 5368036864
transformer/diffusion_pytorch_model-00007-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dbb49ce7603744549ca8776957b6bc9cacdefdd20e5557620522d71c67d10369
3
- size 2257109360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5a767727147f0686615eb3424b5200d2313548b0c1931ab90b88e90c56de197
3
+ size 2284376672
transformer/diffusion_pytorch_model.safetensors.index.json CHANGED
@@ -57,9 +57,8 @@
57
  "transformer_blocks.10.attn.to_v.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
58
  "transformer_blocks.10.encoder_proj.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
59
  "transformer_blocks.10.encoder_proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
60
- "transformer_blocks.10.img_mlp.experts.w1": "diffusion_pytorch_model-00001-of-00007.safetensors",
61
- "transformer_blocks.10.img_mlp.experts.w2": "diffusion_pytorch_model-00001-of-00007.safetensors",
62
- "transformer_blocks.10.img_mlp.experts.w3": "diffusion_pytorch_model-00001-of-00007.safetensors",
63
  "transformer_blocks.10.img_mlp.gate.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
64
  "transformer_blocks.10.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
65
  "transformer_blocks.10.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
@@ -77,9 +76,8 @@
77
  "transformer_blocks.11.attn.to_v.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
78
  "transformer_blocks.11.encoder_proj.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
79
  "transformer_blocks.11.encoder_proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
80
- "transformer_blocks.11.img_mlp.experts.w1": "diffusion_pytorch_model-00001-of-00007.safetensors",
81
- "transformer_blocks.11.img_mlp.experts.w2": "diffusion_pytorch_model-00001-of-00007.safetensors",
82
- "transformer_blocks.11.img_mlp.experts.w3": "diffusion_pytorch_model-00001-of-00007.safetensors",
83
  "transformer_blocks.11.img_mlp.gate.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
84
  "transformer_blocks.11.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
85
  "transformer_blocks.11.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
@@ -97,9 +95,8 @@
97
  "transformer_blocks.12.attn.to_v.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
98
  "transformer_blocks.12.encoder_proj.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
99
  "transformer_blocks.12.encoder_proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
100
- "transformer_blocks.12.img_mlp.experts.w1": "diffusion_pytorch_model-00001-of-00007.safetensors",
101
- "transformer_blocks.12.img_mlp.experts.w2": "diffusion_pytorch_model-00001-of-00007.safetensors",
102
- "transformer_blocks.12.img_mlp.experts.w3": "diffusion_pytorch_model-00001-of-00007.safetensors",
103
  "transformer_blocks.12.img_mlp.gate.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
104
  "transformer_blocks.12.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
105
  "transformer_blocks.12.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
@@ -117,9 +114,8 @@
117
  "transformer_blocks.13.attn.to_v.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
118
  "transformer_blocks.13.encoder_proj.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
119
  "transformer_blocks.13.encoder_proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
120
- "transformer_blocks.13.img_mlp.experts.w1": "diffusion_pytorch_model-00001-of-00007.safetensors",
121
- "transformer_blocks.13.img_mlp.experts.w2": "diffusion_pytorch_model-00001-of-00007.safetensors",
122
- "transformer_blocks.13.img_mlp.experts.w3": "diffusion_pytorch_model-00001-of-00007.safetensors",
123
  "transformer_blocks.13.img_mlp.gate.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
124
  "transformer_blocks.13.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
125
  "transformer_blocks.13.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
@@ -137,9 +133,8 @@
137
  "transformer_blocks.14.attn.to_v.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
138
  "transformer_blocks.14.encoder_proj.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
139
  "transformer_blocks.14.encoder_proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
140
- "transformer_blocks.14.img_mlp.experts.w1": "diffusion_pytorch_model-00001-of-00007.safetensors",
141
- "transformer_blocks.14.img_mlp.experts.w2": "diffusion_pytorch_model-00002-of-00007.safetensors",
142
- "transformer_blocks.14.img_mlp.experts.w3": "diffusion_pytorch_model-00002-of-00007.safetensors",
143
  "transformer_blocks.14.img_mlp.gate.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
144
  "transformer_blocks.14.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
145
  "transformer_blocks.14.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
@@ -157,9 +152,8 @@
157
  "transformer_blocks.15.attn.to_v.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
158
  "transformer_blocks.15.encoder_proj.bias": "diffusion_pytorch_model-00002-of-00007.safetensors",
159
  "transformer_blocks.15.encoder_proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
160
- "transformer_blocks.15.img_mlp.experts.w1": "diffusion_pytorch_model-00002-of-00007.safetensors",
161
- "transformer_blocks.15.img_mlp.experts.w2": "diffusion_pytorch_model-00002-of-00007.safetensors",
162
- "transformer_blocks.15.img_mlp.experts.w3": "diffusion_pytorch_model-00002-of-00007.safetensors",
163
  "transformer_blocks.15.img_mlp.gate.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
164
  "transformer_blocks.15.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
165
  "transformer_blocks.15.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
@@ -177,9 +171,8 @@
177
  "transformer_blocks.16.attn.to_v.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
178
  "transformer_blocks.16.encoder_proj.bias": "diffusion_pytorch_model-00002-of-00007.safetensors",
179
  "transformer_blocks.16.encoder_proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
180
- "transformer_blocks.16.img_mlp.experts.w1": "diffusion_pytorch_model-00002-of-00007.safetensors",
181
- "transformer_blocks.16.img_mlp.experts.w2": "diffusion_pytorch_model-00002-of-00007.safetensors",
182
- "transformer_blocks.16.img_mlp.experts.w3": "diffusion_pytorch_model-00002-of-00007.safetensors",
183
  "transformer_blocks.16.img_mlp.gate.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
184
  "transformer_blocks.16.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
185
  "transformer_blocks.16.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
@@ -197,9 +190,8 @@
197
  "transformer_blocks.17.attn.to_v.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
198
  "transformer_blocks.17.encoder_proj.bias": "diffusion_pytorch_model-00002-of-00007.safetensors",
199
  "transformer_blocks.17.encoder_proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
200
- "transformer_blocks.17.img_mlp.experts.w1": "diffusion_pytorch_model-00002-of-00007.safetensors",
201
- "transformer_blocks.17.img_mlp.experts.w2": "diffusion_pytorch_model-00002-of-00007.safetensors",
202
- "transformer_blocks.17.img_mlp.experts.w3": "diffusion_pytorch_model-00002-of-00007.safetensors",
203
  "transformer_blocks.17.img_mlp.gate.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
204
  "transformer_blocks.17.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
205
  "transformer_blocks.17.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
@@ -217,9 +209,8 @@
217
  "transformer_blocks.18.attn.to_v.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
218
  "transformer_blocks.18.encoder_proj.bias": "diffusion_pytorch_model-00002-of-00007.safetensors",
219
  "transformer_blocks.18.encoder_proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
220
- "transformer_blocks.18.img_mlp.experts.w1": "diffusion_pytorch_model-00002-of-00007.safetensors",
221
- "transformer_blocks.18.img_mlp.experts.w2": "diffusion_pytorch_model-00002-of-00007.safetensors",
222
- "transformer_blocks.18.img_mlp.experts.w3": "diffusion_pytorch_model-00002-of-00007.safetensors",
223
  "transformer_blocks.18.img_mlp.gate.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
224
  "transformer_blocks.18.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
225
  "transformer_blocks.18.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
@@ -237,9 +228,8 @@
237
  "transformer_blocks.19.attn.to_v.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
238
  "transformer_blocks.19.encoder_proj.bias": "diffusion_pytorch_model-00003-of-00007.safetensors",
239
  "transformer_blocks.19.encoder_proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
240
- "transformer_blocks.19.img_mlp.experts.w1": "diffusion_pytorch_model-00003-of-00007.safetensors",
241
- "transformer_blocks.19.img_mlp.experts.w2": "diffusion_pytorch_model-00003-of-00007.safetensors",
242
- "transformer_blocks.19.img_mlp.experts.w3": "diffusion_pytorch_model-00003-of-00007.safetensors",
243
  "transformer_blocks.19.img_mlp.gate.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
244
  "transformer_blocks.19.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
245
  "transformer_blocks.19.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
@@ -273,9 +263,8 @@
273
  "transformer_blocks.20.attn.to_v.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
274
  "transformer_blocks.20.encoder_proj.bias": "diffusion_pytorch_model-00003-of-00007.safetensors",
275
  "transformer_blocks.20.encoder_proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
276
- "transformer_blocks.20.img_mlp.experts.w1": "diffusion_pytorch_model-00003-of-00007.safetensors",
277
- "transformer_blocks.20.img_mlp.experts.w2": "diffusion_pytorch_model-00003-of-00007.safetensors",
278
- "transformer_blocks.20.img_mlp.experts.w3": "diffusion_pytorch_model-00003-of-00007.safetensors",
279
  "transformer_blocks.20.img_mlp.gate.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
280
  "transformer_blocks.20.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
281
  "transformer_blocks.20.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
@@ -293,9 +282,8 @@
293
  "transformer_blocks.21.attn.to_v.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
294
  "transformer_blocks.21.encoder_proj.bias": "diffusion_pytorch_model-00003-of-00007.safetensors",
295
  "transformer_blocks.21.encoder_proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
296
- "transformer_blocks.21.img_mlp.experts.w1": "diffusion_pytorch_model-00003-of-00007.safetensors",
297
- "transformer_blocks.21.img_mlp.experts.w2": "diffusion_pytorch_model-00003-of-00007.safetensors",
298
- "transformer_blocks.21.img_mlp.experts.w3": "diffusion_pytorch_model-00003-of-00007.safetensors",
299
  "transformer_blocks.21.img_mlp.gate.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
300
  "transformer_blocks.21.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
301
  "transformer_blocks.21.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
@@ -313,9 +301,8 @@
313
  "transformer_blocks.22.attn.to_v.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
314
  "transformer_blocks.22.encoder_proj.bias": "diffusion_pytorch_model-00003-of-00007.safetensors",
315
  "transformer_blocks.22.encoder_proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
316
- "transformer_blocks.22.img_mlp.experts.w1": "diffusion_pytorch_model-00003-of-00007.safetensors",
317
- "transformer_blocks.22.img_mlp.experts.w2": "diffusion_pytorch_model-00003-of-00007.safetensors",
318
- "transformer_blocks.22.img_mlp.experts.w3": "diffusion_pytorch_model-00003-of-00007.safetensors",
319
  "transformer_blocks.22.img_mlp.gate.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
320
  "transformer_blocks.22.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
321
  "transformer_blocks.22.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
@@ -333,9 +320,8 @@
333
  "transformer_blocks.23.attn.to_v.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
334
  "transformer_blocks.23.encoder_proj.bias": "diffusion_pytorch_model-00003-of-00007.safetensors",
335
  "transformer_blocks.23.encoder_proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
336
- "transformer_blocks.23.img_mlp.experts.w1": "diffusion_pytorch_model-00003-of-00007.safetensors",
337
- "transformer_blocks.23.img_mlp.experts.w2": "diffusion_pytorch_model-00004-of-00007.safetensors",
338
- "transformer_blocks.23.img_mlp.experts.w3": "diffusion_pytorch_model-00004-of-00007.safetensors",
339
  "transformer_blocks.23.img_mlp.gate.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
340
  "transformer_blocks.23.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
341
  "transformer_blocks.23.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
@@ -353,9 +339,8 @@
353
  "transformer_blocks.24.attn.to_v.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
354
  "transformer_blocks.24.encoder_proj.bias": "diffusion_pytorch_model-00004-of-00007.safetensors",
355
  "transformer_blocks.24.encoder_proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
356
- "transformer_blocks.24.img_mlp.experts.w1": "diffusion_pytorch_model-00004-of-00007.safetensors",
357
- "transformer_blocks.24.img_mlp.experts.w2": "diffusion_pytorch_model-00004-of-00007.safetensors",
358
- "transformer_blocks.24.img_mlp.experts.w3": "diffusion_pytorch_model-00004-of-00007.safetensors",
359
  "transformer_blocks.24.img_mlp.gate.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
360
  "transformer_blocks.24.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
361
  "transformer_blocks.24.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
@@ -373,9 +358,8 @@
373
  "transformer_blocks.25.attn.to_v.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
374
  "transformer_blocks.25.encoder_proj.bias": "diffusion_pytorch_model-00004-of-00007.safetensors",
375
  "transformer_blocks.25.encoder_proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
376
- "transformer_blocks.25.img_mlp.experts.w1": "diffusion_pytorch_model-00004-of-00007.safetensors",
377
- "transformer_blocks.25.img_mlp.experts.w2": "diffusion_pytorch_model-00004-of-00007.safetensors",
378
- "transformer_blocks.25.img_mlp.experts.w3": "diffusion_pytorch_model-00004-of-00007.safetensors",
379
  "transformer_blocks.25.img_mlp.gate.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
380
  "transformer_blocks.25.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
381
  "transformer_blocks.25.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
@@ -393,9 +377,8 @@
393
  "transformer_blocks.26.attn.to_v.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
394
  "transformer_blocks.26.encoder_proj.bias": "diffusion_pytorch_model-00004-of-00007.safetensors",
395
  "transformer_blocks.26.encoder_proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
396
- "transformer_blocks.26.img_mlp.experts.w1": "diffusion_pytorch_model-00004-of-00007.safetensors",
397
- "transformer_blocks.26.img_mlp.experts.w2": "diffusion_pytorch_model-00004-of-00007.safetensors",
398
- "transformer_blocks.26.img_mlp.experts.w3": "diffusion_pytorch_model-00004-of-00007.safetensors",
399
  "transformer_blocks.26.img_mlp.gate.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
400
  "transformer_blocks.26.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
401
  "transformer_blocks.26.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
@@ -413,9 +396,8 @@
413
  "transformer_blocks.27.attn.to_v.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
414
  "transformer_blocks.27.encoder_proj.bias": "diffusion_pytorch_model-00004-of-00007.safetensors",
415
  "transformer_blocks.27.encoder_proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
416
- "transformer_blocks.27.img_mlp.experts.w1": "diffusion_pytorch_model-00004-of-00007.safetensors",
417
- "transformer_blocks.27.img_mlp.experts.w2": "diffusion_pytorch_model-00004-of-00007.safetensors",
418
- "transformer_blocks.27.img_mlp.experts.w3": "diffusion_pytorch_model-00004-of-00007.safetensors",
419
  "transformer_blocks.27.img_mlp.gate.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
420
  "transformer_blocks.27.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
421
  "transformer_blocks.27.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
@@ -433,9 +415,8 @@
433
  "transformer_blocks.28.attn.to_v.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
434
  "transformer_blocks.28.encoder_proj.bias": "diffusion_pytorch_model-00005-of-00007.safetensors",
435
  "transformer_blocks.28.encoder_proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
436
- "transformer_blocks.28.img_mlp.experts.w1": "diffusion_pytorch_model-00005-of-00007.safetensors",
437
- "transformer_blocks.28.img_mlp.experts.w2": "diffusion_pytorch_model-00005-of-00007.safetensors",
438
- "transformer_blocks.28.img_mlp.experts.w3": "diffusion_pytorch_model-00005-of-00007.safetensors",
439
  "transformer_blocks.28.img_mlp.gate.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
440
  "transformer_blocks.28.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
441
  "transformer_blocks.28.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
@@ -453,9 +434,8 @@
453
  "transformer_blocks.29.attn.to_v.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
454
  "transformer_blocks.29.encoder_proj.bias": "diffusion_pytorch_model-00005-of-00007.safetensors",
455
  "transformer_blocks.29.encoder_proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
456
- "transformer_blocks.29.img_mlp.experts.w1": "diffusion_pytorch_model-00005-of-00007.safetensors",
457
- "transformer_blocks.29.img_mlp.experts.w2": "diffusion_pytorch_model-00005-of-00007.safetensors",
458
- "transformer_blocks.29.img_mlp.experts.w3": "diffusion_pytorch_model-00005-of-00007.safetensors",
459
  "transformer_blocks.29.img_mlp.gate.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
460
  "transformer_blocks.29.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
461
  "transformer_blocks.29.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
@@ -473,9 +453,8 @@
473
  "transformer_blocks.3.attn.to_v.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
474
  "transformer_blocks.3.encoder_proj.bias": "diffusion_pytorch_model-00005-of-00007.safetensors",
475
  "transformer_blocks.3.encoder_proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
476
- "transformer_blocks.3.img_mlp.experts.w1": "diffusion_pytorch_model-00005-of-00007.safetensors",
477
- "transformer_blocks.3.img_mlp.experts.w2": "diffusion_pytorch_model-00005-of-00007.safetensors",
478
- "transformer_blocks.3.img_mlp.experts.w3": "diffusion_pytorch_model-00005-of-00007.safetensors",
479
  "transformer_blocks.3.img_mlp.gate.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
480
  "transformer_blocks.3.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
481
  "transformer_blocks.3.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
@@ -493,9 +472,8 @@
493
  "transformer_blocks.30.attn.to_v.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
494
  "transformer_blocks.30.encoder_proj.bias": "diffusion_pytorch_model-00005-of-00007.safetensors",
495
  "transformer_blocks.30.encoder_proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
496
- "transformer_blocks.30.img_mlp.experts.w1": "diffusion_pytorch_model-00005-of-00007.safetensors",
497
- "transformer_blocks.30.img_mlp.experts.w2": "diffusion_pytorch_model-00005-of-00007.safetensors",
498
- "transformer_blocks.30.img_mlp.experts.w3": "diffusion_pytorch_model-00005-of-00007.safetensors",
499
  "transformer_blocks.30.img_mlp.gate.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
500
  "transformer_blocks.30.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
501
  "transformer_blocks.30.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
@@ -513,9 +491,8 @@
513
  "transformer_blocks.31.attn.to_v.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
514
  "transformer_blocks.31.encoder_proj.bias": "diffusion_pytorch_model-00005-of-00007.safetensors",
515
  "transformer_blocks.31.encoder_proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
516
- "transformer_blocks.31.img_mlp.experts.w1": "diffusion_pytorch_model-00005-of-00007.safetensors",
517
- "transformer_blocks.31.img_mlp.experts.w2": "diffusion_pytorch_model-00005-of-00007.safetensors",
518
- "transformer_blocks.31.img_mlp.experts.w3": "diffusion_pytorch_model-00006-of-00007.safetensors",
519
  "transformer_blocks.31.img_mlp.gate.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
520
  "transformer_blocks.31.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
521
  "transformer_blocks.31.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
@@ -533,9 +510,8 @@
533
  "transformer_blocks.4.attn.to_v.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
534
  "transformer_blocks.4.encoder_proj.bias": "diffusion_pytorch_model-00006-of-00007.safetensors",
535
  "transformer_blocks.4.encoder_proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
536
- "transformer_blocks.4.img_mlp.experts.w1": "diffusion_pytorch_model-00006-of-00007.safetensors",
537
- "transformer_blocks.4.img_mlp.experts.w2": "diffusion_pytorch_model-00006-of-00007.safetensors",
538
- "transformer_blocks.4.img_mlp.experts.w3": "diffusion_pytorch_model-00006-of-00007.safetensors",
539
  "transformer_blocks.4.img_mlp.gate.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
540
  "transformer_blocks.4.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
541
  "transformer_blocks.4.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
@@ -553,9 +529,8 @@
553
  "transformer_blocks.5.attn.to_v.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
554
  "transformer_blocks.5.encoder_proj.bias": "diffusion_pytorch_model-00006-of-00007.safetensors",
555
  "transformer_blocks.5.encoder_proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
556
- "transformer_blocks.5.img_mlp.experts.w1": "diffusion_pytorch_model-00006-of-00007.safetensors",
557
- "transformer_blocks.5.img_mlp.experts.w2": "diffusion_pytorch_model-00006-of-00007.safetensors",
558
- "transformer_blocks.5.img_mlp.experts.w3": "diffusion_pytorch_model-00006-of-00007.safetensors",
559
  "transformer_blocks.5.img_mlp.gate.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
560
  "transformer_blocks.5.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
561
  "transformer_blocks.5.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
@@ -573,9 +548,8 @@
573
  "transformer_blocks.6.attn.to_v.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
574
  "transformer_blocks.6.encoder_proj.bias": "diffusion_pytorch_model-00006-of-00007.safetensors",
575
  "transformer_blocks.6.encoder_proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
576
- "transformer_blocks.6.img_mlp.experts.w1": "diffusion_pytorch_model-00006-of-00007.safetensors",
577
- "transformer_blocks.6.img_mlp.experts.w2": "diffusion_pytorch_model-00006-of-00007.safetensors",
578
- "transformer_blocks.6.img_mlp.experts.w3": "diffusion_pytorch_model-00006-of-00007.safetensors",
579
  "transformer_blocks.6.img_mlp.gate.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
580
  "transformer_blocks.6.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
581
  "transformer_blocks.6.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
@@ -593,9 +567,8 @@
593
  "transformer_blocks.7.attn.to_v.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
594
  "transformer_blocks.7.encoder_proj.bias": "diffusion_pytorch_model-00006-of-00007.safetensors",
595
  "transformer_blocks.7.encoder_proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
596
- "transformer_blocks.7.img_mlp.experts.w1": "diffusion_pytorch_model-00006-of-00007.safetensors",
597
- "transformer_blocks.7.img_mlp.experts.w2": "diffusion_pytorch_model-00006-of-00007.safetensors",
598
- "transformer_blocks.7.img_mlp.experts.w3": "diffusion_pytorch_model-00006-of-00007.safetensors",
599
  "transformer_blocks.7.img_mlp.gate.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
600
  "transformer_blocks.7.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
601
  "transformer_blocks.7.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
@@ -609,13 +582,12 @@
609
  "transformer_blocks.8.attn.norm_q.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
610
  "transformer_blocks.8.attn.to_k.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
611
  "transformer_blocks.8.attn.to_out.0.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
612
- "transformer_blocks.8.attn.to_q.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
613
- "transformer_blocks.8.attn.to_v.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
614
- "transformer_blocks.8.encoder_proj.bias": "diffusion_pytorch_model-00006-of-00007.safetensors",
615
- "transformer_blocks.8.encoder_proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
616
- "transformer_blocks.8.img_mlp.experts.w1": "diffusion_pytorch_model-00007-of-00007.safetensors",
617
- "transformer_blocks.8.img_mlp.experts.w2": "diffusion_pytorch_model-00007-of-00007.safetensors",
618
- "transformer_blocks.8.img_mlp.experts.w3": "diffusion_pytorch_model-00007-of-00007.safetensors",
619
  "transformer_blocks.8.img_mlp.gate.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
620
  "transformer_blocks.8.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
621
  "transformer_blocks.8.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
@@ -633,9 +605,8 @@
633
  "transformer_blocks.9.attn.to_v.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
634
  "transformer_blocks.9.encoder_proj.bias": "diffusion_pytorch_model-00007-of-00007.safetensors",
635
  "transformer_blocks.9.encoder_proj.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
636
- "transformer_blocks.9.img_mlp.experts.w1": "diffusion_pytorch_model-00007-of-00007.safetensors",
637
- "transformer_blocks.9.img_mlp.experts.w2": "diffusion_pytorch_model-00007-of-00007.safetensors",
638
- "transformer_blocks.9.img_mlp.experts.w3": "diffusion_pytorch_model-00007-of-00007.safetensors",
639
  "transformer_blocks.9.img_mlp.gate.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
640
  "transformer_blocks.9.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
641
  "transformer_blocks.9.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
 
57
  "transformer_blocks.10.attn.to_v.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
58
  "transformer_blocks.10.encoder_proj.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
59
  "transformer_blocks.10.encoder_proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
60
+ "transformer_blocks.10.img_mlp.experts.down_proj": "diffusion_pytorch_model-00001-of-00007.safetensors",
61
+ "transformer_blocks.10.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00001-of-00007.safetensors",
 
62
  "transformer_blocks.10.img_mlp.gate.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
63
  "transformer_blocks.10.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
64
  "transformer_blocks.10.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
 
76
  "transformer_blocks.11.attn.to_v.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
77
  "transformer_blocks.11.encoder_proj.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
78
  "transformer_blocks.11.encoder_proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
79
+ "transformer_blocks.11.img_mlp.experts.down_proj": "diffusion_pytorch_model-00001-of-00007.safetensors",
80
+ "transformer_blocks.11.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00001-of-00007.safetensors",
 
81
  "transformer_blocks.11.img_mlp.gate.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
82
  "transformer_blocks.11.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
83
  "transformer_blocks.11.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
 
95
  "transformer_blocks.12.attn.to_v.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
96
  "transformer_blocks.12.encoder_proj.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
97
  "transformer_blocks.12.encoder_proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
98
+ "transformer_blocks.12.img_mlp.experts.down_proj": "diffusion_pytorch_model-00001-of-00007.safetensors",
99
+ "transformer_blocks.12.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00001-of-00007.safetensors",
 
100
  "transformer_blocks.12.img_mlp.gate.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
101
  "transformer_blocks.12.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
102
  "transformer_blocks.12.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
 
114
  "transformer_blocks.13.attn.to_v.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
115
  "transformer_blocks.13.encoder_proj.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
116
  "transformer_blocks.13.encoder_proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
117
+ "transformer_blocks.13.img_mlp.experts.down_proj": "diffusion_pytorch_model-00001-of-00007.safetensors",
118
+ "transformer_blocks.13.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00001-of-00007.safetensors",
 
119
  "transformer_blocks.13.img_mlp.gate.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
120
  "transformer_blocks.13.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
121
  "transformer_blocks.13.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
 
133
  "transformer_blocks.14.attn.to_v.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
134
  "transformer_blocks.14.encoder_proj.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
135
  "transformer_blocks.14.encoder_proj.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
136
+ "transformer_blocks.14.img_mlp.experts.down_proj": "diffusion_pytorch_model-00001-of-00007.safetensors",
137
+ "transformer_blocks.14.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00002-of-00007.safetensors",
 
138
  "transformer_blocks.14.img_mlp.gate.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
139
  "transformer_blocks.14.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
140
  "transformer_blocks.14.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
 
152
  "transformer_blocks.15.attn.to_v.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
153
  "transformer_blocks.15.encoder_proj.bias": "diffusion_pytorch_model-00002-of-00007.safetensors",
154
  "transformer_blocks.15.encoder_proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
155
+ "transformer_blocks.15.img_mlp.experts.down_proj": "diffusion_pytorch_model-00002-of-00007.safetensors",
156
+ "transformer_blocks.15.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00002-of-00007.safetensors",
 
157
  "transformer_blocks.15.img_mlp.gate.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
158
  "transformer_blocks.15.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
159
  "transformer_blocks.15.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
 
171
  "transformer_blocks.16.attn.to_v.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
172
  "transformer_blocks.16.encoder_proj.bias": "diffusion_pytorch_model-00002-of-00007.safetensors",
173
  "transformer_blocks.16.encoder_proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
174
+ "transformer_blocks.16.img_mlp.experts.down_proj": "diffusion_pytorch_model-00002-of-00007.safetensors",
175
+ "transformer_blocks.16.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00002-of-00007.safetensors",
 
176
  "transformer_blocks.16.img_mlp.gate.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
177
  "transformer_blocks.16.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
178
  "transformer_blocks.16.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
 
190
  "transformer_blocks.17.attn.to_v.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
191
  "transformer_blocks.17.encoder_proj.bias": "diffusion_pytorch_model-00002-of-00007.safetensors",
192
  "transformer_blocks.17.encoder_proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
193
+ "transformer_blocks.17.img_mlp.experts.down_proj": "diffusion_pytorch_model-00002-of-00007.safetensors",
194
+ "transformer_blocks.17.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00002-of-00007.safetensors",
 
195
  "transformer_blocks.17.img_mlp.gate.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
196
  "transformer_blocks.17.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
197
  "transformer_blocks.17.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
 
209
  "transformer_blocks.18.attn.to_v.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
210
  "transformer_blocks.18.encoder_proj.bias": "diffusion_pytorch_model-00002-of-00007.safetensors",
211
  "transformer_blocks.18.encoder_proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
212
+ "transformer_blocks.18.img_mlp.experts.down_proj": "diffusion_pytorch_model-00002-of-00007.safetensors",
213
+ "transformer_blocks.18.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00002-of-00007.safetensors",
 
214
  "transformer_blocks.18.img_mlp.gate.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
215
  "transformer_blocks.18.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
216
  "transformer_blocks.18.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
 
228
  "transformer_blocks.19.attn.to_v.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
229
  "transformer_blocks.19.encoder_proj.bias": "diffusion_pytorch_model-00003-of-00007.safetensors",
230
  "transformer_blocks.19.encoder_proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
231
+ "transformer_blocks.19.img_mlp.experts.down_proj": "diffusion_pytorch_model-00003-of-00007.safetensors",
232
+ "transformer_blocks.19.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00003-of-00007.safetensors",
 
233
  "transformer_blocks.19.img_mlp.gate.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
234
  "transformer_blocks.19.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
235
  "transformer_blocks.19.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
 
263
  "transformer_blocks.20.attn.to_v.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
264
  "transformer_blocks.20.encoder_proj.bias": "diffusion_pytorch_model-00003-of-00007.safetensors",
265
  "transformer_blocks.20.encoder_proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
266
+ "transformer_blocks.20.img_mlp.experts.down_proj": "diffusion_pytorch_model-00003-of-00007.safetensors",
267
+ "transformer_blocks.20.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00003-of-00007.safetensors",
 
268
  "transformer_blocks.20.img_mlp.gate.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
269
  "transformer_blocks.20.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
270
  "transformer_blocks.20.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
 
282
  "transformer_blocks.21.attn.to_v.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
283
  "transformer_blocks.21.encoder_proj.bias": "diffusion_pytorch_model-00003-of-00007.safetensors",
284
  "transformer_blocks.21.encoder_proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
285
+ "transformer_blocks.21.img_mlp.experts.down_proj": "diffusion_pytorch_model-00003-of-00007.safetensors",
286
+ "transformer_blocks.21.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00003-of-00007.safetensors",
 
287
  "transformer_blocks.21.img_mlp.gate.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
288
  "transformer_blocks.21.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
289
  "transformer_blocks.21.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
 
301
  "transformer_blocks.22.attn.to_v.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
302
  "transformer_blocks.22.encoder_proj.bias": "diffusion_pytorch_model-00003-of-00007.safetensors",
303
  "transformer_blocks.22.encoder_proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
304
+ "transformer_blocks.22.img_mlp.experts.down_proj": "diffusion_pytorch_model-00003-of-00007.safetensors",
305
+ "transformer_blocks.22.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00003-of-00007.safetensors",
 
306
  "transformer_blocks.22.img_mlp.gate.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
307
  "transformer_blocks.22.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
308
  "transformer_blocks.22.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
 
320
  "transformer_blocks.23.attn.to_v.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
321
  "transformer_blocks.23.encoder_proj.bias": "diffusion_pytorch_model-00003-of-00007.safetensors",
322
  "transformer_blocks.23.encoder_proj.weight": "diffusion_pytorch_model-00003-of-00007.safetensors",
323
+ "transformer_blocks.23.img_mlp.experts.down_proj": "diffusion_pytorch_model-00003-of-00007.safetensors",
324
+ "transformer_blocks.23.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00004-of-00007.safetensors",
 
325
  "transformer_blocks.23.img_mlp.gate.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
326
  "transformer_blocks.23.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
327
  "transformer_blocks.23.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
 
339
  "transformer_blocks.24.attn.to_v.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
340
  "transformer_blocks.24.encoder_proj.bias": "diffusion_pytorch_model-00004-of-00007.safetensors",
341
  "transformer_blocks.24.encoder_proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
342
+ "transformer_blocks.24.img_mlp.experts.down_proj": "diffusion_pytorch_model-00004-of-00007.safetensors",
343
+ "transformer_blocks.24.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00004-of-00007.safetensors",
 
344
  "transformer_blocks.24.img_mlp.gate.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
345
  "transformer_blocks.24.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
346
  "transformer_blocks.24.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
 
358
  "transformer_blocks.25.attn.to_v.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
359
  "transformer_blocks.25.encoder_proj.bias": "diffusion_pytorch_model-00004-of-00007.safetensors",
360
  "transformer_blocks.25.encoder_proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
361
+ "transformer_blocks.25.img_mlp.experts.down_proj": "diffusion_pytorch_model-00004-of-00007.safetensors",
362
+ "transformer_blocks.25.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00004-of-00007.safetensors",
 
363
  "transformer_blocks.25.img_mlp.gate.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
364
  "transformer_blocks.25.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
365
  "transformer_blocks.25.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
 
377
  "transformer_blocks.26.attn.to_v.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
378
  "transformer_blocks.26.encoder_proj.bias": "diffusion_pytorch_model-00004-of-00007.safetensors",
379
  "transformer_blocks.26.encoder_proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
380
+ "transformer_blocks.26.img_mlp.experts.down_proj": "diffusion_pytorch_model-00004-of-00007.safetensors",
381
+ "transformer_blocks.26.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00004-of-00007.safetensors",
 
382
  "transformer_blocks.26.img_mlp.gate.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
383
  "transformer_blocks.26.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
384
  "transformer_blocks.26.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
 
396
  "transformer_blocks.27.attn.to_v.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
397
  "transformer_blocks.27.encoder_proj.bias": "diffusion_pytorch_model-00004-of-00007.safetensors",
398
  "transformer_blocks.27.encoder_proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
399
+ "transformer_blocks.27.img_mlp.experts.down_proj": "diffusion_pytorch_model-00004-of-00007.safetensors",
400
+ "transformer_blocks.27.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00004-of-00007.safetensors",
 
401
  "transformer_blocks.27.img_mlp.gate.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
402
  "transformer_blocks.27.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
403
  "transformer_blocks.27.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00004-of-00007.safetensors",
 
415
  "transformer_blocks.28.attn.to_v.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
416
  "transformer_blocks.28.encoder_proj.bias": "diffusion_pytorch_model-00005-of-00007.safetensors",
417
  "transformer_blocks.28.encoder_proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
418
+ "transformer_blocks.28.img_mlp.experts.down_proj": "diffusion_pytorch_model-00005-of-00007.safetensors",
419
+ "transformer_blocks.28.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00005-of-00007.safetensors",
 
420
  "transformer_blocks.28.img_mlp.gate.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
421
  "transformer_blocks.28.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
422
  "transformer_blocks.28.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
 
434
  "transformer_blocks.29.attn.to_v.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
435
  "transformer_blocks.29.encoder_proj.bias": "diffusion_pytorch_model-00005-of-00007.safetensors",
436
  "transformer_blocks.29.encoder_proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
437
+ "transformer_blocks.29.img_mlp.experts.down_proj": "diffusion_pytorch_model-00005-of-00007.safetensors",
438
+ "transformer_blocks.29.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00005-of-00007.safetensors",
 
439
  "transformer_blocks.29.img_mlp.gate.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
440
  "transformer_blocks.29.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
441
  "transformer_blocks.29.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
 
453
  "transformer_blocks.3.attn.to_v.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
454
  "transformer_blocks.3.encoder_proj.bias": "diffusion_pytorch_model-00005-of-00007.safetensors",
455
  "transformer_blocks.3.encoder_proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
456
+ "transformer_blocks.3.img_mlp.experts.down_proj": "diffusion_pytorch_model-00005-of-00007.safetensors",
457
+ "transformer_blocks.3.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00005-of-00007.safetensors",
 
458
  "transformer_blocks.3.img_mlp.gate.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
459
  "transformer_blocks.3.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
460
  "transformer_blocks.3.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
 
472
  "transformer_blocks.30.attn.to_v.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
473
  "transformer_blocks.30.encoder_proj.bias": "diffusion_pytorch_model-00005-of-00007.safetensors",
474
  "transformer_blocks.30.encoder_proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
475
+ "transformer_blocks.30.img_mlp.experts.down_proj": "diffusion_pytorch_model-00005-of-00007.safetensors",
476
+ "transformer_blocks.30.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00005-of-00007.safetensors",
 
477
  "transformer_blocks.30.img_mlp.gate.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
478
  "transformer_blocks.30.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
479
  "transformer_blocks.30.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
 
491
  "transformer_blocks.31.attn.to_v.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
492
  "transformer_blocks.31.encoder_proj.bias": "diffusion_pytorch_model-00005-of-00007.safetensors",
493
  "transformer_blocks.31.encoder_proj.weight": "diffusion_pytorch_model-00005-of-00007.safetensors",
494
+ "transformer_blocks.31.img_mlp.experts.down_proj": "diffusion_pytorch_model-00005-of-00007.safetensors",
495
+ "transformer_blocks.31.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00006-of-00007.safetensors",
 
496
  "transformer_blocks.31.img_mlp.gate.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
497
  "transformer_blocks.31.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
498
  "transformer_blocks.31.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
 
510
  "transformer_blocks.4.attn.to_v.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
511
  "transformer_blocks.4.encoder_proj.bias": "diffusion_pytorch_model-00006-of-00007.safetensors",
512
  "transformer_blocks.4.encoder_proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
513
+ "transformer_blocks.4.img_mlp.experts.down_proj": "diffusion_pytorch_model-00006-of-00007.safetensors",
514
+ "transformer_blocks.4.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00006-of-00007.safetensors",
 
515
  "transformer_blocks.4.img_mlp.gate.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
516
  "transformer_blocks.4.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
517
  "transformer_blocks.4.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
 
529
  "transformer_blocks.5.attn.to_v.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
530
  "transformer_blocks.5.encoder_proj.bias": "diffusion_pytorch_model-00006-of-00007.safetensors",
531
  "transformer_blocks.5.encoder_proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
532
+ "transformer_blocks.5.img_mlp.experts.down_proj": "diffusion_pytorch_model-00006-of-00007.safetensors",
533
+ "transformer_blocks.5.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00006-of-00007.safetensors",
 
534
  "transformer_blocks.5.img_mlp.gate.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
535
  "transformer_blocks.5.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
536
  "transformer_blocks.5.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
 
548
  "transformer_blocks.6.attn.to_v.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
549
  "transformer_blocks.6.encoder_proj.bias": "diffusion_pytorch_model-00006-of-00007.safetensors",
550
  "transformer_blocks.6.encoder_proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
551
+ "transformer_blocks.6.img_mlp.experts.down_proj": "diffusion_pytorch_model-00006-of-00007.safetensors",
552
+ "transformer_blocks.6.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00006-of-00007.safetensors",
 
553
  "transformer_blocks.6.img_mlp.gate.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
554
  "transformer_blocks.6.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
555
  "transformer_blocks.6.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
 
567
  "transformer_blocks.7.attn.to_v.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
568
  "transformer_blocks.7.encoder_proj.bias": "diffusion_pytorch_model-00006-of-00007.safetensors",
569
  "transformer_blocks.7.encoder_proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
570
+ "transformer_blocks.7.img_mlp.experts.down_proj": "diffusion_pytorch_model-00006-of-00007.safetensors",
571
+ "transformer_blocks.7.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00006-of-00007.safetensors",
 
572
  "transformer_blocks.7.img_mlp.gate.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
573
  "transformer_blocks.7.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
574
  "transformer_blocks.7.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
 
582
  "transformer_blocks.8.attn.norm_q.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
583
  "transformer_blocks.8.attn.to_k.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
584
  "transformer_blocks.8.attn.to_out.0.weight": "diffusion_pytorch_model-00006-of-00007.safetensors",
585
+ "transformer_blocks.8.attn.to_q.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
586
+ "transformer_blocks.8.attn.to_v.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
587
+ "transformer_blocks.8.encoder_proj.bias": "diffusion_pytorch_model-00007-of-00007.safetensors",
588
+ "transformer_blocks.8.encoder_proj.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
589
+ "transformer_blocks.8.img_mlp.experts.down_proj": "diffusion_pytorch_model-00007-of-00007.safetensors",
590
+ "transformer_blocks.8.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00007-of-00007.safetensors",
 
591
  "transformer_blocks.8.img_mlp.gate.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
592
  "transformer_blocks.8.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
593
  "transformer_blocks.8.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
 
605
  "transformer_blocks.9.attn.to_v.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
606
  "transformer_blocks.9.encoder_proj.bias": "diffusion_pytorch_model-00007-of-00007.safetensors",
607
  "transformer_blocks.9.encoder_proj.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
608
+ "transformer_blocks.9.img_mlp.experts.down_proj": "diffusion_pytorch_model-00007-of-00007.safetensors",
609
+ "transformer_blocks.9.img_mlp.experts.gate_up_proj": "diffusion_pytorch_model-00007-of-00007.safetensors",
 
610
  "transformer_blocks.9.img_mlp.gate.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
611
  "transformer_blocks.9.img_mlp.shared_expert.net.0.proj.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",
612
  "transformer_blocks.9.img_mlp.shared_expert.net.2.weight": "diffusion_pytorch_model-00007-of-00007.safetensors",