bodhisativa commited on
Commit
7e5c508
·
verified ·
1 Parent(s): f323e93

Add files using upload-large-folder tool

Browse files
Files changed (38) hide show
  1. .gitattributes +35 -35
  2. README.md +7 -0
  3. example workflows_Wan2.1/image_to_video_wan_480p_example.json +697 -0
  4. example workflows_Wan2.1/image_to_video_wan_720p_example.json +697 -0
  5. example workflows_Wan2.1/text_to_video_wan.json +481 -0
  6. split_files/clip_vision/clip_vision_h.safetensors +3 -0
  7. split_files/diffusion_models/wan2.1_flf2v_720p_14B_fp16.safetensors +3 -0
  8. split_files/diffusion_models/wan2.1_flf2v_720p_14B_fp8_e4m3fn.safetensors +3 -0
  9. split_files/diffusion_models/wan2.1_fun_camera_v1.1_1.3B_bf16.safetensors +3 -0
  10. split_files/diffusion_models/wan2.1_fun_camera_v1.1_14B_bf16.safetensors +3 -0
  11. split_files/diffusion_models/wan2.1_fun_control_1.3B_bf16.safetensors +3 -0
  12. split_files/diffusion_models/wan2.1_fun_inp_1.3B_bf16.safetensors +3 -0
  13. split_files/diffusion_models/wan2.1_i2v_480p_14B_bf16.safetensors +3 -0
  14. split_files/diffusion_models/wan2.1_i2v_480p_14B_fp16.safetensors +3 -0
  15. split_files/diffusion_models/wan2.1_i2v_480p_14B_fp8_e4m3fn.safetensors +3 -0
  16. split_files/diffusion_models/wan2.1_i2v_480p_14B_fp8_scaled.safetensors +3 -0
  17. split_files/diffusion_models/wan2.1_i2v_720p_14B_bf16.safetensors +3 -0
  18. split_files/diffusion_models/wan2.1_i2v_720p_14B_fp16.safetensors +3 -0
  19. split_files/diffusion_models/wan2.1_i2v_720p_14B_fp8_e4m3fn.safetensors +3 -0
  20. split_files/diffusion_models/wan2.1_i2v_720p_14B_fp8_scaled.safetensors +3 -0
  21. split_files/diffusion_models/wan2.1_magref_14B_fp16.safetensors +3 -0
  22. split_files/diffusion_models/wan2.1_t2v_1.3B_bf16.safetensors +3 -0
  23. split_files/diffusion_models/wan2.1_t2v_1.3B_fp16.safetensors +3 -0
  24. split_files/diffusion_models/wan2.1_t2v_14B_bf16.safetensors +3 -0
  25. split_files/diffusion_models/wan2.1_t2v_14B_fp16.safetensors +3 -0
  26. split_files/diffusion_models/wan2.1_t2v_14B_fp8_e4m3fn.safetensors +3 -0
  27. split_files/diffusion_models/wan2.1_t2v_14B_fp8_scaled.safetensors +3 -0
  28. split_files/diffusion_models/wan2.1_vace_1.3B_fp16.safetensors +3 -0
  29. split_files/diffusion_models/wan2.1_vace_1.3B_preview_fp16.safetensors +3 -0
  30. split_files/diffusion_models/wan2.1_vace_14B_fp16.safetensors +3 -0
  31. split_files/loras/wan_alpha_2.1_rgba_lora.safetensors +3 -0
  32. split_files/model_patches/wan2.1_infiniteTalk_multi_fp16.safetensors +3 -0
  33. split_files/model_patches/wan2.1_infiniteTalk_single_fp16.safetensors +3 -0
  34. split_files/text_encoders/umt5_xxl_fp16.safetensors +3 -0
  35. split_files/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensors +3 -0
  36. split_files/vae/wan_2.1_vae.safetensors +3 -0
  37. split_files/vae/wan_alpha_2.1_vae_alpha_channel.safetensors +3 -0
  38. split_files/vae/wan_alpha_2.1_vae_rgb_channel.safetensors +3 -0
.gitattributes CHANGED
@@ -1,35 +1,35 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - diffusion-single-file
4
+ - comfyui
5
+ ---
6
+
7
+ Wan 2.1 repackaged for ComfyUI use. For examples see: https://comfyanonymous.github.io/ComfyUI_examples/wan
example workflows_Wan2.1/image_to_video_wan_480p_example.json ADDED
@@ -0,0 +1,697 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "last_node_id": 53,
3
+ "last_link_id": 109,
4
+ "nodes": [
5
+ {
6
+ "id": 8,
7
+ "type": "VAEDecode",
8
+ "pos": [
9
+ 1210,
10
+ 190
11
+ ],
12
+ "size": [
13
+ 210,
14
+ 46
15
+ ],
16
+ "flags": {},
17
+ "order": 10,
18
+ "mode": 0,
19
+ "inputs": [
20
+ {
21
+ "name": "samples",
22
+ "type": "LATENT",
23
+ "link": 35
24
+ },
25
+ {
26
+ "name": "vae",
27
+ "type": "VAE",
28
+ "link": 76
29
+ }
30
+ ],
31
+ "outputs": [
32
+ {
33
+ "name": "IMAGE",
34
+ "type": "IMAGE",
35
+ "links": [
36
+ 56,
37
+ 93
38
+ ],
39
+ "slot_index": 0
40
+ }
41
+ ],
42
+ "properties": {
43
+ "Node name for S&R": "VAEDecode"
44
+ },
45
+ "widgets_values": []
46
+ },
47
+ {
48
+ "id": 39,
49
+ "type": "VAELoader",
50
+ "pos": [
51
+ 866.3932495117188,
52
+ 499.18597412109375
53
+ ],
54
+ "size": [
55
+ 306.36004638671875,
56
+ 58
57
+ ],
58
+ "flags": {},
59
+ "order": 0,
60
+ "mode": 0,
61
+ "inputs": [],
62
+ "outputs": [
63
+ {
64
+ "name": "VAE",
65
+ "type": "VAE",
66
+ "links": [
67
+ 76,
68
+ 99
69
+ ],
70
+ "slot_index": 0
71
+ }
72
+ ],
73
+ "properties": {
74
+ "Node name for S&R": "VAELoader"
75
+ },
76
+ "widgets_values": [
77
+ "wan_2.1_vae.safetensors"
78
+ ]
79
+ },
80
+ {
81
+ "id": 28,
82
+ "type": "SaveAnimatedWEBP",
83
+ "pos": [
84
+ 1460,
85
+ 190
86
+ ],
87
+ "size": [
88
+ 870.8511352539062,
89
+ 643.7430419921875
90
+ ],
91
+ "flags": {},
92
+ "order": 11,
93
+ "mode": 0,
94
+ "inputs": [
95
+ {
96
+ "name": "images",
97
+ "type": "IMAGE",
98
+ "link": 56
99
+ }
100
+ ],
101
+ "outputs": [],
102
+ "properties": {},
103
+ "widgets_values": [
104
+ "ComfyUI",
105
+ 16,
106
+ false,
107
+ 90,
108
+ "default"
109
+ ]
110
+ },
111
+ {
112
+ "id": 38,
113
+ "type": "CLIPLoader",
114
+ "pos": [
115
+ 12.94982624053955,
116
+ 184.6981658935547
117
+ ],
118
+ "size": [
119
+ 390,
120
+ 82
121
+ ],
122
+ "flags": {},
123
+ "order": 1,
124
+ "mode": 0,
125
+ "inputs": [],
126
+ "outputs": [
127
+ {
128
+ "name": "CLIP",
129
+ "type": "CLIP",
130
+ "links": [
131
+ 74,
132
+ 75
133
+ ],
134
+ "slot_index": 0
135
+ }
136
+ ],
137
+ "properties": {
138
+ "Node name for S&R": "CLIPLoader"
139
+ },
140
+ "widgets_values": [
141
+ "umt5_xxl_fp8_e4m3fn_scaled.safetensors",
142
+ "wan",
143
+ "default"
144
+ ]
145
+ },
146
+ {
147
+ "id": 47,
148
+ "type": "SaveWEBM",
149
+ "pos": [
150
+ 2367.213134765625,
151
+ 193.6114959716797
152
+ ],
153
+ "size": [
154
+ 315,
155
+ 130
156
+ ],
157
+ "flags": {},
158
+ "order": 12,
159
+ "mode": 4,
160
+ "inputs": [
161
+ {
162
+ "name": "images",
163
+ "type": "IMAGE",
164
+ "link": 93
165
+ }
166
+ ],
167
+ "outputs": [],
168
+ "properties": {
169
+ "Node name for S&R": "SaveWEBM"
170
+ },
171
+ "widgets_values": [
172
+ "ComfyUI",
173
+ "vp9",
174
+ 24,
175
+ 32
176
+ ]
177
+ },
178
+ {
179
+ "id": 37,
180
+ "type": "UNETLoader",
181
+ "pos": [
182
+ 485.1220397949219,
183
+ 57.094566345214844
184
+ ],
185
+ "size": [
186
+ 346.7470703125,
187
+ 82
188
+ ],
189
+ "flags": {},
190
+ "order": 2,
191
+ "mode": 0,
192
+ "inputs": [],
193
+ "outputs": [
194
+ {
195
+ "name": "MODEL",
196
+ "type": "MODEL",
197
+ "links": [
198
+ 92
199
+ ],
200
+ "slot_index": 0
201
+ }
202
+ ],
203
+ "properties": {
204
+ "Node name for S&R": "UNETLoader"
205
+ },
206
+ "widgets_values": [
207
+ "wan2.1_i2v_480p_14B_bf16.safetensors",
208
+ "default"
209
+ ]
210
+ },
211
+ {
212
+ "id": 51,
213
+ "type": "CLIPVisionEncode",
214
+ "pos": [
215
+ 340,
216
+ 650
217
+ ],
218
+ "size": [
219
+ 253.60000610351562,
220
+ 78
221
+ ],
222
+ "flags": {},
223
+ "order": 7,
224
+ "mode": 0,
225
+ "inputs": [
226
+ {
227
+ "name": "clip_vision",
228
+ "type": "CLIP_VISION",
229
+ "link": 94
230
+ },
231
+ {
232
+ "name": "image",
233
+ "type": "IMAGE",
234
+ "link": 109
235
+ }
236
+ ],
237
+ "outputs": [
238
+ {
239
+ "name": "CLIP_VISION_OUTPUT",
240
+ "type": "CLIP_VISION_OUTPUT",
241
+ "links": [
242
+ 107
243
+ ],
244
+ "slot_index": 0
245
+ }
246
+ ],
247
+ "properties": {
248
+ "Node name for S&R": "CLIPVisionEncode"
249
+ },
250
+ "widgets_values": [
251
+ "none"
252
+ ]
253
+ },
254
+ {
255
+ "id": 49,
256
+ "type": "CLIPVisionLoader",
257
+ "pos": [
258
+ -20,
259
+ 650
260
+ ],
261
+ "size": [
262
+ 315,
263
+ 58
264
+ ],
265
+ "flags": {},
266
+ "order": 3,
267
+ "mode": 0,
268
+ "inputs": [],
269
+ "outputs": [
270
+ {
271
+ "name": "CLIP_VISION",
272
+ "type": "CLIP_VISION",
273
+ "links": [
274
+ 94
275
+ ],
276
+ "slot_index": 0
277
+ }
278
+ ],
279
+ "properties": {
280
+ "Node name for S&R": "CLIPVisionLoader"
281
+ },
282
+ "widgets_values": [
283
+ "clip_vision_h.safetensors"
284
+ ]
285
+ },
286
+ {
287
+ "id": 52,
288
+ "type": "LoadImage",
289
+ "pos": [
290
+ -30,
291
+ 760
292
+ ],
293
+ "size": [
294
+ 315,
295
+ 314
296
+ ],
297
+ "flags": {},
298
+ "order": 4,
299
+ "mode": 0,
300
+ "inputs": [],
301
+ "outputs": [
302
+ {
303
+ "name": "IMAGE",
304
+ "type": "IMAGE",
305
+ "links": [
306
+ 106,
307
+ 109
308
+ ],
309
+ "slot_index": 0
310
+ },
311
+ {
312
+ "name": "MASK",
313
+ "type": "MASK",
314
+ "links": null,
315
+ "slot_index": 1
316
+ }
317
+ ],
318
+ "properties": {
319
+ "Node name for S&R": "LoadImage"
320
+ },
321
+ "widgets_values": [
322
+ "flux_dev_example.png",
323
+ "image"
324
+ ]
325
+ },
326
+ {
327
+ "id": 50,
328
+ "type": "WanImageToVideo",
329
+ "pos": [
330
+ 673.0507202148438,
331
+ 627.272705078125
332
+ ],
333
+ "size": [
334
+ 342.5999755859375,
335
+ 210
336
+ ],
337
+ "flags": {},
338
+ "order": 8,
339
+ "mode": 0,
340
+ "inputs": [
341
+ {
342
+ "name": "positive",
343
+ "type": "CONDITIONING",
344
+ "link": 97
345
+ },
346
+ {
347
+ "name": "negative",
348
+ "type": "CONDITIONING",
349
+ "link": 98
350
+ },
351
+ {
352
+ "name": "vae",
353
+ "type": "VAE",
354
+ "link": 99
355
+ },
356
+ {
357
+ "name": "clip_vision_output",
358
+ "type": "CLIP_VISION_OUTPUT",
359
+ "shape": 7,
360
+ "link": 107
361
+ },
362
+ {
363
+ "name": "start_image",
364
+ "type": "IMAGE",
365
+ "shape": 7,
366
+ "link": 106
367
+ }
368
+ ],
369
+ "outputs": [
370
+ {
371
+ "name": "positive",
372
+ "type": "CONDITIONING",
373
+ "links": [
374
+ 101
375
+ ],
376
+ "slot_index": 0
377
+ },
378
+ {
379
+ "name": "negative",
380
+ "type": "CONDITIONING",
381
+ "links": [
382
+ 102
383
+ ],
384
+ "slot_index": 1
385
+ },
386
+ {
387
+ "name": "latent",
388
+ "type": "LATENT",
389
+ "links": [
390
+ 103
391
+ ],
392
+ "slot_index": 2
393
+ }
394
+ ],
395
+ "properties": {
396
+ "Node name for S&R": "WanImageToVideo"
397
+ },
398
+ "widgets_values": [
399
+ 512,
400
+ 512,
401
+ 33,
402
+ 1
403
+ ]
404
+ },
405
+ {
406
+ "id": 6,
407
+ "type": "CLIPTextEncode",
408
+ "pos": [
409
+ 415,
410
+ 186
411
+ ],
412
+ "size": [
413
+ 422.84503173828125,
414
+ 164.31304931640625
415
+ ],
416
+ "flags": {},
417
+ "order": 5,
418
+ "mode": 0,
419
+ "inputs": [
420
+ {
421
+ "name": "clip",
422
+ "type": "CLIP",
423
+ "link": 74
424
+ }
425
+ ],
426
+ "outputs": [
427
+ {
428
+ "name": "CONDITIONING",
429
+ "type": "CONDITIONING",
430
+ "links": [
431
+ 97
432
+ ],
433
+ "slot_index": 0
434
+ }
435
+ ],
436
+ "title": "CLIP Text Encode (Positive Prompt)",
437
+ "properties": {
438
+ "Node name for S&R": "CLIPTextEncode"
439
+ },
440
+ "widgets_values": [
441
+ "a cute anime girl with massive fennec ears and a big fluffy tail wearing a maid outfit turning around"
442
+ ],
443
+ "color": "#232",
444
+ "bgcolor": "#353"
445
+ },
446
+ {
447
+ "id": 3,
448
+ "type": "KSampler",
449
+ "pos": [
450
+ 863,
451
+ 187
452
+ ],
453
+ "size": [
454
+ 315,
455
+ 262
456
+ ],
457
+ "flags": {},
458
+ "order": 9,
459
+ "mode": 0,
460
+ "inputs": [
461
+ {
462
+ "name": "model",
463
+ "type": "MODEL",
464
+ "link": 92
465
+ },
466
+ {
467
+ "name": "positive",
468
+ "type": "CONDITIONING",
469
+ "link": 101
470
+ },
471
+ {
472
+ "name": "negative",
473
+ "type": "CONDITIONING",
474
+ "link": 102
475
+ },
476
+ {
477
+ "name": "latent_image",
478
+ "type": "LATENT",
479
+ "link": 103
480
+ }
481
+ ],
482
+ "outputs": [
483
+ {
484
+ "name": "LATENT",
485
+ "type": "LATENT",
486
+ "links": [
487
+ 35
488
+ ],
489
+ "slot_index": 0
490
+ }
491
+ ],
492
+ "properties": {
493
+ "Node name for S&R": "KSampler"
494
+ },
495
+ "widgets_values": [
496
+ 987948718394761,
497
+ "randomize",
498
+ 20,
499
+ 6,
500
+ "uni_pc",
501
+ "simple",
502
+ 1
503
+ ]
504
+ },
505
+ {
506
+ "id": 7,
507
+ "type": "CLIPTextEncode",
508
+ "pos": [
509
+ 413,
510
+ 389
511
+ ],
512
+ "size": [
513
+ 425.27801513671875,
514
+ 180.6060791015625
515
+ ],
516
+ "flags": {},
517
+ "order": 6,
518
+ "mode": 0,
519
+ "inputs": [
520
+ {
521
+ "name": "clip",
522
+ "type": "CLIP",
523
+ "link": 75
524
+ }
525
+ ],
526
+ "outputs": [
527
+ {
528
+ "name": "CONDITIONING",
529
+ "type": "CONDITIONING",
530
+ "links": [
531
+ 98
532
+ ],
533
+ "slot_index": 0
534
+ }
535
+ ],
536
+ "title": "CLIP Text Encode (Negative Prompt)",
537
+ "properties": {
538
+ "Node name for S&R": "CLIPTextEncode"
539
+ },
540
+ "widgets_values": [
541
+ "Overexposure, static, blurred details, subtitles, paintings, pictures, still, overall gray, worst quality, low quality, JPEG compression residue, ugly, mutilated, redundant fingers, poorly painted hands, poorly painted faces, deformed, disfigured, deformed limbs, fused fingers, cluttered background, three legs, a lot of people in the background, upside down"
542
+ ],
543
+ "color": "#322",
544
+ "bgcolor": "#533"
545
+ }
546
+ ],
547
+ "links": [
548
+ [
549
+ 35,
550
+ 3,
551
+ 0,
552
+ 8,
553
+ 0,
554
+ "LATENT"
555
+ ],
556
+ [
557
+ 56,
558
+ 8,
559
+ 0,
560
+ 28,
561
+ 0,
562
+ "IMAGE"
563
+ ],
564
+ [
565
+ 74,
566
+ 38,
567
+ 0,
568
+ 6,
569
+ 0,
570
+ "CLIP"
571
+ ],
572
+ [
573
+ 75,
574
+ 38,
575
+ 0,
576
+ 7,
577
+ 0,
578
+ "CLIP"
579
+ ],
580
+ [
581
+ 76,
582
+ 39,
583
+ 0,
584
+ 8,
585
+ 1,
586
+ "VAE"
587
+ ],
588
+ [
589
+ 92,
590
+ 37,
591
+ 0,
592
+ 3,
593
+ 0,
594
+ "MODEL"
595
+ ],
596
+ [
597
+ 93,
598
+ 8,
599
+ 0,
600
+ 47,
601
+ 0,
602
+ "IMAGE"
603
+ ],
604
+ [
605
+ 94,
606
+ 49,
607
+ 0,
608
+ 51,
609
+ 0,
610
+ "CLIP_VISION"
611
+ ],
612
+ [
613
+ 97,
614
+ 6,
615
+ 0,
616
+ 50,
617
+ 0,
618
+ "CONDITIONING"
619
+ ],
620
+ [
621
+ 98,
622
+ 7,
623
+ 0,
624
+ 50,
625
+ 1,
626
+ "CONDITIONING"
627
+ ],
628
+ [
629
+ 99,
630
+ 39,
631
+ 0,
632
+ 50,
633
+ 2,
634
+ "VAE"
635
+ ],
636
+ [
637
+ 101,
638
+ 50,
639
+ 0,
640
+ 3,
641
+ 1,
642
+ "CONDITIONING"
643
+ ],
644
+ [
645
+ 102,
646
+ 50,
647
+ 1,
648
+ 3,
649
+ 2,
650
+ "CONDITIONING"
651
+ ],
652
+ [
653
+ 103,
654
+ 50,
655
+ 2,
656
+ 3,
657
+ 3,
658
+ "LATENT"
659
+ ],
660
+ [
661
+ 106,
662
+ 52,
663
+ 0,
664
+ 50,
665
+ 4,
666
+ "IMAGE"
667
+ ],
668
+ [
669
+ 107,
670
+ 51,
671
+ 0,
672
+ 50,
673
+ 3,
674
+ "CLIP_VISION_OUTPUT"
675
+ ],
676
+ [
677
+ 109,
678
+ 52,
679
+ 0,
680
+ 51,
681
+ 1,
682
+ "IMAGE"
683
+ ]
684
+ ],
685
+ "groups": [],
686
+ "config": {},
687
+ "extra": {
688
+ "ds": {
689
+ "scale": 1.015255979947749,
690
+ "offset": [
691
+ 615.7681203368851,
692
+ 205.10773519106684
693
+ ]
694
+ }
695
+ },
696
+ "version": 0.4
697
+ }
example workflows_Wan2.1/image_to_video_wan_720p_example.json ADDED
@@ -0,0 +1,697 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "last_node_id": 53,
3
+ "last_link_id": 109,
4
+ "nodes": [
5
+ {
6
+ "id": 8,
7
+ "type": "VAEDecode",
8
+ "pos": [
9
+ 1210,
10
+ 190
11
+ ],
12
+ "size": [
13
+ 210,
14
+ 46
15
+ ],
16
+ "flags": {},
17
+ "order": 10,
18
+ "mode": 0,
19
+ "inputs": [
20
+ {
21
+ "name": "samples",
22
+ "type": "LATENT",
23
+ "link": 35
24
+ },
25
+ {
26
+ "name": "vae",
27
+ "type": "VAE",
28
+ "link": 76
29
+ }
30
+ ],
31
+ "outputs": [
32
+ {
33
+ "name": "IMAGE",
34
+ "type": "IMAGE",
35
+ "links": [
36
+ 56,
37
+ 93
38
+ ],
39
+ "slot_index": 0
40
+ }
41
+ ],
42
+ "properties": {
43
+ "Node name for S&R": "VAEDecode"
44
+ },
45
+ "widgets_values": []
46
+ },
47
+ {
48
+ "id": 39,
49
+ "type": "VAELoader",
50
+ "pos": [
51
+ 866.3932495117188,
52
+ 499.18597412109375
53
+ ],
54
+ "size": [
55
+ 306.36004638671875,
56
+ 58
57
+ ],
58
+ "flags": {},
59
+ "order": 0,
60
+ "mode": 0,
61
+ "inputs": [],
62
+ "outputs": [
63
+ {
64
+ "name": "VAE",
65
+ "type": "VAE",
66
+ "links": [
67
+ 76,
68
+ 99
69
+ ],
70
+ "slot_index": 0
71
+ }
72
+ ],
73
+ "properties": {
74
+ "Node name for S&R": "VAELoader"
75
+ },
76
+ "widgets_values": [
77
+ "wan_2.1_vae.safetensors"
78
+ ]
79
+ },
80
+ {
81
+ "id": 28,
82
+ "type": "SaveAnimatedWEBP",
83
+ "pos": [
84
+ 1460,
85
+ 190
86
+ ],
87
+ "size": [
88
+ 870.8511352539062,
89
+ 643.7430419921875
90
+ ],
91
+ "flags": {},
92
+ "order": 11,
93
+ "mode": 0,
94
+ "inputs": [
95
+ {
96
+ "name": "images",
97
+ "type": "IMAGE",
98
+ "link": 56
99
+ }
100
+ ],
101
+ "outputs": [],
102
+ "properties": {},
103
+ "widgets_values": [
104
+ "ComfyUI",
105
+ 16,
106
+ false,
107
+ 90,
108
+ "default"
109
+ ]
110
+ },
111
+ {
112
+ "id": 51,
113
+ "type": "CLIPVisionEncode",
114
+ "pos": [
115
+ 340,
116
+ 650
117
+ ],
118
+ "size": [
119
+ 253.60000610351562,
120
+ 78
121
+ ],
122
+ "flags": {},
123
+ "order": 7,
124
+ "mode": 0,
125
+ "inputs": [
126
+ {
127
+ "name": "clip_vision",
128
+ "type": "CLIP_VISION",
129
+ "link": 94
130
+ },
131
+ {
132
+ "name": "image",
133
+ "type": "IMAGE",
134
+ "link": 109
135
+ }
136
+ ],
137
+ "outputs": [
138
+ {
139
+ "name": "CLIP_VISION_OUTPUT",
140
+ "type": "CLIP_VISION_OUTPUT",
141
+ "links": [
142
+ 107
143
+ ],
144
+ "slot_index": 0
145
+ }
146
+ ],
147
+ "properties": {
148
+ "Node name for S&R": "CLIPVisionEncode"
149
+ },
150
+ "widgets_values": [
151
+ "none"
152
+ ]
153
+ },
154
+ {
155
+ "id": 49,
156
+ "type": "CLIPVisionLoader",
157
+ "pos": [
158
+ -20,
159
+ 650
160
+ ],
161
+ "size": [
162
+ 315,
163
+ 58
164
+ ],
165
+ "flags": {},
166
+ "order": 1,
167
+ "mode": 0,
168
+ "inputs": [],
169
+ "outputs": [
170
+ {
171
+ "name": "CLIP_VISION",
172
+ "type": "CLIP_VISION",
173
+ "links": [
174
+ 94
175
+ ],
176
+ "slot_index": 0
177
+ }
178
+ ],
179
+ "properties": {
180
+ "Node name for S&R": "CLIPVisionLoader"
181
+ },
182
+ "widgets_values": [
183
+ "clip_vision_h.safetensors"
184
+ ]
185
+ },
186
+ {
187
+ "id": 3,
188
+ "type": "KSampler",
189
+ "pos": [
190
+ 863,
191
+ 187
192
+ ],
193
+ "size": [
194
+ 315,
195
+ 262
196
+ ],
197
+ "flags": {},
198
+ "order": 9,
199
+ "mode": 0,
200
+ "inputs": [
201
+ {
202
+ "name": "model",
203
+ "type": "MODEL",
204
+ "link": 92
205
+ },
206
+ {
207
+ "name": "positive",
208
+ "type": "CONDITIONING",
209
+ "link": 101
210
+ },
211
+ {
212
+ "name": "negative",
213
+ "type": "CONDITIONING",
214
+ "link": 102
215
+ },
216
+ {
217
+ "name": "latent_image",
218
+ "type": "LATENT",
219
+ "link": 103
220
+ }
221
+ ],
222
+ "outputs": [
223
+ {
224
+ "name": "LATENT",
225
+ "type": "LATENT",
226
+ "links": [
227
+ 35
228
+ ],
229
+ "slot_index": 0
230
+ }
231
+ ],
232
+ "properties": {
233
+ "Node name for S&R": "KSampler"
234
+ },
235
+ "widgets_values": [
236
+ 353470491069365,
237
+ "randomize",
238
+ 20,
239
+ 6,
240
+ "uni_pc",
241
+ "simple",
242
+ 1
243
+ ]
244
+ },
245
+ {
246
+ "id": 50,
247
+ "type": "WanImageToVideo",
248
+ "pos": [
249
+ 673.0507202148438,
250
+ 627.272705078125
251
+ ],
252
+ "size": [
253
+ 342.5999755859375,
254
+ 210
255
+ ],
256
+ "flags": {},
257
+ "order": 8,
258
+ "mode": 0,
259
+ "inputs": [
260
+ {
261
+ "name": "positive",
262
+ "type": "CONDITIONING",
263
+ "link": 97
264
+ },
265
+ {
266
+ "name": "negative",
267
+ "type": "CONDITIONING",
268
+ "link": 98
269
+ },
270
+ {
271
+ "name": "vae",
272
+ "type": "VAE",
273
+ "link": 99
274
+ },
275
+ {
276
+ "name": "clip_vision_output",
277
+ "type": "CLIP_VISION_OUTPUT",
278
+ "shape": 7,
279
+ "link": 107
280
+ },
281
+ {
282
+ "name": "start_image",
283
+ "type": "IMAGE",
284
+ "shape": 7,
285
+ "link": 106
286
+ }
287
+ ],
288
+ "outputs": [
289
+ {
290
+ "name": "positive",
291
+ "type": "CONDITIONING",
292
+ "links": [
293
+ 101
294
+ ],
295
+ "slot_index": 0
296
+ },
297
+ {
298
+ "name": "negative",
299
+ "type": "CONDITIONING",
300
+ "links": [
301
+ 102
302
+ ],
303
+ "slot_index": 1
304
+ },
305
+ {
306
+ "name": "latent",
307
+ "type": "LATENT",
308
+ "links": [
309
+ 103
310
+ ],
311
+ "slot_index": 2
312
+ }
313
+ ],
314
+ "properties": {
315
+ "Node name for S&R": "WanImageToVideo"
316
+ },
317
+ "widgets_values": [
318
+ 768,
319
+ 768,
320
+ 53,
321
+ 1
322
+ ]
323
+ },
324
+ {
325
+ "id": 38,
326
+ "type": "CLIPLoader",
327
+ "pos": [
328
+ 12.94982624053955,
329
+ 184.6981658935547
330
+ ],
331
+ "size": [
332
+ 390,
333
+ 82
334
+ ],
335
+ "flags": {},
336
+ "order": 2,
337
+ "mode": 0,
338
+ "inputs": [],
339
+ "outputs": [
340
+ {
341
+ "name": "CLIP",
342
+ "type": "CLIP",
343
+ "links": [
344
+ 74,
345
+ 75
346
+ ],
347
+ "slot_index": 0
348
+ }
349
+ ],
350
+ "properties": {
351
+ "Node name for S&R": "CLIPLoader"
352
+ },
353
+ "widgets_values": [
354
+ "umt5_xxl_fp8_e4m3fn_scaled.safetensors",
355
+ "wan",
356
+ "default"
357
+ ]
358
+ },
359
+ {
360
+ "id": 37,
361
+ "type": "UNETLoader",
362
+ "pos": [
363
+ 485.1220397949219,
364
+ 57.094566345214844
365
+ ],
366
+ "size": [
367
+ 346.7470703125,
368
+ 82
369
+ ],
370
+ "flags": {},
371
+ "order": 3,
372
+ "mode": 0,
373
+ "inputs": [],
374
+ "outputs": [
375
+ {
376
+ "name": "MODEL",
377
+ "type": "MODEL",
378
+ "links": [
379
+ 92
380
+ ],
381
+ "slot_index": 0
382
+ }
383
+ ],
384
+ "properties": {
385
+ "Node name for S&R": "UNETLoader"
386
+ },
387
+ "widgets_values": [
388
+ "wan2.1_i2v_720p_14B_bf16.safetensors",
389
+ "default"
390
+ ]
391
+ },
392
+ {
393
+ "id": 47,
394
+ "type": "SaveWEBM",
395
+ "pos": [
396
+ 2371.950927734375,
397
+ 213.44703674316406
398
+ ],
399
+ "size": [
400
+ 315,
401
+ 130
402
+ ],
403
+ "flags": {},
404
+ "order": 12,
405
+ "mode": 4,
406
+ "inputs": [
407
+ {
408
+ "name": "images",
409
+ "type": "IMAGE",
410
+ "link": 93
411
+ }
412
+ ],
413
+ "outputs": [],
414
+ "properties": {
415
+ "Node name for S&R": "SaveWEBM"
416
+ },
417
+ "widgets_values": [
418
+ "ComfyUI",
419
+ "vp9",
420
+ 24,
421
+ 32
422
+ ]
423
+ },
424
+ {
425
+ "id": 7,
426
+ "type": "CLIPTextEncode",
427
+ "pos": [
428
+ 413,
429
+ 389
430
+ ],
431
+ "size": [
432
+ 425.27801513671875,
433
+ 180.6060791015625
434
+ ],
435
+ "flags": {},
436
+ "order": 6,
437
+ "mode": 0,
438
+ "inputs": [
439
+ {
440
+ "name": "clip",
441
+ "type": "CLIP",
442
+ "link": 75
443
+ }
444
+ ],
445
+ "outputs": [
446
+ {
447
+ "name": "CONDITIONING",
448
+ "type": "CONDITIONING",
449
+ "links": [
450
+ 98
451
+ ],
452
+ "slot_index": 0
453
+ }
454
+ ],
455
+ "title": "CLIP Text Encode (Negative Prompt)",
456
+ "properties": {
457
+ "Node name for S&R": "CLIPTextEncode"
458
+ },
459
+ "widgets_values": [
460
+ "Overexposure, static, blurred details, subtitles, paintings, pictures, still, overall gray, worst quality, low quality, JPEG compression residue, ugly, mutilated, redundant fingers, poorly painted hands, poorly painted faces, deformed, disfigured, deformed limbs, fused fingers, cluttered background, three legs, a lot of people in the background, upside down"
461
+ ],
462
+ "color": "#322",
463
+ "bgcolor": "#533"
464
+ },
465
+ {
466
+ "id": 52,
467
+ "type": "LoadImage",
468
+ "pos": [
469
+ -30,
470
+ 760
471
+ ],
472
+ "size": [
473
+ 315,
474
+ 314
475
+ ],
476
+ "flags": {},
477
+ "order": 4,
478
+ "mode": 0,
479
+ "inputs": [],
480
+ "outputs": [
481
+ {
482
+ "name": "IMAGE",
483
+ "type": "IMAGE",
484
+ "links": [
485
+ 106,
486
+ 109
487
+ ],
488
+ "slot_index": 0
489
+ },
490
+ {
491
+ "name": "MASK",
492
+ "type": "MASK",
493
+ "links": null,
494
+ "slot_index": 1
495
+ }
496
+ ],
497
+ "properties": {
498
+ "Node name for S&R": "LoadImage"
499
+ },
500
+ "widgets_values": [
501
+ "0_2 (3).png",
502
+ "image"
503
+ ]
504
+ },
505
+ {
506
+ "id": 6,
507
+ "type": "CLIPTextEncode",
508
+ "pos": [
509
+ 415,
510
+ 186
511
+ ],
512
+ "size": [
513
+ 422.84503173828125,
514
+ 164.31304931640625
515
+ ],
516
+ "flags": {},
517
+ "order": 5,
518
+ "mode": 0,
519
+ "inputs": [
520
+ {
521
+ "name": "clip",
522
+ "type": "CLIP",
523
+ "link": 74
524
+ }
525
+ ],
526
+ "outputs": [
527
+ {
528
+ "name": "CONDITIONING",
529
+ "type": "CONDITIONING",
530
+ "links": [
531
+ 97
532
+ ],
533
+ "slot_index": 0
534
+ }
535
+ ],
536
+ "title": "CLIP Text Encode (Positive Prompt)",
537
+ "properties": {
538
+ "Node name for S&R": "CLIPTextEncode"
539
+ },
540
+ "widgets_values": [
541
+ "A magician made flying cards in his hand, a lot of colored light and colored paper and confetti"
542
+ ],
543
+ "color": "#232",
544
+ "bgcolor": "#353"
545
+ }
546
+ ],
547
+ "links": [
548
+ [
549
+ 35,
550
+ 3,
551
+ 0,
552
+ 8,
553
+ 0,
554
+ "LATENT"
555
+ ],
556
+ [
557
+ 56,
558
+ 8,
559
+ 0,
560
+ 28,
561
+ 0,
562
+ "IMAGE"
563
+ ],
564
+ [
565
+ 74,
566
+ 38,
567
+ 0,
568
+ 6,
569
+ 0,
570
+ "CLIP"
571
+ ],
572
+ [
573
+ 75,
574
+ 38,
575
+ 0,
576
+ 7,
577
+ 0,
578
+ "CLIP"
579
+ ],
580
+ [
581
+ 76,
582
+ 39,
583
+ 0,
584
+ 8,
585
+ 1,
586
+ "VAE"
587
+ ],
588
+ [
589
+ 92,
590
+ 37,
591
+ 0,
592
+ 3,
593
+ 0,
594
+ "MODEL"
595
+ ],
596
+ [
597
+ 93,
598
+ 8,
599
+ 0,
600
+ 47,
601
+ 0,
602
+ "IMAGE"
603
+ ],
604
+ [
605
+ 94,
606
+ 49,
607
+ 0,
608
+ 51,
609
+ 0,
610
+ "CLIP_VISION"
611
+ ],
612
+ [
613
+ 97,
614
+ 6,
615
+ 0,
616
+ 50,
617
+ 0,
618
+ "CONDITIONING"
619
+ ],
620
+ [
621
+ 98,
622
+ 7,
623
+ 0,
624
+ 50,
625
+ 1,
626
+ "CONDITIONING"
627
+ ],
628
+ [
629
+ 99,
630
+ 39,
631
+ 0,
632
+ 50,
633
+ 2,
634
+ "VAE"
635
+ ],
636
+ [
637
+ 101,
638
+ 50,
639
+ 0,
640
+ 3,
641
+ 1,
642
+ "CONDITIONING"
643
+ ],
644
+ [
645
+ 102,
646
+ 50,
647
+ 1,
648
+ 3,
649
+ 2,
650
+ "CONDITIONING"
651
+ ],
652
+ [
653
+ 103,
654
+ 50,
655
+ 2,
656
+ 3,
657
+ 3,
658
+ "LATENT"
659
+ ],
660
+ [
661
+ 106,
662
+ 52,
663
+ 0,
664
+ 50,
665
+ 4,
666
+ "IMAGE"
667
+ ],
668
+ [
669
+ 107,
670
+ 51,
671
+ 0,
672
+ 50,
673
+ 3,
674
+ "CLIP_VISION_OUTPUT"
675
+ ],
676
+ [
677
+ 109,
678
+ 52,
679
+ 0,
680
+ 51,
681
+ 1,
682
+ "IMAGE"
683
+ ]
684
+ ],
685
+ "groups": [],
686
+ "config": {},
687
+ "extra": {
688
+ "ds": {
689
+ "scale": 1.01525597994775,
690
+ "offset": [
691
+ 509.5627989671756,
692
+ 312.04429170999896
693
+ ]
694
+ }
695
+ },
696
+ "version": 0.4
697
+ }
example workflows_Wan2.1/text_to_video_wan.json ADDED
@@ -0,0 +1,481 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "last_node_id": 47,
3
+ "last_link_id": 93,
4
+ "nodes": [
5
+ {
6
+ "id": 8,
7
+ "type": "VAEDecode",
8
+ "pos": [
9
+ 1210,
10
+ 190
11
+ ],
12
+ "size": [
13
+ 210,
14
+ 46
15
+ ],
16
+ "flags": {},
17
+ "order": 7,
18
+ "mode": 0,
19
+ "inputs": [
20
+ {
21
+ "name": "samples",
22
+ "type": "LATENT",
23
+ "link": 35
24
+ },
25
+ {
26
+ "name": "vae",
27
+ "type": "VAE",
28
+ "link": 76
29
+ }
30
+ ],
31
+ "outputs": [
32
+ {
33
+ "name": "IMAGE",
34
+ "type": "IMAGE",
35
+ "links": [
36
+ 56,
37
+ 93
38
+ ],
39
+ "slot_index": 0
40
+ }
41
+ ],
42
+ "properties": {
43
+ "Node name for S&R": "VAEDecode"
44
+ },
45
+ "widgets_values": []
46
+ },
47
+ {
48
+ "id": 39,
49
+ "type": "VAELoader",
50
+ "pos": [
51
+ 866.3932495117188,
52
+ 499.18597412109375
53
+ ],
54
+ "size": [
55
+ 306.36004638671875,
56
+ 58
57
+ ],
58
+ "flags": {},
59
+ "order": 0,
60
+ "mode": 0,
61
+ "inputs": [],
62
+ "outputs": [
63
+ {
64
+ "name": "VAE",
65
+ "type": "VAE",
66
+ "links": [
67
+ 76
68
+ ],
69
+ "slot_index": 0
70
+ }
71
+ ],
72
+ "properties": {
73
+ "Node name for S&R": "VAELoader"
74
+ },
75
+ "widgets_values": [
76
+ "wan_2.1_vae.safetensors"
77
+ ]
78
+ },
79
+ {
80
+ "id": 28,
81
+ "type": "SaveAnimatedWEBP",
82
+ "pos": [
83
+ 1460,
84
+ 190
85
+ ],
86
+ "size": [
87
+ 870.8511352539062,
88
+ 643.7430419921875
89
+ ],
90
+ "flags": {},
91
+ "order": 8,
92
+ "mode": 0,
93
+ "inputs": [
94
+ {
95
+ "name": "images",
96
+ "type": "IMAGE",
97
+ "link": 56
98
+ }
99
+ ],
100
+ "outputs": [],
101
+ "properties": {},
102
+ "widgets_values": [
103
+ "ComfyUI",
104
+ 16,
105
+ false,
106
+ 90,
107
+ "default"
108
+ ]
109
+ },
110
+ {
111
+ "id": 6,
112
+ "type": "CLIPTextEncode",
113
+ "pos": [
114
+ 415,
115
+ 186
116
+ ],
117
+ "size": [
118
+ 422.84503173828125,
119
+ 164.31304931640625
120
+ ],
121
+ "flags": {},
122
+ "order": 4,
123
+ "mode": 0,
124
+ "inputs": [
125
+ {
126
+ "name": "clip",
127
+ "type": "CLIP",
128
+ "link": 74
129
+ }
130
+ ],
131
+ "outputs": [
132
+ {
133
+ "name": "CONDITIONING",
134
+ "type": "CONDITIONING",
135
+ "links": [
136
+ 46
137
+ ],
138
+ "slot_index": 0
139
+ }
140
+ ],
141
+ "title": "CLIP Text Encode (Positive Prompt)",
142
+ "properties": {
143
+ "Node name for S&R": "CLIPTextEncode"
144
+ },
145
+ "widgets_values": [
146
+ "a fox moving quickly in a beautiful winter scenery nature trees sunset tracking camera"
147
+ ],
148
+ "color": "#232",
149
+ "bgcolor": "#353"
150
+ },
151
+ {
152
+ "id": 38,
153
+ "type": "CLIPLoader",
154
+ "pos": [
155
+ 12.94982624053955,
156
+ 184.6981658935547
157
+ ],
158
+ "size": [
159
+ 390,
160
+ 82
161
+ ],
162
+ "flags": {},
163
+ "order": 1,
164
+ "mode": 0,
165
+ "inputs": [],
166
+ "outputs": [
167
+ {
168
+ "name": "CLIP",
169
+ "type": "CLIP",
170
+ "links": [
171
+ 74,
172
+ 75
173
+ ],
174
+ "slot_index": 0
175
+ }
176
+ ],
177
+ "properties": {
178
+ "Node name for S&R": "CLIPLoader"
179
+ },
180
+ "widgets_values": [
181
+ "umt5_xxl_fp8_e4m3fn_scaled.safetensors",
182
+ "wan",
183
+ "default"
184
+ ]
185
+ },
186
+ {
187
+ "id": 40,
188
+ "type": "EmptyHunyuanLatentVideo",
189
+ "pos": [
190
+ 520,
191
+ 620
192
+ ],
193
+ "size": [
194
+ 315,
195
+ 130
196
+ ],
197
+ "flags": {},
198
+ "order": 2,
199
+ "mode": 0,
200
+ "inputs": [],
201
+ "outputs": [
202
+ {
203
+ "name": "LATENT",
204
+ "type": "LATENT",
205
+ "links": [
206
+ 91
207
+ ],
208
+ "slot_index": 0
209
+ }
210
+ ],
211
+ "properties": {
212
+ "Node name for S&R": "EmptyHunyuanLatentVideo"
213
+ },
214
+ "widgets_values": [
215
+ 832,
216
+ 480,
217
+ 33,
218
+ 1
219
+ ]
220
+ },
221
+ {
222
+ "id": 37,
223
+ "type": "UNETLoader",
224
+ "pos": [
225
+ 485.1220397949219,
226
+ 57.094566345214844
227
+ ],
228
+ "size": [
229
+ 346.7470703125,
230
+ 82
231
+ ],
232
+ "flags": {},
233
+ "order": 3,
234
+ "mode": 0,
235
+ "inputs": [],
236
+ "outputs": [
237
+ {
238
+ "name": "MODEL",
239
+ "type": "MODEL",
240
+ "links": [
241
+ 92
242
+ ],
243
+ "slot_index": 0
244
+ }
245
+ ],
246
+ "properties": {
247
+ "Node name for S&R": "UNETLoader"
248
+ },
249
+ "widgets_values": [
250
+ "wan2.1_t2v_1.3B_bf16.safetensors",
251
+ "default"
252
+ ]
253
+ },
254
+ {
255
+ "id": 47,
256
+ "type": "SaveWEBM",
257
+ "pos": [
258
+ 2367.213134765625,
259
+ 193.6114959716797
260
+ ],
261
+ "size": [
262
+ 315,
263
+ 130
264
+ ],
265
+ "flags": {},
266
+ "order": 9,
267
+ "mode": 4,
268
+ "inputs": [
269
+ {
270
+ "name": "images",
271
+ "type": "IMAGE",
272
+ "link": 93
273
+ }
274
+ ],
275
+ "outputs": [],
276
+ "properties": {
277
+ "Node name for S&R": "SaveWEBM"
278
+ },
279
+ "widgets_values": [
280
+ "ComfyUI",
281
+ "vp9",
282
+ 24,
283
+ 32
284
+ ]
285
+ },
286
+ {
287
+ "id": 3,
288
+ "type": "KSampler",
289
+ "pos": [
290
+ 863,
291
+ 187
292
+ ],
293
+ "size": [
294
+ 315,
295
+ 262
296
+ ],
297
+ "flags": {},
298
+ "order": 6,
299
+ "mode": 0,
300
+ "inputs": [
301
+ {
302
+ "name": "model",
303
+ "type": "MODEL",
304
+ "link": 92
305
+ },
306
+ {
307
+ "name": "positive",
308
+ "type": "CONDITIONING",
309
+ "link": 46
310
+ },
311
+ {
312
+ "name": "negative",
313
+ "type": "CONDITIONING",
314
+ "link": 52
315
+ },
316
+ {
317
+ "name": "latent_image",
318
+ "type": "LATENT",
319
+ "link": 91
320
+ }
321
+ ],
322
+ "outputs": [
323
+ {
324
+ "name": "LATENT",
325
+ "type": "LATENT",
326
+ "links": [
327
+ 35
328
+ ],
329
+ "slot_index": 0
330
+ }
331
+ ],
332
+ "properties": {
333
+ "Node name for S&R": "KSampler"
334
+ },
335
+ "widgets_values": [
336
+ 878361741413693,
337
+ "randomize",
338
+ 30,
339
+ 6,
340
+ "uni_pc",
341
+ "simple",
342
+ 1
343
+ ]
344
+ },
345
+ {
346
+ "id": 7,
347
+ "type": "CLIPTextEncode",
348
+ "pos": [
349
+ 413,
350
+ 389
351
+ ],
352
+ "size": [
353
+ 425.27801513671875,
354
+ 180.6060791015625
355
+ ],
356
+ "flags": {},
357
+ "order": 5,
358
+ "mode": 0,
359
+ "inputs": [
360
+ {
361
+ "name": "clip",
362
+ "type": "CLIP",
363
+ "link": 75
364
+ }
365
+ ],
366
+ "outputs": [
367
+ {
368
+ "name": "CONDITIONING",
369
+ "type": "CONDITIONING",
370
+ "links": [
371
+ 52
372
+ ],
373
+ "slot_index": 0
374
+ }
375
+ ],
376
+ "title": "CLIP Text Encode (Negative Prompt)",
377
+ "properties": {
378
+ "Node name for S&R": "CLIPTextEncode"
379
+ },
380
+ "widgets_values": [
381
+ "Overexposure, static, blurred details, subtitles, paintings, pictures, still, overall gray, worst quality, low quality, JPEG compression residue, ugly, mutilated, redundant fingers, poorly painted hands, poorly painted faces, deformed, disfigured, deformed limbs, fused fingers, cluttered background, three legs, a lot of people in the background, upside down"
382
+ ],
383
+ "color": "#322",
384
+ "bgcolor": "#533"
385
+ }
386
+ ],
387
+ "links": [
388
+ [
389
+ 35,
390
+ 3,
391
+ 0,
392
+ 8,
393
+ 0,
394
+ "LATENT"
395
+ ],
396
+ [
397
+ 46,
398
+ 6,
399
+ 0,
400
+ 3,
401
+ 1,
402
+ "CONDITIONING"
403
+ ],
404
+ [
405
+ 52,
406
+ 7,
407
+ 0,
408
+ 3,
409
+ 2,
410
+ "CONDITIONING"
411
+ ],
412
+ [
413
+ 56,
414
+ 8,
415
+ 0,
416
+ 28,
417
+ 0,
418
+ "IMAGE"
419
+ ],
420
+ [
421
+ 74,
422
+ 38,
423
+ 0,
424
+ 6,
425
+ 0,
426
+ "CLIP"
427
+ ],
428
+ [
429
+ 75,
430
+ 38,
431
+ 0,
432
+ 7,
433
+ 0,
434
+ "CLIP"
435
+ ],
436
+ [
437
+ 76,
438
+ 39,
439
+ 0,
440
+ 8,
441
+ 1,
442
+ "VAE"
443
+ ],
444
+ [
445
+ 91,
446
+ 40,
447
+ 0,
448
+ 3,
449
+ 3,
450
+ "LATENT"
451
+ ],
452
+ [
453
+ 92,
454
+ 37,
455
+ 0,
456
+ 3,
457
+ 0,
458
+ "MODEL"
459
+ ],
460
+ [
461
+ 93,
462
+ 8,
463
+ 0,
464
+ 47,
465
+ 0,
466
+ "IMAGE"
467
+ ]
468
+ ],
469
+ "groups": [],
470
+ "config": {},
471
+ "extra": {
472
+ "ds": {
473
+ "scale": 1.2284597357367726,
474
+ "offset": [
475
+ 445.486325470362,
476
+ 133.8941997640868
477
+ ]
478
+ }
479
+ },
480
+ "version": 0.4
481
+ }
split_files/clip_vision/clip_vision_h.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d90f32c6262505ca9b84adb4d3203ed4b2d3dd03410dc2799ea418ab08c19096
3
+ size 135
split_files/diffusion_models/wan2.1_flf2v_720p_14B_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8b0e025b4e5a81f1f281f12897d1f51024758a750ea602f2e8eb79adda4dfb7
3
+ size 136
split_files/diffusion_models/wan2.1_flf2v_720p_14B_fp8_e4m3fn.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4ebed4b6457ce895e970cc0e35f05240b1d9c5dfa11b6806f40b025a4c8144e
3
+ size 136
split_files/diffusion_models/wan2.1_fun_camera_v1.1_1.3B_bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc26801402119cc93096724871280ce273c523360ad09229116781cc4ad60419
3
+ size 135
split_files/diffusion_models/wan2.1_fun_camera_v1.1_14B_bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:412b9f12473c082ac576ccf08b8211fb09088b870c415ad07500b4bb9da9b641
3
+ size 136
split_files/diffusion_models/wan2.1_fun_control_1.3B_bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3216ddcb688a9c211cf00eefb3ded38070082bc0c91cb5da446a4c53a5db0502
3
+ size 135
split_files/diffusion_models/wan2.1_fun_inp_1.3B_bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bcad86a91c003cdcb091a0f4e0beb3c8038815e08faa0a13f175f300ae67112
3
+ size 135
split_files/diffusion_models/wan2.1_i2v_480p_14B_bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21f9e9515c456b247376256787d0b01c658852b687678f5c0997a8e1e7a99091
3
+ size 136
split_files/diffusion_models/wan2.1_i2v_480p_14B_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aedd939256a93efa3df4731e2222d8d95dea95ec772c3f5f97a886f458cb48ab
3
+ size 136
split_files/diffusion_models/wan2.1_i2v_480p_14B_fp8_e4m3fn.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83f880ff2454d4f2882a3d2e7f30dbe923c6c14d297a5e23a9771e52990932ea
3
+ size 136
split_files/diffusion_models/wan2.1_i2v_480p_14B_fp8_scaled.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33352964f5e63268affa05d356bfcb7373715903572293bee469cbbe0c178b81
3
+ size 136
split_files/diffusion_models/wan2.1_i2v_720p_14B_bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4acfa97eb8db6d20d8851e62852a87e6643a4177304c5d391622a7022dddefb
3
+ size 136
split_files/diffusion_models/wan2.1_i2v_720p_14B_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b6dac9086326e76a686f898f7fcf5b86a54fdac125d72bd8800658bf9686f5a
3
+ size 136
split_files/diffusion_models/wan2.1_i2v_720p_14B_fp8_e4m3fn.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:faa6309c23fbe3ba4a2a61eac0f28605bbee347254639a85a7fd9d15987e9cea
3
+ size 136
split_files/diffusion_models/wan2.1_i2v_720p_14B_fp8_scaled.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2645f174e201ec9054ccd0abcb0900553e842f0decb8f68f924378be5d2c12f
3
+ size 136
split_files/diffusion_models/wan2.1_magref_14B_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afa20e86ab2d88cd5978fc7e46fdb8319d396d4b26156c48a10a7ef3d147d997
3
+ size 136
split_files/diffusion_models/wan2.1_t2v_1.3B_bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bbac68ea651a3b5ad53a74bc4b47721a611440f1a3761419c37644824d8f97e
3
+ size 135
split_files/diffusion_models/wan2.1_t2v_1.3B_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e56127a429bfab2a0572254795ab425708c32d3eb61ae68c205b72c213dc8b96
3
+ size 135
split_files/diffusion_models/wan2.1_t2v_14B_bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bd2cad024bd3e8cb311f7c5c62cd3fa21cb56abbc8ea0f965eb3955ac5f0523
3
+ size 136
split_files/diffusion_models/wan2.1_t2v_14B_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46c27d3693bf2475990a912e08bf67fc6e6cd5396eab87b5e8dd1fcd3651364a
3
+ size 28577761176
split_files/diffusion_models/wan2.1_t2v_14B_fp8_e4m3fn.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9dfb81c259e1a689b7f1d935e1572387fbdbcb842307f41b110ba19850cab59
3
+ size 136
split_files/diffusion_models/wan2.1_t2v_14B_fp8_scaled.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8578b7b8f724e69f3c617bbf6c4a323e50839419d19c13900bca1fae2ad52839
3
+ size 136
split_files/diffusion_models/wan2.1_vace_1.3B_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6c57f00bce46bb58c1882edde0b1e9fb538fe77cae765b8751f990d5ac36047
3
+ size 135
split_files/diffusion_models/wan2.1_vace_1.3B_preview_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38c603383b42b7ec48abfa0cdd48d82bef3d47a92394dd870333d519932c8fac
3
+ size 135
split_files/diffusion_models/wan2.1_vace_14B_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e685ac6206c23fc60e66df7ad9340bfde4f778043d633a60d3a4eb7d98f2cda8
3
+ size 136
split_files/loras/wan_alpha_2.1_rgba_lora.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e79dc18f39fcaa9bc944e8c49e4d2112bdae37121d9926ba1e9c0264a60bf436
3
+ size 134
split_files/model_patches/wan2.1_infiniteTalk_multi_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b0e3f83164ec9a17246b94e4941f1ece03cbe4599b553990f1f4aaf1a40f65f
3
+ size 135
split_files/model_patches/wan2.1_infiniteTalk_single_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:030a7f0c1dd45346bb777ac64eb6e040e8ccae6068fc6b80b2dca77b08743bbf
3
+ size 135
split_files/text_encoders/umt5_xxl_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63ab3e79727e4f6fb39fbe543e936f272546ad1091b92fdf9b7606a893149342
3
+ size 136
split_files/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f8dbfdaf11b6c1bf2f63458a771dd59fd64ada6e0765f14de6fa851af050324
3
+ size 135
split_files/vae/wan_2.1_vae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff210a5e5b1b7b5ea0391cc56b8de1c83a982699da322a9986dd08f44eca5aeb
3
+ size 134
split_files/vae/wan_alpha_2.1_vae_alpha_channel.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00173cc77311eb1366cc614124a20a0e27eca87cefa717a11f74216abc9005fa
3
+ size 134
split_files/vae/wan_alpha_2.1_vae_rgb_channel.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1726019e155a918e728b199fbda0a7b824bdc260075f20c504c50861da640b1f
3
+ size 134