RedbeardNZ comfyanonymous commited on
Commit
ca8bf47
·
verified ·
0 Parent(s):

Duplicate from Comfy-Org/Wan_2.1_ComfyUI_repackaged

Browse files

Co-authored-by: comfy <comfyanonymous@users.noreply.huggingface.co>

Files changed (32) hide show
  1. .gitattributes +35 -0
  2. README.md +1 -0
  3. example workflows_Wan2.1/image_to_video_wan_480p_example.json +697 -0
  4. example workflows_Wan2.1/image_to_video_wan_720p_example.json +697 -0
  5. example workflows_Wan2.1/text_to_video_wan.json +481 -0
  6. split_files/clip_vision/clip_vision_h.safetensors +3 -0
  7. split_files/diffusion_models/wan2.1_flf2v_720p_14B_fp16.safetensors +3 -0
  8. split_files/diffusion_models/wan2.1_flf2v_720p_14B_fp8_e4m3fn.safetensors +3 -0
  9. split_files/diffusion_models/wan2.1_fun_camera_v1.1_1.3B_bf16.safetensors +3 -0
  10. split_files/diffusion_models/wan2.1_fun_camera_v1.1_14B_bf16.safetensors +3 -0
  11. split_files/diffusion_models/wan2.1_fun_control_1.3B_bf16.safetensors +3 -0
  12. split_files/diffusion_models/wan2.1_fun_inp_1.3B_bf16.safetensors +3 -0
  13. split_files/diffusion_models/wan2.1_i2v_480p_14B_bf16.safetensors +3 -0
  14. split_files/diffusion_models/wan2.1_i2v_480p_14B_fp16.safetensors +3 -0
  15. split_files/diffusion_models/wan2.1_i2v_480p_14B_fp8_e4m3fn.safetensors +3 -0
  16. split_files/diffusion_models/wan2.1_i2v_480p_14B_fp8_scaled.safetensors +3 -0
  17. split_files/diffusion_models/wan2.1_i2v_720p_14B_bf16.safetensors +3 -0
  18. split_files/diffusion_models/wan2.1_i2v_720p_14B_fp16.safetensors +3 -0
  19. split_files/diffusion_models/wan2.1_i2v_720p_14B_fp8_e4m3fn.safetensors +3 -0
  20. split_files/diffusion_models/wan2.1_i2v_720p_14B_fp8_scaled.safetensors +3 -0
  21. split_files/diffusion_models/wan2.1_t2v_1.3B_bf16.safetensors +3 -0
  22. split_files/diffusion_models/wan2.1_t2v_1.3B_fp16.safetensors +3 -0
  23. split_files/diffusion_models/wan2.1_t2v_14B_bf16.safetensors +3 -0
  24. split_files/diffusion_models/wan2.1_t2v_14B_fp16.safetensors +3 -0
  25. split_files/diffusion_models/wan2.1_t2v_14B_fp8_e4m3fn.safetensors +3 -0
  26. split_files/diffusion_models/wan2.1_t2v_14B_fp8_scaled.safetensors +3 -0
  27. split_files/diffusion_models/wan2.1_vace_1.3B_fp16.safetensors +3 -0
  28. split_files/diffusion_models/wan2.1_vace_1.3B_preview_fp16.safetensors +3 -0
  29. split_files/diffusion_models/wan2.1_vace_14B_fp16.safetensors +3 -0
  30. split_files/text_encoders/umt5_xxl_fp16.safetensors +3 -0
  31. split_files/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensors +3 -0
  32. split_files/vae/wan_2.1_vae.safetensors +3 -0
.gitattributes ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1 @@
 
 
1
+ Wan 2.1 repackaged for ComfyUI use. For examples see: https://comfyanonymous.github.io/ComfyUI_examples/wan
example workflows_Wan2.1/image_to_video_wan_480p_example.json ADDED
@@ -0,0 +1,697 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "last_node_id": 53,
3
+ "last_link_id": 109,
4
+ "nodes": [
5
+ {
6
+ "id": 8,
7
+ "type": "VAEDecode",
8
+ "pos": [
9
+ 1210,
10
+ 190
11
+ ],
12
+ "size": [
13
+ 210,
14
+ 46
15
+ ],
16
+ "flags": {},
17
+ "order": 10,
18
+ "mode": 0,
19
+ "inputs": [
20
+ {
21
+ "name": "samples",
22
+ "type": "LATENT",
23
+ "link": 35
24
+ },
25
+ {
26
+ "name": "vae",
27
+ "type": "VAE",
28
+ "link": 76
29
+ }
30
+ ],
31
+ "outputs": [
32
+ {
33
+ "name": "IMAGE",
34
+ "type": "IMAGE",
35
+ "links": [
36
+ 56,
37
+ 93
38
+ ],
39
+ "slot_index": 0
40
+ }
41
+ ],
42
+ "properties": {
43
+ "Node name for S&R": "VAEDecode"
44
+ },
45
+ "widgets_values": []
46
+ },
47
+ {
48
+ "id": 39,
49
+ "type": "VAELoader",
50
+ "pos": [
51
+ 866.3932495117188,
52
+ 499.18597412109375
53
+ ],
54
+ "size": [
55
+ 306.36004638671875,
56
+ 58
57
+ ],
58
+ "flags": {},
59
+ "order": 0,
60
+ "mode": 0,
61
+ "inputs": [],
62
+ "outputs": [
63
+ {
64
+ "name": "VAE",
65
+ "type": "VAE",
66
+ "links": [
67
+ 76,
68
+ 99
69
+ ],
70
+ "slot_index": 0
71
+ }
72
+ ],
73
+ "properties": {
74
+ "Node name for S&R": "VAELoader"
75
+ },
76
+ "widgets_values": [
77
+ "wan_2.1_vae.safetensors"
78
+ ]
79
+ },
80
+ {
81
+ "id": 28,
82
+ "type": "SaveAnimatedWEBP",
83
+ "pos": [
84
+ 1460,
85
+ 190
86
+ ],
87
+ "size": [
88
+ 870.8511352539062,
89
+ 643.7430419921875
90
+ ],
91
+ "flags": {},
92
+ "order": 11,
93
+ "mode": 0,
94
+ "inputs": [
95
+ {
96
+ "name": "images",
97
+ "type": "IMAGE",
98
+ "link": 56
99
+ }
100
+ ],
101
+ "outputs": [],
102
+ "properties": {},
103
+ "widgets_values": [
104
+ "ComfyUI",
105
+ 16,
106
+ false,
107
+ 90,
108
+ "default"
109
+ ]
110
+ },
111
+ {
112
+ "id": 38,
113
+ "type": "CLIPLoader",
114
+ "pos": [
115
+ 12.94982624053955,
116
+ 184.6981658935547
117
+ ],
118
+ "size": [
119
+ 390,
120
+ 82
121
+ ],
122
+ "flags": {},
123
+ "order": 1,
124
+ "mode": 0,
125
+ "inputs": [],
126
+ "outputs": [
127
+ {
128
+ "name": "CLIP",
129
+ "type": "CLIP",
130
+ "links": [
131
+ 74,
132
+ 75
133
+ ],
134
+ "slot_index": 0
135
+ }
136
+ ],
137
+ "properties": {
138
+ "Node name for S&R": "CLIPLoader"
139
+ },
140
+ "widgets_values": [
141
+ "umt5_xxl_fp8_e4m3fn_scaled.safetensors",
142
+ "wan",
143
+ "default"
144
+ ]
145
+ },
146
+ {
147
+ "id": 47,
148
+ "type": "SaveWEBM",
149
+ "pos": [
150
+ 2367.213134765625,
151
+ 193.6114959716797
152
+ ],
153
+ "size": [
154
+ 315,
155
+ 130
156
+ ],
157
+ "flags": {},
158
+ "order": 12,
159
+ "mode": 4,
160
+ "inputs": [
161
+ {
162
+ "name": "images",
163
+ "type": "IMAGE",
164
+ "link": 93
165
+ }
166
+ ],
167
+ "outputs": [],
168
+ "properties": {
169
+ "Node name for S&R": "SaveWEBM"
170
+ },
171
+ "widgets_values": [
172
+ "ComfyUI",
173
+ "vp9",
174
+ 24,
175
+ 32
176
+ ]
177
+ },
178
+ {
179
+ "id": 37,
180
+ "type": "UNETLoader",
181
+ "pos": [
182
+ 485.1220397949219,
183
+ 57.094566345214844
184
+ ],
185
+ "size": [
186
+ 346.7470703125,
187
+ 82
188
+ ],
189
+ "flags": {},
190
+ "order": 2,
191
+ "mode": 0,
192
+ "inputs": [],
193
+ "outputs": [
194
+ {
195
+ "name": "MODEL",
196
+ "type": "MODEL",
197
+ "links": [
198
+ 92
199
+ ],
200
+ "slot_index": 0
201
+ }
202
+ ],
203
+ "properties": {
204
+ "Node name for S&R": "UNETLoader"
205
+ },
206
+ "widgets_values": [
207
+ "wan2.1_i2v_480p_14B_bf16.safetensors",
208
+ "default"
209
+ ]
210
+ },
211
+ {
212
+ "id": 51,
213
+ "type": "CLIPVisionEncode",
214
+ "pos": [
215
+ 340,
216
+ 650
217
+ ],
218
+ "size": [
219
+ 253.60000610351562,
220
+ 78
221
+ ],
222
+ "flags": {},
223
+ "order": 7,
224
+ "mode": 0,
225
+ "inputs": [
226
+ {
227
+ "name": "clip_vision",
228
+ "type": "CLIP_VISION",
229
+ "link": 94
230
+ },
231
+ {
232
+ "name": "image",
233
+ "type": "IMAGE",
234
+ "link": 109
235
+ }
236
+ ],
237
+ "outputs": [
238
+ {
239
+ "name": "CLIP_VISION_OUTPUT",
240
+ "type": "CLIP_VISION_OUTPUT",
241
+ "links": [
242
+ 107
243
+ ],
244
+ "slot_index": 0
245
+ }
246
+ ],
247
+ "properties": {
248
+ "Node name for S&R": "CLIPVisionEncode"
249
+ },
250
+ "widgets_values": [
251
+ "none"
252
+ ]
253
+ },
254
+ {
255
+ "id": 49,
256
+ "type": "CLIPVisionLoader",
257
+ "pos": [
258
+ -20,
259
+ 650
260
+ ],
261
+ "size": [
262
+ 315,
263
+ 58
264
+ ],
265
+ "flags": {},
266
+ "order": 3,
267
+ "mode": 0,
268
+ "inputs": [],
269
+ "outputs": [
270
+ {
271
+ "name": "CLIP_VISION",
272
+ "type": "CLIP_VISION",
273
+ "links": [
274
+ 94
275
+ ],
276
+ "slot_index": 0
277
+ }
278
+ ],
279
+ "properties": {
280
+ "Node name for S&R": "CLIPVisionLoader"
281
+ },
282
+ "widgets_values": [
283
+ "clip_vision_h.safetensors"
284
+ ]
285
+ },
286
+ {
287
+ "id": 52,
288
+ "type": "LoadImage",
289
+ "pos": [
290
+ -30,
291
+ 760
292
+ ],
293
+ "size": [
294
+ 315,
295
+ 314
296
+ ],
297
+ "flags": {},
298
+ "order": 4,
299
+ "mode": 0,
300
+ "inputs": [],
301
+ "outputs": [
302
+ {
303
+ "name": "IMAGE",
304
+ "type": "IMAGE",
305
+ "links": [
306
+ 106,
307
+ 109
308
+ ],
309
+ "slot_index": 0
310
+ },
311
+ {
312
+ "name": "MASK",
313
+ "type": "MASK",
314
+ "links": null,
315
+ "slot_index": 1
316
+ }
317
+ ],
318
+ "properties": {
319
+ "Node name for S&R": "LoadImage"
320
+ },
321
+ "widgets_values": [
322
+ "flux_dev_example.png",
323
+ "image"
324
+ ]
325
+ },
326
+ {
327
+ "id": 50,
328
+ "type": "WanImageToVideo",
329
+ "pos": [
330
+ 673.0507202148438,
331
+ 627.272705078125
332
+ ],
333
+ "size": [
334
+ 342.5999755859375,
335
+ 210
336
+ ],
337
+ "flags": {},
338
+ "order": 8,
339
+ "mode": 0,
340
+ "inputs": [
341
+ {
342
+ "name": "positive",
343
+ "type": "CONDITIONING",
344
+ "link": 97
345
+ },
346
+ {
347
+ "name": "negative",
348
+ "type": "CONDITIONING",
349
+ "link": 98
350
+ },
351
+ {
352
+ "name": "vae",
353
+ "type": "VAE",
354
+ "link": 99
355
+ },
356
+ {
357
+ "name": "clip_vision_output",
358
+ "type": "CLIP_VISION_OUTPUT",
359
+ "shape": 7,
360
+ "link": 107
361
+ },
362
+ {
363
+ "name": "start_image",
364
+ "type": "IMAGE",
365
+ "shape": 7,
366
+ "link": 106
367
+ }
368
+ ],
369
+ "outputs": [
370
+ {
371
+ "name": "positive",
372
+ "type": "CONDITIONING",
373
+ "links": [
374
+ 101
375
+ ],
376
+ "slot_index": 0
377
+ },
378
+ {
379
+ "name": "negative",
380
+ "type": "CONDITIONING",
381
+ "links": [
382
+ 102
383
+ ],
384
+ "slot_index": 1
385
+ },
386
+ {
387
+ "name": "latent",
388
+ "type": "LATENT",
389
+ "links": [
390
+ 103
391
+ ],
392
+ "slot_index": 2
393
+ }
394
+ ],
395
+ "properties": {
396
+ "Node name for S&R": "WanImageToVideo"
397
+ },
398
+ "widgets_values": [
399
+ 512,
400
+ 512,
401
+ 33,
402
+ 1
403
+ ]
404
+ },
405
+ {
406
+ "id": 6,
407
+ "type": "CLIPTextEncode",
408
+ "pos": [
409
+ 415,
410
+ 186
411
+ ],
412
+ "size": [
413
+ 422.84503173828125,
414
+ 164.31304931640625
415
+ ],
416
+ "flags": {},
417
+ "order": 5,
418
+ "mode": 0,
419
+ "inputs": [
420
+ {
421
+ "name": "clip",
422
+ "type": "CLIP",
423
+ "link": 74
424
+ }
425
+ ],
426
+ "outputs": [
427
+ {
428
+ "name": "CONDITIONING",
429
+ "type": "CONDITIONING",
430
+ "links": [
431
+ 97
432
+ ],
433
+ "slot_index": 0
434
+ }
435
+ ],
436
+ "title": "CLIP Text Encode (Positive Prompt)",
437
+ "properties": {
438
+ "Node name for S&R": "CLIPTextEncode"
439
+ },
440
+ "widgets_values": [
441
+ "a cute anime girl with massive fennec ears and a big fluffy tail wearing a maid outfit turning around"
442
+ ],
443
+ "color": "#232",
444
+ "bgcolor": "#353"
445
+ },
446
+ {
447
+ "id": 3,
448
+ "type": "KSampler",
449
+ "pos": [
450
+ 863,
451
+ 187
452
+ ],
453
+ "size": [
454
+ 315,
455
+ 262
456
+ ],
457
+ "flags": {},
458
+ "order": 9,
459
+ "mode": 0,
460
+ "inputs": [
461
+ {
462
+ "name": "model",
463
+ "type": "MODEL",
464
+ "link": 92
465
+ },
466
+ {
467
+ "name": "positive",
468
+ "type": "CONDITIONING",
469
+ "link": 101
470
+ },
471
+ {
472
+ "name": "negative",
473
+ "type": "CONDITIONING",
474
+ "link": 102
475
+ },
476
+ {
477
+ "name": "latent_image",
478
+ "type": "LATENT",
479
+ "link": 103
480
+ }
481
+ ],
482
+ "outputs": [
483
+ {
484
+ "name": "LATENT",
485
+ "type": "LATENT",
486
+ "links": [
487
+ 35
488
+ ],
489
+ "slot_index": 0
490
+ }
491
+ ],
492
+ "properties": {
493
+ "Node name for S&R": "KSampler"
494
+ },
495
+ "widgets_values": [
496
+ 987948718394761,
497
+ "randomize",
498
+ 20,
499
+ 6,
500
+ "uni_pc",
501
+ "simple",
502
+ 1
503
+ ]
504
+ },
505
+ {
506
+ "id": 7,
507
+ "type": "CLIPTextEncode",
508
+ "pos": [
509
+ 413,
510
+ 389
511
+ ],
512
+ "size": [
513
+ 425.27801513671875,
514
+ 180.6060791015625
515
+ ],
516
+ "flags": {},
517
+ "order": 6,
518
+ "mode": 0,
519
+ "inputs": [
520
+ {
521
+ "name": "clip",
522
+ "type": "CLIP",
523
+ "link": 75
524
+ }
525
+ ],
526
+ "outputs": [
527
+ {
528
+ "name": "CONDITIONING",
529
+ "type": "CONDITIONING",
530
+ "links": [
531
+ 98
532
+ ],
533
+ "slot_index": 0
534
+ }
535
+ ],
536
+ "title": "CLIP Text Encode (Negative Prompt)",
537
+ "properties": {
538
+ "Node name for S&R": "CLIPTextEncode"
539
+ },
540
+ "widgets_values": [
541
+ "Overexposure, static, blurred details, subtitles, paintings, pictures, still, overall gray, worst quality, low quality, JPEG compression residue, ugly, mutilated, redundant fingers, poorly painted hands, poorly painted faces, deformed, disfigured, deformed limbs, fused fingers, cluttered background, three legs, a lot of people in the background, upside down"
542
+ ],
543
+ "color": "#322",
544
+ "bgcolor": "#533"
545
+ }
546
+ ],
547
+ "links": [
548
+ [
549
+ 35,
550
+ 3,
551
+ 0,
552
+ 8,
553
+ 0,
554
+ "LATENT"
555
+ ],
556
+ [
557
+ 56,
558
+ 8,
559
+ 0,
560
+ 28,
561
+ 0,
562
+ "IMAGE"
563
+ ],
564
+ [
565
+ 74,
566
+ 38,
567
+ 0,
568
+ 6,
569
+ 0,
570
+ "CLIP"
571
+ ],
572
+ [
573
+ 75,
574
+ 38,
575
+ 0,
576
+ 7,
577
+ 0,
578
+ "CLIP"
579
+ ],
580
+ [
581
+ 76,
582
+ 39,
583
+ 0,
584
+ 8,
585
+ 1,
586
+ "VAE"
587
+ ],
588
+ [
589
+ 92,
590
+ 37,
591
+ 0,
592
+ 3,
593
+ 0,
594
+ "MODEL"
595
+ ],
596
+ [
597
+ 93,
598
+ 8,
599
+ 0,
600
+ 47,
601
+ 0,
602
+ "IMAGE"
603
+ ],
604
+ [
605
+ 94,
606
+ 49,
607
+ 0,
608
+ 51,
609
+ 0,
610
+ "CLIP_VISION"
611
+ ],
612
+ [
613
+ 97,
614
+ 6,
615
+ 0,
616
+ 50,
617
+ 0,
618
+ "CONDITIONING"
619
+ ],
620
+ [
621
+ 98,
622
+ 7,
623
+ 0,
624
+ 50,
625
+ 1,
626
+ "CONDITIONING"
627
+ ],
628
+ [
629
+ 99,
630
+ 39,
631
+ 0,
632
+ 50,
633
+ 2,
634
+ "VAE"
635
+ ],
636
+ [
637
+ 101,
638
+ 50,
639
+ 0,
640
+ 3,
641
+ 1,
642
+ "CONDITIONING"
643
+ ],
644
+ [
645
+ 102,
646
+ 50,
647
+ 1,
648
+ 3,
649
+ 2,
650
+ "CONDITIONING"
651
+ ],
652
+ [
653
+ 103,
654
+ 50,
655
+ 2,
656
+ 3,
657
+ 3,
658
+ "LATENT"
659
+ ],
660
+ [
661
+ 106,
662
+ 52,
663
+ 0,
664
+ 50,
665
+ 4,
666
+ "IMAGE"
667
+ ],
668
+ [
669
+ 107,
670
+ 51,
671
+ 0,
672
+ 50,
673
+ 3,
674
+ "CLIP_VISION_OUTPUT"
675
+ ],
676
+ [
677
+ 109,
678
+ 52,
679
+ 0,
680
+ 51,
681
+ 1,
682
+ "IMAGE"
683
+ ]
684
+ ],
685
+ "groups": [],
686
+ "config": {},
687
+ "extra": {
688
+ "ds": {
689
+ "scale": 1.015255979947749,
690
+ "offset": [
691
+ 615.7681203368851,
692
+ 205.10773519106684
693
+ ]
694
+ }
695
+ },
696
+ "version": 0.4
697
+ }
example workflows_Wan2.1/image_to_video_wan_720p_example.json ADDED
@@ -0,0 +1,697 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "last_node_id": 53,
3
+ "last_link_id": 109,
4
+ "nodes": [
5
+ {
6
+ "id": 8,
7
+ "type": "VAEDecode",
8
+ "pos": [
9
+ 1210,
10
+ 190
11
+ ],
12
+ "size": [
13
+ 210,
14
+ 46
15
+ ],
16
+ "flags": {},
17
+ "order": 10,
18
+ "mode": 0,
19
+ "inputs": [
20
+ {
21
+ "name": "samples",
22
+ "type": "LATENT",
23
+ "link": 35
24
+ },
25
+ {
26
+ "name": "vae",
27
+ "type": "VAE",
28
+ "link": 76
29
+ }
30
+ ],
31
+ "outputs": [
32
+ {
33
+ "name": "IMAGE",
34
+ "type": "IMAGE",
35
+ "links": [
36
+ 56,
37
+ 93
38
+ ],
39
+ "slot_index": 0
40
+ }
41
+ ],
42
+ "properties": {
43
+ "Node name for S&R": "VAEDecode"
44
+ },
45
+ "widgets_values": []
46
+ },
47
+ {
48
+ "id": 39,
49
+ "type": "VAELoader",
50
+ "pos": [
51
+ 866.3932495117188,
52
+ 499.18597412109375
53
+ ],
54
+ "size": [
55
+ 306.36004638671875,
56
+ 58
57
+ ],
58
+ "flags": {},
59
+ "order": 0,
60
+ "mode": 0,
61
+ "inputs": [],
62
+ "outputs": [
63
+ {
64
+ "name": "VAE",
65
+ "type": "VAE",
66
+ "links": [
67
+ 76,
68
+ 99
69
+ ],
70
+ "slot_index": 0
71
+ }
72
+ ],
73
+ "properties": {
74
+ "Node name for S&R": "VAELoader"
75
+ },
76
+ "widgets_values": [
77
+ "wan_2.1_vae.safetensors"
78
+ ]
79
+ },
80
+ {
81
+ "id": 28,
82
+ "type": "SaveAnimatedWEBP",
83
+ "pos": [
84
+ 1460,
85
+ 190
86
+ ],
87
+ "size": [
88
+ 870.8511352539062,
89
+ 643.7430419921875
90
+ ],
91
+ "flags": {},
92
+ "order": 11,
93
+ "mode": 0,
94
+ "inputs": [
95
+ {
96
+ "name": "images",
97
+ "type": "IMAGE",
98
+ "link": 56
99
+ }
100
+ ],
101
+ "outputs": [],
102
+ "properties": {},
103
+ "widgets_values": [
104
+ "ComfyUI",
105
+ 16,
106
+ false,
107
+ 90,
108
+ "default"
109
+ ]
110
+ },
111
+ {
112
+ "id": 51,
113
+ "type": "CLIPVisionEncode",
114
+ "pos": [
115
+ 340,
116
+ 650
117
+ ],
118
+ "size": [
119
+ 253.60000610351562,
120
+ 78
121
+ ],
122
+ "flags": {},
123
+ "order": 7,
124
+ "mode": 0,
125
+ "inputs": [
126
+ {
127
+ "name": "clip_vision",
128
+ "type": "CLIP_VISION",
129
+ "link": 94
130
+ },
131
+ {
132
+ "name": "image",
133
+ "type": "IMAGE",
134
+ "link": 109
135
+ }
136
+ ],
137
+ "outputs": [
138
+ {
139
+ "name": "CLIP_VISION_OUTPUT",
140
+ "type": "CLIP_VISION_OUTPUT",
141
+ "links": [
142
+ 107
143
+ ],
144
+ "slot_index": 0
145
+ }
146
+ ],
147
+ "properties": {
148
+ "Node name for S&R": "CLIPVisionEncode"
149
+ },
150
+ "widgets_values": [
151
+ "none"
152
+ ]
153
+ },
154
+ {
155
+ "id": 49,
156
+ "type": "CLIPVisionLoader",
157
+ "pos": [
158
+ -20,
159
+ 650
160
+ ],
161
+ "size": [
162
+ 315,
163
+ 58
164
+ ],
165
+ "flags": {},
166
+ "order": 1,
167
+ "mode": 0,
168
+ "inputs": [],
169
+ "outputs": [
170
+ {
171
+ "name": "CLIP_VISION",
172
+ "type": "CLIP_VISION",
173
+ "links": [
174
+ 94
175
+ ],
176
+ "slot_index": 0
177
+ }
178
+ ],
179
+ "properties": {
180
+ "Node name for S&R": "CLIPVisionLoader"
181
+ },
182
+ "widgets_values": [
183
+ "clip_vision_h.safetensors"
184
+ ]
185
+ },
186
+ {
187
+ "id": 3,
188
+ "type": "KSampler",
189
+ "pos": [
190
+ 863,
191
+ 187
192
+ ],
193
+ "size": [
194
+ 315,
195
+ 262
196
+ ],
197
+ "flags": {},
198
+ "order": 9,
199
+ "mode": 0,
200
+ "inputs": [
201
+ {
202
+ "name": "model",
203
+ "type": "MODEL",
204
+ "link": 92
205
+ },
206
+ {
207
+ "name": "positive",
208
+ "type": "CONDITIONING",
209
+ "link": 101
210
+ },
211
+ {
212
+ "name": "negative",
213
+ "type": "CONDITIONING",
214
+ "link": 102
215
+ },
216
+ {
217
+ "name": "latent_image",
218
+ "type": "LATENT",
219
+ "link": 103
220
+ }
221
+ ],
222
+ "outputs": [
223
+ {
224
+ "name": "LATENT",
225
+ "type": "LATENT",
226
+ "links": [
227
+ 35
228
+ ],
229
+ "slot_index": 0
230
+ }
231
+ ],
232
+ "properties": {
233
+ "Node name for S&R": "KSampler"
234
+ },
235
+ "widgets_values": [
236
+ 353470491069365,
237
+ "randomize",
238
+ 20,
239
+ 6,
240
+ "uni_pc",
241
+ "simple",
242
+ 1
243
+ ]
244
+ },
245
+ {
246
+ "id": 50,
247
+ "type": "WanImageToVideo",
248
+ "pos": [
249
+ 673.0507202148438,
250
+ 627.272705078125
251
+ ],
252
+ "size": [
253
+ 342.5999755859375,
254
+ 210
255
+ ],
256
+ "flags": {},
257
+ "order": 8,
258
+ "mode": 0,
259
+ "inputs": [
260
+ {
261
+ "name": "positive",
262
+ "type": "CONDITIONING",
263
+ "link": 97
264
+ },
265
+ {
266
+ "name": "negative",
267
+ "type": "CONDITIONING",
268
+ "link": 98
269
+ },
270
+ {
271
+ "name": "vae",
272
+ "type": "VAE",
273
+ "link": 99
274
+ },
275
+ {
276
+ "name": "clip_vision_output",
277
+ "type": "CLIP_VISION_OUTPUT",
278
+ "shape": 7,
279
+ "link": 107
280
+ },
281
+ {
282
+ "name": "start_image",
283
+ "type": "IMAGE",
284
+ "shape": 7,
285
+ "link": 106
286
+ }
287
+ ],
288
+ "outputs": [
289
+ {
290
+ "name": "positive",
291
+ "type": "CONDITIONING",
292
+ "links": [
293
+ 101
294
+ ],
295
+ "slot_index": 0
296
+ },
297
+ {
298
+ "name": "negative",
299
+ "type": "CONDITIONING",
300
+ "links": [
301
+ 102
302
+ ],
303
+ "slot_index": 1
304
+ },
305
+ {
306
+ "name": "latent",
307
+ "type": "LATENT",
308
+ "links": [
309
+ 103
310
+ ],
311
+ "slot_index": 2
312
+ }
313
+ ],
314
+ "properties": {
315
+ "Node name for S&R": "WanImageToVideo"
316
+ },
317
+ "widgets_values": [
318
+ 768,
319
+ 768,
320
+ 53,
321
+ 1
322
+ ]
323
+ },
324
+ {
325
+ "id": 38,
326
+ "type": "CLIPLoader",
327
+ "pos": [
328
+ 12.94982624053955,
329
+ 184.6981658935547
330
+ ],
331
+ "size": [
332
+ 390,
333
+ 82
334
+ ],
335
+ "flags": {},
336
+ "order": 2,
337
+ "mode": 0,
338
+ "inputs": [],
339
+ "outputs": [
340
+ {
341
+ "name": "CLIP",
342
+ "type": "CLIP",
343
+ "links": [
344
+ 74,
345
+ 75
346
+ ],
347
+ "slot_index": 0
348
+ }
349
+ ],
350
+ "properties": {
351
+ "Node name for S&R": "CLIPLoader"
352
+ },
353
+ "widgets_values": [
354
+ "umt5_xxl_fp8_e4m3fn_scaled.safetensors",
355
+ "wan",
356
+ "default"
357
+ ]
358
+ },
359
+ {
360
+ "id": 37,
361
+ "type": "UNETLoader",
362
+ "pos": [
363
+ 485.1220397949219,
364
+ 57.094566345214844
365
+ ],
366
+ "size": [
367
+ 346.7470703125,
368
+ 82
369
+ ],
370
+ "flags": {},
371
+ "order": 3,
372
+ "mode": 0,
373
+ "inputs": [],
374
+ "outputs": [
375
+ {
376
+ "name": "MODEL",
377
+ "type": "MODEL",
378
+ "links": [
379
+ 92
380
+ ],
381
+ "slot_index": 0
382
+ }
383
+ ],
384
+ "properties": {
385
+ "Node name for S&R": "UNETLoader"
386
+ },
387
+ "widgets_values": [
388
+ "wan2.1_i2v_720p_14B_bf16.safetensors",
389
+ "default"
390
+ ]
391
+ },
392
+ {
393
+ "id": 47,
394
+ "type": "SaveWEBM",
395
+ "pos": [
396
+ 2371.950927734375,
397
+ 213.44703674316406
398
+ ],
399
+ "size": [
400
+ 315,
401
+ 130
402
+ ],
403
+ "flags": {},
404
+ "order": 12,
405
+ "mode": 4,
406
+ "inputs": [
407
+ {
408
+ "name": "images",
409
+ "type": "IMAGE",
410
+ "link": 93
411
+ }
412
+ ],
413
+ "outputs": [],
414
+ "properties": {
415
+ "Node name for S&R": "SaveWEBM"
416
+ },
417
+ "widgets_values": [
418
+ "ComfyUI",
419
+ "vp9",
420
+ 24,
421
+ 32
422
+ ]
423
+ },
424
+ {
425
+ "id": 7,
426
+ "type": "CLIPTextEncode",
427
+ "pos": [
428
+ 413,
429
+ 389
430
+ ],
431
+ "size": [
432
+ 425.27801513671875,
433
+ 180.6060791015625
434
+ ],
435
+ "flags": {},
436
+ "order": 6,
437
+ "mode": 0,
438
+ "inputs": [
439
+ {
440
+ "name": "clip",
441
+ "type": "CLIP",
442
+ "link": 75
443
+ }
444
+ ],
445
+ "outputs": [
446
+ {
447
+ "name": "CONDITIONING",
448
+ "type": "CONDITIONING",
449
+ "links": [
450
+ 98
451
+ ],
452
+ "slot_index": 0
453
+ }
454
+ ],
455
+ "title": "CLIP Text Encode (Negative Prompt)",
456
+ "properties": {
457
+ "Node name for S&R": "CLIPTextEncode"
458
+ },
459
+ "widgets_values": [
460
+ "Overexposure, static, blurred details, subtitles, paintings, pictures, still, overall gray, worst quality, low quality, JPEG compression residue, ugly, mutilated, redundant fingers, poorly painted hands, poorly painted faces, deformed, disfigured, deformed limbs, fused fingers, cluttered background, three legs, a lot of people in the background, upside down"
461
+ ],
462
+ "color": "#322",
463
+ "bgcolor": "#533"
464
+ },
465
+ {
466
+ "id": 52,
467
+ "type": "LoadImage",
468
+ "pos": [
469
+ -30,
470
+ 760
471
+ ],
472
+ "size": [
473
+ 315,
474
+ 314
475
+ ],
476
+ "flags": {},
477
+ "order": 4,
478
+ "mode": 0,
479
+ "inputs": [],
480
+ "outputs": [
481
+ {
482
+ "name": "IMAGE",
483
+ "type": "IMAGE",
484
+ "links": [
485
+ 106,
486
+ 109
487
+ ],
488
+ "slot_index": 0
489
+ },
490
+ {
491
+ "name": "MASK",
492
+ "type": "MASK",
493
+ "links": null,
494
+ "slot_index": 1
495
+ }
496
+ ],
497
+ "properties": {
498
+ "Node name for S&R": "LoadImage"
499
+ },
500
+ "widgets_values": [
501
+ "0_2 (3).png",
502
+ "image"
503
+ ]
504
+ },
505
+ {
506
+ "id": 6,
507
+ "type": "CLIPTextEncode",
508
+ "pos": [
509
+ 415,
510
+ 186
511
+ ],
512
+ "size": [
513
+ 422.84503173828125,
514
+ 164.31304931640625
515
+ ],
516
+ "flags": {},
517
+ "order": 5,
518
+ "mode": 0,
519
+ "inputs": [
520
+ {
521
+ "name": "clip",
522
+ "type": "CLIP",
523
+ "link": 74
524
+ }
525
+ ],
526
+ "outputs": [
527
+ {
528
+ "name": "CONDITIONING",
529
+ "type": "CONDITIONING",
530
+ "links": [
531
+ 97
532
+ ],
533
+ "slot_index": 0
534
+ }
535
+ ],
536
+ "title": "CLIP Text Encode (Positive Prompt)",
537
+ "properties": {
538
+ "Node name for S&R": "CLIPTextEncode"
539
+ },
540
+ "widgets_values": [
541
+ "A magician made flying cards in his hand, a lot of colored light and colored paper and confetti"
542
+ ],
543
+ "color": "#232",
544
+ "bgcolor": "#353"
545
+ }
546
+ ],
547
+ "links": [
548
+ [
549
+ 35,
550
+ 3,
551
+ 0,
552
+ 8,
553
+ 0,
554
+ "LATENT"
555
+ ],
556
+ [
557
+ 56,
558
+ 8,
559
+ 0,
560
+ 28,
561
+ 0,
562
+ "IMAGE"
563
+ ],
564
+ [
565
+ 74,
566
+ 38,
567
+ 0,
568
+ 6,
569
+ 0,
570
+ "CLIP"
571
+ ],
572
+ [
573
+ 75,
574
+ 38,
575
+ 0,
576
+ 7,
577
+ 0,
578
+ "CLIP"
579
+ ],
580
+ [
581
+ 76,
582
+ 39,
583
+ 0,
584
+ 8,
585
+ 1,
586
+ "VAE"
587
+ ],
588
+ [
589
+ 92,
590
+ 37,
591
+ 0,
592
+ 3,
593
+ 0,
594
+ "MODEL"
595
+ ],
596
+ [
597
+ 93,
598
+ 8,
599
+ 0,
600
+ 47,
601
+ 0,
602
+ "IMAGE"
603
+ ],
604
+ [
605
+ 94,
606
+ 49,
607
+ 0,
608
+ 51,
609
+ 0,
610
+ "CLIP_VISION"
611
+ ],
612
+ [
613
+ 97,
614
+ 6,
615
+ 0,
616
+ 50,
617
+ 0,
618
+ "CONDITIONING"
619
+ ],
620
+ [
621
+ 98,
622
+ 7,
623
+ 0,
624
+ 50,
625
+ 1,
626
+ "CONDITIONING"
627
+ ],
628
+ [
629
+ 99,
630
+ 39,
631
+ 0,
632
+ 50,
633
+ 2,
634
+ "VAE"
635
+ ],
636
+ [
637
+ 101,
638
+ 50,
639
+ 0,
640
+ 3,
641
+ 1,
642
+ "CONDITIONING"
643
+ ],
644
+ [
645
+ 102,
646
+ 50,
647
+ 1,
648
+ 3,
649
+ 2,
650
+ "CONDITIONING"
651
+ ],
652
+ [
653
+ 103,
654
+ 50,
655
+ 2,
656
+ 3,
657
+ 3,
658
+ "LATENT"
659
+ ],
660
+ [
661
+ 106,
662
+ 52,
663
+ 0,
664
+ 50,
665
+ 4,
666
+ "IMAGE"
667
+ ],
668
+ [
669
+ 107,
670
+ 51,
671
+ 0,
672
+ 50,
673
+ 3,
674
+ "CLIP_VISION_OUTPUT"
675
+ ],
676
+ [
677
+ 109,
678
+ 52,
679
+ 0,
680
+ 51,
681
+ 1,
682
+ "IMAGE"
683
+ ]
684
+ ],
685
+ "groups": [],
686
+ "config": {},
687
+ "extra": {
688
+ "ds": {
689
+ "scale": 1.01525597994775,
690
+ "offset": [
691
+ 509.5627989671756,
692
+ 312.04429170999896
693
+ ]
694
+ }
695
+ },
696
+ "version": 0.4
697
+ }
example workflows_Wan2.1/text_to_video_wan.json ADDED
@@ -0,0 +1,481 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "last_node_id": 47,
3
+ "last_link_id": 93,
4
+ "nodes": [
5
+ {
6
+ "id": 8,
7
+ "type": "VAEDecode",
8
+ "pos": [
9
+ 1210,
10
+ 190
11
+ ],
12
+ "size": [
13
+ 210,
14
+ 46
15
+ ],
16
+ "flags": {},
17
+ "order": 7,
18
+ "mode": 0,
19
+ "inputs": [
20
+ {
21
+ "name": "samples",
22
+ "type": "LATENT",
23
+ "link": 35
24
+ },
25
+ {
26
+ "name": "vae",
27
+ "type": "VAE",
28
+ "link": 76
29
+ }
30
+ ],
31
+ "outputs": [
32
+ {
33
+ "name": "IMAGE",
34
+ "type": "IMAGE",
35
+ "links": [
36
+ 56,
37
+ 93
38
+ ],
39
+ "slot_index": 0
40
+ }
41
+ ],
42
+ "properties": {
43
+ "Node name for S&R": "VAEDecode"
44
+ },
45
+ "widgets_values": []
46
+ },
47
+ {
48
+ "id": 39,
49
+ "type": "VAELoader",
50
+ "pos": [
51
+ 866.3932495117188,
52
+ 499.18597412109375
53
+ ],
54
+ "size": [
55
+ 306.36004638671875,
56
+ 58
57
+ ],
58
+ "flags": {},
59
+ "order": 0,
60
+ "mode": 0,
61
+ "inputs": [],
62
+ "outputs": [
63
+ {
64
+ "name": "VAE",
65
+ "type": "VAE",
66
+ "links": [
67
+ 76
68
+ ],
69
+ "slot_index": 0
70
+ }
71
+ ],
72
+ "properties": {
73
+ "Node name for S&R": "VAELoader"
74
+ },
75
+ "widgets_values": [
76
+ "wan_2.1_vae.safetensors"
77
+ ]
78
+ },
79
+ {
80
+ "id": 28,
81
+ "type": "SaveAnimatedWEBP",
82
+ "pos": [
83
+ 1460,
84
+ 190
85
+ ],
86
+ "size": [
87
+ 870.8511352539062,
88
+ 643.7430419921875
89
+ ],
90
+ "flags": {},
91
+ "order": 8,
92
+ "mode": 0,
93
+ "inputs": [
94
+ {
95
+ "name": "images",
96
+ "type": "IMAGE",
97
+ "link": 56
98
+ }
99
+ ],
100
+ "outputs": [],
101
+ "properties": {},
102
+ "widgets_values": [
103
+ "ComfyUI",
104
+ 16,
105
+ false,
106
+ 90,
107
+ "default"
108
+ ]
109
+ },
110
+ {
111
+ "id": 6,
112
+ "type": "CLIPTextEncode",
113
+ "pos": [
114
+ 415,
115
+ 186
116
+ ],
117
+ "size": [
118
+ 422.84503173828125,
119
+ 164.31304931640625
120
+ ],
121
+ "flags": {},
122
+ "order": 4,
123
+ "mode": 0,
124
+ "inputs": [
125
+ {
126
+ "name": "clip",
127
+ "type": "CLIP",
128
+ "link": 74
129
+ }
130
+ ],
131
+ "outputs": [
132
+ {
133
+ "name": "CONDITIONING",
134
+ "type": "CONDITIONING",
135
+ "links": [
136
+ 46
137
+ ],
138
+ "slot_index": 0
139
+ }
140
+ ],
141
+ "title": "CLIP Text Encode (Positive Prompt)",
142
+ "properties": {
143
+ "Node name for S&R": "CLIPTextEncode"
144
+ },
145
+ "widgets_values": [
146
+ "a fox moving quickly in a beautiful winter scenery nature trees sunset tracking camera"
147
+ ],
148
+ "color": "#232",
149
+ "bgcolor": "#353"
150
+ },
151
+ {
152
+ "id": 38,
153
+ "type": "CLIPLoader",
154
+ "pos": [
155
+ 12.94982624053955,
156
+ 184.6981658935547
157
+ ],
158
+ "size": [
159
+ 390,
160
+ 82
161
+ ],
162
+ "flags": {},
163
+ "order": 1,
164
+ "mode": 0,
165
+ "inputs": [],
166
+ "outputs": [
167
+ {
168
+ "name": "CLIP",
169
+ "type": "CLIP",
170
+ "links": [
171
+ 74,
172
+ 75
173
+ ],
174
+ "slot_index": 0
175
+ }
176
+ ],
177
+ "properties": {
178
+ "Node name for S&R": "CLIPLoader"
179
+ },
180
+ "widgets_values": [
181
+ "umt5_xxl_fp8_e4m3fn_scaled.safetensors",
182
+ "wan",
183
+ "default"
184
+ ]
185
+ },
186
+ {
187
+ "id": 40,
188
+ "type": "EmptyHunyuanLatentVideo",
189
+ "pos": [
190
+ 520,
191
+ 620
192
+ ],
193
+ "size": [
194
+ 315,
195
+ 130
196
+ ],
197
+ "flags": {},
198
+ "order": 2,
199
+ "mode": 0,
200
+ "inputs": [],
201
+ "outputs": [
202
+ {
203
+ "name": "LATENT",
204
+ "type": "LATENT",
205
+ "links": [
206
+ 91
207
+ ],
208
+ "slot_index": 0
209
+ }
210
+ ],
211
+ "properties": {
212
+ "Node name for S&R": "EmptyHunyuanLatentVideo"
213
+ },
214
+ "widgets_values": [
215
+ 832,
216
+ 480,
217
+ 33,
218
+ 1
219
+ ]
220
+ },
221
+ {
222
+ "id": 37,
223
+ "type": "UNETLoader",
224
+ "pos": [
225
+ 485.1220397949219,
226
+ 57.094566345214844
227
+ ],
228
+ "size": [
229
+ 346.7470703125,
230
+ 82
231
+ ],
232
+ "flags": {},
233
+ "order": 3,
234
+ "mode": 0,
235
+ "inputs": [],
236
+ "outputs": [
237
+ {
238
+ "name": "MODEL",
239
+ "type": "MODEL",
240
+ "links": [
241
+ 92
242
+ ],
243
+ "slot_index": 0
244
+ }
245
+ ],
246
+ "properties": {
247
+ "Node name for S&R": "UNETLoader"
248
+ },
249
+ "widgets_values": [
250
+ "wan2.1_t2v_1.3B_bf16.safetensors",
251
+ "default"
252
+ ]
253
+ },
254
+ {
255
+ "id": 47,
256
+ "type": "SaveWEBM",
257
+ "pos": [
258
+ 2367.213134765625,
259
+ 193.6114959716797
260
+ ],
261
+ "size": [
262
+ 315,
263
+ 130
264
+ ],
265
+ "flags": {},
266
+ "order": 9,
267
+ "mode": 4,
268
+ "inputs": [
269
+ {
270
+ "name": "images",
271
+ "type": "IMAGE",
272
+ "link": 93
273
+ }
274
+ ],
275
+ "outputs": [],
276
+ "properties": {
277
+ "Node name for S&R": "SaveWEBM"
278
+ },
279
+ "widgets_values": [
280
+ "ComfyUI",
281
+ "vp9",
282
+ 24,
283
+ 32
284
+ ]
285
+ },
286
+ {
287
+ "id": 3,
288
+ "type": "KSampler",
289
+ "pos": [
290
+ 863,
291
+ 187
292
+ ],
293
+ "size": [
294
+ 315,
295
+ 262
296
+ ],
297
+ "flags": {},
298
+ "order": 6,
299
+ "mode": 0,
300
+ "inputs": [
301
+ {
302
+ "name": "model",
303
+ "type": "MODEL",
304
+ "link": 92
305
+ },
306
+ {
307
+ "name": "positive",
308
+ "type": "CONDITIONING",
309
+ "link": 46
310
+ },
311
+ {
312
+ "name": "negative",
313
+ "type": "CONDITIONING",
314
+ "link": 52
315
+ },
316
+ {
317
+ "name": "latent_image",
318
+ "type": "LATENT",
319
+ "link": 91
320
+ }
321
+ ],
322
+ "outputs": [
323
+ {
324
+ "name": "LATENT",
325
+ "type": "LATENT",
326
+ "links": [
327
+ 35
328
+ ],
329
+ "slot_index": 0
330
+ }
331
+ ],
332
+ "properties": {
333
+ "Node name for S&R": "KSampler"
334
+ },
335
+ "widgets_values": [
336
+ 878361741413693,
337
+ "randomize",
338
+ 30,
339
+ 6,
340
+ "uni_pc",
341
+ "simple",
342
+ 1
343
+ ]
344
+ },
345
+ {
346
+ "id": 7,
347
+ "type": "CLIPTextEncode",
348
+ "pos": [
349
+ 413,
350
+ 389
351
+ ],
352
+ "size": [
353
+ 425.27801513671875,
354
+ 180.6060791015625
355
+ ],
356
+ "flags": {},
357
+ "order": 5,
358
+ "mode": 0,
359
+ "inputs": [
360
+ {
361
+ "name": "clip",
362
+ "type": "CLIP",
363
+ "link": 75
364
+ }
365
+ ],
366
+ "outputs": [
367
+ {
368
+ "name": "CONDITIONING",
369
+ "type": "CONDITIONING",
370
+ "links": [
371
+ 52
372
+ ],
373
+ "slot_index": 0
374
+ }
375
+ ],
376
+ "title": "CLIP Text Encode (Negative Prompt)",
377
+ "properties": {
378
+ "Node name for S&R": "CLIPTextEncode"
379
+ },
380
+ "widgets_values": [
381
+ "Overexposure, static, blurred details, subtitles, paintings, pictures, still, overall gray, worst quality, low quality, JPEG compression residue, ugly, mutilated, redundant fingers, poorly painted hands, poorly painted faces, deformed, disfigured, deformed limbs, fused fingers, cluttered background, three legs, a lot of people in the background, upside down"
382
+ ],
383
+ "color": "#322",
384
+ "bgcolor": "#533"
385
+ }
386
+ ],
387
+ "links": [
388
+ [
389
+ 35,
390
+ 3,
391
+ 0,
392
+ 8,
393
+ 0,
394
+ "LATENT"
395
+ ],
396
+ [
397
+ 46,
398
+ 6,
399
+ 0,
400
+ 3,
401
+ 1,
402
+ "CONDITIONING"
403
+ ],
404
+ [
405
+ 52,
406
+ 7,
407
+ 0,
408
+ 3,
409
+ 2,
410
+ "CONDITIONING"
411
+ ],
412
+ [
413
+ 56,
414
+ 8,
415
+ 0,
416
+ 28,
417
+ 0,
418
+ "IMAGE"
419
+ ],
420
+ [
421
+ 74,
422
+ 38,
423
+ 0,
424
+ 6,
425
+ 0,
426
+ "CLIP"
427
+ ],
428
+ [
429
+ 75,
430
+ 38,
431
+ 0,
432
+ 7,
433
+ 0,
434
+ "CLIP"
435
+ ],
436
+ [
437
+ 76,
438
+ 39,
439
+ 0,
440
+ 8,
441
+ 1,
442
+ "VAE"
443
+ ],
444
+ [
445
+ 91,
446
+ 40,
447
+ 0,
448
+ 3,
449
+ 3,
450
+ "LATENT"
451
+ ],
452
+ [
453
+ 92,
454
+ 37,
455
+ 0,
456
+ 3,
457
+ 0,
458
+ "MODEL"
459
+ ],
460
+ [
461
+ 93,
462
+ 8,
463
+ 0,
464
+ 47,
465
+ 0,
466
+ "IMAGE"
467
+ ]
468
+ ],
469
+ "groups": [],
470
+ "config": {},
471
+ "extra": {
472
+ "ds": {
473
+ "scale": 1.2284597357367726,
474
+ "offset": [
475
+ 445.486325470362,
476
+ 133.8941997640868
477
+ ]
478
+ }
479
+ },
480
+ "version": 0.4
481
+ }
split_files/clip_vision/clip_vision_h.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64a7ef761bfccbadbaa3da77366aac4185a6c58fa5de5f589b42a65bcc21f161
3
+ size 1264219396
split_files/diffusion_models/wan2.1_flf2v_720p_14B_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf4ac25667d00f53f49df02c5771f5aa7801c1dcb9b3ccade1407687c426d030
3
+ size 32792693440
split_files/diffusion_models/wan2.1_flf2v_720p_14B_fp8_e4m3fn.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d68ca694a695274e48e00974128337e06e497d95a1dc09e86fd2a01a405f455f
3
+ size 16397952536
split_files/diffusion_models/wan2.1_fun_camera_v1.1_1.3B_bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a648f6fbb543dd6ac9f2f9af1ca4bc1bbe460482ee4d7984d50565699e97398
3
+ size 3232727784
split_files/diffusion_models/wan2.1_fun_camera_v1.1_14B_bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f519d276502dcdb0e657d32faca64a5bf7d47b3c2b2c9b12bb849e9e0f3c495f
3
+ size 33796394608
split_files/diffusion_models/wan2.1_fun_control_1.3B_bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ff6289322b41bf187206eac2a57e85ce85c9ee5bfe8bc44eabeaeb86b44129a
3
+ size 3129105448
split_files/diffusion_models/wan2.1_fun_inp_1.3B_bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8495d2b1673ffb18abb548a64ff3b0e4bd367734f653096f7a8a3ad46954d511
3
+ size 3128957992
split_files/diffusion_models/wan2.1_i2v_480p_14B_bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c058a4ac5363c35d1ab4dd3bdec788c23b267fa42a0d7c68aba599f2f74600c9
3
+ size 32789894024
split_files/diffusion_models/wan2.1_i2v_480p_14B_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27988f6b510eb8d5fdd7485671b54897f8683f2bba7a772c5671be21d3491253
3
+ size 32791377504
split_files/diffusion_models/wan2.1_i2v_480p_14B_fp8_e4m3fn.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ca75338e7a47ca7cacddb7e626647e65829c497387f718ecb6ea0bae456944a
3
+ size 16397245448
split_files/diffusion_models/wan2.1_i2v_480p_14B_fp8_scaled.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2de21b99b2e72cb0ff15253b07e926f26e7cf1b7e229efc32f94ad1f1ed9395
3
+ size 16401356938
split_files/diffusion_models/wan2.1_i2v_720p_14B_bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ef7ea5bf9eea636b9b3ebd84c40671b4a18ae2704cb4c8595cb5b25c1d8e8b9
3
+ size 32789894024
split_files/diffusion_models/wan2.1_i2v_720p_14B_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7664fe075b3c82dcecf89012ad3429eee41ee9f10d476f60bc2d2ae3c4ca986c
3
+ size 32791377504
split_files/diffusion_models/wan2.1_i2v_720p_14B_fp8_e4m3fn.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2051cd29d6b2f0c924fa7a3e78a4772f0134d7b059f21590dcce416f4f6cbe8
3
+ size 16397245448
split_files/diffusion_models/wan2.1_i2v_720p_14B_fp8_scaled.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d7a34b63b70eb608324e546d979167a5e787ac6bca3528e63f54a11572d66aa
3
+ size 16401356938
split_files/diffusion_models/wan2.1_t2v_1.3B_bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f999b0d6cb9a72b3d98ac386ed96f57f8cecae13994a69232514ea4974ad5fd
3
+ size 2838104528
split_files/diffusion_models/wan2.1_t2v_1.3B_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be531024cd9018cb5b48c40cfbb6a6191645b1c792eb8bf4f8c1c6e10f924dc5
3
+ size 2838303560
split_files/diffusion_models/wan2.1_t2v_14B_bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:193535c6450045f718df5f011de6d94d49bd9b13f37ca0412500f050dbbb01a8
3
+ size 28577096680
split_files/diffusion_models/wan2.1_t2v_14B_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46c27d3693bf2475990a912e08bf67fc6e6cd5396eab87b5e8dd1fcd3651364a
3
+ size 28577761176
split_files/diffusion_models/wan2.1_t2v_14B_fp8_e4m3fn.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ee88ab18d7ed7691c5b7f8bdc3d0a9815e6efe75499287564830fd209d3cdfb
3
+ size 14289605920
split_files/diffusion_models/wan2.1_t2v_14B_fp8_scaled.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e39adde59c5e0e90edbb35873126b0d67928b5c11c501e384e976d6dc597cce
3
+ size 14293896178
split_files/diffusion_models/wan2.1_vace_1.3B_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:640ccc0577e6a5d4bb15cd91b11b699ef914fc55f126c5a1c544e152130784f2
3
+ size 4309519800
split_files/diffusion_models/wan2.1_vace_1.3B_preview_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:654693bf2a93a27cd67c3bcee238bc1d0cbb0dd9a74928ed7155fb21a2a1900a
3
+ size 4309497120
split_files/diffusion_models/wan2.1_vace_14B_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f202a5c59b8a91ada1862c46a038214f1f7f216c61ec8350d25f69b919da4307
3
+ size 34675323640
split_files/text_encoders/umt5_xxl_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b8850f1961e1cf8a77cca4c964a358d303f490833c6c087d0cff4b2f99db2af
3
+ size 11366399385
split_files/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3355d30191f1f066b26d93fba017ae9809dce6c627dda5f6a66eaa651204f68
3
+ size 6735906897
split_files/vae/wan_2.1_vae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fc39d31359a4b0a64f55876d8ff7fa8d780956ae2cb13463b0223e15148976b
3
+ size 253815318