{ "id": "cb96220b-6d31-4734-8321-8173168195ca", "revision": 0, "last_node_id": 54, "last_link_id": 111, "nodes": [ { "id": 8, "type": "VAEDecode", "pos": [ 1210, 190 ], "size": [ 210, 46 ], "flags": {}, "order": 11, "mode": 0, "inputs": [ { "name": "samples", "type": "LATENT", "link": 35 }, { "name": "vae", "type": "VAE", "link": 76 } ], "outputs": [ { "name": "IMAGE", "type": "IMAGE", "slot_index": 0, "links": [ 56, 93 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "VAEDecode" }, "widgets_values": [] }, { "id": 39, "type": "VAELoader", "pos": [ 866.3932495117188, 499.18597412109375 ], "size": [ 306.36004638671875, 58 ], "flags": {}, "order": 0, "mode": 0, "inputs": [], "outputs": [ { "name": "VAE", "type": "VAE", "slot_index": 0, "links": [ 76, 99 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "VAELoader", "models": [ { "name": "wan_2.1_vae.safetensors", "url": "https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/vae/wan_2.1_vae.safetensors?download=true", "directory": "vae" } ] }, "widgets_values": [ "wan_2.1_vae.safetensors" ] }, { "id": 47, "type": "SaveWEBM", "pos": [ 2367.213134765625, 193.6114959716797 ], "size": [ 315, 130 ], "flags": {}, "order": 13, "mode": 4, "inputs": [ { "name": "images", "type": "IMAGE", "link": 93 } ], "outputs": [], "properties": { "cnr_id": "comfy-core", "ver": "0.3.26", "Node name for S&R": "SaveWEBM" }, "widgets_values": [ "ComfyUI", "vp9", 24, 32 ] }, { "id": 7, "type": "CLIPTextEncode", "pos": [ 413, 389 ], "size": [ 425.27801513671875, 180.6060791015625 ], "flags": {}, "order": 8, "mode": 0, "inputs": [ { "name": "clip", "type": "CLIP", "link": 75 } ], "outputs": [ { "name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [ 98 ] } ], "title": "CLIP Text Encode (Negative Prompt)", "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "CLIPTextEncode" }, "widgets_values": [ "色调艳丽,过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多,倒着走" ], "color": "#322", "bgcolor": "#533" }, { "id": 50, "type": "WanImageToVideo", "pos": [ 673.0507202148438, 627.272705078125 ], "size": [ 342.5999755859375, 210 ], "flags": {}, "order": 9, "mode": 0, "inputs": [ { "name": "positive", "type": "CONDITIONING", "link": 97 }, { "name": "negative", "type": "CONDITIONING", "link": 98 }, { "name": "vae", "type": "VAE", "link": 99 }, { "name": "clip_vision_output", "shape": 7, "type": "CLIP_VISION_OUTPUT", "link": 107 }, { "name": "start_image", "shape": 7, "type": "IMAGE", "link": 106 } ], "outputs": [ { "name": "positive", "type": "CONDITIONING", "slot_index": 0, "links": [ 101 ] }, { "name": "negative", "type": "CONDITIONING", "slot_index": 1, "links": [ 102 ] }, { "name": "latent", "type": "LATENT", "slot_index": 2, "links": [ 103 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "WanImageToVideo" }, "widgets_values": [ 512, 512, 33, 1 ] }, { "id": 6, "type": "CLIPTextEncode", "pos": [ 415, 186 ], "size": [ 422.84503173828125, 164.31304931640625 ], "flags": {}, "order": 7, "mode": 0, "inputs": [ { "name": "clip", "type": "CLIP", "link": 74 } ], "outputs": [ { "name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [ 97 ] } ], "title": "CLIP Text Encode (Positive Prompt)", "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "CLIPTextEncode" }, "widgets_values": [ "a cute anime girl with massive fennec ears and a big fluffy tail wearing a maid outfit turning around" ], "color": "#232", "bgcolor": "#353" }, { "id": 3, "type": "KSampler", "pos": [ 863, 187 ], "size": [ 315, 262 ], "flags": {}, "order": 10, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 111 }, { "name": "positive", "type": "CONDITIONING", "link": 101 }, { "name": "negative", "type": "CONDITIONING", "link": 102 }, { "name": "latent_image", "type": "LATENT", "link": 103 } ], "outputs": [ { "name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [ 35 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "KSampler" }, "widgets_values": [ 1045519799525262, "randomize", 20, 6, "uni_pc", "simple", 1 ] }, { "id": 49, "type": "CLIPVisionLoader", "pos": [ 20, 640 ], "size": [ 315, 58 ], "flags": {}, "order": 1, "mode": 0, "inputs": [], "outputs": [ { "name": "CLIP_VISION", "type": "CLIP_VISION", "slot_index": 0, "links": [ 94 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "CLIPVisionLoader", "models": [ { "name": "clip_vision_h.safetensors", "url": "https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/clip_vision/clip_vision_h.safetensors?download=true", "directory": "clip_vision" } ] }, "widgets_values": [ "clip_vision_h.safetensors" ] }, { "id": 51, "type": "CLIPVisionEncode", "pos": [ 360, 640 ], "size": [ 253.60000610351562, 78 ], "flags": {}, "order": 5, "mode": 0, "inputs": [ { "name": "clip_vision", "type": "CLIP_VISION", "link": 94 }, { "name": "image", "type": "IMAGE", "link": 109 } ], "outputs": [ { "name": "CLIP_VISION_OUTPUT", "type": "CLIP_VISION_OUTPUT", "slot_index": 0, "links": [ 107 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "CLIPVisionEncode" }, "widgets_values": [ "none" ] }, { "id": 52, "type": "LoadImage", "pos": [ 20, 760 ], "size": [ 315, 314 ], "flags": {}, "order": 2, "mode": 0, "inputs": [], "outputs": [ { "name": "IMAGE", "type": "IMAGE", "slot_index": 0, "links": [ 106, 109 ] }, { "name": "MASK", "type": "MASK", "slot_index": 1, "links": null } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "LoadImage" }, "widgets_values": [ "flux_dev_example.png", "image" ] }, { "id": 54, "type": "ModelSamplingSD3", "pos": [ 510, 70 ], "size": [ 315, 58 ], "flags": {}, "order": 6, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 110 } ], "outputs": [ { "name": "MODEL", "type": "MODEL", "slot_index": 0, "links": [ 111 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "ModelSamplingSD3" }, "widgets_values": [ 8 ] }, { "id": 37, "type": "UNETLoader", "pos": [ 20, 70 ], "size": [ 346.7470703125, 82 ], "flags": {}, "order": 3, "mode": 0, "inputs": [], "outputs": [ { "name": "MODEL", "type": "MODEL", "slot_index": 0, "links": [ 110 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "UNETLoader", "models": [ { "name": "wan2.1_i2v_480p_14B_fp16.safetensors", "url": "https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/diffusion_models/wan2.1_i2v_480p_14B_fp16.safetensors?download=true", "directory": "diffusion_models" } ] }, "widgets_values": [ "wan2.1_i2v_480p_14B_fp16.safetensors", "default" ] }, { "id": 38, "type": "CLIPLoader", "pos": [ 20, 190 ], "size": [ 390, 106 ], "flags": {}, "order": 4, "mode": 0, "inputs": [], "outputs": [ { "name": "CLIP", "type": "CLIP", "slot_index": 0, "links": [ 74, 75 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "CLIPLoader", "models": [ { "name": "umt5_xxl_fp8_e4m3fn_scaled.safetensors", "url": "https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensors?download=true", "directory": "text_encoders" } ] }, "widgets_values": [ "umt5_xxl_fp8_e4m3fn_scaled.safetensors", "wan", "default" ] }, { "id": 28, "type": "SaveAnimatedWEBP", "pos": [ 1460, 190 ], "size": [ 578.561279296875, 636.9578247070312 ], "flags": {}, "order": 12, "mode": 0, "inputs": [ { "name": "images", "type": "IMAGE", "link": 56 } ], "outputs": [], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36" }, "widgets_values": [ "ComfyUI", 16, false, 90, "default" ] } ], "links": [ [ 35, 3, 0, 8, 0, "LATENT" ], [ 56, 8, 0, 28, 0, "IMAGE" ], [ 74, 38, 0, 6, 0, "CLIP" ], [ 75, 38, 0, 7, 0, "CLIP" ], [ 76, 39, 0, 8, 1, "VAE" ], [ 93, 8, 0, 47, 0, "IMAGE" ], [ 94, 49, 0, 51, 0, "CLIP_VISION" ], [ 97, 6, 0, 50, 0, "CONDITIONING" ], [ 98, 7, 0, 50, 1, "CONDITIONING" ], [ 99, 39, 0, 50, 2, "VAE" ], [ 101, 50, 0, 3, 1, "CONDITIONING" ], [ 102, 50, 1, 3, 2, "CONDITIONING" ], [ 103, 50, 2, 3, 3, "LATENT" ], [ 106, 52, 0, 50, 4, "IMAGE" ], [ 107, 51, 0, 50, 3, "CLIP_VISION_OUTPUT" ], [ 109, 52, 0, 51, 1, "IMAGE" ], [ 110, 37, 0, 54, 0, "MODEL" ], [ 111, 54, 0, 3, 0, "MODEL" ] ], "groups": [], "config": {}, "extra": { "ds": { "scale": 0.7627768444385802, "offset": [ 80.64364428726601, -27.0147851915508 ] }, "frontendVersion": "1.20.5" }, "version": 0.4 }