{ "last_node_id": 89, "last_link_id": 230, "nodes": [ { "id": 22, "type": "BasicGuider", "pos": [ 600, 0 ], "size": [ 222.3482666015625, 46 ], "flags": {}, "order": 17, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 195, "slot_index": 0 }, { "name": "conditioning", "type": "CONDITIONING", "link": 129, "slot_index": 1 } ], "outputs": [ { "name": "GUIDER", "type": "GUIDER", "shape": 3, "links": [ 30 ], "slot_index": 0 } ], "properties": { "Node name for S&R": "BasicGuider" }, "widgets_values": [] }, { "id": 8, "type": "VAEDecode", "pos": [ 1150, 90 ], "size": [ 210, 46 ], "flags": {}, "order": 19, "mode": 2, "inputs": [ { "name": "samples", "type": "LATENT", "link": 181 }, { "name": "vae", "type": "VAE", "link": 206 } ], "outputs": [ { "name": "IMAGE", "type": "IMAGE", "links": [], "slot_index": 0 } ], "properties": { "Node name for S&R": "VAEDecode" }, "widgets_values": [] }, { "id": 13, "type": "SamplerCustomAdvanced", "pos": [ 860, 200 ], "size": [ 272.3617858886719, 124.53733825683594 ], "flags": {}, "order": 18, "mode": 0, "inputs": [ { "name": "noise", "type": "NOISE", "link": 37, "slot_index": 0 }, { "name": "guider", "type": "GUIDER", "link": 30, "slot_index": 1 }, { "name": "sampler", "type": "SAMPLER", "link": 19, "slot_index": 2 }, { "name": "sigmas", "type": "SIGMAS", "link": 20, "slot_index": 3 }, { "name": "latent_image", "type": "LATENT", "link": 216, "slot_index": 4 } ], "outputs": [ { "name": "output", "type": "LATENT", "shape": 3, "links": [ 181, 210 ], "slot_index": 0 }, { "name": "denoised_output", "type": "LATENT", "shape": 3, "links": null } ], "properties": { "Node name for S&R": "SamplerCustomAdvanced" }, "widgets_values": [] }, { "id": 74, "type": "Note", "pos": [ 1147.7459716796875, 405.0789489746094 ], "size": [ 210, 170 ], "flags": {}, "order": 0, "mode": 0, "inputs": [], "outputs": [], "properties": {}, "widgets_values": [ "Use the tiled decode node by default because most people will need it.\n\nLower the tile_size and overlap if you run out of memory." ], "color": "#432", "bgcolor": "#653" }, { "id": 26, "type": "FluxGuidance", "pos": [ 514.2149047851562, 86.77685546875 ], "size": [ 317.4000244140625, 58 ], "flags": {}, "order": 16, "mode": 0, "inputs": [ { "name": "conditioning", "type": "CONDITIONING", "link": 225 } ], "outputs": [ { "name": "CONDITIONING", "type": "CONDITIONING", "shape": 3, "links": [ 129 ], "slot_index": 0 } ], "properties": { "Node name for S&R": "FluxGuidance" }, "widgets_values": [ 6 ], "color": "#233", "bgcolor": "#355" }, { "id": 67, "type": "ModelSamplingSD3", "pos": [ 360, 0 ], "size": [ 210, 58 ], "flags": {}, "order": 13, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 209 } ], "outputs": [ { "name": "MODEL", "type": "MODEL", "links": [ 195 ], "slot_index": 0 } ], "properties": { "Node name for S&R": "ModelSamplingSD3" }, "widgets_values": [ 7 ] }, { "id": 17, "type": "BasicScheduler", "pos": [ 510, 660 ], "size": [ 315, 106 ], "flags": {}, "order": 12, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 190, "slot_index": 0 } ], "outputs": [ { "name": "SIGMAS", "type": "SIGMAS", "shape": 3, "links": [ 20 ] } ], "properties": { "Node name for S&R": "BasicScheduler" }, "widgets_values": [ "simple", 20, 1 ] }, { "id": 16, "type": "KSamplerSelect", "pos": [ 520, 550 ], "size": [ 315, 58 ], "flags": {}, "order": 1, "mode": 0, "inputs": [], "outputs": [ { "name": "SAMPLER", "type": "SAMPLER", "shape": 3, "links": [ 19 ] } ], "properties": { "Node name for S&R": "KSamplerSelect" }, "widgets_values": [ "euler" ] }, { "id": 78, "type": "HunyuanImageToVideo", "pos": [ 510, 820 ], "size": [ 315, 170 ], "flags": {}, "order": 15, "mode": 0, "inputs": [ { "name": "positive", "type": "CONDITIONING", "link": 218 }, { "name": "vae", "type": "VAE", "link": 223 }, { "name": "start_image", "type": "IMAGE", "shape": 7, "link": 222 } ], "outputs": [ { "name": "positive", "type": "CONDITIONING", "links": [ 225 ], "slot_index": 0 }, { "name": "latent", "type": "LATENT", "links": [ 216 ] } ], "properties": { "Node name for S&R": "HunyuanImageToVideo" }, "widgets_values": [ 720, 720, 57, 1 ] }, { "id": 73, "type": "VAEDecodeTiled", "pos": [ 1150, 200 ], "size": [ 210, 150 ], "flags": {}, "order": 20, "mode": 0, "inputs": [ { "name": "samples", "type": "LATENT", "link": 210 }, { "name": "vae", "type": "VAE", "link": 211 } ], "outputs": [ { "name": "IMAGE", "type": "IMAGE", "links": [ 230 ], "slot_index": 0 } ], "properties": { "Node name for S&R": "VAEDecodeTiled" }, "widgets_values": [ 256, 64, 64, 8 ] }, { "id": 75, "type": "SaveAnimatedWEBP", "pos": [ 1640, 200 ], "size": [ 621.495361328125, 587.12451171875 ], "flags": {}, "order": 22, "mode": 0, "inputs": [ { "name": "images", "type": "IMAGE", "link": 228 } ], "outputs": [], "properties": {}, "widgets_values": [ "ComfyUI", 24, false, 90, "default" ] }, { "id": 84, "type": "SaveWEBM", "pos": [ 2280, 200 ], "size": [ 315, 130 ], "flags": {}, "order": 23, "mode": 4, "inputs": [ { "name": "images", "type": "IMAGE", "link": 229 } ], "outputs": [], "properties": { "Node name for S&R": "SaveWEBM" }, "widgets_values": [ "ComfyUI", "vp9", 24, 12 ] }, { "id": 87, "type": "Note", "pos": [ 1410, 340 ], "size": [ 210, 170 ], "flags": {}, "order": 2, "mode": 0, "inputs": [], "outputs": [], "properties": {}, "widgets_values": [ "The official code removes the first 4 frames because they are sometimes bad. You can bypass (CTRL-B) this node if you don't want this." ], "color": "#432", "bgcolor": "#653" }, { "id": 80, "type": "TextEncodeHunyuanVideo_ImageToVideo", "pos": [ 390, 180 ], "size": [ 441, 200 ], "flags": {}, "order": 14, "mode": 0, "inputs": [ { "name": "clip", "type": "CLIP", "link": 224 }, { "name": "clip_vision_output", "type": "CLIP_VISION_OUTPUT", "link": 219 } ], "outputs": [ { "name": "CONDITIONING", "type": "CONDITIONING", "links": [ 218 ] } ], "title": "Text Encode Hunyuan Video (ImageToVideo)", "properties": { "Node name for S&R": "TextEncodeHunyuanVideo_ImageToVideo" }, "widgets_values": [ "a cute anime girl with massive fennec ears and a big fluffy tail wearing a maid outfit running" ], "color": "#232", "bgcolor": "#353" }, { "id": 81, "type": "CLIPVisionEncode", "pos": [ 200, 530 ], "size": [ 253.60000610351562, 78 ], "flags": {}, "order": 11, "mode": 0, "inputs": [ { "name": "clip_vision", "type": "CLIP_VISION", "link": 220 }, { "name": "image", "type": "IMAGE", "link": 221 } ], "outputs": [ { "name": "CLIP_VISION_OUTPUT", "type": "CLIP_VISION_OUTPUT", "links": [ 219 ] } ], "properties": { "Node name for S&R": "CLIPVisionEncode" }, "widgets_values": [ "none" ] }, { "id": 89, "type": "Note", "pos": [ 190, 660 ], "size": [ 260, 210 ], "flags": {}, "order": 3, "mode": 0, "inputs": [], "outputs": [], "properties": {}, "widgets_values": [ "Select your image here. The image is fed to both the text encoder and directly to the model.\n\nYou can set the resolution and length of the video using the HunyuanImageToVideo node." ], "color": "#432", "bgcolor": "#653" }, { "id": 83, "type": "LoadImage", "pos": [ -190, 700 ], "size": [ 365.4132080078125, 471.8512268066406 ], "flags": {}, "order": 4, "mode": 0, "inputs": [], "outputs": [ { "name": "IMAGE", "type": "IMAGE", "links": [ 221, 222 ], "slot_index": 0 }, { "name": "MASK", "type": "MASK", "links": null } ], "properties": { "Node name for S&R": "LoadImage" }, "widgets_values": [ "flux_dev_example.png", "image" ] }, { "id": 82, "type": "CLIPVisionLoader", "pos": [ -190, 580 ], "size": [ 315, 58 ], "flags": {}, "order": 5, "mode": 0, "inputs": [], "outputs": [ { "name": "CLIP_VISION", "type": "CLIP_VISION", "links": [ 220 ] } ], "properties": { "Node name for S&R": "CLIPVisionLoader" }, "widgets_values": [ "llava_llama3_vision.safetensors" ], "color": "#223", "bgcolor": "#335" }, { "id": 10, "type": "VAELoader", "pos": [ -190, 470 ], "size": [ 350, 60 ], "flags": {}, "order": 6, "mode": 0, "inputs": [], "outputs": [ { "name": "VAE", "type": "VAE", "shape": 3, "links": [ 206, 211, 223 ], "slot_index": 0 } ], "properties": { "Node name for S&R": "VAELoader" }, "widgets_values": [ "hunyuan_video_vae_bf16.safetensors" ], "color": "#223", "bgcolor": "#335" }, { "id": 11, "type": "DualCLIPLoader", "pos": [ -190, 290 ], "size": [ 350, 122 ], "flags": {}, "order": 7, "mode": 0, "inputs": [], "outputs": [ { "name": "CLIP", "type": "CLIP", "shape": 3, "links": [ 224 ], "slot_index": 0 } ], "properties": { "Node name for S&R": "DualCLIPLoader" }, "widgets_values": [ "clip_l.safetensors", "llava_llama3_fp8_scaled.safetensors", "hunyuan_video", "default" ], "color": "#223", "bgcolor": "#335" }, { "id": 12, "type": "UNETLoader", "pos": [ -190, 160 ], "size": [ 404.6181640625, 82 ], "flags": {}, "order": 8, "mode": 0, "inputs": [], "outputs": [ { "name": "MODEL", "type": "MODEL", "shape": 3, "links": [ 190, 209 ], "slot_index": 0 } ], "properties": { "Node name for S&R": "UNETLoader" }, "widgets_values": [ "hunyuan_video_image_to_video_720p_bf16.safetensors", "default" ], "color": "#223", "bgcolor": "#335" }, { "id": 77, "type": "Note", "pos": [ -140, 0 ], "size": [ 350, 110 ], "flags": {}, "order": 9, "mode": 0, "inputs": [], "outputs": [], "properties": {}, "widgets_values": [ "Select a fp8 weight_dtype if you are running out of memory." ], "color": "#432", "bgcolor": "#653" }, { "id": 86, "type": "ImageFromBatch", "pos": [ 1410, 200 ], "size": [ 210, 82 ], "flags": {}, "order": 21, "mode": 0, "inputs": [ { "name": "image", "type": "IMAGE", "link": 230 } ], "outputs": [ { "name": "IMAGE", "type": "IMAGE", "links": [ 228, 229 ] } ], "properties": { "Node name for S&R": "ImageFromBatch" }, "widgets_values": [ 4, 4096 ] }, { "id": 25, "type": "RandomNoise", "pos": [ 520, 420 ], "size": [ 315, 82 ], "flags": {}, "order": 10, "mode": 0, "inputs": [], "outputs": [ { "name": "NOISE", "type": "NOISE", "shape": 3, "links": [ 37 ] } ], "properties": { "Node name for S&R": "RandomNoise" }, "widgets_values": [ 187330947843879, "randomize" ], "color": "#2a363b", "bgcolor": "#3f5159" } ], "links": [ [ 19, 16, 0, 13, 2, "SAMPLER" ], [ 20, 17, 0, 13, 3, "SIGMAS" ], [ 30, 22, 0, 13, 1, "GUIDER" ], [ 37, 25, 0, 13, 0, "NOISE" ], [ 129, 26, 0, 22, 1, "CONDITIONING" ], [ 181, 13, 0, 8, 0, "LATENT" ], [ 190, 12, 0, 17, 0, "MODEL" ], [ 195, 67, 0, 22, 0, "MODEL" ], [ 206, 10, 0, 8, 1, "VAE" ], [ 209, 12, 0, 67, 0, "MODEL" ], [ 210, 13, 0, 73, 0, "LATENT" ], [ 211, 10, 0, 73, 1, "VAE" ], [ 216, 78, 1, 13, 4, "LATENT" ], [ 218, 80, 0, 78, 0, "CONDITIONING" ], [ 219, 81, 0, 80, 1, "CLIP_VISION_OUTPUT" ], [ 220, 82, 0, 81, 0, "CLIP_VISION" ], [ 221, 83, 0, 81, 1, "IMAGE" ], [ 222, 83, 0, 78, 2, "IMAGE" ], [ 223, 10, 0, 78, 1, "VAE" ], [ 224, 11, 0, 80, 0, "CLIP" ], [ 225, 78, 0, 26, 0, "CONDITIONING" ], [ 228, 86, 0, 75, 0, "IMAGE" ], [ 229, 86, 0, 84, 0, "IMAGE" ], [ 230, 73, 0, 86, 0, "IMAGE" ] ], "groups": [], "config": {}, "extra": { "groupNodes": {} }, "models": [ { "name": "llava_llama3_vision.safetensors", "url": "https://huggingface.co/Comfy-Org/HunyuanVideo_repackaged/resolve/main/split_files/clip_vision/llava_llama3_vision.safetensors?download=true", "directory": "clip_vision" }, { "name": "clip_l.safetensors", "url": "https://huggingface.co/Comfy-Org/HunyuanVideo_repackaged/resolve/main/split_files/text_encoders/clip_l.safetensors?download=true", "directory": "text_encoders" }, { "name": "llava_llama3_fp8_scaled.safetensors", "url": "https://huggingface.co/Comfy-Org/HunyuanVideo_repackaged/resolve/main/split_files/text_encoders/llava_llama3_fp8_scaled.safetensors?download=true", "directory": "text_encoders" }, { "name": "hunyuan_video_vae_bf16.safetensors", "url": "https://huggingface.co/Comfy-Org/HunyuanVideo_repackaged/resolve/main/split_files/vae/hunyuan_video_vae_bf16.safetensors?download=true", "directory": "vae" }, { "name": "hunyuan_video_image_to_video_720p_bf16.safetensors", "url": "https://huggingface.co/Comfy-Org/HunyuanVideo_repackaged/resolve/main/split_files/diffusion_models/hunyuan_video_image_to_video_720p_bf16.safetensors?download=true", "directory": "diffusion_models" } ], "version": 0.4 }