workflows / comfyui-workflows1 /AnimateDiff+LoRAstack+Upscale-v04b.json
Luis-Filipe's picture
Upload 424 files
3339706 verified
{
"last_node_id": 181,
"last_link_id": 351,
"nodes": [
{
"id": 10,
"type": "VAEDecode",
"pos": [
1181.3890400078121,
-509.13167196875025
],
"size": {
"0": 210,
"1": 46
},
"flags": {},
"order": 33,
"mode": 0,
"inputs": [
{
"name": "samples",
"type": "LATENT",
"link": 294
},
{
"name": "vae",
"type": "VAE",
"link": 10
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"links": [
295,
300
],
"shape": 3,
"slot_index": 0
}
],
"properties": {
"Node name for S&R": "VAEDecode"
}
},
{
"id": 137,
"type": "CR Upscale Image",
"pos": [
1122.3890400078121,
-416.1316719687504
],
"size": {
"0": 318.5529479980469,
"1": 222
},
"flags": {},
"order": 34,
"mode": 0,
"inputs": [
{
"name": "image",
"type": "IMAGE",
"link": 295
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"links": [
296,
297
],
"shape": 3,
"slot_index": 0
},
{
"name": "show_help",
"type": "STRING",
"links": [],
"shape": 3,
"slot_index": 1
}
],
"properties": {
"Node name for S&R": "CR Upscale Image"
},
"widgets_values": [
"4x_foolhardy_Remacri.pth",
"rescale",
2,
1024,
"lanczos",
"true",
8
]
},
{
"id": 93,
"type": "ADE_AnimateDiffLoaderWithContext",
"pos": [
204.00000000000006,
-1687.5194515546877
],
"size": {
"0": 317.0839538574219,
"1": 315.5437316894531
},
"flags": {},
"order": 20,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 305
},
{
"name": "context_options",
"type": "CONTEXT_OPTIONS",
"link": 206,
"slot_index": 1
},
{
"name": "motion_lora",
"type": "MOTION_LORA",
"link": null
},
{
"name": "motion_model_settings",
"type": "MOTION_MODEL_SETTINGS",
"link": null
},
{
"name": "sample_settings",
"type": "sample_settings",
"link": null
}
],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
205
],
"shape": 3
}
],
"properties": {
"Node name for S&R": "ADE_AnimateDiffLoaderWithContext"
},
"widgets_values": [
"mm_sd_v15_v2.ckpt",
"sqrt_linear (AnimateDiff)",
1,
true
]
},
{
"id": 143,
"type": "CR Apply LoRA Stack",
"pos": [
298,
-674
],
"size": {
"0": 254.40000915527344,
"1": 66
},
"flags": {},
"order": 18,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 303
},
{
"name": "clip",
"type": "CLIP",
"link": 304
},
{
"name": "lora_stack",
"type": "LORA_STACK",
"link": 302
}
],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
305
],
"shape": 3,
"slot_index": 0
},
{
"name": "CLIP",
"type": "CLIP",
"links": [
306,
307
],
"shape": 3,
"slot_index": 1
},
{
"name": "show_help",
"type": "STRING",
"links": null,
"shape": 3
}
],
"properties": {
"Node name for S&R": "CR Apply LoRA Stack"
}
},
{
"id": 142,
"type": "CR LoRA Stack",
"pos": [
-201,
-584
],
"size": {
"0": 483.5227355957031,
"1": 385.16314697265625
},
"flags": {},
"order": 0,
"mode": 0,
"inputs": [
{
"name": "lora_stack",
"type": "LORA_STACK",
"link": null
}
],
"outputs": [
{
"name": "LORA_STACK",
"type": "LORA_STACK",
"links": [
302
],
"shape": 3,
"slot_index": 0
},
{
"name": "show_help",
"type": "STRING",
"links": null,
"shape": 3
}
],
"properties": {
"Node name for S&R": "CR LoRA Stack"
},
"widgets_values": [
"Off",
"None",
1,
1,
"Off",
"None",
1,
1,
"Off",
"None",
1,
1
]
},
{
"id": 173,
"type": "Reroute",
"pos": [
588.5432895156247,
-1725.417354304687
],
"size": [
75,
26
],
"flags": {},
"order": 25,
"mode": 0,
"inputs": [
{
"name": "",
"type": "*",
"link": 351
}
],
"outputs": [
{
"name": "",
"type": "IMAGE",
"links": [
325,
326
],
"slot_index": 0
}
],
"title": "RerouteInput Images",
"properties": {
"showOutputText": false,
"horizontal": false
}
},
{
"id": 12,
"type": "SaveImage",
"pos": [
1478,
-884
],
"size": {
"0": 330.0085754394531,
"1": 320.7124938964844
},
"flags": {},
"order": 36,
"mode": 0,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 296
}
],
"properties": {},
"widgets_values": [
"Images\\image"
]
},
{
"id": 140,
"type": "PreviewImage",
"pos": [
1142,
-859
],
"size": {
"0": 281.9832458496094,
"1": 296.9180603027344
},
"flags": {},
"order": 35,
"mode": 0,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 300
}
],
"title": "Preview pre upscale",
"properties": {
"Node name for S&R": ""
}
},
{
"id": 178,
"type": "Note",
"pos": [
834,
-903
],
"size": {
"0": 243.64602661132812,
"1": 127.71123504638672
},
"flags": {},
"order": 1,
"mode": 0,
"properties": {
"text": ""
},
"widgets_values": [
"Use Euler ancestral or DPM samplers for smoother and more stable animations.\n\nFind a comparison of samplers here:\nhttps://www.reddit.com/r/StableDiffusion/comments/15qmjxy/animatediff_sampler_stats/"
],
"color": "#222",
"bgcolor": "#000"
},
{
"id": 106,
"type": "PrimitiveNode",
"pos": [
598,
-874
],
"size": {
"0": 210,
"1": 82
},
"flags": {},
"order": 2,
"mode": 0,
"outputs": [
{
"name": "INT",
"type": "INT",
"links": [
213
],
"slot_index": 0,
"widget": {
"name": "seed"
}
}
],
"properties": {
"Run widget replace on values": false
},
"widgets_values": [
981423447252632,
"randomize"
]
},
{
"id": 170,
"type": "ControlNetLoaderAdvanced",
"pos": [
640,
-1640
],
"size": {
"0": 381.58392333984375,
"1": 58
},
"flags": {},
"order": 3,
"mode": 0,
"inputs": [
{
"name": "timestep_keyframe",
"type": "TIMESTEP_KEYFRAME",
"link": null
}
],
"outputs": [
{
"name": "CONTROL_NET",
"type": "CONTROL_NET",
"links": [
337
],
"shape": 3,
"slot_index": 0
}
],
"properties": {
"Node name for S&R": "ControlNetLoaderAdvanced"
},
"widgets_values": [
"control_v11p_sd15_openpose.pth"
],
"color": "#2a363b",
"bgcolor": "#3f5159"
},
{
"id": 72,
"type": "ControlNetApplyAdvanced",
"pos": [
622,
-1243
],
"size": {
"0": 427.7201232910156,
"1": 239.03790283203125
},
"flags": {},
"order": 30,
"mode": 0,
"inputs": [
{
"name": "positive",
"type": "CONDITIONING",
"link": 132
},
{
"name": "negative",
"type": "CONDITIONING",
"link": 133
},
{
"name": "control_net",
"type": "CONTROL_NET",
"link": 337
},
{
"name": "image",
"type": "IMAGE",
"link": 341
}
],
"outputs": [
{
"name": "positive",
"type": "CONDITIONING",
"links": [
342
],
"shape": 3,
"slot_index": 0
},
{
"name": "negative",
"type": "CONDITIONING",
"links": [
343
],
"shape": 3,
"slot_index": 1
}
],
"properties": {
"Node name for S&R": "ControlNetApplyAdvanced"
},
"widgets_values": [
0.5,
0,
1
],
"color": "#2a363b",
"bgcolor": "#3f5159"
},
{
"id": 177,
"type": "PreviewImage",
"pos": [
1820,
-1640
],
"size": {
"0": 321.4112548828125,
"1": 246
},
"flags": {},
"order": 28,
"mode": 0,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 349
}
],
"properties": {
"Node name for S&R": "PreviewImage"
}
},
{
"id": 176,
"type": "ControlNetApplyAdvanced",
"pos": [
1346,
-1244
],
"size": {
"0": 427.7201232910156,
"1": 239.03790283203125
},
"flags": {},
"order": 31,
"mode": 0,
"inputs": [
{
"name": "positive",
"type": "CONDITIONING",
"link": 342
},
{
"name": "negative",
"type": "CONDITIONING",
"link": 343
},
{
"name": "control_net",
"type": "CONTROL_NET",
"link": 345
},
{
"name": "image",
"type": "IMAGE",
"link": 344
}
],
"outputs": [
{
"name": "positive",
"type": "CONDITIONING",
"links": [
347
],
"shape": 3,
"slot_index": 0
},
{
"name": "negative",
"type": "CONDITIONING",
"links": [
348
],
"shape": 3,
"slot_index": 1
}
],
"properties": {
"Node name for S&R": "ControlNetApplyAdvanced"
},
"widgets_values": [
0.7000000000000001,
0,
1
],
"color": "#2a363b",
"bgcolor": "#3f5159"
},
{
"id": 160,
"type": "Note",
"pos": [
1813,
-1263
],
"size": {
"0": 416.1223449707031,
"1": 266.65899658203125
},
"flags": {},
"order": 4,
"mode": 0,
"properties": {
"text": ""
},
"widgets_values": [
"You must install 'ComfyUI's ControlNet Auxiliary Preprocessors' custom nodes manually or via the Manager (->Install Missing Custom Nodes) \nmore info here: https://github.com/Fannovel16/comfyui_controlnet_aux\n\nLineart and HED Lines work best for faces/expressions and if you want to keep details of the input sequence.\n\nOpenPose works best for full body movement (i.e. dancing, walking etc)\n\nAs shown here, you can mix multiple ControlNets for the cost of longer render time and higher usage of VRAM.\n\nControlNet examples:\nhttps://comfyanonymous.github.io/ComfyUI_examples/controlnet/"
],
"color": "#222",
"bgcolor": "#000"
},
{
"id": 94,
"type": "ADE_AnimateDiffUniformContextOptions",
"pos": [
-187.99999999999994,
-1691.5194515546877
],
"size": {
"0": 315,
"1": 154
},
"flags": {
"collapsed": false
},
"order": 5,
"mode": 0,
"outputs": [
{
"name": "CONTEXT_OPTIONS",
"type": "CONTEXT_OPTIONS",
"links": [
206
],
"shape": 3
}
],
"properties": {
"Node name for S&R": "ADE_AnimateDiffUniformContextOptions"
},
"widgets_values": [
16,
1,
6,
"uniform",
false
]
},
{
"id": 156,
"type": "Note",
"pos": [
272,
-1237
],
"size": {
"0": 240.375,
"1": 181.19326782226562
},
"flags": {},
"order": 6,
"mode": 0,
"properties": {
"text": ""
},
"widgets_values": [
"If you get strange results try to reduce the length of your positive & negative prompts. Keep your prompts below 75 tokens.\n\nThe negative prompt should not be longer than the positive prompt!"
],
"color": "#222",
"bgcolor": "#000"
},
{
"id": 50,
"type": "PreviewImage",
"pos": [
-596.9493750961091,
-1677.8411773868754
],
"size": {
"0": 350.0137634277344,
"1": 326.18408203125
},
"flags": {},
"order": 23,
"mode": 0,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 86
}
],
"properties": {
"Node name for S&R": "PreviewImage"
}
},
{
"id": 96,
"type": "PrimitiveNode",
"pos": [
-670,
-970
],
"size": {
"0": 210,
"1": 82
},
"flags": {},
"order": 7,
"mode": 0,
"outputs": [
{
"name": "INT",
"type": "INT",
"links": [
209
],
"slot_index": 0,
"widget": {
"name": "width"
}
}
],
"title": "Width",
"properties": {
"Run widget replace on values": false
},
"widgets_values": [
512,
"fixed"
],
"color": "#332922",
"bgcolor": "#593930"
},
{
"id": 97,
"type": "PrimitiveNode",
"pos": [
-449,
-969
],
"size": {
"0": 210,
"1": 82
},
"flags": {},
"order": 8,
"mode": 0,
"outputs": [
{
"name": "INT",
"type": "INT",
"links": [
210
],
"slot_index": 0,
"widget": {
"name": "height"
}
}
],
"title": "Height",
"properties": {
"Run widget replace on values": false
},
"widgets_values": [
512,
"fixed"
],
"color": "#332922",
"bgcolor": "#593930"
},
{
"id": 56,
"type": "VAEEncode",
"pos": [
-491,
-611
],
"size": {
"0": 241.0370635986328,
"1": 81.81757354736328
},
"flags": {},
"order": 24,
"mode": 0,
"inputs": [
{
"name": "pixels",
"type": "IMAGE",
"link": 87
},
{
"name": "vae",
"type": "VAE",
"link": 88
}
],
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"links": [
208
],
"shape": 3,
"slot_index": 0
}
],
"properties": {
"Node name for S&R": "VAEEncode"
}
},
{
"id": 1,
"type": "CheckpointLoaderSimpleWithNoiseSelect",
"pos": [
-1052,
-842
],
"size": {
"0": 802.761474609375,
"1": 182.4549102783203
},
"flags": {},
"order": 9,
"mode": 0,
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
303
],
"shape": 3,
"slot_index": 0
},
{
"name": "CLIP",
"type": "CLIP",
"links": [
304
],
"shape": 3,
"slot_index": 1
},
{
"name": "VAE",
"type": "VAE",
"links": null,
"shape": 3,
"slot_index": 2
}
],
"properties": {
"Node name for S&R": "CheckpointLoaderSimpleWithNoiseSelect"
},
"widgets_values": [
"Illu\\revAnimated_v11.safetensors",
"sqrt_linear (AnimateDiff)",
false,
1
],
"color": "#432",
"bgcolor": "#653"
},
{
"id": 53,
"type": "ImageScale",
"pos": [
-595.4545454545456,
-1170.9090909090908
],
"size": {
"0": 351.3953857421875,
"1": 161.42391967773438
},
"flags": {},
"order": 19,
"mode": 0,
"inputs": [
{
"name": "image",
"type": "IMAGE",
"link": 318
},
{
"name": "width",
"type": "INT",
"link": 209,
"widget": {
"name": "width"
},
"slot_index": 1
},
{
"name": "height",
"type": "INT",
"link": 210,
"widget": {
"name": "height"
},
"slot_index": 2
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"links": [
86,
87,
351
],
"shape": 3,
"slot_index": 0
}
],
"properties": {
"Node name for S&R": "ImageScale"
},
"widgets_values": [
"nearest-exact",
512,
512,
"center"
]
},
{
"id": 157,
"type": "Note",
"pos": [
300,
-559
],
"size": {
"0": 250.90220642089844,
"1": 351.9985046386719
},
"flags": {},
"order": 10,
"mode": 0,
"properties": {
"text": ""
},
"widgets_values": [
"You can add any LoRA to the stack. There are also 'Motion LoRAs' for camera control, but they don't make much sense here, as we want the result to be somewhat close to the source.\n\nBut sometimes I got interesting effects from using them so I suggest you just try by yourself. \n\nYou can download the motion LoRAs here:\n\nhttps://huggingface.co/conrevo/AnimateDiff-A1111/tree/main/lora"
],
"color": "#222",
"bgcolor": "#000"
},
{
"id": 153,
"type": "Note",
"pos": [
-1072,
-1032
],
"size": {
"0": 389.503662109375,
"1": 147.4991455078125
},
"flags": {},
"order": 11,
"mode": 0,
"properties": {
"text": ""
},
"widgets_values": [
"You need to prepare a folder that contains an image sequence and load it here.\n\nUse this tool to create the sequence from a video or gif file https://www.shutterencoder.com\n\nThe SD1.5 motion modules are trained with 16 frames, so it’ll give the best results when the number of frames is 16 - but it should also work for longer animations."
],
"color": "#222",
"bgcolor": "#000"
},
{
"id": 181,
"type": "Note",
"pos": [
-597,
-1307
],
"size": {
"0": 348.4784240722656,
"1": 92.85890197753906
},
"flags": {},
"order": 12,
"mode": 0,
"properties": {
"text": ""
},
"widgets_values": [
"The images are scaled (and optionally cropped) to the correct size of 512x512px here. \n\nAnimateDiff is trained on 512x512 images so it works best with 512x512 output."
],
"color": "#222",
"bgcolor": "#000"
},
{
"id": 159,
"type": "Note",
"pos": [
-195,
-1476
],
"size": {
"0": 373.4175109863281,
"1": 105.35885620117188
},
"flags": {},
"order": 13,
"mode": 0,
"properties": {
"text": ""
},
"widgets_values": [
"Download the models here\nhttps://huggingface.co/guoyww/animatediff/tree/main\n\nand drop them in\n[ComfyUI folder]/custom_nodes/comfyui-animatediff/models/"
],
"color": "#222",
"bgcolor": "#000"
},
{
"id": 103,
"type": "VHS_LoadImages",
"pos": [
-1059.9493750961087,
-1683.8411773868754
],
"size": [
443.2850646972656,
611.2850646972656
],
"flags": {},
"order": 14,
"mode": 0,
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"links": [
318
],
"shape": 3,
"slot_index": 0
},
{
"name": "MASK",
"type": "MASK",
"links": null,
"shape": 3
},
{
"name": "INT",
"type": "INT",
"links": null,
"shape": 3
}
],
"properties": {
"Node name for S&R": "VHS_LoadImages"
},
"widgets_values": {
"directory": "lol16",
"image_load_cap": 0,
"skip_first_images": 0,
"select_every_nth": 1,
"choose folder to upload": 1,
"videopreview": {
"hidden": false,
"paused": false,
"params": {
"frame_load_cap": 0,
"skip_first_images": 0,
"filename": "lol16",
"type": "input",
"format": "folder",
"select_every_nth": 1
}
}
},
"color": "#323",
"bgcolor": "#535"
},
{
"id": 104,
"type": "VHS_VideoCombine",
"pos": [
1822.9135430390625,
-879.8344292578128
],
"size": [
429.1750793457031,
673.7808078424341
],
"flags": {},
"order": 37,
"mode": 0,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 297
},
{
"name": "audio",
"type": "VHS_AUDIO",
"link": null
}
],
"outputs": [
{
"name": "Filenames",
"type": "VHS_FILENAMES",
"links": null,
"shape": 3
}
],
"properties": {
"Node name for S&R": "VHS_VideoCombine"
},
"widgets_values": {
"frame_rate": 12,
"loop_count": 0,
"filename_prefix": "AnimateDiff",
"format": "video/h264-mp4",
"pix_fmt": "yuv420p",
"crf": 19,
"save_metadata": true,
"pingpong": true,
"save_output": true,
"videopreview": {
"hidden": false,
"paused": false,
"params": {
"filename": "AnimateDiff_00180.mp4",
"subfolder": "",
"type": "output",
"format": "video/h264-mp4"
}
}
}
},
{
"id": 6,
"type": "CLIPTextEncode",
"pos": [
78.43446388127765,
-986.5605387679963
],
"size": {
"0": 412.9415283203125,
"1": 186.01625061035156
},
"flags": {},
"order": 21,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 306
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"links": [
133
],
"shape": 3,
"slot_index": 0
}
],
"properties": {
"Node name for S&R": "CLIPTextEncode"
},
"widgets_values": [
"bad quality,blurry, monochrome, (mutation:1.2), (deformed:1.2), ugly, blurry"
],
"color": "#322",
"bgcolor": "#533"
},
{
"id": 119,
"type": "ControlNetLoaderAdvanced",
"pos": [
1464,
-1636
],
"size": {
"0": 327.6000061035156,
"1": 80.41925048828125
},
"flags": {},
"order": 15,
"mode": 0,
"inputs": [
{
"name": "timestep_keyframe",
"type": "TIMESTEP_KEYFRAME",
"link": null
}
],
"outputs": [
{
"name": "CONTROL_NET",
"type": "CONTROL_NET",
"links": [
345
],
"shape": 3,
"slot_index": 0
}
],
"properties": {
"Node name for S&R": "ControlNetLoaderAdvanced"
},
"widgets_values": [
"control_v11p_sd15_softedge.pth"
],
"color": "#2a363b",
"bgcolor": "#3f5159"
},
{
"id": 121,
"type": "HEDPreprocessor",
"pos": [
1469,
-1510
],
"size": {
"0": 321.9792785644531,
"1": 118.91194152832031
},
"flags": {},
"order": 26,
"mode": 0,
"inputs": [
{
"name": "image",
"type": "IMAGE",
"link": 325
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"links": [
344,
349
],
"shape": 3,
"slot_index": 0
}
],
"properties": {
"Node name for S&R": "HEDPreprocessor"
},
"widgets_values": [
"enable",
512
]
},
{
"id": 2,
"type": "VAELoader",
"pos": [
-1050,
-610
],
"size": {
"0": 527.2886962890625,
"1": 82.21073150634766
},
"flags": {},
"order": 16,
"mode": 0,
"outputs": [
{
"name": "VAE",
"type": "VAE",
"links": [
10,
88
],
"shape": 3,
"slot_index": 0
}
],
"properties": {
"Node name for S&R": "VAELoader"
},
"widgets_values": [
"anythingKlF8Anime2VaeFtMse840000_klF8Anime2.safetensors"
],
"color": "#332922",
"bgcolor": "#593930"
},
{
"id": 92,
"type": "PreviewImage",
"pos": [
1099,
-1626
],
"size": {
"0": 315.4920654296875,
"1": 251.81893920898438
},
"flags": {},
"order": 29,
"mode": 0,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 335
}
],
"properties": {
"Node name for S&R": "PreviewImage"
}
},
{
"id": 171,
"type": "OpenposePreprocessor",
"pos": [
645,
-1538
],
"size": {
"0": 378.2839050292969,
"1": 150
},
"flags": {},
"order": 27,
"mode": 0,
"inputs": [
{
"name": "image",
"type": "IMAGE",
"link": 326
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"links": [
335,
341
],
"shape": 3,
"slot_index": 0
},
{
"name": "POSE_KEYPOINT",
"type": "POSE_KEYPOINT",
"links": null,
"shape": 3
}
],
"properties": {
"Node name for S&R": "OpenposePreprocessor"
},
"widgets_values": [
"disable",
"enable",
"disable",
512
]
},
{
"id": 174,
"type": "Note",
"pos": [
1484,
-509
],
"size": {
"0": 318.6959228515625,
"1": 300.1888427734375
},
"flags": {},
"order": 17,
"mode": 0,
"properties": {
"text": ""
},
"widgets_values": [
""
],
"color": "#222",
"bgcolor": "#000"
},
{
"id": 3,
"type": "CLIPTextEncode",
"pos": [
-191.56553611872263,
-1236.5605387679964
],
"size": {
"0": 424.5931396484375,
"1": 197.06280517578125
},
"flags": {},
"order": 22,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 307
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"links": [
132
],
"shape": 3,
"slot_index": 0
}
],
"properties": {
"Node name for S&R": "CLIPTextEncode"
},
"widgets_values": [
"(fairy princess:1.2),hi-tech,vivid hues,digital painting"
],
"color": "#232",
"bgcolor": "#353"
},
{
"id": 7,
"type": "KSampler",
"pos": [
604.6292581953124,
-731.8115887685935
],
"size": {
"0": 464.5998840332031,
"1": 535.4624633789062
},
"flags": {
"pinned": false
},
"order": 32,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 205
},
{
"name": "positive",
"type": "CONDITIONING",
"link": 347
},
{
"name": "negative",
"type": "CONDITIONING",
"link": 348
},
{
"name": "latent_image",
"type": "LATENT",
"link": 208
},
{
"name": "seed",
"type": "INT",
"link": 213,
"widget": {
"name": "seed"
}
}
],
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"links": [
294
],
"shape": 3,
"slot_index": 0
}
],
"properties": {
"Node name for S&R": "KSampler"
},
"widgets_values": [
981423447252632,
"fixed",
20,
7,
"dpm_2_ancestral",
"karras",
1
],
"color": "#223",
"bgcolor": "#335"
}
],
"links": [
[
10,
2,
0,
10,
1,
"VAE"
],
[
86,
53,
0,
50,
0,
"IMAGE"
],
[
87,
53,
0,
56,
0,
"IMAGE"
],
[
88,
2,
0,
56,
1,
"VAE"
],
[
132,
3,
0,
72,
0,
"CONDITIONING"
],
[
133,
6,
0,
72,
1,
"CONDITIONING"
],
[
205,
93,
0,
7,
0,
"MODEL"
],
[
206,
94,
0,
93,
1,
"CONTEXT_OPTIONS"
],
[
208,
56,
0,
7,
3,
"LATENT"
],
[
209,
96,
0,
53,
1,
"INT"
],
[
210,
97,
0,
53,
2,
"INT"
],
[
213,
106,
0,
7,
4,
"INT"
],
[
294,
7,
0,
10,
0,
"LATENT"
],
[
295,
10,
0,
137,
0,
"IMAGE"
],
[
296,
137,
0,
12,
0,
"IMAGE"
],
[
297,
137,
0,
104,
0,
"IMAGE"
],
[
300,
10,
0,
140,
0,
"IMAGE"
],
[
302,
142,
0,
143,
2,
"LORA_STACK"
],
[
303,
1,
0,
143,
0,
"MODEL"
],
[
304,
1,
1,
143,
1,
"CLIP"
],
[
305,
143,
0,
93,
0,
"MODEL"
],
[
306,
143,
1,
6,
0,
"CLIP"
],
[
307,
143,
1,
3,
0,
"CLIP"
],
[
318,
103,
0,
53,
0,
"IMAGE"
],
[
325,
173,
0,
121,
0,
"IMAGE"
],
[
326,
173,
0,
171,
0,
"IMAGE"
],
[
335,
171,
0,
92,
0,
"IMAGE"
],
[
337,
170,
0,
72,
2,
"CONTROL_NET"
],
[
341,
171,
0,
72,
3,
"IMAGE"
],
[
342,
72,
0,
176,
0,
"CONDITIONING"
],
[
343,
72,
1,
176,
1,
"CONDITIONING"
],
[
344,
121,
0,
176,
3,
"IMAGE"
],
[
345,
119,
0,
176,
2,
"CONTROL_NET"
],
[
347,
176,
0,
7,
1,
"CONDITIONING"
],
[
348,
176,
1,
7,
2,
"CONDITIONING"
],
[
349,
121,
0,
177,
0,
"IMAGE"
],
[
351,
53,
0,
173,
0,
"*"
]
],
"groups": [
{
"title": "Inputs",
"bounding": [
-1083,
-1776,
853,
1267
],
"color": "#8A8",
"font_size": 24
},
{
"title": "Outputs",
"bounding": [
1466,
-961,
792,
778
],
"color": "#8AA",
"font_size": 24
},
{
"title": "ControlNet",
"bounding": [
577,
-1775,
1681,
805
],
"color": "#b58b2a",
"font_size": 24
},
{
"title": "Prompt",
"bounding": [
-221,
-1335,
788,
562
],
"color": "#a1309b",
"font_size": 24
},
{
"title": "Animate Diff Nodes",
"bounding": [
-220,
-1777,
790,
432
],
"color": "#3f789e",
"font_size": 24
},
{
"title": "LoRA",
"bounding": [
-223,
-766,
791,
586
],
"color": "#3f789e",
"font_size": 24
},
{
"title": "Sampler",
"bounding": [
575,
-957,
519,
777
],
"color": "#A88",
"font_size": 24
},
{
"title": "Upscale",
"bounding": [
1100,
-959,
361,
778
],
"color": "#3f789e",
"font_size": 24
},
{
"title": "Group",
"bounding": [
-1082,
-500,
851,
321
],
"color": "#444",
"font_size": 24
}
],
"config": {},
"extra": {},
"version": 0.4
}