| { |
| "id": 991916, |
| "modelId": 502468, |
| "name": "v2.0", |
| "createdAt": "2024-10-25T03:50:28.686Z", |
| "updatedAt": "2024-10-25T04:07:13.500Z", |
| "status": "Published", |
| "publishedAt": "2024-10-25T04:07:13.492Z", |
| "trainedWords": [], |
| "trainingStatus": null, |
| "trainingDetails": null, |
| "baseModel": "SDXL 1.0", |
| "baseModelType": "Standard", |
| "earlyAccessEndsAt": null, |
| "earlyAccessConfig": null, |
| "description": "<p>Version 2</p>", |
| "uploadType": "Created", |
| "usageControl": "Download", |
| "air": "urn:air:sdxl:checkpoint:civitai:502468@991916", |
| "stats": { |
| "downloadCount": 15832, |
| "ratingCount": 0, |
| "rating": 0, |
| "thumbsUpCount": 838 |
| }, |
| "model": { |
| "name": "bigASP \ud83d\udc0d", |
| "type": "Checkpoint", |
| "nsfw": true, |
| "poi": false |
| }, |
| "files": [ |
| { |
| "id": 897883, |
| "sizeKB": 6775430.353515625, |
| "name": "bigasp_v20.safetensors", |
| "type": "Model", |
| "pickleScanResult": "Success", |
| "pickleScanMessage": "No Pickle imports", |
| "virusScanResult": "Success", |
| "virusScanMessage": null, |
| "scannedAt": "2024-10-25T04:03:20.609Z", |
| "metadata": { |
| "format": "SafeTensor", |
| "size": "pruned", |
| "fp": "fp16" |
| }, |
| "hashes": { |
| "AutoV1": "7731E63E", |
| "AutoV2": "6C77BE501B", |
| "SHA256": "6C77BE501B95DA35528431C224CCFB51B0AE948857431B3556DB64E213CC1EDF", |
| "CRC32": "F496D401", |
| "BLAKE3": "33B82FAACF8E7848F4B9C124256AACAEA54F256EC5CF9679E37A4D22E2930D29", |
| "AutoV3": "55EC553D7B23" |
| }, |
| "primary": true, |
| "downloadUrl": "https://civitai.com/api/download/models/991916" |
| } |
| ], |
| "images": [ |
| { |
| "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/88c910ee-0c3b-4593-9b58-5b80bcdb1699/width=1024/36427260.jpeg", |
| "nsfwLevel": 1, |
| "width": 1024, |
| "height": 1024, |
| "hash": "UFEo71}@=x^jIT=|w]kB%NITIUIpivxat7sm", |
| "type": "image", |
| "metadata": { |
| "hash": "UFEo71}@=x^jIT=|w]kB%NITIUIpivxat7sm", |
| "size": 1761592, |
| "width": 1024, |
| "height": 1024 |
| }, |
| "meta": { |
| "seed": 219058663085791, |
| "vaes": [], |
| "Model": "cuwm1gxo-complete-20241024a", |
| "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 219058663085791, \"steps\": 40, \"cfg\": 3.0, \"sampler_name\": \"dpmpp_2m_sde_gpu\", \"scheduler\": \"karras\", \"denoise\": 1.0, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\"}, \"4\": {\"inputs\": {\"ckpt_name\": \"cuwm1gxo-complete-20241024a.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\"}, \"5\": {\"inputs\": {\"width\": 1024, \"height\": 1024, \"batch_size\": 16}, \"class_type\": \"EmptyLatentImage\"}, \"6\": {\"inputs\": {\"text\": \"score_7_up, A stunning photograph of a snake sitting in front of a roaring fireplace in a cozy cabin. The snake is on a fur rug as the warm light from the fire glows around it. The dim room is lit only by the flames. Depth of field highlights the subject. The photo has a cozy yet sensual feel to it. The composition of the photo is perfectly balanced. The title text \\\"BigAsp v2\\\" appears in bold font at the top center of the image.\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"7\": {\"inputs\": {\"text\": \"score_1, score_2, score_3\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\"}, \"9\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\"}, \"10\": {\"inputs\": {\"model_type\": \"SDXL\", \"steps\": 40, \"denoise\": 1}, \"class_type\": \"AlignYourStepsScheduler\"}, \"12\": {\"inputs\": {\"add_noise\": true, \"noise_seed\": 485400483572008, \"cfg\": 5, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"sampler\": [\"13\", 0], \"sigmas\": [\"10\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"SamplerCustom\"}, \"13\": {\"inputs\": {\"sampler_name\": \"dpmpp_3m_sde_gpu\"}, \"class_type\": \"KSamplerSelect\"}, \"16\": {\"inputs\": {\"model_name\": \"bbox/face_yolov8m.pt\"}, \"class_type\": \"UltralyticsDetectorProvider\"}, \"17\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\"}, \"class_type\": \"SaveImage\"}, \"18\": {\"inputs\": {\"scale\": 2, \"model\": [\"4\", 0]}, \"class_type\": \"PerturbedAttentionGuidance\"}}, \"workflow\": {\"last_node_id\": 18, \"last_link_id\": 40, \"nodes\": [{\"id\": 3, \"type\": \"KSampler\", \"pos\": [868, 167], \"size\": {\"0\": 315, \"1\": 262}, \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 40}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [17], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [219058663085791, \"randomize\", 40, 3, \"dpmpp_2m_sde_gpu\", \"karras\", 1]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [9, 371], \"size\": {\"0\": 315, \"1\": 98}, \"flags\": {}, \"order\": 4, \"mode\": 0, \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [10, 19, 26, 40], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [3, 5, 21], \"slot_index\": 1}, {\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [8, 20], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"CheckpointLoaderSimple\"}, \"widgets_values\": [\"cuwm1gxo-complete-20241024a.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [473, 609], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 0, \"mode\": 0, \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [2, 15], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"EmptyLatentImage\"}, \"widgets_values\": [1024, 1024, 16]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [415, 186], \"size\": {\"0\": 422.84503173828125, \"1\": 164.31304931640625}, \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 3}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4, 11, 22], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_7_up, A stunning photograph of a snake sitting in front of a roaring fireplace in a cozy cabin. The snake is on a fur rug as the warm light from the fire glows around it. The dim room is lit only by the flames. Depth of field highlights the subject. The photo has a cozy yet sensual feel to it. The composition of the photo is perfectly balanced. The title text \\\"BigAsp v2\\\" appears in bold font at the top center of the image.\"]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [413, 389], \"size\": {\"0\": 425.27801513671875, \"1\": 180.6060791015625}, \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 5}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [6, 12, 23], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_1, score_2, score_3\"]}, {\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1209, 188], \"size\": {\"0\": 210, \"1\": 46}, \"flags\": {}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 17}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9, 18], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1451, 189], \"size\": {\"0\": 210, \"1\": 270}, \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 10, \"type\": \"AlignYourStepsScheduler\", \"pos\": [-6, 888], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 1, \"mode\": 0, \"outputs\": [{\"name\": \"SIGMAS\", \"type\": \"SIGMAS\", \"links\": [13], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"AlignYourStepsScheduler\"}, \"widgets_values\": [\"SDXL\", 40, 1]}, {\"id\": 12, \"type\": \"SamplerCustom\", \"pos\": [418, 820], \"size\": {\"0\": 355.20001220703125, \"1\": 230}, \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 10}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 11}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 12}, {\"name\": \"sampler\", \"type\": \"SAMPLER\", \"link\": 14}, {\"name\": \"sigmas\", \"type\": \"SIGMAS\", \"link\": 13}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 15}], \"outputs\": [{\"name\": \"output\", \"type\": \"LATENT\", \"links\": [], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"denoised_output\", \"type\": \"LATENT\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"SamplerCustom\"}, \"widgets_values\": [true, 485400483572008, \"randomize\", 5]}, {\"id\": 13, \"type\": \"KSamplerSelect\", \"pos\": [7, 738], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 2, \"mode\": 0, \"outputs\": [{\"name\": \"SAMPLER\", \"type\": \"SAMPLER\", \"links\": [14], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"KSamplerSelect\"}, \"widgets_values\": [\"dpmpp_3m_sde_gpu\"]}, {\"id\": 15, \"type\": \"FaceDetailer\", \"pos\": [1139, 602], \"size\": {\"0\": 506.4000244140625, \"1\": 900}, \"flags\": {}, \"order\": 12, \"mode\": 2, \"inputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"link\": 18}, {\"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 21}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 20}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 22}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 23}, {\"name\": \"bbox_detector\", \"type\": \"BBOX_DETECTOR\", \"link\": 24}, {\"name\": \"sam_model_opt\", \"type\": \"SAM_MODEL\", \"link\": null}, {\"name\": \"segm_detector_opt\", \"type\": \"SEGM_DETECTOR\", \"link\": null}, {\"name\": \"detailer_hook\", \"type\": \"DETAILER_HOOK\", \"link\": null}, {\"name\": \"scheduler_func_opt\", \"type\": \"SCHEDULER_FUNC\", \"link\": null}], \"outputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"links\": [25], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"cropped_refined\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"cropped_enhanced_alpha\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"mask\", \"type\": \"MASK\", \"links\": null, \"shape\": 3}, {\"name\": \"detailer_pipe\", \"type\": \"DETAILER_PIPE\", \"links\": null, \"shape\": 3}, {\"name\": \"cnet_images\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}], \"properties\": {\"Node name for S&R\": \"FaceDetailer\"}, \"widgets_values\": [384, true, 1024, 429362487444475, \"randomize\", 20, 8, \"euler\", \"normal\", 0.5, 5, true, true, 0.5, 10, 3, \"center-1\", 0, 0.93, 0, 0.7, \"False\", 10, \"\", 1, false, 20]}, {\"id\": 16, \"type\": \"UltralyticsDetectorProvider\", \"pos\": [788, 796], \"size\": {\"0\": 315, \"1\": 78}, \"flags\": {}, \"order\": 3, \"mode\": 0, \"outputs\": [{\"name\": \"BBOX_DETECTOR\", \"type\": \"BBOX_DETECTOR\", \"links\": [24], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"SEGM_DETECTOR\", \"type\": \"SEGM_DETECTOR\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"UltralyticsDetectorProvider\"}, \"widgets_values\": [\"bbox/face_yolov8m.pt\"]}, {\"id\": 17, \"type\": \"SaveImage\", \"pos\": [1744, 566], \"size\": {\"0\": 315, \"1\": 270}, \"flags\": {}, \"order\": 13, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 25}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 18, \"type\": \"PerturbedAttentionGuidance\", \"pos\": [468, 87], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 26}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"PerturbedAttentionGuidance\"}, \"widgets_values\": [2]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [3, 4, 1, 6, 0, \"CLIP\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [5, 4, 1, 7, 0, \"CLIP\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [10, 4, 0, 12, 0, \"MODEL\"], [11, 6, 0, 12, 1, \"CONDITIONING\"], [12, 7, 0, 12, 2, \"CONDITIONING\"], [13, 10, 0, 12, 4, \"SIGMAS\"], [14, 13, 0, 12, 3, \"SAMPLER\"], [15, 5, 0, 12, 5, \"LATENT\"], [17, 3, 0, 8, 0, \"LATENT\"], [18, 8, 0, 15, 0, \"IMAGE\"], [19, 4, 0, 15, 1, \"MODEL\"], [20, 4, 2, 15, 3, \"VAE\"], [21, 4, 1, 15, 2, \"CLIP\"], [22, 6, 0, 15, 4, \"CONDITIONING\"], [23, 7, 0, 15, 5, \"CONDITIONING\"], [24, 16, 0, 15, 6, \"BBOX_DETECTOR\"], [25, 15, 0, 17, 0, \"IMAGE\"], [26, 4, 0, 18, 0, \"MODEL\"], [40, 4, 0, 3, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 1.2100000000000006, \"offset\": [29.157246416591963, -66.9923150654007]}, \"workspace_info\": {\"id\": \"TNOwVzKHGn6FUMraMM-Y7\"}}, \"version\": 0.4}}", |
| "steps": 40, |
| "width": 1024, |
| "height": 1024, |
| "models": [ |
| "cuwm1gxo-complete-20241024a.safetensors" |
| ], |
| "prompt": "score_7_up, A stunning photograph of a snake sitting in front of a roaring fireplace in a cozy cabin. The snake is on a fur rug as the warm light from the fire glows around it. The dim room is lit only by the flames. Depth of field highlights the subject. The photo has a cozy yet sensual feel to it. The composition of the photo is perfectly balanced. The title text \"BigAsp v2\" appears in bold font at the top center of the image.", |
| "denoise": 1, |
| "sampler": "dpmpp_2m_sde_gpu", |
| "cfgScale": 3, |
| "modelIds": [], |
| "scheduler": "karras", |
| "upscalers": [], |
| "versionIds": [], |
| "controlNets": [], |
| "negativePrompt": "score_1, score_2, score_3", |
| "additionalResources": [] |
| }, |
| "availability": "Public", |
| "hasMeta": true, |
| "hasPositivePrompt": true, |
| "onSite": false, |
| "remixOfId": null |
| }, |
| { |
| "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/bed4b260-8b31-4204-bc06-c7f61f929b8c/width=1024/36427371.jpeg", |
| "nsfwLevel": 2, |
| "width": 1024, |
| "height": 1024, |
| "hash": "U9G8.l_N00DO5+Mx}[J68_RjtlRP~qIUD%-;", |
| "type": "image", |
| "metadata": { |
| "hash": "U9G8.l_N00DO5+Mx}[J68_RjtlRP~qIUD%-;", |
| "size": 1577930, |
| "width": 1024, |
| "height": 1024 |
| }, |
| "meta": { |
| "seed": 831882560098235, |
| "vaes": [], |
| "Model": "cuwm1gxo-complete-20241024a", |
| "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 831882560098235, \"steps\": 40, \"cfg\": 2.0, \"sampler_name\": \"dpmpp_2m_sde_gpu\", \"scheduler\": \"karras\", \"denoise\": 1.0, \"model\": [\"18\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\"}, \"4\": {\"inputs\": {\"ckpt_name\": \"cuwm1gxo-complete-20241024a.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\"}, \"5\": {\"inputs\": {\"width\": 1024, \"height\": 1024, \"batch_size\": 16}, \"class_type\": \"EmptyLatentImage\"}, \"6\": {\"inputs\": {\"text\": \"score_7_up, Photograph of a woman wearing a red cocktail dress that fits tightly to her feminine curves, showing off her ample bosom and wide hips. The lighting in the photograph is professional with soft shadows adding depth. The background is dim and blurry, showing a fancy event, keeping the focus on the beautiful woman. Her curly brown hair falls past her shoulders, framing the woman's gorgeous face.\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"7\": {\"inputs\": {\"text\": \"score_1, score_2, score_3\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\"}, \"9\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\"}, \"10\": {\"inputs\": {\"model_type\": \"SDXL\", \"steps\": 40, \"denoise\": 1}, \"class_type\": \"AlignYourStepsScheduler\"}, \"12\": {\"inputs\": {\"add_noise\": true, \"noise_seed\": 214992713802638, \"cfg\": 5, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"sampler\": [\"13\", 0], \"sigmas\": [\"10\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"SamplerCustom\"}, \"13\": {\"inputs\": {\"sampler_name\": \"dpmpp_3m_sde_gpu\"}, \"class_type\": \"KSamplerSelect\"}, \"16\": {\"inputs\": {\"model_name\": \"bbox/face_yolov8m.pt\"}, \"class_type\": \"UltralyticsDetectorProvider\"}, \"17\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\"}, \"class_type\": \"SaveImage\"}, \"18\": {\"inputs\": {\"scale\": 2.0, \"model\": [\"4\", 0]}, \"class_type\": \"PerturbedAttentionGuidance\"}}, \"workflow\": {\"last_node_id\": 18, \"last_link_id\": 37, \"nodes\": [{\"id\": 3, \"type\": \"KSampler\", \"pos\": [868, 167], \"size\": {\"0\": 315, \"1\": 262}, \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [17], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [831882560098235, \"randomize\", 40, 2, \"dpmpp_2m_sde_gpu\", \"karras\", 1]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [9, 371], \"size\": {\"0\": 315, \"1\": 98}, \"flags\": {}, \"order\": 4, \"mode\": 0, \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [10, 19, 26], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [3, 5, 21], \"slot_index\": 1}, {\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [8, 20], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"CheckpointLoaderSimple\"}, \"widgets_values\": [\"cuwm1gxo-complete-20241024a.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [473, 609], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 0, \"mode\": 0, \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [2, 15], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"EmptyLatentImage\"}, \"widgets_values\": [1024, 1024, 16]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [415, 186], \"size\": {\"0\": 422.84503173828125, \"1\": 164.31304931640625}, \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 3}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4, 11, 22], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_7_up, Photograph of a woman wearing a red cocktail dress that fits tightly to her feminine curves, showing off her ample bosom and wide hips. The lighting in the photograph is professional with soft shadows adding depth. The background is dim and blurry, showing a fancy event, keeping the focus on the beautiful woman. Her curly brown hair falls past her shoulders, framing the woman's gorgeous face.\"]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [413, 389], \"size\": {\"0\": 425.27801513671875, \"1\": 180.6060791015625}, \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 5}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [6, 12, 23], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_1, score_2, score_3\"]}, {\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1209, 188], \"size\": {\"0\": 210, \"1\": 46}, \"flags\": {}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 17}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9, 18], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1451, 189], \"size\": {\"0\": 210, \"1\": 270}, \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 10, \"type\": \"AlignYourStepsScheduler\", \"pos\": [-6, 888], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 1, \"mode\": 0, \"outputs\": [{\"name\": \"SIGMAS\", \"type\": \"SIGMAS\", \"links\": [13], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"AlignYourStepsScheduler\"}, \"widgets_values\": [\"SDXL\", 40, 1]}, {\"id\": 12, \"type\": \"SamplerCustom\", \"pos\": [418, 820], \"size\": {\"0\": 355.20001220703125, \"1\": 230}, \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 10}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 11}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 12}, {\"name\": \"sampler\", \"type\": \"SAMPLER\", \"link\": 14}, {\"name\": \"sigmas\", \"type\": \"SIGMAS\", \"link\": 13}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 15}], \"outputs\": [{\"name\": \"output\", \"type\": \"LATENT\", \"links\": [], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"denoised_output\", \"type\": \"LATENT\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"SamplerCustom\"}, \"widgets_values\": [true, 214992713802638, \"randomize\", 5]}, {\"id\": 13, \"type\": \"KSamplerSelect\", \"pos\": [7, 738], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 2, \"mode\": 0, \"outputs\": [{\"name\": \"SAMPLER\", \"type\": \"SAMPLER\", \"links\": [14], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"KSamplerSelect\"}, \"widgets_values\": [\"dpmpp_3m_sde_gpu\"]}, {\"id\": 15, \"type\": \"FaceDetailer\", \"pos\": [1139, 602], \"size\": {\"0\": 506.4000244140625, \"1\": 900}, \"flags\": {}, \"order\": 12, \"mode\": 2, \"inputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"link\": 18}, {\"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 21}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 20}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 22}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 23}, {\"name\": \"bbox_detector\", \"type\": \"BBOX_DETECTOR\", \"link\": 24}, {\"name\": \"sam_model_opt\", \"type\": \"SAM_MODEL\", \"link\": null}, {\"name\": \"segm_detector_opt\", \"type\": \"SEGM_DETECTOR\", \"link\": null}, {\"name\": \"detailer_hook\", \"type\": \"DETAILER_HOOK\", \"link\": null}, {\"name\": \"scheduler_func_opt\", \"type\": \"SCHEDULER_FUNC\", \"link\": null}], \"outputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"links\": [25], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"cropped_refined\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"cropped_enhanced_alpha\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"mask\", \"type\": \"MASK\", \"links\": null, \"shape\": 3}, {\"name\": \"detailer_pipe\", \"type\": \"DETAILER_PIPE\", \"links\": null, \"shape\": 3}, {\"name\": \"cnet_images\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}], \"properties\": {\"Node name for S&R\": \"FaceDetailer\"}, \"widgets_values\": [384, true, 1024, 589703604624929, \"randomize\", 20, 8, \"euler\", \"normal\", 0.5, 5, true, true, 0.5, 10, 3, \"center-1\", 0, 0.93, 0, 0.7, \"False\", 10, \"\", 1, false, 20]}, {\"id\": 16, \"type\": \"UltralyticsDetectorProvider\", \"pos\": [788, 796], \"size\": {\"0\": 315, \"1\": 78}, \"flags\": {}, \"order\": 3, \"mode\": 0, \"outputs\": [{\"name\": \"BBOX_DETECTOR\", \"type\": \"BBOX_DETECTOR\", \"links\": [24], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"SEGM_DETECTOR\", \"type\": \"SEGM_DETECTOR\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"UltralyticsDetectorProvider\"}, \"widgets_values\": [\"bbox/face_yolov8m.pt\"]}, {\"id\": 17, \"type\": \"SaveImage\", \"pos\": [1744, 566], \"size\": {\"0\": 315, \"1\": 270}, \"flags\": {}, \"order\": 13, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 25}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 18, \"type\": \"PerturbedAttentionGuidance\", \"pos\": [468, 87], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 26}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"PerturbedAttentionGuidance\"}, \"widgets_values\": [2]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [3, 4, 1, 6, 0, \"CLIP\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [5, 4, 1, 7, 0, \"CLIP\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [10, 4, 0, 12, 0, \"MODEL\"], [11, 6, 0, 12, 1, \"CONDITIONING\"], [12, 7, 0, 12, 2, \"CONDITIONING\"], [13, 10, 0, 12, 4, \"SIGMAS\"], [14, 13, 0, 12, 3, \"SAMPLER\"], [15, 5, 0, 12, 5, \"LATENT\"], [17, 3, 0, 8, 0, \"LATENT\"], [18, 8, 0, 15, 0, \"IMAGE\"], [19, 4, 0, 15, 1, \"MODEL\"], [20, 4, 2, 15, 3, \"VAE\"], [21, 4, 1, 15, 2, \"CLIP\"], [22, 6, 0, 15, 4, \"CONDITIONING\"], [23, 7, 0, 15, 5, \"CONDITIONING\"], [24, 16, 0, 15, 6, \"BBOX_DETECTOR\"], [25, 15, 0, 17, 0, \"IMAGE\"], [26, 4, 0, 18, 0, \"MODEL\"], [37, 18, 0, 3, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 1.3310000000000006, \"offset\": [-28.6939932528295, -29.576837980502148]}, \"workspace_info\": {\"id\": \"TNOwVzKHGn6FUMraMM-Y7\"}}, \"version\": 0.4}}", |
| "steps": 40, |
| "width": 1024, |
| "height": 1024, |
| "models": [ |
| "cuwm1gxo-complete-20241024a.safetensors" |
| ], |
| "prompt": "score_7_up, Photograph of a woman wearing a red cocktail dress that fits tightly to her feminine curves, showing off her ample bosom and wide hips. The lighting in the photograph is professional with soft shadows adding depth. The background is dim and blurry, showing a fancy event, keeping the focus on the beautiful woman. Her curly brown hair falls past her shoulders, framing the woman's gorgeous face.", |
| "denoise": 1, |
| "sampler": "dpmpp_2m_sde_gpu", |
| "cfgScale": 2, |
| "modelIds": [], |
| "scheduler": "karras", |
| "upscalers": [], |
| "versionIds": [], |
| "controlNets": [], |
| "negativePrompt": "score_1, score_2, score_3", |
| "additionalResources": [] |
| }, |
| "availability": "Public", |
| "hasMeta": true, |
| "hasPositivePrompt": true, |
| "onSite": false, |
| "remixOfId": null |
| }, |
| { |
| "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/1a6da6df-9f83-49d5-8963-23256a1549e0/width=1024/36427324.jpeg", |
| "nsfwLevel": 1, |
| "width": 1024, |
| "height": 1024, |
| "hash": "UJIhHQj?xuadxwt6t7IVoxWAR+j?~pD%WCM{", |
| "type": "image", |
| "metadata": { |
| "hash": "UJIhHQj?xuadxwt6t7IVoxWAR+j?~pD%WCM{", |
| "size": 1829779, |
| "width": 1024, |
| "height": 1024 |
| }, |
| "meta": { |
| "seed": 244771154348248, |
| "vaes": [], |
| "Model": "cuwm1gxo-complete-20241024a", |
| "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 244771154348248, \"steps\": 40, \"cfg\": 3.0, \"sampler_name\": \"dpmpp_2m_sde_gpu\", \"scheduler\": \"karras\", \"denoise\": 1.0, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\"}, \"4\": {\"inputs\": {\"ckpt_name\": \"cuwm1gxo-complete-20241024a.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\"}, \"5\": {\"inputs\": {\"width\": 1024, \"height\": 1024, \"batch_size\": 16}, \"class_type\": \"EmptyLatentImage\"}, \"6\": {\"inputs\": {\"text\": \"score_7_up, A stunning photograph of a horse trotting along a hilltop as the sun sets behind him, light rays streaming. The horse's fur is highly detailed. The golden background gives the image a surreal feel.\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"7\": {\"inputs\": {\"text\": \"score_1, score_2, score_3\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\"}, \"9\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\"}, \"10\": {\"inputs\": {\"model_type\": \"SDXL\", \"steps\": 40, \"denoise\": 1}, \"class_type\": \"AlignYourStepsScheduler\"}, \"12\": {\"inputs\": {\"add_noise\": true, \"noise_seed\": 508575067978939, \"cfg\": 5, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"sampler\": [\"13\", 0], \"sigmas\": [\"10\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"SamplerCustom\"}, \"13\": {\"inputs\": {\"sampler_name\": \"dpmpp_3m_sde_gpu\"}, \"class_type\": \"KSamplerSelect\"}, \"16\": {\"inputs\": {\"model_name\": \"bbox/face_yolov8m.pt\"}, \"class_type\": \"UltralyticsDetectorProvider\"}, \"17\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\"}, \"class_type\": \"SaveImage\"}, \"18\": {\"inputs\": {\"scale\": 2, \"model\": [\"4\", 0]}, \"class_type\": \"PerturbedAttentionGuidance\"}}, \"workflow\": {\"last_node_id\": 18, \"last_link_id\": 38, \"nodes\": [{\"id\": 3, \"type\": \"KSampler\", \"pos\": [868, 167], \"size\": {\"0\": 315, \"1\": 262}, \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 38}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [17], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [244771154348248, \"randomize\", 40, 3, \"dpmpp_2m_sde_gpu\", \"karras\", 1]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [9, 371], \"size\": {\"0\": 315, \"1\": 98}, \"flags\": {}, \"order\": 4, \"mode\": 0, \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [10, 19, 26, 38], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [3, 5, 21], \"slot_index\": 1}, {\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [8, 20], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"CheckpointLoaderSimple\"}, \"widgets_values\": [\"cuwm1gxo-complete-20241024a.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [473, 609], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 0, \"mode\": 0, \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [2, 15], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"EmptyLatentImage\"}, \"widgets_values\": [1024, 1024, 16]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [415, 186], \"size\": {\"0\": 422.84503173828125, \"1\": 164.31304931640625}, \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 3}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4, 11, 22], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_7_up, A stunning photograph of a horse trotting along a hilltop as the sun sets behind him, light rays streaming. The horse's fur is highly detailed. The golden background gives the image a surreal feel.\"]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [413, 389], \"size\": {\"0\": 425.27801513671875, \"1\": 180.6060791015625}, \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 5}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [6, 12, 23], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_1, score_2, score_3\"]}, {\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1209, 188], \"size\": {\"0\": 210, \"1\": 46}, \"flags\": {}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 17}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9, 18], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1451, 189], \"size\": {\"0\": 210, \"1\": 270}, \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 10, \"type\": \"AlignYourStepsScheduler\", \"pos\": [-6, 888], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 1, \"mode\": 0, \"outputs\": [{\"name\": \"SIGMAS\", \"type\": \"SIGMAS\", \"links\": [13], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"AlignYourStepsScheduler\"}, \"widgets_values\": [\"SDXL\", 40, 1]}, {\"id\": 12, \"type\": \"SamplerCustom\", \"pos\": [418, 820], \"size\": {\"0\": 355.20001220703125, \"1\": 230}, \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 10}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 11}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 12}, {\"name\": \"sampler\", \"type\": \"SAMPLER\", \"link\": 14}, {\"name\": \"sigmas\", \"type\": \"SIGMAS\", \"link\": 13}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 15}], \"outputs\": [{\"name\": \"output\", \"type\": \"LATENT\", \"links\": [], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"denoised_output\", \"type\": \"LATENT\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"SamplerCustom\"}, \"widgets_values\": [true, 508575067978939, \"randomize\", 5]}, {\"id\": 13, \"type\": \"KSamplerSelect\", \"pos\": [7, 738], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 2, \"mode\": 0, \"outputs\": [{\"name\": \"SAMPLER\", \"type\": \"SAMPLER\", \"links\": [14], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"KSamplerSelect\"}, \"widgets_values\": [\"dpmpp_3m_sde_gpu\"]}, {\"id\": 15, \"type\": \"FaceDetailer\", \"pos\": [1139, 602], \"size\": {\"0\": 506.4000244140625, \"1\": 900}, \"flags\": {}, \"order\": 12, \"mode\": 2, \"inputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"link\": 18}, {\"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 21}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 20}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 22}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 23}, {\"name\": \"bbox_detector\", \"type\": \"BBOX_DETECTOR\", \"link\": 24}, {\"name\": \"sam_model_opt\", \"type\": \"SAM_MODEL\", \"link\": null}, {\"name\": \"segm_detector_opt\", \"type\": \"SEGM_DETECTOR\", \"link\": null}, {\"name\": \"detailer_hook\", \"type\": \"DETAILER_HOOK\", \"link\": null}, {\"name\": \"scheduler_func_opt\", \"type\": \"SCHEDULER_FUNC\", \"link\": null}], \"outputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"links\": [25], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"cropped_refined\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"cropped_enhanced_alpha\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"mask\", \"type\": \"MASK\", \"links\": null, \"shape\": 3}, {\"name\": \"detailer_pipe\", \"type\": \"DETAILER_PIPE\", \"links\": null, \"shape\": 3}, {\"name\": \"cnet_images\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}], \"properties\": {\"Node name for S&R\": \"FaceDetailer\"}, \"widgets_values\": [384, true, 1024, 1013624614925806, \"randomize\", 20, 8, \"euler\", \"normal\", 0.5, 5, true, true, 0.5, 10, 3, \"center-1\", 0, 0.93, 0, 0.7, \"False\", 10, \"\", 1, false, 20]}, {\"id\": 16, \"type\": \"UltralyticsDetectorProvider\", \"pos\": [788, 796], \"size\": {\"0\": 315, \"1\": 78}, \"flags\": {}, \"order\": 3, \"mode\": 0, \"outputs\": [{\"name\": \"BBOX_DETECTOR\", \"type\": \"BBOX_DETECTOR\", \"links\": [24], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"SEGM_DETECTOR\", \"type\": \"SEGM_DETECTOR\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"UltralyticsDetectorProvider\"}, \"widgets_values\": [\"bbox/face_yolov8m.pt\"]}, {\"id\": 17, \"type\": \"SaveImage\", \"pos\": [1744, 566], \"size\": {\"0\": 315, \"1\": 270}, \"flags\": {}, \"order\": 13, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 25}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 18, \"type\": \"PerturbedAttentionGuidance\", \"pos\": [468, 87], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 26}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"PerturbedAttentionGuidance\"}, \"widgets_values\": [2]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [3, 4, 1, 6, 0, \"CLIP\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [5, 4, 1, 7, 0, \"CLIP\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [10, 4, 0, 12, 0, \"MODEL\"], [11, 6, 0, 12, 1, \"CONDITIONING\"], [12, 7, 0, 12, 2, \"CONDITIONING\"], [13, 10, 0, 12, 4, \"SIGMAS\"], [14, 13, 0, 12, 3, \"SAMPLER\"], [15, 5, 0, 12, 5, \"LATENT\"], [17, 3, 0, 8, 0, \"LATENT\"], [18, 8, 0, 15, 0, \"IMAGE\"], [19, 4, 0, 15, 1, \"MODEL\"], [20, 4, 2, 15, 3, \"VAE\"], [21, 4, 1, 15, 2, \"CLIP\"], [22, 6, 0, 15, 4, \"CONDITIONING\"], [23, 7, 0, 15, 5, \"CONDITIONING\"], [24, 16, 0, 15, 6, \"BBOX_DETECTOR\"], [25, 15, 0, 17, 0, \"IMAGE\"], [26, 4, 0, 18, 0, \"MODEL\"], [38, 4, 0, 3, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 1.2100000000000006, \"offset\": [29.157246416591963, -66.9923150654007]}, \"workspace_info\": {\"id\": \"TNOwVzKHGn6FUMraMM-Y7\"}}, \"version\": 0.4}}", |
| "steps": 40, |
| "width": 1024, |
| "height": 1024, |
| "models": [ |
| "cuwm1gxo-complete-20241024a.safetensors" |
| ], |
| "prompt": "score_7_up, A stunning photograph of a horse trotting along a hilltop as the sun sets behind him, light rays streaming. The horse's fur is highly detailed. The golden background gives the image a surreal feel.", |
| "denoise": 1, |
| "sampler": "dpmpp_2m_sde_gpu", |
| "cfgScale": 3, |
| "modelIds": [], |
| "scheduler": "karras", |
| "upscalers": [], |
| "versionIds": [], |
| "controlNets": [], |
| "negativePrompt": "score_1, score_2, score_3", |
| "additionalResources": [] |
| }, |
| "availability": "Public", |
| "hasMeta": true, |
| "hasPositivePrompt": true, |
| "onSite": false, |
| "remixOfId": null |
| }, |
| { |
| "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/96df59a9-44bc-457c-a742-38b212d78b0b/width=1024/36427421.jpeg", |
| "nsfwLevel": 8, |
| "width": 1024, |
| "height": 1024, |
| "hash": "ULL:_-_NELDi%hxGjFW;o}kCr=WBIADiX9Wr", |
| "type": "image", |
| "metadata": { |
| "hash": "ULL:_-_NELDi%hxGjFW;o}kCr=WBIADiX9Wr", |
| "size": 1210920, |
| "width": 1024, |
| "height": 1024 |
| }, |
| "meta": { |
| "seed": 45888131981240, |
| "vaes": [], |
| "Model": "cuwm1gxo-complete-20241024a", |
| "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 45888131981240, \"steps\": 40, \"cfg\": 2.0, \"sampler_name\": \"dpmpp_2m_sde_gpu\", \"scheduler\": \"karras\", \"denoise\": 1.0, \"model\": [\"18\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\"}, \"4\": {\"inputs\": {\"ckpt_name\": \"cuwm1gxo-complete-20241024a.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\"}, \"5\": {\"inputs\": {\"width\": 1024, \"height\": 1024, \"batch_size\": 16}, \"class_type\": \"EmptyLatentImage\"}, \"6\": {\"inputs\": {\"text\": \"score_6_up, Artistic and striking photograph of a woman cosplaying as a cat. She expresses sensuality with her curvy body and fluffy ears. The details of her skin are highlighted by the dynamic and expert lighting of the photo. The background shows an intimate scene.\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"7\": {\"inputs\": {\"text\": \"score_1, score_2, score_3, multiple views\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\"}, \"9\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\"}, \"10\": {\"inputs\": {\"model_type\": \"SDXL\", \"steps\": 40, \"denoise\": 1}, \"class_type\": \"AlignYourStepsScheduler\"}, \"12\": {\"inputs\": {\"add_noise\": true, \"noise_seed\": 385441023207091, \"cfg\": 5, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"sampler\": [\"13\", 0], \"sigmas\": [\"10\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"SamplerCustom\"}, \"13\": {\"inputs\": {\"sampler_name\": \"dpmpp_3m_sde_gpu\"}, \"class_type\": \"KSamplerSelect\"}, \"16\": {\"inputs\": {\"model_name\": \"bbox/face_yolov8m.pt\"}, \"class_type\": \"UltralyticsDetectorProvider\"}, \"17\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\"}, \"class_type\": \"SaveImage\"}, \"18\": {\"inputs\": {\"scale\": 2.0, \"model\": [\"4\", 0]}, \"class_type\": \"PerturbedAttentionGuidance\"}}, \"workflow\": {\"last_node_id\": 18, \"last_link_id\": 35, \"nodes\": [{\"id\": 3, \"type\": \"KSampler\", \"pos\": [868, 167], \"size\": {\"0\": 315, \"1\": 262}, \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 35}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [17], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [45888131981240, \"randomize\", 40, 2, \"dpmpp_2m_sde_gpu\", \"karras\", 1]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [9, 371], \"size\": {\"0\": 315, \"1\": 98}, \"flags\": {}, \"order\": 0, \"mode\": 0, \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [10, 19, 26], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [3, 5, 21], \"slot_index\": 1}, {\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [8, 20], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"CheckpointLoaderSimple\"}, \"widgets_values\": [\"cuwm1gxo-complete-20241024a.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [473, 609], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 1, \"mode\": 0, \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [2, 15], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"EmptyLatentImage\"}, \"widgets_values\": [1024, 1024, 16]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [415, 186], \"size\": {\"0\": 422.84503173828125, \"1\": 164.31304931640625}, \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 3}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4, 11, 22], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_6_up, Artistic and striking photograph of a woman cosplaying as a cat. She expresses sensuality with her curvy body and fluffy ears. The details of her skin are highlighted by the dynamic and expert lighting of the photo. The background shows an intimate scene.\"]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [413, 389], \"size\": {\"0\": 425.27801513671875, \"1\": 180.6060791015625}, \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 5}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [6, 12, 23], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_1, score_2, score_3, multiple views\"]}, {\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1209, 188], \"size\": {\"0\": 210, \"1\": 46}, \"flags\": {}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 17}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9, 18], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1451, 189], \"size\": {\"0\": 210, \"1\": 270}, \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 10, \"type\": \"AlignYourStepsScheduler\", \"pos\": [-6, 888], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 2, \"mode\": 0, \"outputs\": [{\"name\": \"SIGMAS\", \"type\": \"SIGMAS\", \"links\": [13], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"AlignYourStepsScheduler\"}, \"widgets_values\": [\"SDXL\", 40, 1]}, {\"id\": 12, \"type\": \"SamplerCustom\", \"pos\": [418, 820], \"size\": {\"0\": 355.20001220703125, \"1\": 230}, \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 10}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 11}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 12}, {\"name\": \"sampler\", \"type\": \"SAMPLER\", \"link\": 14}, {\"name\": \"sigmas\", \"type\": \"SIGMAS\", \"link\": 13}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 15}], \"outputs\": [{\"name\": \"output\", \"type\": \"LATENT\", \"links\": [], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"denoised_output\", \"type\": \"LATENT\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"SamplerCustom\"}, \"widgets_values\": [true, 385441023207091, \"randomize\", 5]}, {\"id\": 13, \"type\": \"KSamplerSelect\", \"pos\": [7, 738], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 3, \"mode\": 0, \"outputs\": [{\"name\": \"SAMPLER\", \"type\": \"SAMPLER\", \"links\": [14], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"KSamplerSelect\"}, \"widgets_values\": [\"dpmpp_3m_sde_gpu\"]}, {\"id\": 15, \"type\": \"FaceDetailer\", \"pos\": [1139, 602], \"size\": {\"0\": 506.4000244140625, \"1\": 900}, \"flags\": {}, \"order\": 12, \"mode\": 2, \"inputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"link\": 18}, {\"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 21}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 20}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 22}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 23}, {\"name\": \"bbox_detector\", \"type\": \"BBOX_DETECTOR\", \"link\": 24}, {\"name\": \"sam_model_opt\", \"type\": \"SAM_MODEL\", \"link\": null}, {\"name\": \"segm_detector_opt\", \"type\": \"SEGM_DETECTOR\", \"link\": null}, {\"name\": \"detailer_hook\", \"type\": \"DETAILER_HOOK\", \"link\": null}, {\"name\": \"scheduler_func_opt\", \"type\": \"SCHEDULER_FUNC\", \"link\": null}], \"outputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"links\": [25], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"cropped_refined\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"cropped_enhanced_alpha\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"mask\", \"type\": \"MASK\", \"links\": null, \"shape\": 3}, {\"name\": \"detailer_pipe\", \"type\": \"DETAILER_PIPE\", \"links\": null, \"shape\": 3}, {\"name\": \"cnet_images\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}], \"properties\": {\"Node name for S&R\": \"FaceDetailer\"}, \"widgets_values\": [384, true, 1024, 67323658749798, \"randomize\", 20, 8, \"euler\", \"normal\", 0.5, 5, true, true, 0.5, 10, 3, \"center-1\", 0, 0.93, 0, 0.7, \"False\", 10, \"\", 1, false, 20]}, {\"id\": 16, \"type\": \"UltralyticsDetectorProvider\", \"pos\": [788, 796], \"size\": {\"0\": 315, \"1\": 78}, \"flags\": {}, \"order\": 4, \"mode\": 0, \"outputs\": [{\"name\": \"BBOX_DETECTOR\", \"type\": \"BBOX_DETECTOR\", \"links\": [24], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"SEGM_DETECTOR\", \"type\": \"SEGM_DETECTOR\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"UltralyticsDetectorProvider\"}, \"widgets_values\": [\"bbox/face_yolov8m.pt\"]}, {\"id\": 17, \"type\": \"SaveImage\", \"pos\": [1744, 566], \"size\": {\"0\": 315, \"1\": 270}, \"flags\": {}, \"order\": 13, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 25}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 18, \"type\": \"PerturbedAttentionGuidance\", \"pos\": [468, 87], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 26}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [35], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"PerturbedAttentionGuidance\"}, \"widgets_values\": [2]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [3, 4, 1, 6, 0, \"CLIP\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [5, 4, 1, 7, 0, \"CLIP\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [10, 4, 0, 12, 0, \"MODEL\"], [11, 6, 0, 12, 1, \"CONDITIONING\"], [12, 7, 0, 12, 2, \"CONDITIONING\"], [13, 10, 0, 12, 4, \"SIGMAS\"], [14, 13, 0, 12, 3, \"SAMPLER\"], [15, 5, 0, 12, 5, \"LATENT\"], [17, 3, 0, 8, 0, \"LATENT\"], [18, 8, 0, 15, 0, \"IMAGE\"], [19, 4, 0, 15, 1, \"MODEL\"], [20, 4, 2, 15, 3, \"VAE\"], [21, 4, 1, 15, 2, \"CLIP\"], [22, 6, 0, 15, 4, \"CONDITIONING\"], [23, 7, 0, 15, 5, \"CONDITIONING\"], [24, 16, 0, 15, 6, \"BBOX_DETECTOR\"], [25, 15, 0, 17, 0, \"IMAGE\"], [26, 4, 0, 18, 0, \"MODEL\"], [35, 18, 0, 3, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 1.3310000000000008, \"offset\": [-28.693993252829536, -29.576837980502162]}, \"workspace_info\": {\"id\": \"TNOwVzKHGn6FUMraMM-Y7\"}}, \"version\": 0.4}}", |
| "steps": 40, |
| "width": 1024, |
| "height": 1024, |
| "models": [ |
| "cuwm1gxo-complete-20241024a.safetensors" |
| ], |
| "prompt": "score_6_up, Artistic and striking photograph of a woman cosplaying as a cat. She expresses sensuality with her curvy body and fluffy ears. The details of her skin are highlighted by the dynamic and expert lighting of the photo. The background shows an intimate scene.", |
| "denoise": 1, |
| "sampler": "dpmpp_2m_sde_gpu", |
| "cfgScale": 2, |
| "modelIds": [], |
| "scheduler": "karras", |
| "upscalers": [], |
| "versionIds": [], |
| "controlNets": [], |
| "negativePrompt": "score_1, score_2, score_3, multiple views", |
| "additionalResources": [] |
| }, |
| "availability": "Public", |
| "hasMeta": true, |
| "hasPositivePrompt": true, |
| "onSite": false, |
| "remixOfId": null |
| }, |
| { |
| "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/32f68ebe-6d7b-47ae-ba10-a49268663c2d/width=1024/36427403.jpeg", |
| "nsfwLevel": 1, |
| "width": 1024, |
| "height": 1024, |
| "hash": "U6F=s__3-V~W$%~qt6?H00M{ozD%-pIU?HoM", |
| "type": "image", |
| "metadata": { |
| "hash": "U6F=s__3-V~W$%~qt6?H00M{ozD%-pIU?HoM", |
| "size": 1783681, |
| "width": 1024, |
| "height": 1024 |
| }, |
| "meta": { |
| "seed": 90865607806140, |
| "vaes": [], |
| "Model": "cuwm1gxo-complete-20241024a", |
| "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 90865607806140, \"steps\": 40, \"cfg\": 2.0, \"sampler_name\": \"dpmpp_2m_sde_gpu\", \"scheduler\": \"karras\", \"denoise\": 1.0, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\"}, \"4\": {\"inputs\": {\"ckpt_name\": \"cuwm1gxo-complete-20241024a.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\"}, \"5\": {\"inputs\": {\"width\": 1024, \"height\": 1024, \"batch_size\": 16}, \"class_type\": \"EmptyLatentImage\"}, \"6\": {\"inputs\": {\"text\": \"score_6_up, Photograph of an Indian woman wearing a traditional dress.\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"7\": {\"inputs\": {\"text\": \"score_1, score_2, score_3\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\"}, \"9\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\"}, \"10\": {\"inputs\": {\"model_type\": \"SDXL\", \"steps\": 40, \"denoise\": 1}, \"class_type\": \"AlignYourStepsScheduler\"}, \"12\": {\"inputs\": {\"add_noise\": true, \"noise_seed\": 850619760009950, \"cfg\": 5, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"sampler\": [\"13\", 0], \"sigmas\": [\"10\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"SamplerCustom\"}, \"13\": {\"inputs\": {\"sampler_name\": \"dpmpp_3m_sde_gpu\"}, \"class_type\": \"KSamplerSelect\"}, \"16\": {\"inputs\": {\"model_name\": \"bbox/face_yolov8m.pt\"}, \"class_type\": \"UltralyticsDetectorProvider\"}, \"17\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\"}, \"class_type\": \"SaveImage\"}, \"18\": {\"inputs\": {\"scale\": 2, \"model\": [\"4\", 0]}, \"class_type\": \"PerturbedAttentionGuidance\"}}, \"workflow\": {\"last_node_id\": 18, \"last_link_id\": 36, \"nodes\": [{\"id\": 3, \"type\": \"KSampler\", \"pos\": [868, 167], \"size\": {\"0\": 315, \"1\": 262}, \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 36}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [17], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [90865607806140, \"randomize\", 40, 2, \"dpmpp_2m_sde_gpu\", \"karras\", 1]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [9, 371], \"size\": {\"0\": 315, \"1\": 98}, \"flags\": {}, \"order\": 4, \"mode\": 0, \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [10, 19, 26, 36], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [3, 5, 21], \"slot_index\": 1}, {\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [8, 20], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"CheckpointLoaderSimple\"}, \"widgets_values\": [\"cuwm1gxo-complete-20241024a.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [473, 609], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 0, \"mode\": 0, \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [2, 15], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"EmptyLatentImage\"}, \"widgets_values\": [1024, 1024, 16]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [415, 186], \"size\": {\"0\": 422.84503173828125, \"1\": 164.31304931640625}, \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 3}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4, 11, 22], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_6_up, Photograph of an Indian woman wearing a traditional dress.\"]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [413, 389], \"size\": {\"0\": 425.27801513671875, \"1\": 180.6060791015625}, \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 5}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [6, 12, 23], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_1, score_2, score_3\"]}, {\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1209, 188], \"size\": {\"0\": 210, \"1\": 46}, \"flags\": {}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 17}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9, 18], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1451, 189], \"size\": {\"0\": 210, \"1\": 270}, \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 10, \"type\": \"AlignYourStepsScheduler\", \"pos\": [-6, 888], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 1, \"mode\": 0, \"outputs\": [{\"name\": \"SIGMAS\", \"type\": \"SIGMAS\", \"links\": [13], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"AlignYourStepsScheduler\"}, \"widgets_values\": [\"SDXL\", 40, 1]}, {\"id\": 12, \"type\": \"SamplerCustom\", \"pos\": [418, 820], \"size\": {\"0\": 355.20001220703125, \"1\": 230}, \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 10}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 11}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 12}, {\"name\": \"sampler\", \"type\": \"SAMPLER\", \"link\": 14}, {\"name\": \"sigmas\", \"type\": \"SIGMAS\", \"link\": 13}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 15}], \"outputs\": [{\"name\": \"output\", \"type\": \"LATENT\", \"links\": [], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"denoised_output\", \"type\": \"LATENT\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"SamplerCustom\"}, \"widgets_values\": [true, 850619760009950, \"randomize\", 5]}, {\"id\": 13, \"type\": \"KSamplerSelect\", \"pos\": [7, 738], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 2, \"mode\": 0, \"outputs\": [{\"name\": \"SAMPLER\", \"type\": \"SAMPLER\", \"links\": [14], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"KSamplerSelect\"}, \"widgets_values\": [\"dpmpp_3m_sde_gpu\"]}, {\"id\": 15, \"type\": \"FaceDetailer\", \"pos\": [1139, 602], \"size\": {\"0\": 506.4000244140625, \"1\": 900}, \"flags\": {}, \"order\": 12, \"mode\": 2, \"inputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"link\": 18}, {\"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 21}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 20}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 22}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 23}, {\"name\": \"bbox_detector\", \"type\": \"BBOX_DETECTOR\", \"link\": 24}, {\"name\": \"sam_model_opt\", \"type\": \"SAM_MODEL\", \"link\": null}, {\"name\": \"segm_detector_opt\", \"type\": \"SEGM_DETECTOR\", \"link\": null}, {\"name\": \"detailer_hook\", \"type\": \"DETAILER_HOOK\", \"link\": null}, {\"name\": \"scheduler_func_opt\", \"type\": \"SCHEDULER_FUNC\", \"link\": null}], \"outputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"links\": [25], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"cropped_refined\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"cropped_enhanced_alpha\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"mask\", \"type\": \"MASK\", \"links\": null, \"shape\": 3}, {\"name\": \"detailer_pipe\", \"type\": \"DETAILER_PIPE\", \"links\": null, \"shape\": 3}, {\"name\": \"cnet_images\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}], \"properties\": {\"Node name for S&R\": \"FaceDetailer\"}, \"widgets_values\": [384, true, 1024, 164715747660623, \"randomize\", 20, 8, \"euler\", \"normal\", 0.5, 5, true, true, 0.5, 10, 3, \"center-1\", 0, 0.93, 0, 0.7, \"False\", 10, \"\", 1, false, 20]}, {\"id\": 16, \"type\": \"UltralyticsDetectorProvider\", \"pos\": [788, 796], \"size\": {\"0\": 315, \"1\": 78}, \"flags\": {}, \"order\": 3, \"mode\": 0, \"outputs\": [{\"name\": \"BBOX_DETECTOR\", \"type\": \"BBOX_DETECTOR\", \"links\": [24], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"SEGM_DETECTOR\", \"type\": \"SEGM_DETECTOR\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"UltralyticsDetectorProvider\"}, \"widgets_values\": [\"bbox/face_yolov8m.pt\"]}, {\"id\": 17, \"type\": \"SaveImage\", \"pos\": [1744, 566], \"size\": {\"0\": 315, \"1\": 270}, \"flags\": {}, \"order\": 13, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 25}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 18, \"type\": \"PerturbedAttentionGuidance\", \"pos\": [468, 87], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 26}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"PerturbedAttentionGuidance\"}, \"widgets_values\": [2]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [3, 4, 1, 6, 0, \"CLIP\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [5, 4, 1, 7, 0, \"CLIP\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [10, 4, 0, 12, 0, \"MODEL\"], [11, 6, 0, 12, 1, \"CONDITIONING\"], [12, 7, 0, 12, 2, \"CONDITIONING\"], [13, 10, 0, 12, 4, \"SIGMAS\"], [14, 13, 0, 12, 3, \"SAMPLER\"], [15, 5, 0, 12, 5, \"LATENT\"], [17, 3, 0, 8, 0, \"LATENT\"], [18, 8, 0, 15, 0, \"IMAGE\"], [19, 4, 0, 15, 1, \"MODEL\"], [20, 4, 2, 15, 3, \"VAE\"], [21, 4, 1, 15, 2, \"CLIP\"], [22, 6, 0, 15, 4, \"CONDITIONING\"], [23, 7, 0, 15, 5, \"CONDITIONING\"], [24, 16, 0, 15, 6, \"BBOX_DETECTOR\"], [25, 15, 0, 17, 0, \"IMAGE\"], [26, 4, 0, 18, 0, \"MODEL\"], [36, 4, 0, 3, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 1.3310000000000006, \"offset\": [-28.6939932528295, -29.576837980502148]}, \"workspace_info\": {\"id\": \"TNOwVzKHGn6FUMraMM-Y7\"}}, \"version\": 0.4}}", |
| "steps": 40, |
| "width": 1024, |
| "height": 1024, |
| "models": [ |
| "cuwm1gxo-complete-20241024a.safetensors" |
| ], |
| "prompt": "score_6_up, Photograph of an Indian woman wearing a traditional dress.", |
| "denoise": 1, |
| "sampler": "dpmpp_2m_sde_gpu", |
| "cfgScale": 2, |
| "modelIds": [], |
| "scheduler": "karras", |
| "upscalers": [], |
| "versionIds": [], |
| "controlNets": [], |
| "negativePrompt": "score_1, score_2, score_3", |
| "additionalResources": [] |
| }, |
| "availability": "Public", |
| "hasMeta": true, |
| "hasPositivePrompt": true, |
| "onSite": false, |
| "remixOfId": null |
| }, |
| { |
| "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/7fef27a5-d356-48ed-a6a7-edfcde3e25ed/width=1024/36427948.jpeg", |
| "nsfwLevel": 8, |
| "width": 1024, |
| "height": 1024, |
| "hash": "UGI4@aI[yZ-p_3jE-;ofD%ngROV?9F-;t8kX", |
| "type": "image", |
| "metadata": { |
| "hash": "UGI4@aI[yZ-p_3jE-;ofD%ngROV?9F-;t8kX", |
| "size": 1131606, |
| "width": 1024, |
| "height": 1024 |
| }, |
| "meta": { |
| "seed": 738629668769063, |
| "vaes": [], |
| "Model": "cuwm1gxo-complete-20241024a", |
| "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 738629668769063, \"steps\": 40, \"cfg\": 3.0, \"sampler_name\": \"dpmpp_2m_sde_gpu\", \"scheduler\": \"karras\", \"denoise\": 1.0, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\"}, \"4\": {\"inputs\": {\"ckpt_name\": \"cuwm1gxo-complete-20241024a.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\"}, \"5\": {\"inputs\": {\"width\": 1024, \"height\": 1024, \"batch_size\": 16}, \"class_type\": \"EmptyLatentImage\"}, \"6\": {\"inputs\": {\"text\": \"score_6_up, Voyeur photograph looking down a woman's blouse, showing her perky nipples, highly detailed skin textures, downblouse, accidental exposure.\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"7\": {\"inputs\": {\"text\": \"score_1, score_2, score_3\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\"}, \"9\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\"}, \"10\": {\"inputs\": {\"model_type\": \"SDXL\", \"steps\": 40, \"denoise\": 1}, \"class_type\": \"AlignYourStepsScheduler\"}, \"12\": {\"inputs\": {\"add_noise\": true, \"noise_seed\": 830502167753162, \"cfg\": 5, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"sampler\": [\"13\", 0], \"sigmas\": [\"10\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"SamplerCustom\"}, \"13\": {\"inputs\": {\"sampler_name\": \"dpmpp_3m_sde_gpu\"}, \"class_type\": \"KSamplerSelect\"}, \"16\": {\"inputs\": {\"model_name\": \"bbox/face_yolov8m.pt\"}, \"class_type\": \"UltralyticsDetectorProvider\"}, \"17\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\"}, \"class_type\": \"SaveImage\"}, \"18\": {\"inputs\": {\"scale\": 2, \"model\": [\"4\", 0]}, \"class_type\": \"PerturbedAttentionGuidance\"}}, \"workflow\": {\"last_node_id\": 18, \"last_link_id\": 34, \"nodes\": [{\"id\": 3, \"type\": \"KSampler\", \"pos\": [868, 167], \"size\": {\"0\": 315, \"1\": 262}, \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 34}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [17], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [738629668769063, \"randomize\", 40, 3, \"dpmpp_2m_sde_gpu\", \"karras\", 1]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [9, 371], \"size\": {\"0\": 315, \"1\": 98}, \"flags\": {}, \"order\": 4, \"mode\": 0, \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [10, 19, 26, 34], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [3, 5, 21], \"slot_index\": 1}, {\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [8, 20], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"CheckpointLoaderSimple\"}, \"widgets_values\": [\"cuwm1gxo-complete-20241024a.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [473, 609], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 3, \"mode\": 0, \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [2, 15], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"EmptyLatentImage\"}, \"widgets_values\": [1024, 1024, 16]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [415, 186], \"size\": {\"0\": 422.84503173828125, \"1\": 164.31304931640625}, \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 3}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4, 11, 22], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_6_up, Voyeur photograph looking down a woman's blouse, showing her perky nipples, highly detailed skin textures, downblouse, accidental exposure.\"]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [413, 389], \"size\": {\"0\": 425.27801513671875, \"1\": 180.6060791015625}, \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 5}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [6, 12, 23], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_1, score_2, score_3\"]}, {\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1209, 188], \"size\": {\"0\": 210, \"1\": 46}, \"flags\": {}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 17}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9, 18], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1451, 189], \"size\": {\"0\": 210, \"1\": 270}, \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 10, \"type\": \"AlignYourStepsScheduler\", \"pos\": [-6, 888], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 0, \"mode\": 0, \"outputs\": [{\"name\": \"SIGMAS\", \"type\": \"SIGMAS\", \"links\": [13], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"AlignYourStepsScheduler\"}, \"widgets_values\": [\"SDXL\", 40, 1]}, {\"id\": 12, \"type\": \"SamplerCustom\", \"pos\": [418, 820], \"size\": {\"0\": 355.20001220703125, \"1\": 230}, \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 10}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 11}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 12}, {\"name\": \"sampler\", \"type\": \"SAMPLER\", \"link\": 14}, {\"name\": \"sigmas\", \"type\": \"SIGMAS\", \"link\": 13}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 15}], \"outputs\": [{\"name\": \"output\", \"type\": \"LATENT\", \"links\": [], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"denoised_output\", \"type\": \"LATENT\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"SamplerCustom\"}, \"widgets_values\": [true, 830502167753162, \"randomize\", 5]}, {\"id\": 13, \"type\": \"KSamplerSelect\", \"pos\": [7, 738], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 1, \"mode\": 0, \"outputs\": [{\"name\": \"SAMPLER\", \"type\": \"SAMPLER\", \"links\": [14], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"KSamplerSelect\"}, \"widgets_values\": [\"dpmpp_3m_sde_gpu\"]}, {\"id\": 15, \"type\": \"FaceDetailer\", \"pos\": [1139, 602], \"size\": {\"0\": 506.4000244140625, \"1\": 900}, \"flags\": {}, \"order\": 12, \"mode\": 2, \"inputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"link\": 18}, {\"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 21}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 20}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 22}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 23}, {\"name\": \"bbox_detector\", \"type\": \"BBOX_DETECTOR\", \"link\": 24}, {\"name\": \"sam_model_opt\", \"type\": \"SAM_MODEL\", \"link\": null}, {\"name\": \"segm_detector_opt\", \"type\": \"SEGM_DETECTOR\", \"link\": null}, {\"name\": \"detailer_hook\", \"type\": \"DETAILER_HOOK\", \"link\": null}, {\"name\": \"scheduler_func_opt\", \"type\": \"SCHEDULER_FUNC\", \"link\": null}], \"outputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"links\": [25], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"cropped_refined\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"cropped_enhanced_alpha\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"mask\", \"type\": \"MASK\", \"links\": null, \"shape\": 3}, {\"name\": \"detailer_pipe\", \"type\": \"DETAILER_PIPE\", \"links\": null, \"shape\": 3}, {\"name\": \"cnet_images\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}], \"properties\": {\"Node name for S&R\": \"FaceDetailer\"}, \"widgets_values\": [384, true, 1024, 225928200077487, \"randomize\", 20, 8, \"euler\", \"normal\", 0.5, 5, true, true, 0.5, 10, 3, \"center-1\", 0, 0.93, 0, 0.7, \"False\", 10, \"\", 1, false, 20]}, {\"id\": 16, \"type\": \"UltralyticsDetectorProvider\", \"pos\": [788, 796], \"size\": {\"0\": 315, \"1\": 78}, \"flags\": {}, \"order\": 2, \"mode\": 0, \"outputs\": [{\"name\": \"BBOX_DETECTOR\", \"type\": \"BBOX_DETECTOR\", \"links\": [24], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"SEGM_DETECTOR\", \"type\": \"SEGM_DETECTOR\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"UltralyticsDetectorProvider\"}, \"widgets_values\": [\"bbox/face_yolov8m.pt\"]}, {\"id\": 17, \"type\": \"SaveImage\", \"pos\": [1744, 566], \"size\": {\"0\": 315, \"1\": 270}, \"flags\": {}, \"order\": 13, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 25}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 18, \"type\": \"PerturbedAttentionGuidance\", \"pos\": [468, 87], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 26}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"PerturbedAttentionGuidance\"}, \"widgets_values\": [2]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [3, 4, 1, 6, 0, \"CLIP\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [5, 4, 1, 7, 0, \"CLIP\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [10, 4, 0, 12, 0, \"MODEL\"], [11, 6, 0, 12, 1, \"CONDITIONING\"], [12, 7, 0, 12, 2, \"CONDITIONING\"], [13, 10, 0, 12, 4, \"SIGMAS\"], [14, 13, 0, 12, 3, \"SAMPLER\"], [15, 5, 0, 12, 5, \"LATENT\"], [17, 3, 0, 8, 0, \"LATENT\"], [18, 8, 0, 15, 0, \"IMAGE\"], [19, 4, 0, 15, 1, \"MODEL\"], [20, 4, 2, 15, 3, \"VAE\"], [21, 4, 1, 15, 2, \"CLIP\"], [22, 6, 0, 15, 4, \"CONDITIONING\"], [23, 7, 0, 15, 5, \"CONDITIONING\"], [24, 16, 0, 15, 6, \"BBOX_DETECTOR\"], [25, 15, 0, 17, 0, \"IMAGE\"], [26, 4, 0, 18, 0, \"MODEL\"], [34, 4, 0, 3, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 1.3310000000000008, \"offset\": [119.3150225247812, -29.576837980502162]}, \"workspace_info\": {\"id\": \"TNOwVzKHGn6FUMraMM-Y7\"}}, \"version\": 0.4}}", |
| "steps": 40, |
| "width": 1024, |
| "height": 1024, |
| "models": [ |
| "cuwm1gxo-complete-20241024a.safetensors" |
| ], |
| "prompt": "score_6_up, Voyeur photograph looking down a woman's blouse, showing her perky nipples, highly detailed skin textures, downblouse, accidental exposure.", |
| "denoise": 1, |
| "sampler": "dpmpp_2m_sde_gpu", |
| "cfgScale": 3, |
| "modelIds": [], |
| "scheduler": "karras", |
| "upscalers": [], |
| "versionIds": [], |
| "controlNets": [], |
| "negativePrompt": "score_1, score_2, score_3", |
| "additionalResources": [] |
| }, |
| "availability": "Public", |
| "hasMeta": true, |
| "hasPositivePrompt": true, |
| "onSite": false, |
| "remixOfId": null |
| }, |
| { |
| "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/d5a63499-c8ab-4f15-8bf0-251b52ef1919/width=896/36428205.jpeg", |
| "nsfwLevel": 1, |
| "width": 896, |
| "height": 1152, |
| "hash": "UIK0~P_2~p~VE3_2?b.8?baeM|Io9F-;%2IU", |
| "type": "image", |
| "metadata": { |
| "hash": "UIK0~P_2~p~VE3_2?b.8?baeM|Io9F-;%2IU", |
| "size": 1810399, |
| "width": 896, |
| "height": 1152 |
| }, |
| "meta": { |
| "seed": 833289305935711, |
| "vaes": [], |
| "Model": "cuwm1gxo-complete-20241024a", |
| "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 833289305935711, \"steps\": 40, \"cfg\": 4.0, \"sampler_name\": \"dpmpp_3m_sde_gpu\", \"scheduler\": \"karras\", \"denoise\": 1.0, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\"}, \"4\": {\"inputs\": {\"ckpt_name\": \"cuwm1gxo-complete-20241024a.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\"}, \"5\": {\"inputs\": {\"width\": 896, \"height\": 1152, \"batch_size\": 8}, \"class_type\": \"EmptyLatentImage\"}, \"6\": {\"inputs\": {\"text\": \"score_7_up, Photograph of a handsome man, posing, dappled sunlight, trimmed beard, skin details, highly detailed, DSLR, depth of field, Instagram model, smiling seductively at the camera.\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"7\": {\"inputs\": {\"text\": \"score_1, score_2, score_3\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\"}, \"9\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\"}, \"10\": {\"inputs\": {\"model_type\": \"SDXL\", \"steps\": 40, \"denoise\": 1}, \"class_type\": \"AlignYourStepsScheduler\"}, \"12\": {\"inputs\": {\"add_noise\": true, \"noise_seed\": 1004801377935024, \"cfg\": 5, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"sampler\": [\"13\", 0], \"sigmas\": [\"10\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"SamplerCustom\"}, \"13\": {\"inputs\": {\"sampler_name\": \"dpmpp_3m_sde_gpu\"}, \"class_type\": \"KSamplerSelect\"}, \"16\": {\"inputs\": {\"model_name\": \"bbox/face_yolov8m.pt\"}, \"class_type\": \"UltralyticsDetectorProvider\"}, \"17\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\"}, \"class_type\": \"SaveImage\"}}, \"workflow\": {\"last_node_id\": 17, \"last_link_id\": 25, \"nodes\": [{\"id\": 3, \"type\": \"KSampler\", \"pos\": [863, 186], \"size\": {\"0\": 315, \"1\": 262}, \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 1}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [17], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [833289305935711, \"randomize\", 40, 4, \"dpmpp_3m_sde_gpu\", \"karras\", 1]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [9, 371], \"size\": {\"0\": 315, \"1\": 98}, \"flags\": {}, \"order\": 4, \"mode\": 0, \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [1, 10, 19], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [3, 5, 21], \"slot_index\": 1}, {\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [8, 20], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"CheckpointLoaderSimple\"}, \"widgets_values\": [\"cuwm1gxo-complete-20241024a.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [473, 609], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 2, \"mode\": 0, \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [2, 15], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"EmptyLatentImage\"}, \"widgets_values\": [896, 1152, 8]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [415, 186], \"size\": {\"0\": 422.84503173828125, \"1\": 164.31304931640625}, \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 3}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4, 11, 22], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_7_up, Photograph of a handsome man, posing, dappled sunlight, trimmed beard, skin details, highly detailed, DSLR, depth of field, Instagram model, smiling seductively at the camera.\"]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [413, 389], \"size\": {\"0\": 425.27801513671875, \"1\": 180.6060791015625}, \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 5}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [6, 12, 23], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_1, score_2, score_3\"]}, {\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1209, 188], \"size\": {\"0\": 210, \"1\": 46}, \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 17}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9, 18], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1451, 189], \"size\": {\"0\": 210, \"1\": 270}, \"flags\": {}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 10, \"type\": \"AlignYourStepsScheduler\", \"pos\": [6, 611], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 0, \"mode\": 0, \"outputs\": [{\"name\": \"SIGMAS\", \"type\": \"SIGMAS\", \"links\": [13], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"AlignYourStepsScheduler\"}, \"widgets_values\": [\"SDXL\", 40, 1]}, {\"id\": 12, \"type\": \"SamplerCustom\", \"pos\": [506, 797], \"size\": {\"0\": 355.20001220703125, \"1\": 230}, \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 10}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 11}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 12}, {\"name\": \"sampler\", \"type\": \"SAMPLER\", \"link\": 14}, {\"name\": \"sigmas\", \"type\": \"SIGMAS\", \"link\": 13}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 15}], \"outputs\": [{\"name\": \"output\", \"type\": \"LATENT\", \"links\": [], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"denoised_output\", \"type\": \"LATENT\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"SamplerCustom\"}, \"widgets_values\": [true, 1004801377935024, \"randomize\", 5]}, {\"id\": 13, \"type\": \"KSamplerSelect\", \"pos\": [7, 738], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 1, \"mode\": 0, \"outputs\": [{\"name\": \"SAMPLER\", \"type\": \"SAMPLER\", \"links\": [14], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"KSamplerSelect\"}, \"widgets_values\": [\"dpmpp_3m_sde_gpu\"]}, {\"id\": 15, \"type\": \"FaceDetailer\", \"pos\": [1139, 602], \"size\": {\"0\": 506.4000244140625, \"1\": 900}, \"flags\": {}, \"order\": 11, \"mode\": 2, \"inputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"link\": 18}, {\"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 21}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 20}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 22}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 23}, {\"name\": \"bbox_detector\", \"type\": \"BBOX_DETECTOR\", \"link\": 24}, {\"name\": \"sam_model_opt\", \"type\": \"SAM_MODEL\", \"link\": null}, {\"name\": \"segm_detector_opt\", \"type\": \"SEGM_DETECTOR\", \"link\": null}, {\"name\": \"detailer_hook\", \"type\": \"DETAILER_HOOK\", \"link\": null}, {\"name\": \"scheduler_func_opt\", \"type\": \"SCHEDULER_FUNC\", \"link\": null}], \"outputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"links\": [25], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"cropped_refined\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"cropped_enhanced_alpha\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"mask\", \"type\": \"MASK\", \"links\": null, \"shape\": 3}, {\"name\": \"detailer_pipe\", \"type\": \"DETAILER_PIPE\", \"links\": null, \"shape\": 3}, {\"name\": \"cnet_images\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}], \"properties\": {\"Node name for S&R\": \"FaceDetailer\"}, \"widgets_values\": [384, true, 1024, 307353808373916, \"randomize\", 20, 8, \"euler\", \"normal\", 0.5, 5, true, true, 0.5, 10, 3, \"center-1\", 0, 0.93, 0, 0.7, \"False\", 10, \"\", 1, false, 20]}, {\"id\": 16, \"type\": \"UltralyticsDetectorProvider\", \"pos\": [788, 796], \"size\": {\"0\": 315, \"1\": 78}, \"flags\": {}, \"order\": 3, \"mode\": 0, \"outputs\": [{\"name\": \"BBOX_DETECTOR\", \"type\": \"BBOX_DETECTOR\", \"links\": [24], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"SEGM_DETECTOR\", \"type\": \"SEGM_DETECTOR\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"UltralyticsDetectorProvider\"}, \"widgets_values\": [\"bbox/face_yolov8m.pt\"]}, {\"id\": 17, \"type\": \"SaveImage\", \"pos\": [1744, 566], \"size\": [315, 270], \"flags\": {}, \"order\": 12, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 25}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}], \"links\": [[1, 4, 0, 3, 0, \"MODEL\"], [2, 5, 0, 3, 3, \"LATENT\"], [3, 4, 1, 6, 0, \"CLIP\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [5, 4, 1, 7, 0, \"CLIP\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [10, 4, 0, 12, 0, \"MODEL\"], [11, 6, 0, 12, 1, \"CONDITIONING\"], [12, 7, 0, 12, 2, \"CONDITIONING\"], [13, 10, 0, 12, 4, \"SIGMAS\"], [14, 13, 0, 12, 3, \"SAMPLER\"], [15, 5, 0, 12, 5, \"LATENT\"], [17, 3, 0, 8, 0, \"LATENT\"], [18, 8, 0, 15, 0, \"IMAGE\"], [19, 4, 0, 15, 1, \"MODEL\"], [20, 4, 2, 15, 3, \"VAE\"], [21, 4, 1, 15, 2, \"CLIP\"], [22, 6, 0, 15, 4, \"CONDITIONING\"], [23, 7, 0, 15, 5, \"CONDITIONING\"], [24, 16, 0, 15, 6, \"BBOX_DETECTOR\"], [25, 15, 0, 17, 0, \"IMAGE\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 1.1, \"offset\": [209.48971257327597, -22.927613200773784]}, \"workspace_info\": {\"id\": \"4BkoIhhObUUR_t7Kt7_qu\"}}, \"version\": 0.4}}", |
| "steps": 40, |
| "width": 896, |
| "height": 1152, |
| "models": [ |
| "cuwm1gxo-complete-20241024a.safetensors" |
| ], |
| "prompt": "score_7_up, Photograph of a handsome man, posing, dappled sunlight, trimmed beard, skin details, highly detailed, DSLR, depth of field, Instagram model, smiling seductively at the camera.", |
| "denoise": 1, |
| "sampler": "dpmpp_3m_sde_gpu", |
| "cfgScale": 4, |
| "modelIds": [], |
| "scheduler": "karras", |
| "upscalers": [], |
| "versionIds": [], |
| "controlNets": [], |
| "negativePrompt": "score_1, score_2, score_3", |
| "additionalResources": [] |
| }, |
| "availability": "Public", |
| "hasMeta": true, |
| "hasPositivePrompt": true, |
| "onSite": false, |
| "remixOfId": null |
| }, |
| { |
| "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/97b4171c-5e81-4672-833a-a063a97e5672/width=896/36428369.jpeg", |
| "nsfwLevel": 1, |
| "width": 896, |
| "height": 1152, |
| "hash": "UHIqr{^i^*s;#l$%-Pxu0LNGIV-=_NN_tmx^", |
| "type": "image", |
| "metadata": { |
| "hash": "UHIqr{^i^*s;#l$%-Pxu0LNGIV-=_NN_tmx^", |
| "size": 1341383, |
| "width": 896, |
| "height": 1152 |
| }, |
| "meta": { |
| "seed": 713343998787738, |
| "vaes": [], |
| "Model": "cuwm1gxo-complete-20241024a", |
| "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 713343998787738, \"steps\": 40, \"cfg\": 5.0, \"sampler_name\": \"dpmpp_3m_sde_gpu\", \"scheduler\": \"exponential\", \"denoise\": 1.0, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\"}, \"4\": {\"inputs\": {\"ckpt_name\": \"cuwm1gxo-complete-20241024a.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\"}, \"5\": {\"inputs\": {\"width\": 896, \"height\": 1152, \"batch_size\": 8}, \"class_type\": \"EmptyLatentImage\"}, \"6\": {\"inputs\": {\"text\": \"score_7_up, Close-up photograph of a redheaded woman with a nose piercing, crossed arms, dappled sunlight, freckles on her face, skin details, highly detailed, DSLR, depth of field, Instagram model\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"7\": {\"inputs\": {\"text\": \"score_1, score_2, score_3\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\"}, \"9\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\"}, \"10\": {\"inputs\": {\"model_type\": \"SDXL\", \"steps\": 40, \"denoise\": 1}, \"class_type\": \"AlignYourStepsScheduler\"}, \"12\": {\"inputs\": {\"add_noise\": true, \"noise_seed\": 431439307717558, \"cfg\": 5, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"sampler\": [\"13\", 0], \"sigmas\": [\"10\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"SamplerCustom\"}, \"13\": {\"inputs\": {\"sampler_name\": \"dpmpp_3m_sde_gpu\"}, \"class_type\": \"KSamplerSelect\"}, \"15\": {\"inputs\": {\"guide_size\": 384.0, \"guide_size_for\": true, \"max_size\": 1024.0, \"seed\": 0, \"steps\": 20, \"cfg\": 8.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 0.5, \"feather\": 5, \"noise_mask\": true, \"force_inpaint\": true, \"bbox_threshold\": 0.5, \"bbox_dilation\": 10, \"bbox_crop_factor\": 3.0, \"sam_detection_hint\": \"center-1\", \"sam_dilation\": 0, \"sam_threshold\": 0.93, \"sam_bbox_expansion\": 0, \"sam_mask_hint_threshold\": 0.7, \"sam_mask_hint_use_negative\": \"False\", \"drop_size\": 10, \"wildcard\": \"\", \"cycle\": 1, \"inpaint_model\": false, \"noise_mask_feather\": 20, \"image\": [\"8\", 0], \"model\": [\"4\", 0], \"clip\": [\"4\", 1], \"vae\": [\"4\", 2], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"bbox_detector\": [\"16\", 0]}, \"class_type\": \"FaceDetailer\"}, \"16\": {\"inputs\": {\"model_name\": \"bbox/face_yolov8m.pt\"}, \"class_type\": \"UltralyticsDetectorProvider\"}, \"17\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"15\", 0]}, \"class_type\": \"SaveImage\"}}, \"workflow\": {\"last_node_id\": 17, \"last_link_id\": 25, \"nodes\": [{\"id\": 3, \"type\": \"KSampler\", \"pos\": [863, 186], \"size\": {\"0\": 315, \"1\": 262}, \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 1}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [17], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [713343998787738, \"randomize\", 40, 5, \"dpmpp_3m_sde_gpu\", \"exponential\", 1]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [9, 371], \"size\": {\"0\": 315, \"1\": 98}, \"flags\": {}, \"order\": 4, \"mode\": 0, \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [1, 10, 19], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [3, 5, 21], \"slot_index\": 1}, {\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [8, 20], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"CheckpointLoaderSimple\"}, \"widgets_values\": [\"cuwm1gxo-complete-20241024a.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [473, 609], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 2, \"mode\": 0, \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [2, 15], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"EmptyLatentImage\"}, \"widgets_values\": [896, 1152, 8]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [415, 186], \"size\": {\"0\": 422.84503173828125, \"1\": 164.31304931640625}, \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 3}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4, 11, 22], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_7_up, Close-up photograph of a redheaded woman with a nose piercing, crossed arms, dappled sunlight, freckles on her face, skin details, highly detailed, DSLR, depth of field, Instagram model\"]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [413, 389], \"size\": {\"0\": 425.27801513671875, \"1\": 180.6060791015625}, \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 5}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [6, 12, 23], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_1, score_2, score_3\"]}, {\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1209, 188], \"size\": {\"0\": 210, \"1\": 46}, \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 17}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9, 18], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1451, 189], \"size\": {\"0\": 210, \"1\": 270}, \"flags\": {}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 10, \"type\": \"AlignYourStepsScheduler\", \"pos\": [6, 611], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 0, \"mode\": 0, \"outputs\": [{\"name\": \"SIGMAS\", \"type\": \"SIGMAS\", \"links\": [13], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"AlignYourStepsScheduler\"}, \"widgets_values\": [\"SDXL\", 40, 1]}, {\"id\": 12, \"type\": \"SamplerCustom\", \"pos\": [506, 797], \"size\": {\"0\": 355.20001220703125, \"1\": 230}, \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 10}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 11}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 12}, {\"name\": \"sampler\", \"type\": \"SAMPLER\", \"link\": 14}, {\"name\": \"sigmas\", \"type\": \"SIGMAS\", \"link\": 13}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 15}], \"outputs\": [{\"name\": \"output\", \"type\": \"LATENT\", \"links\": [], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"denoised_output\", \"type\": \"LATENT\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"SamplerCustom\"}, \"widgets_values\": [true, 431439307717558, \"randomize\", 5]}, {\"id\": 13, \"type\": \"KSamplerSelect\", \"pos\": [7, 738], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 1, \"mode\": 0, \"outputs\": [{\"name\": \"SAMPLER\", \"type\": \"SAMPLER\", \"links\": [14], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"KSamplerSelect\"}, \"widgets_values\": [\"dpmpp_3m_sde_gpu\"]}, {\"id\": 15, \"type\": \"FaceDetailer\", \"pos\": [1139, 602], \"size\": {\"0\": 506.4000244140625, \"1\": 900}, \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"link\": 18}, {\"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 21}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 20}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 22}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 23}, {\"name\": \"bbox_detector\", \"type\": \"BBOX_DETECTOR\", \"link\": 24}, {\"name\": \"sam_model_opt\", \"type\": \"SAM_MODEL\", \"link\": null}, {\"name\": \"segm_detector_opt\", \"type\": \"SEGM_DETECTOR\", \"link\": null}, {\"name\": \"detailer_hook\", \"type\": \"DETAILER_HOOK\", \"link\": null}, {\"name\": \"scheduler_func_opt\", \"type\": \"SCHEDULER_FUNC\", \"link\": null}], \"outputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"links\": [25], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"cropped_refined\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"cropped_enhanced_alpha\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"mask\", \"type\": \"MASK\", \"links\": null, \"shape\": 3}, {\"name\": \"detailer_pipe\", \"type\": \"DETAILER_PIPE\", \"links\": null, \"shape\": 3}, {\"name\": \"cnet_images\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}], \"properties\": {\"Node name for S&R\": \"FaceDetailer\"}, \"widgets_values\": [384, true, 1024, 0, \"randomize\", 20, 8, \"euler\", \"normal\", 0.5, 5, true, true, 0.5, 10, 3, \"center-1\", 0, 0.93, 0, 0.7, \"False\", 10, \"\", 1, false, 20]}, {\"id\": 16, \"type\": \"UltralyticsDetectorProvider\", \"pos\": [788, 796], \"size\": {\"0\": 315, \"1\": 78}, \"flags\": {}, \"order\": 3, \"mode\": 0, \"outputs\": [{\"name\": \"BBOX_DETECTOR\", \"type\": \"BBOX_DETECTOR\", \"links\": [24], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"SEGM_DETECTOR\", \"type\": \"SEGM_DETECTOR\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"UltralyticsDetectorProvider\"}, \"widgets_values\": [\"bbox/face_yolov8m.pt\"]}, {\"id\": 17, \"type\": \"SaveImage\", \"pos\": [1744, 566], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 12, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 25}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}], \"links\": [[1, 4, 0, 3, 0, \"MODEL\"], [2, 5, 0, 3, 3, \"LATENT\"], [3, 4, 1, 6, 0, \"CLIP\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [5, 4, 1, 7, 0, \"CLIP\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [10, 4, 0, 12, 0, \"MODEL\"], [11, 6, 0, 12, 1, \"CONDITIONING\"], [12, 7, 0, 12, 2, \"CONDITIONING\"], [13, 10, 0, 12, 4, \"SIGMAS\"], [14, 13, 0, 12, 3, \"SAMPLER\"], [15, 5, 0, 12, 5, \"LATENT\"], [17, 3, 0, 8, 0, \"LATENT\"], [18, 8, 0, 15, 0, \"IMAGE\"], [19, 4, 0, 15, 1, \"MODEL\"], [20, 4, 2, 15, 3, \"VAE\"], [21, 4, 1, 15, 2, \"CLIP\"], [22, 6, 0, 15, 4, \"CONDITIONING\"], [23, 7, 0, 15, 5, \"CONDITIONING\"], [24, 16, 0, 15, 6, \"BBOX_DETECTOR\"], [25, 15, 0, 17, 0, \"IMAGE\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.8264462809917354, \"offset\": [11.761042492999877, -56.49801817020733]}, \"workspace_info\": {\"id\": \"4BkoIhhObUUR_t7Kt7_qu\"}}, \"version\": 0.4}}", |
| "steps": 40, |
| "width": 896, |
| "height": 1152, |
| "models": [ |
| "cuwm1gxo-complete-20241024a.safetensors" |
| ], |
| "prompt": "score_7_up, Close-up photograph of a redheaded woman with a nose piercing, crossed arms, dappled sunlight, freckles on her face, skin details, highly detailed, DSLR, depth of field, Instagram model", |
| "denoise": 1, |
| "sampler": "dpmpp_3m_sde_gpu", |
| "cfgScale": 5, |
| "modelIds": [], |
| "scheduler": "exponential", |
| "upscalers": [], |
| "versionIds": [], |
| "controlNets": [], |
| "negativePrompt": "score_1, score_2, score_3", |
| "additionalResources": [] |
| }, |
| "availability": "Public", |
| "hasMeta": true, |
| "hasPositivePrompt": true, |
| "onSite": false, |
| "remixOfId": null |
| }, |
| { |
| "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e4f45114-15ee-4fa6-ac92-7b36122e9751/width=896/36428116.jpeg", |
| "nsfwLevel": 1, |
| "width": 896, |
| "height": 1152, |
| "hash": "UIHoIBWBxu_4_NWBWBxu.8IAD%%MM_t8ofWB", |
| "type": "image", |
| "metadata": { |
| "hash": "UIHoIBWBxu_4_NWBWBxu.8IAD%%MM_t8ofWB", |
| "size": 1531859, |
| "width": 896, |
| "height": 1152 |
| }, |
| "meta": { |
| "seed": 127646902657452, |
| "vaes": [], |
| "Model": "cuwm1gxo-complete-20241024a", |
| "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 127646902657452, \"steps\": 40, \"cfg\": 2.0, \"sampler_name\": \"dpmpp_2m_sde_gpu\", \"scheduler\": \"karras\", \"denoise\": 1.0, \"model\": [\"18\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\"}, \"4\": {\"inputs\": {\"ckpt_name\": \"cuwm1gxo-complete-20241024a.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\"}, \"5\": {\"inputs\": {\"width\": 896, \"height\": 1152, \"batch_size\": 8}, \"class_type\": \"EmptyLatentImage\"}, \"6\": {\"inputs\": {\"text\": \"score_7_up, Photo of an owl perched on a branch amidst a foggy forest. Shallow depth of field, taken with a Canon DSLR. Detailed feathers and patterns.\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"7\": {\"inputs\": {\"text\": \"score_1, score_2, score_3\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\"}, \"9\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\"}, \"10\": {\"inputs\": {\"model_type\": \"SDXL\", \"steps\": 40, \"denoise\": 1}, \"class_type\": \"AlignYourStepsScheduler\"}, \"12\": {\"inputs\": {\"add_noise\": true, \"noise_seed\": 316084924494979, \"cfg\": 5, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"sampler\": [\"13\", 0], \"sigmas\": [\"10\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"SamplerCustom\"}, \"13\": {\"inputs\": {\"sampler_name\": \"dpmpp_3m_sde_gpu\"}, \"class_type\": \"KSamplerSelect\"}, \"16\": {\"inputs\": {\"model_name\": \"bbox/face_yolov8m.pt\"}, \"class_type\": \"UltralyticsDetectorProvider\"}, \"17\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\"}, \"class_type\": \"SaveImage\"}, \"18\": {\"inputs\": {\"scale\": 3.0, \"model\": [\"4\", 0]}, \"class_type\": \"PerturbedAttentionGuidance\"}}, \"workflow\": {\"last_node_id\": 18, \"last_link_id\": 27, \"nodes\": [{\"id\": 3, \"type\": \"KSampler\", \"pos\": [868, 167], \"size\": {\"0\": 315, \"1\": 262}, \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 27}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [17], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [127646902657452, \"randomize\", 40, 2, \"dpmpp_2m_sde_gpu\", \"karras\", 1]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [9, 371], \"size\": {\"0\": 315, \"1\": 98}, \"flags\": {}, \"order\": 4, \"mode\": 0, \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [10, 19, 26], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [3, 5, 21], \"slot_index\": 1}, {\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [8, 20], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"CheckpointLoaderSimple\"}, \"widgets_values\": [\"cuwm1gxo-complete-20241024a.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [473, 609], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 1, \"mode\": 0, \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [2, 15], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"EmptyLatentImage\"}, \"widgets_values\": [896, 1152, 8]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [415, 186], \"size\": {\"0\": 422.84503173828125, \"1\": 164.31304931640625}, \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 3}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4, 11, 22], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_7_up, Photo of an owl perched on a branch amidst a foggy forest. Shallow depth of field, taken with a Canon DSLR. Detailed feathers and patterns.\"]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [413, 389], \"size\": {\"0\": 425.27801513671875, \"1\": 180.6060791015625}, \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 5}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [6, 12, 23], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_1, score_2, score_3\"]}, {\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1209, 188], \"size\": {\"0\": 210, \"1\": 46}, \"flags\": {}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 17}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9, 18], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1451, 189], \"size\": {\"0\": 210, \"1\": 270}, \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 10, \"type\": \"AlignYourStepsScheduler\", \"pos\": [-6, 888], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 3, \"mode\": 0, \"outputs\": [{\"name\": \"SIGMAS\", \"type\": \"SIGMAS\", \"links\": [13], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"AlignYourStepsScheduler\"}, \"widgets_values\": [\"SDXL\", 40, 1]}, {\"id\": 12, \"type\": \"SamplerCustom\", \"pos\": [418, 820], \"size\": {\"0\": 355.20001220703125, \"1\": 230}, \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 10}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 11}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 12}, {\"name\": \"sampler\", \"type\": \"SAMPLER\", \"link\": 14}, {\"name\": \"sigmas\", \"type\": \"SIGMAS\", \"link\": 13}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 15}], \"outputs\": [{\"name\": \"output\", \"type\": \"LATENT\", \"links\": [], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"denoised_output\", \"type\": \"LATENT\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"SamplerCustom\"}, \"widgets_values\": [true, 316084924494979, \"randomize\", 5]}, {\"id\": 13, \"type\": \"KSamplerSelect\", \"pos\": [7, 738], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 0, \"mode\": 0, \"outputs\": [{\"name\": \"SAMPLER\", \"type\": \"SAMPLER\", \"links\": [14], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"KSamplerSelect\"}, \"widgets_values\": [\"dpmpp_3m_sde_gpu\"]}, {\"id\": 15, \"type\": \"FaceDetailer\", \"pos\": [1139, 602], \"size\": {\"0\": 506.4000244140625, \"1\": 900}, \"flags\": {}, \"order\": 12, \"mode\": 2, \"inputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"link\": 18}, {\"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 21}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 20}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 22}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 23}, {\"name\": \"bbox_detector\", \"type\": \"BBOX_DETECTOR\", \"link\": 24}, {\"name\": \"sam_model_opt\", \"type\": \"SAM_MODEL\", \"link\": null}, {\"name\": \"segm_detector_opt\", \"type\": \"SEGM_DETECTOR\", \"link\": null}, {\"name\": \"detailer_hook\", \"type\": \"DETAILER_HOOK\", \"link\": null}, {\"name\": \"scheduler_func_opt\", \"type\": \"SCHEDULER_FUNC\", \"link\": null}], \"outputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"links\": [25], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"cropped_refined\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"cropped_enhanced_alpha\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"mask\", \"type\": \"MASK\", \"links\": null, \"shape\": 3}, {\"name\": \"detailer_pipe\", \"type\": \"DETAILER_PIPE\", \"links\": null, \"shape\": 3}, {\"name\": \"cnet_images\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}], \"properties\": {\"Node name for S&R\": \"FaceDetailer\"}, \"widgets_values\": [384, true, 1024, 778750805372175, \"randomize\", 20, 8, \"euler\", \"normal\", 0.5, 5, true, true, 0.5, 10, 3, \"center-1\", 0, 0.93, 0, 0.7, \"False\", 10, \"\", 1, false, 20]}, {\"id\": 16, \"type\": \"UltralyticsDetectorProvider\", \"pos\": [788, 796], \"size\": {\"0\": 315, \"1\": 78}, \"flags\": {}, \"order\": 2, \"mode\": 0, \"outputs\": [{\"name\": \"BBOX_DETECTOR\", \"type\": \"BBOX_DETECTOR\", \"links\": [24], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"SEGM_DETECTOR\", \"type\": \"SEGM_DETECTOR\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"UltralyticsDetectorProvider\"}, \"widgets_values\": [\"bbox/face_yolov8m.pt\"]}, {\"id\": 17, \"type\": \"SaveImage\", \"pos\": [1744, 566], \"size\": [315, 270], \"flags\": {}, \"order\": 13, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 25}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 18, \"type\": \"PerturbedAttentionGuidance\", \"pos\": [468, 87], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 26}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [27], \"shape\": 3, \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"PerturbedAttentionGuidance\"}, \"widgets_values\": [3]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [3, 4, 1, 6, 0, \"CLIP\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [5, 4, 1, 7, 0, \"CLIP\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [10, 4, 0, 12, 0, \"MODEL\"], [11, 6, 0, 12, 1, \"CONDITIONING\"], [12, 7, 0, 12, 2, \"CONDITIONING\"], [13, 10, 0, 12, 4, \"SIGMAS\"], [14, 13, 0, 12, 3, \"SAMPLER\"], [15, 5, 0, 12, 5, \"LATENT\"], [17, 3, 0, 8, 0, \"LATENT\"], [18, 8, 0, 15, 0, \"IMAGE\"], [19, 4, 0, 15, 1, \"MODEL\"], [20, 4, 2, 15, 3, \"VAE\"], [21, 4, 1, 15, 2, \"CLIP\"], [22, 6, 0, 15, 4, \"CONDITIONING\"], [23, 7, 0, 15, 5, \"CONDITIONING\"], [24, 16, 0, 15, 6, \"BBOX_DETECTOR\"], [25, 15, 0, 17, 0, \"IMAGE\"], [26, 4, 0, 18, 0, \"MODEL\"], [27, 18, 0, 3, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 1.2100000000000002, \"offset\": [-23.468940319286656, 42.23519671658151]}, \"workspace_info\": {\"id\": \"alxrbIVbToHFtdDV0ok62\"}}, \"version\": 0.4}}", |
| "steps": 40, |
| "width": 896, |
| "height": 1152, |
| "models": [ |
| "cuwm1gxo-complete-20241024a.safetensors" |
| ], |
| "prompt": "score_7_up, Photo of an owl perched on a branch amidst a foggy forest. Shallow depth of field, taken with a Canon DSLR. Detailed feathers and patterns.", |
| "denoise": 1, |
| "sampler": "dpmpp_2m_sde_gpu", |
| "cfgScale": 2, |
| "modelIds": [], |
| "scheduler": "karras", |
| "upscalers": [], |
| "versionIds": [], |
| "controlNets": [], |
| "negativePrompt": "score_1, score_2, score_3", |
| "additionalResources": [] |
| }, |
| "availability": "Public", |
| "hasMeta": true, |
| "hasPositivePrompt": true, |
| "onSite": false, |
| "remixOfId": null |
| }, |
| { |
| "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/5de3d574-abc9-4b54-87c1-d7437bdc81ec/width=896/36428187.jpeg", |
| "nsfwLevel": 1, |
| "width": 896, |
| "height": 1152, |
| "hash": "UKKdJD9Fx^IA~q%fx]VsS3-WI;NZ%f-;M{WC", |
| "type": "image", |
| "metadata": { |
| "hash": "UKKdJD9Fx^IA~q%fx]VsS3-WI;NZ%f-;M{WC", |
| "size": 1535232, |
| "width": 896, |
| "height": 1152 |
| }, |
| "meta": { |
| "seed": 432973723624219, |
| "vaes": [], |
| "Model": "cuwm1gxo-complete-20241024a", |
| "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 432973723624219, \"steps\": 40, \"cfg\": 4.0, \"sampler_name\": \"dpmpp_3m_sde_gpu\", \"scheduler\": \"karras\", \"denoise\": 1.0, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\"}, \"4\": {\"inputs\": {\"ckpt_name\": \"cuwm1gxo-complete-20241024a.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\"}, \"5\": {\"inputs\": {\"width\": 896, \"height\": 1152, \"batch_size\": 8}, \"class_type\": \"EmptyLatentImage\"}, \"6\": {\"inputs\": {\"text\": \"score_7_up, Photograph of a handsome man, posing, dappled sunlight, trimmed beard, skin details, highly detailed, DSLR, depth of field, Instagram model, smiling seductively at the camera, wearing a firefighting outfit.\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"7\": {\"inputs\": {\"text\": \"score_1, score_2, score_3\", \"clip\": [\"4\", 1]}, \"class_type\": \"CLIPTextEncode\"}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\"}, \"9\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\"}, \"10\": {\"inputs\": {\"model_type\": \"SDXL\", \"steps\": 40, \"denoise\": 1}, \"class_type\": \"AlignYourStepsScheduler\"}, \"12\": {\"inputs\": {\"add_noise\": true, \"noise_seed\": 626998365859564, \"cfg\": 5, \"model\": [\"4\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"sampler\": [\"13\", 0], \"sigmas\": [\"10\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"SamplerCustom\"}, \"13\": {\"inputs\": {\"sampler_name\": \"dpmpp_3m_sde_gpu\"}, \"class_type\": \"KSamplerSelect\"}, \"16\": {\"inputs\": {\"model_name\": \"bbox/face_yolov8m.pt\"}, \"class_type\": \"UltralyticsDetectorProvider\"}, \"17\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\"}, \"class_type\": \"SaveImage\"}}, \"workflow\": {\"last_node_id\": 17, \"last_link_id\": 25, \"nodes\": [{\"id\": 3, \"type\": \"KSampler\", \"pos\": [863, 186], \"size\": {\"0\": 315, \"1\": 262}, \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 1}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [17], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [432973723624219, \"randomize\", 40, 4, \"dpmpp_3m_sde_gpu\", \"karras\", 1]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [9, 371], \"size\": {\"0\": 315, \"1\": 98}, \"flags\": {}, \"order\": 4, \"mode\": 0, \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [1, 10, 19], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [3, 5, 21], \"slot_index\": 1}, {\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [8, 20], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"CheckpointLoaderSimple\"}, \"widgets_values\": [\"cuwm1gxo-complete-20241024a.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [473, 609], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 2, \"mode\": 0, \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [2, 15], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"EmptyLatentImage\"}, \"widgets_values\": [896, 1152, 8]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [415, 186], \"size\": {\"0\": 422.84503173828125, \"1\": 164.31304931640625}, \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 3}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4, 11, 22], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_7_up, Photograph of a handsome man, posing, dappled sunlight, trimmed beard, skin details, highly detailed, DSLR, depth of field, Instagram model, smiling seductively at the camera, wearing a firefighting outfit.\"]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [413, 389], \"size\": {\"0\": 425.27801513671875, \"1\": 180.6060791015625}, \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 5}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [6, 12, 23], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"score_1, score_2, score_3\"]}, {\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1209, 188], \"size\": {\"0\": 210, \"1\": 46}, \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 17}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9, 18], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1451, 189], \"size\": {\"0\": 210, \"1\": 270}, \"flags\": {}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 10, \"type\": \"AlignYourStepsScheduler\", \"pos\": [6, 611], \"size\": {\"0\": 315, \"1\": 106}, \"flags\": {}, \"order\": 0, \"mode\": 0, \"outputs\": [{\"name\": \"SIGMAS\", \"type\": \"SIGMAS\", \"links\": [13], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"AlignYourStepsScheduler\"}, \"widgets_values\": [\"SDXL\", 40, 1]}, {\"id\": 12, \"type\": \"SamplerCustom\", \"pos\": [506, 797], \"size\": {\"0\": 355.20001220703125, \"1\": 230}, \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 10}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 11}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 12}, {\"name\": \"sampler\", \"type\": \"SAMPLER\", \"link\": 14}, {\"name\": \"sigmas\", \"type\": \"SIGMAS\", \"link\": 13}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 15}], \"outputs\": [{\"name\": \"output\", \"type\": \"LATENT\", \"links\": [], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"denoised_output\", \"type\": \"LATENT\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"SamplerCustom\"}, \"widgets_values\": [true, 626998365859564, \"randomize\", 5]}, {\"id\": 13, \"type\": \"KSamplerSelect\", \"pos\": [7, 738], \"size\": {\"0\": 315, \"1\": 58}, \"flags\": {}, \"order\": 1, \"mode\": 0, \"outputs\": [{\"name\": \"SAMPLER\", \"type\": \"SAMPLER\", \"links\": [14], \"slot_index\": 0, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"KSamplerSelect\"}, \"widgets_values\": [\"dpmpp_3m_sde_gpu\"]}, {\"id\": 15, \"type\": \"FaceDetailer\", \"pos\": [1139, 602], \"size\": {\"0\": 506.4000244140625, \"1\": 900}, \"flags\": {}, \"order\": 11, \"mode\": 2, \"inputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"link\": 18}, {\"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 21}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 20}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 22}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 23}, {\"name\": \"bbox_detector\", \"type\": \"BBOX_DETECTOR\", \"link\": 24}, {\"name\": \"sam_model_opt\", \"type\": \"SAM_MODEL\", \"link\": null}, {\"name\": \"segm_detector_opt\", \"type\": \"SEGM_DETECTOR\", \"link\": null}, {\"name\": \"detailer_hook\", \"type\": \"DETAILER_HOOK\", \"link\": null}, {\"name\": \"scheduler_func_opt\", \"type\": \"SCHEDULER_FUNC\", \"link\": null}], \"outputs\": [{\"name\": \"image\", \"type\": \"IMAGE\", \"links\": [25], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"cropped_refined\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"cropped_enhanced_alpha\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}, {\"name\": \"mask\", \"type\": \"MASK\", \"links\": null, \"shape\": 3}, {\"name\": \"detailer_pipe\", \"type\": \"DETAILER_PIPE\", \"links\": null, \"shape\": 3}, {\"name\": \"cnet_images\", \"type\": \"IMAGE\", \"links\": null, \"shape\": 6}], \"properties\": {\"Node name for S&R\": \"FaceDetailer\"}, \"widgets_values\": [384, true, 1024, 617748877836126, \"randomize\", 20, 8, \"euler\", \"normal\", 0.5, 5, true, true, 0.5, 10, 3, \"center-1\", 0, 0.93, 0, 0.7, \"False\", 10, \"\", 1, false, 20]}, {\"id\": 16, \"type\": \"UltralyticsDetectorProvider\", \"pos\": [788, 796], \"size\": {\"0\": 315, \"1\": 78}, \"flags\": {}, \"order\": 3, \"mode\": 0, \"outputs\": [{\"name\": \"BBOX_DETECTOR\", \"type\": \"BBOX_DETECTOR\", \"links\": [24], \"slot_index\": 0, \"shape\": 3}, {\"name\": \"SEGM_DETECTOR\", \"type\": \"SEGM_DETECTOR\", \"links\": null, \"shape\": 3}], \"properties\": {\"Node name for S&R\": \"UltralyticsDetectorProvider\"}, \"widgets_values\": [\"bbox/face_yolov8m.pt\"]}, {\"id\": 17, \"type\": \"SaveImage\", \"pos\": [1744, 566], \"size\": [315, 270], \"flags\": {}, \"order\": 12, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 25}], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}], \"links\": [[1, 4, 0, 3, 0, \"MODEL\"], [2, 5, 0, 3, 3, \"LATENT\"], [3, 4, 1, 6, 0, \"CLIP\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [5, 4, 1, 7, 0, \"CLIP\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [10, 4, 0, 12, 0, \"MODEL\"], [11, 6, 0, 12, 1, \"CONDITIONING\"], [12, 7, 0, 12, 2, \"CONDITIONING\"], [13, 10, 0, 12, 4, \"SIGMAS\"], [14, 13, 0, 12, 3, \"SAMPLER\"], [15, 5, 0, 12, 5, \"LATENT\"], [17, 3, 0, 8, 0, \"LATENT\"], [18, 8, 0, 15, 0, \"IMAGE\"], [19, 4, 0, 15, 1, \"MODEL\"], [20, 4, 2, 15, 3, \"VAE\"], [21, 4, 1, 15, 2, \"CLIP\"], [22, 6, 0, 15, 4, \"CONDITIONING\"], [23, 7, 0, 15, 5, \"CONDITIONING\"], [24, 16, 0, 15, 6, \"BBOX_DETECTOR\"], [25, 15, 0, 17, 0, \"IMAGE\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 1.1, \"offset\": [209.48971257327597, -22.927613200773784]}, \"workspace_info\": {\"id\": \"4BkoIhhObUUR_t7Kt7_qu\"}}, \"version\": 0.4}}", |
| "steps": 40, |
| "width": 896, |
| "height": 1152, |
| "models": [ |
| "cuwm1gxo-complete-20241024a.safetensors" |
| ], |
| "prompt": "score_7_up, Photograph of a handsome man, posing, dappled sunlight, trimmed beard, skin details, highly detailed, DSLR, depth of field, Instagram model, smiling seductively at the camera, wearing a firefighting outfit.", |
| "denoise": 1, |
| "sampler": "dpmpp_3m_sde_gpu", |
| "cfgScale": 4, |
| "modelIds": [], |
| "scheduler": "karras", |
| "upscalers": [], |
| "versionIds": [], |
| "controlNets": [], |
| "negativePrompt": "score_1, score_2, score_3", |
| "additionalResources": [] |
| }, |
| "availability": "Public", |
| "hasMeta": true, |
| "hasPositivePrompt": true, |
| "onSite": false, |
| "remixOfId": null |
| } |
| ], |
| "downloadUrl": "https://civitai.com/api/download/models/991916" |
| } |