diff --git "a/Stable-diffusion/illumiyumeXL_v35VPred.html" "b/Stable-diffusion/illumiyumeXL_v35VPred.html" new file mode 100644--- /dev/null +++ "b/Stable-diffusion/illumiyumeXL_v35VPred.html" @@ -0,0 +1,230 @@ + + + + + + +
+

IllumiYume XL (Illustrious)

+

Uploaded by duongve13112002

+
+
+
Version
+
v3.5 (v-pred)
+
Base Model
+
Illustrious
+
Published
+
2025-06-10
+
Availability
+
Public
+
CivitAI Tags
+
+
+ animebase modelgirlscharactersnsfwsfw +
+
+
Download Link
+
https://civitai.com/api/download/models/1887507
+
+
+
+

Use the model without crediting the creator
Sell images they generate
Run on services that generate images for money
Run on Civitai
Share merges using this model
Sell this model or merges using this model
Have different permissions when sharing merges

+
+
+
+ +
+

Description

+

Introduction

For version 1.0:

For version 2.0:

For version 3.0:

For version 3.1:

For version 3.2:

For version 3.5:

Important Note

Currently, the model is not available for use via Civitai Generation. You can visit the following website to use it:

Suggested settings:

All example images were generated using the following settings:

Note: I don't use any post-processing and Lora to enhance the example images. I only use these settings and a custom prompt with my base model to generate.

Acknowledgments

If you'd like to support my work, you can do so through Ko-fi!

+
+ +
+
+
+
+ + + + +
+ +
+
Prompt
, masterpiece,best quality,amazing quality, Gawr Gura, 1girl, solo, derp face , standing, blue eyes, holding a sign, sign says "I am Gura" , english text
Negative prompt
bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background
Seed
416054014329380
Model
illumiyume_35
Sampler
DPM++ 2S a
Steps
30
CFG scale
5
+
+
+ + +
+
Vaes
[]
Comfy
{"prompt": {"3": {"inputs": {"seed": 416054014329380, "steps": 30, "cfg": 5.0, "sampler_name": "dpmpp_2s_ancestral", "scheduler": "normal", "denoise": 1.0, "model": ["4", 0], "positive": ["6", 0], "negative": ["7", 0], "latent_image": ["5", 0]}, "class_type": "KSampler", "_meta": {"title": "KSampler"}}, "4": {"inputs": {"ckpt_name": "illumiyume_35.safetensors"}, "class_type": "CheckpointLoaderSimple", "_meta": {"title": "Load Checkpoint"}}, "5": {"inputs": {"width": 1024, "height": 1536, "batch_size": 2}, "class_type": "EmptyLatentImage", "_meta": {"title": "Empty Latent Image"}}, "6": {"inputs": {"text": ", masterpiece,best quality,amazing quality, Gawr Gura, 1girl, solo, derp face , standing, blue eyes, holding a sign, sign says \"I am Gura\" , english text", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "7": {"inputs": {"text": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "8": {"inputs": {"samples": ["3", 0], "vae": ["4", 2]}, "class_type": "VAEDecode", "_meta": {"title": "VAE Decode"}}, "9": {"inputs": {"filename_prefix": "ComfyUI", "images": ["8", 0]}, "class_type": "SaveImage", "_meta": {"title": "Save Image"}}, "10": {"inputs": {"stop_at_clip_layer": -2, "clip": ["4", 1]}, "class_type": "CLIPSetLastLayer", "_meta": {"title": "CLIP Set Last Layer"}}}, "workflow": {"id": "7f35ca53-6911-4aba-ade5-a7087933ca8d", "revision": 0, "last_node_id": 10, "last_link_id": 12, "nodes": [{"id": 8, "type": "VAEDecode", "pos": [1209, 188], "size": [210, 46], "flags": {}, "order": 6, "mode": 0, "inputs": [{"name": "samples", "type": "LATENT", "link": 7}, {"name": "vae", "type": "VAE", "link": 8}], "outputs": [{"name": "IMAGE", "type": "IMAGE", "slot_index": 0, "links": [9]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "VAEDecode", "widget_ue_connectable": {}}, "widgets_values": []}, {"id": 9, "type": "SaveImage", "pos": [1451, 189], "size": [210, 58], "flags": {}, "order": 7, "mode": 0, "inputs": [{"name": "images", "type": "IMAGE", "link": 9}], "outputs": [], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "widget_ue_connectable": {}}, "widgets_values": ["ComfyUI"]}, {"id": 4, "type": "CheckpointLoaderSimple", "pos": [-272.2301025390625, 457.4971923828125], "size": [315, 98], "flags": {}, "order": 0, "mode": 0, "inputs": [], "outputs": [{"name": "MODEL", "type": "MODEL", "slot_index": 0, "links": [1]}, {"name": "CLIP", "type": "CLIP", "slot_index": 1, "links": [10]}, {"name": "VAE", "type": "VAE", "slot_index": 2, "links": [8]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CheckpointLoaderSimple", "widget_ue_connectable": {}}, "widgets_values": ["illumiyume_35.safetensors"]}, {"id": 10, "type": "CLIPSetLastLayer", "pos": [41.93305206298828, 215.06910705566406], "size": [270, 58], "flags": {}, "order": 2, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 10}], "outputs": [{"name": "CLIP", "type": "CLIP", "links": [11, 12]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPSetLastLayer", "widget_ue_connectable": {}}, "widgets_values": [-2]}, {"id": 5, "type": "EmptyLatentImage", "pos": [473, 609], "size": [315, 106], "flags": {}, "order": 1, "mode": 0, "inputs": [], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [2]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "EmptyLatentImage", "widget_ue_connectable": {}}, "widgets_values": [1024, 1536, 2]}, {"id": 3, "type": "KSampler", "pos": [863, 186], "size": [315, 262], "flags": {}, "order": 5, "mode": 0, "inputs": [{"name": "model", "type": "MODEL", "link": 1}, {"name": "positive", "type": "CONDITIONING", "link": 4}, {"name": "negative", "type": "CONDITIONING", "link": 6}, {"name": "latent_image", "type": "LATENT", "link": 2}], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [7]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "KSampler", "widget_ue_connectable": {}}, "widgets_values": [416054014329380, "randomize", 30, 5, "dpmpp_2s_ancestral", "normal", 1]}, {"id": 7, "type": "CLIPTextEncode", "pos": [413, 389], "size": [425.27801513671875, 180.6060791015625], "flags": {}, "order": 4, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 12}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [6]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background"]}, {"id": 6, "type": "CLIPTextEncode", "pos": [415, 186], "size": [422.84503173828125, 164.31304931640625], "flags": {}, "order": 3, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 11}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [4]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": [", masterpiece,best quality,amazing quality, Gawr Gura, 1girl, solo, derp face , standing, blue eyes, holding a sign, sign says \"I am Gura\" , english text"]}], "links": [[1, 4, 0, 3, 0, "MODEL"], [2, 5, 0, 3, 3, "LATENT"], [4, 6, 0, 3, 1, "CONDITIONING"], [6, 7, 0, 3, 2, "CONDITIONING"], [7, 3, 0, 8, 0, "LATENT"], [8, 4, 2, 8, 1, "VAE"], [9, 8, 0, 9, 0, "IMAGE"], [10, 4, 1, 10, 0, "CLIP"], [11, 10, 0, 6, 0, "CLIP"], [12, 10, 0, 7, 0, "CLIP"]], "groups": [], "config": {}, "extra": {"ue_links": [], "links_added_by_ue": [], "ds": {"scale": 0.8483383685800604, "offset": [112.48646723646735, 111.45156695156695]}, "frontendVersion": "1.21.7"}, "version": 0.4}}
Width
1024
Height
1536
Models
['illumiyume_35.safetensors']
Denoise
1
Modelids
[]
Scheduler
normal
Upscalers
[]
Versionids
[]
Controlnets
[]
Additionalresources
[]
+
+
+ + + + +
+ +
+
Prompt
masterpiece,best quality,amazing quality, nakano ichika, nakano itsuki, nakano miku, nakano nino, nakano yotsuba, 4-finger heart hands, 5girls, :D, :O, ;D, ;O, ahoge, backlighting, blazer, bloom, blue cardigan, blue eyes, blush, bow, breasts, brown bow, brown hair, brown jacket, cardigan, collared shirt, dress shirt, fingernails, green bow, green hairband, green ribbon, green skirt, hair between eyes, hair bow, hair ornament, hair ribbon, hairband, hands up, headphones, headphones around neck, heart, heart hands, jacket, layered clothes, layered sleeves, lineup, long sleeves, looking at viewer, miniskirt, multiple girls, one eye closed, open clothes, open jacket, open mouth, parted lips, plaid, plaid bow, pleated skirt, quintuplets, red hair, red sweater vest, ribbon, school uniform, shirt, short hair, siblings, sisters, skirt, small breasts, smile, star (symbol), star hair ornament, sweater vest, twitter username, untucked shirt, v-neck, white shirt, wing collar, yellow sweater vest
Negative prompt
bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background
Seed
40217979478388
Model
illumiyume_35
Sampler
DPM++ 2S a
Steps
30
CFG scale
5
+
+
+ + +
+
Vaes
[]
Comfy
{"prompt": {"3": {"inputs": {"seed": 40217979478388, "steps": 30, "cfg": 5.0, "sampler_name": "dpmpp_2s_ancestral", "scheduler": "normal", "denoise": 1.0, "model": ["4", 0], "positive": ["6", 0], "negative": ["7", 0], "latent_image": ["5", 0]}, "class_type": "KSampler", "_meta": {"title": "KSampler"}}, "4": {"inputs": {"ckpt_name": "illumiyume_35.safetensors"}, "class_type": "CheckpointLoaderSimple", "_meta": {"title": "Load Checkpoint"}}, "5": {"inputs": {"width": 1024, "height": 1536, "batch_size": 3}, "class_type": "EmptyLatentImage", "_meta": {"title": "Empty Latent Image"}}, "6": {"inputs": {"text": "masterpiece,best quality,amazing quality, nakano ichika, nakano itsuki, nakano miku, nakano nino, nakano yotsuba, 4-finger heart hands, 5girls, :D, :O, ;D, ;O, ahoge, backlighting, blazer, bloom, blue cardigan, blue eyes, blush, bow, breasts, brown bow, brown hair, brown jacket, cardigan, collared shirt, dress shirt, fingernails, green bow, green hairband, green ribbon, green skirt, hair between eyes, hair bow, hair ornament, hair ribbon, hairband, hands up, headphones, headphones around neck, heart, heart hands, jacket, layered clothes, layered sleeves, lineup, long sleeves, looking at viewer, miniskirt, multiple girls, one eye closed, open clothes, open jacket, open mouth, parted lips, plaid, plaid bow, pleated skirt, quintuplets, red hair, red sweater vest, ribbon, school uniform, shirt, short hair, siblings, sisters, skirt, small breasts, smile, star (symbol), star hair ornament, sweater vest, twitter username, untucked shirt, v-neck, white shirt, wing collar, yellow sweater vest", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "7": {"inputs": {"text": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "8": {"inputs": {"samples": ["3", 0], "vae": ["4", 2]}, "class_type": "VAEDecode", "_meta": {"title": "VAE Decode"}}, "9": {"inputs": {"filename_prefix": "ComfyUI", "images": ["8", 0]}, "class_type": "SaveImage", "_meta": {"title": "Save Image"}}, "10": {"inputs": {"stop_at_clip_layer": -2, "clip": ["4", 1]}, "class_type": "CLIPSetLastLayer", "_meta": {"title": "CLIP Set Last Layer"}}}, "workflow": {"id": "7f35ca53-6911-4aba-ade5-a7087933ca8d", "revision": 0, "last_node_id": 10, "last_link_id": 12, "nodes": [{"id": 8, "type": "VAEDecode", "pos": [1209, 188], "size": [210, 46], "flags": {}, "order": 6, "mode": 0, "inputs": [{"name": "samples", "type": "LATENT", "link": 7}, {"name": "vae", "type": "VAE", "link": 8}], "outputs": [{"name": "IMAGE", "type": "IMAGE", "slot_index": 0, "links": [9]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "VAEDecode", "widget_ue_connectable": {}}, "widgets_values": []}, {"id": 9, "type": "SaveImage", "pos": [1451, 189], "size": [210, 58], "flags": {}, "order": 7, "mode": 0, "inputs": [{"name": "images", "type": "IMAGE", "link": 9}], "outputs": [], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "widget_ue_connectable": {}}, "widgets_values": ["ComfyUI"]}, {"id": 4, "type": "CheckpointLoaderSimple", "pos": [-272.2301025390625, 457.4971923828125], "size": [315, 98], "flags": {}, "order": 0, "mode": 0, "inputs": [], "outputs": [{"name": "MODEL", "type": "MODEL", "slot_index": 0, "links": [1]}, {"name": "CLIP", "type": "CLIP", "slot_index": 1, "links": [10]}, {"name": "VAE", "type": "VAE", "slot_index": 2, "links": [8]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CheckpointLoaderSimple", "widget_ue_connectable": {}}, "widgets_values": ["illumiyume_35.safetensors"]}, {"id": 10, "type": "CLIPSetLastLayer", "pos": [41.93305206298828, 215.06910705566406], "size": [270, 58], "flags": {}, "order": 2, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 10}], "outputs": [{"name": "CLIP", "type": "CLIP", "links": [11, 12]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPSetLastLayer", "widget_ue_connectable": {}}, "widgets_values": [-2]}, {"id": 3, "type": "KSampler", "pos": [863, 186], "size": [315, 262], "flags": {}, "order": 5, "mode": 0, "inputs": [{"name": "model", "type": "MODEL", "link": 1}, {"name": "positive", "type": "CONDITIONING", "link": 4}, {"name": "negative", "type": "CONDITIONING", "link": 6}, {"name": "latent_image", "type": "LATENT", "link": 2}], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [7]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "KSampler", "widget_ue_connectable": {}}, "widgets_values": [40217979478388, "randomize", 30, 5, "dpmpp_2s_ancestral", "normal", 1]}, {"id": 7, "type": "CLIPTextEncode", "pos": [413, 389], "size": [425.27801513671875, 180.6060791015625], "flags": {}, "order": 4, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 12}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [6]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background"]}, {"id": 5, "type": "EmptyLatentImage", "pos": [473, 609], "size": [315, 106], "flags": {}, "order": 1, "mode": 0, "inputs": [], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [2]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "EmptyLatentImage", "widget_ue_connectable": {}}, "widgets_values": [1024, 1536, 3]}, {"id": 6, "type": "CLIPTextEncode", "pos": [415, 186], "size": [422.84503173828125, 164.31304931640625], "flags": {}, "order": 3, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 11}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [4]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["masterpiece,best quality,amazing quality, nakano ichika, nakano itsuki, nakano miku, nakano nino, nakano yotsuba, 4-finger heart hands, 5girls, :D, :O, ;D, ;O, ahoge, backlighting, blazer, bloom, blue cardigan, blue eyes, blush, bow, breasts, brown bow, brown hair, brown jacket, cardigan, collared shirt, dress shirt, fingernails, green bow, green hairband, green ribbon, green skirt, hair between eyes, hair bow, hair ornament, hair ribbon, hairband, hands up, headphones, headphones around neck, heart, heart hands, jacket, layered clothes, layered sleeves, lineup, long sleeves, looking at viewer, miniskirt, multiple girls, one eye closed, open clothes, open jacket, open mouth, parted lips, plaid, plaid bow, pleated skirt, quintuplets, red hair, red sweater vest, ribbon, school uniform, shirt, short hair, siblings, sisters, skirt, small breasts, smile, star (symbol), star hair ornament, sweater vest, twitter username, untucked shirt, v-neck, white shirt, wing collar, yellow sweater vest"]}], "links": [[1, 4, 0, 3, 0, "MODEL"], [2, 5, 0, 3, 3, "LATENT"], [4, 6, 0, 3, 1, "CONDITIONING"], [6, 7, 0, 3, 2, "CONDITIONING"], [7, 3, 0, 8, 0, "LATENT"], [8, 4, 2, 8, 1, "VAE"], [9, 8, 0, 9, 0, "IMAGE"], [10, 4, 1, 10, 0, "CLIP"], [11, 10, 0, 6, 0, "CLIP"], [12, 10, 0, 7, 0, "CLIP"]], "groups": [], "config": {}, "extra": {"ue_links": [], "links_added_by_ue": [], "ds": {"scale": 0.9331722054380678, "offset": [-48.259536584463746, 88.07089210857593]}, "frontendVersion": "1.21.7"}, "version": 0.4}}
Width
1024
Height
1536
Models
['illumiyume_35.safetensors']
Denoise
1
Modelids
[]
Scheduler
normal
Upscalers
[]
Versionids
[]
Controlnets
[]
Additionalresources
[]
+
+
+ + + + +
+ +
+
Prompt
masterpiece, best quality, upper body, indoors, cinematic lighting, amazing quality, very aesthetic, absurdres, best quality, 4k, upscale, varesa (genshin impact), 1girl, solo, horns, pink hair, skirt, ahoge, animal ears, navel, purple eyes, bandaid, long hair, braid, black skirt, multicolored hair, smile, double v, twin braids, looking at viewer, midriff, bangs, hairband, cow horns, hair ornament, cow ears, wink
Negative prompt
bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background
Seed
613204843267553
Model
illumiyume_35
Sampler
DPM++ 2S a
Steps
30
CFG scale
5
+
+
+ + +
+
Vaes
[]
Comfy
{"prompt": {"3": {"inputs": {"seed": 613204843267553, "steps": 30, "cfg": 5.0, "sampler_name": "dpmpp_2s_ancestral", "scheduler": "normal", "denoise": 1.0, "model": ["4", 0], "positive": ["6", 0], "negative": ["7", 0], "latent_image": ["5", 0]}, "class_type": "KSampler", "_meta": {"title": "KSampler"}}, "4": {"inputs": {"ckpt_name": "illumiyume_35.safetensors"}, "class_type": "CheckpointLoaderSimple", "_meta": {"title": "Load Checkpoint"}}, "5": {"inputs": {"width": 1024, "height": 1536, "batch_size": 3}, "class_type": "EmptyLatentImage", "_meta": {"title": "Empty Latent Image"}}, "6": {"inputs": {"text": "masterpiece, best quality, upper body, indoors, cinematic lighting, amazing quality, very aesthetic, absurdres, best quality, 4k, upscale, varesa (genshin impact), 1girl, solo, horns, pink hair, skirt, ahoge, animal ears, navel, purple eyes, bandaid, long hair, braid, black skirt, multicolored hair, smile, double v, twin braids, looking at viewer, midriff, bangs, hairband, cow horns, hair ornament, cow ears, wink", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "7": {"inputs": {"text": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "8": {"inputs": {"samples": ["3", 0], "vae": ["4", 2]}, "class_type": "VAEDecode", "_meta": {"title": "VAE Decode"}}, "9": {"inputs": {"filename_prefix": "ComfyUI", "images": ["8", 0]}, "class_type": "SaveImage", "_meta": {"title": "Save Image"}}, "10": {"inputs": {"stop_at_clip_layer": -2, "clip": ["4", 1]}, "class_type": "CLIPSetLastLayer", "_meta": {"title": "CLIP Set Last Layer"}}}, "workflow": {"id": "7f35ca53-6911-4aba-ade5-a7087933ca8d", "revision": 0, "last_node_id": 10, "last_link_id": 12, "nodes": [{"id": 8, "type": "VAEDecode", "pos": [1209, 188], "size": [210, 46], "flags": {}, "order": 6, "mode": 0, "inputs": [{"name": "samples", "type": "LATENT", "link": 7}, {"name": "vae", "type": "VAE", "link": 8}], "outputs": [{"name": "IMAGE", "type": "IMAGE", "slot_index": 0, "links": [9]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "VAEDecode", "widget_ue_connectable": {}}, "widgets_values": []}, {"id": 9, "type": "SaveImage", "pos": [1451, 189], "size": [210, 58], "flags": {}, "order": 7, "mode": 0, "inputs": [{"name": "images", "type": "IMAGE", "link": 9}], "outputs": [], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "widget_ue_connectable": {}}, "widgets_values": ["ComfyUI"]}, {"id": 4, "type": "CheckpointLoaderSimple", "pos": [-272.2301025390625, 457.4971923828125], "size": [315, 98], "flags": {}, "order": 0, "mode": 0, "inputs": [], "outputs": [{"name": "MODEL", "type": "MODEL", "slot_index": 0, "links": [1]}, {"name": "CLIP", "type": "CLIP", "slot_index": 1, "links": [10]}, {"name": "VAE", "type": "VAE", "slot_index": 2, "links": [8]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CheckpointLoaderSimple", "widget_ue_connectable": {}}, "widgets_values": ["illumiyume_35.safetensors"]}, {"id": 10, "type": "CLIPSetLastLayer", "pos": [41.93305206298828, 215.06910705566406], "size": [270, 58], "flags": {}, "order": 2, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 10}], "outputs": [{"name": "CLIP", "type": "CLIP", "links": [11, 12]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPSetLastLayer", "widget_ue_connectable": {}}, "widgets_values": [-2]}, {"id": 3, "type": "KSampler", "pos": [863, 186], "size": [315, 262], "flags": {}, "order": 5, "mode": 0, "inputs": [{"name": "model", "type": "MODEL", "link": 1}, {"name": "positive", "type": "CONDITIONING", "link": 4}, {"name": "negative", "type": "CONDITIONING", "link": 6}, {"name": "latent_image", "type": "LATENT", "link": 2}], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [7]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "KSampler", "widget_ue_connectable": {}}, "widgets_values": [613204843267553, "randomize", 30, 5, "dpmpp_2s_ancestral", "normal", 1]}, {"id": 5, "type": "EmptyLatentImage", "pos": [473, 609], "size": [315, 106], "flags": {}, "order": 1, "mode": 0, "inputs": [], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [2]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "EmptyLatentImage", "widget_ue_connectable": {}}, "widgets_values": [1024, 1536, 3]}, {"id": 7, "type": "CLIPTextEncode", "pos": [413, 389], "size": [425.27801513671875, 180.6060791015625], "flags": {}, "order": 4, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 12}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [6]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background"]}, {"id": 6, "type": "CLIPTextEncode", "pos": [415, 186], "size": [422.84503173828125, 164.31304931640625], "flags": {}, "order": 3, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 11}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [4]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["masterpiece, best quality, upper body, indoors, cinematic lighting, amazing quality, very aesthetic, absurdres, best quality, 4k, upscale, varesa (genshin impact), 1girl, solo, horns, pink hair, skirt, ahoge, animal ears, navel, purple eyes, bandaid, long hair, braid, black skirt, multicolored hair, smile, double v, twin braids, looking at viewer, midriff, bangs, hairband, cow horns, hair ornament, cow ears, wink"]}], "links": [[1, 4, 0, 3, 0, "MODEL"], [2, 5, 0, 3, 3, "LATENT"], [4, 6, 0, 3, 1, "CONDITIONING"], [6, 7, 0, 3, 2, "CONDITIONING"], [7, 3, 0, 8, 0, "LATENT"], [8, 4, 2, 8, 1, "VAE"], [9, 8, 0, 9, 0, "IMAGE"], [10, 4, 1, 10, 0, "CLIP"], [11, 10, 0, 6, 0, "CLIP"], [12, 10, 0, 7, 0, "CLIP"]], "groups": [], "config": {}, "extra": {"ue_links": [], "links_added_by_ue": [], "ds": {"scale": 1.0264894259818746, "offset": [-95.02085607305597, 60.403778077825564]}, "frontendVersion": "1.21.7"}, "version": 0.4}}
Width
1024
Height
1536
Models
['illumiyume_35.safetensors']
Denoise
1
Modelids
[]
Scheduler
normal
Upscalers
[]
Versionids
[]
Controlnets
[]
Additionalresources
[]
+
+
+ + + + +
+ +
+
Prompt
+masterpiece,best quality,amazing quality, 1girl, crazy mita \(miside\), masterpiece, blue eyes, low twintails, red hairband, blue scrunchie, red hairclip, red choker, finger to mouth, partially shaded face, head tilt, smirk, upper body, blue hair, red shirt, +
Negative prompt
bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background
Seed
423970272988574
Model
illumiyume_35
Sampler
DPM++ 2S a
Steps
30
CFG scale
5
+
+
+ + +
+
Vaes
[]
Comfy
{"prompt": {"3": {"inputs": {"seed": 423970272988574, "steps": 30, "cfg": 5.0, "sampler_name": "dpmpp_2s_ancestral", "scheduler": "normal", "denoise": 1.0, "model": ["4", 0], "positive": ["6", 0], "negative": ["7", 0], "latent_image": ["5", 0]}, "class_type": "KSampler", "_meta": {"title": "KSampler"}}, "4": {"inputs": {"ckpt_name": "illumiyume_35.safetensors"}, "class_type": "CheckpointLoaderSimple", "_meta": {"title": "Load Checkpoint"}}, "5": {"inputs": {"width": 1024, "height": 1536, "batch_size": 3}, "class_type": "EmptyLatentImage", "_meta": {"title": "Empty Latent Image"}}, "6": {"inputs": {"text": "\nmasterpiece,best quality,amazing quality, 1girl, crazy mita \\(miside\\), masterpiece, blue eyes, low twintails, red hairband, blue scrunchie, red hairclip, red choker, finger to mouth, partially shaded face, head tilt, smirk, upper body, blue hair, red shirt,\n", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "7": {"inputs": {"text": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "8": {"inputs": {"samples": ["3", 0], "vae": ["4", 2]}, "class_type": "VAEDecode", "_meta": {"title": "VAE Decode"}}, "9": {"inputs": {"filename_prefix": "ComfyUI", "images": ["8", 0]}, "class_type": "SaveImage", "_meta": {"title": "Save Image"}}, "10": {"inputs": {"stop_at_clip_layer": -2, "clip": ["4", 1]}, "class_type": "CLIPSetLastLayer", "_meta": {"title": "CLIP Set Last Layer"}}}, "workflow": {"id": "7f35ca53-6911-4aba-ade5-a7087933ca8d", "revision": 0, "last_node_id": 10, "last_link_id": 12, "nodes": [{"id": 8, "type": "VAEDecode", "pos": [1209, 188], "size": [210, 46], "flags": {}, "order": 6, "mode": 0, "inputs": [{"name": "samples", "type": "LATENT", "link": 7}, {"name": "vae", "type": "VAE", "link": 8}], "outputs": [{"name": "IMAGE", "type": "IMAGE", "slot_index": 0, "links": [9]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "VAEDecode", "widget_ue_connectable": {}}, "widgets_values": []}, {"id": 9, "type": "SaveImage", "pos": [1451, 189], "size": [210, 58], "flags": {}, "order": 7, "mode": 0, "inputs": [{"name": "images", "type": "IMAGE", "link": 9}], "outputs": [], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "widget_ue_connectable": {}}, "widgets_values": ["ComfyUI"]}, {"id": 4, "type": "CheckpointLoaderSimple", "pos": [-272.2301025390625, 457.4971923828125], "size": [315, 98], "flags": {}, "order": 0, "mode": 0, "inputs": [], "outputs": [{"name": "MODEL", "type": "MODEL", "slot_index": 0, "links": [1]}, {"name": "CLIP", "type": "CLIP", "slot_index": 1, "links": [10]}, {"name": "VAE", "type": "VAE", "slot_index": 2, "links": [8]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CheckpointLoaderSimple", "widget_ue_connectable": {}}, "widgets_values": ["illumiyume_35.safetensors"]}, {"id": 10, "type": "CLIPSetLastLayer", "pos": [41.93305206298828, 215.06910705566406], "size": [270, 58], "flags": {}, "order": 2, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 10}], "outputs": [{"name": "CLIP", "type": "CLIP", "links": [11, 12]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPSetLastLayer", "widget_ue_connectable": {}}, "widgets_values": [-2]}, {"id": 3, "type": "KSampler", "pos": [863, 186], "size": [315, 262], "flags": {}, "order": 5, "mode": 0, "inputs": [{"name": "model", "type": "MODEL", "link": 1}, {"name": "positive", "type": "CONDITIONING", "link": 4}, {"name": "negative", "type": "CONDITIONING", "link": 6}, {"name": "latent_image", "type": "LATENT", "link": 2}], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [7]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "KSampler", "widget_ue_connectable": {}}, "widgets_values": [423970272988574, "randomize", 30, 5, "dpmpp_2s_ancestral", "normal", 1]}, {"id": 5, "type": "EmptyLatentImage", "pos": [473, 609], "size": [315, 106], "flags": {}, "order": 1, "mode": 0, "inputs": [], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [2]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "EmptyLatentImage", "widget_ue_connectable": {}}, "widgets_values": [1024, 1536, 3]}, {"id": 7, "type": "CLIPTextEncode", "pos": [413, 389], "size": [425.27801513671875, 180.6060791015625], "flags": {}, "order": 4, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 12}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [6]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background"]}, {"id": 6, "type": "CLIPTextEncode", "pos": [415, 186], "size": [422.84503173828125, 164.31304931640625], "flags": {}, "order": 3, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 11}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [4]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["\nmasterpiece,best quality,amazing quality, 1girl, crazy mita \\(miside\\), masterpiece, blue eyes, low twintails, red hairband, blue scrunchie, red hairclip, red choker, finger to mouth, partially shaded face, head tilt, smirk, upper body, blue hair, red shirt,\n"]}], "links": [[1, 4, 0, 3, 0, "MODEL"], [2, 5, 0, 3, 3, "LATENT"], [4, 6, 0, 3, 1, "CONDITIONING"], [6, 7, 0, 3, 2, "CONDITIONING"], [7, 3, 0, 8, 0, "LATENT"], [8, 4, 2, 8, 1, "VAE"], [9, 8, 0, 9, 0, "IMAGE"], [10, 4, 1, 10, 0, "CLIP"], [11, 10, 0, 6, 0, "CLIP"], [12, 10, 0, 7, 0, "CLIP"]], "groups": [], "config": {}, "extra": {"ue_links": [], "links_added_by_ue": [], "ds": {"scale": 1.0264894259818746, "offset": [18.98547760748014, 84.52661128132267]}, "frontendVersion": "1.21.7"}, "version": 0.4}}
Width
1024
Height
1536
Models
['illumiyume_35.safetensors']
Denoise
1
Modelids
[]
Scheduler
normal
Upscalers
[]
Versionids
[]
Controlnets
[]
Additionalresources
[]
+
+
+ + + + +
+ +
+
Prompt
+masterpiece,best quality,amazing quality, rosmontis (arknights), 1girl, solo, long hair, blush, bow, animal ears, green eyes, grey hair, lying, hood, cat ears, on back, chibi, pillow, :3, cat girl, under covers, dreaming, thought bubble, pizza,
Negative prompt
bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background
Seed
852675588020754
Model
illumiyume_35
Sampler
DPM++ 2S a
Steps
30
CFG scale
5
+
+
+ + +
+
Vaes
[]
Comfy
{"prompt": {"3": {"inputs": {"seed": 852675588020754, "steps": 30, "cfg": 5.0, "sampler_name": "dpmpp_2s_ancestral", "scheduler": "normal", "denoise": 1.0, "model": ["4", 0], "positive": ["6", 0], "negative": ["7", 0], "latent_image": ["5", 0]}, "class_type": "KSampler", "_meta": {"title": "KSampler"}}, "4": {"inputs": {"ckpt_name": "illumiyume_35.safetensors"}, "class_type": "CheckpointLoaderSimple", "_meta": {"title": "Load Checkpoint"}}, "5": {"inputs": {"width": 1024, "height": 1536, "batch_size": 3}, "class_type": "EmptyLatentImage", "_meta": {"title": "Empty Latent Image"}}, "6": {"inputs": {"text": "\nmasterpiece,best quality,amazing quality, rosmontis (arknights), 1girl, solo, long hair, blush, bow, animal ears, green eyes, grey hair, lying, hood, cat ears, on back, chibi, pillow, :3, cat girl, under covers, dreaming, thought bubble, pizza,", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "7": {"inputs": {"text": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "8": {"inputs": {"samples": ["3", 0], "vae": ["4", 2]}, "class_type": "VAEDecode", "_meta": {"title": "VAE Decode"}}, "9": {"inputs": {"filename_prefix": "ComfyUI", "images": ["8", 0]}, "class_type": "SaveImage", "_meta": {"title": "Save Image"}}, "10": {"inputs": {"stop_at_clip_layer": -2, "clip": ["4", 1]}, "class_type": "CLIPSetLastLayer", "_meta": {"title": "CLIP Set Last Layer"}}}, "workflow": {"id": "7f35ca53-6911-4aba-ade5-a7087933ca8d", "revision": 0, "last_node_id": 10, "last_link_id": 12, "nodes": [{"id": 8, "type": "VAEDecode", "pos": [1209, 188], "size": [210, 46], "flags": {}, "order": 6, "mode": 0, "inputs": [{"name": "samples", "type": "LATENT", "link": 7}, {"name": "vae", "type": "VAE", "link": 8}], "outputs": [{"name": "IMAGE", "type": "IMAGE", "slot_index": 0, "links": [9]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "VAEDecode", "widget_ue_connectable": {}}, "widgets_values": []}, {"id": 9, "type": "SaveImage", "pos": [1451, 189], "size": [210, 58], "flags": {}, "order": 7, "mode": 0, "inputs": [{"name": "images", "type": "IMAGE", "link": 9}], "outputs": [], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "widget_ue_connectable": {}}, "widgets_values": ["ComfyUI"]}, {"id": 4, "type": "CheckpointLoaderSimple", "pos": [-272.2301025390625, 457.4971923828125], "size": [315, 98], "flags": {}, "order": 0, "mode": 0, "inputs": [], "outputs": [{"name": "MODEL", "type": "MODEL", "slot_index": 0, "links": [1]}, {"name": "CLIP", "type": "CLIP", "slot_index": 1, "links": [10]}, {"name": "VAE", "type": "VAE", "slot_index": 2, "links": [8]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CheckpointLoaderSimple", "widget_ue_connectable": {}}, "widgets_values": ["illumiyume_35.safetensors"]}, {"id": 10, "type": "CLIPSetLastLayer", "pos": [41.93305206298828, 215.06910705566406], "size": [270, 58], "flags": {}, "order": 2, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 10}], "outputs": [{"name": "CLIP", "type": "CLIP", "links": [11, 12]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPSetLastLayer", "widget_ue_connectable": {}}, "widgets_values": [-2]}, {"id": 3, "type": "KSampler", "pos": [863, 186], "size": [315, 262], "flags": {}, "order": 5, "mode": 0, "inputs": [{"name": "model", "type": "MODEL", "link": 1}, {"name": "positive", "type": "CONDITIONING", "link": 4}, {"name": "negative", "type": "CONDITIONING", "link": 6}, {"name": "latent_image", "type": "LATENT", "link": 2}], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [7]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "KSampler", "widget_ue_connectable": {}}, "widgets_values": [852675588020754, "randomize", 30, 5, "dpmpp_2s_ancestral", "normal", 1]}, {"id": 5, "type": "EmptyLatentImage", "pos": [473, 609], "size": [315, 106], "flags": {}, "order": 1, "mode": 0, "inputs": [], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [2]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "EmptyLatentImage", "widget_ue_connectable": {}}, "widgets_values": [1024, 1536, 3]}, {"id": 7, "type": "CLIPTextEncode", "pos": [413, 389], "size": [425.27801513671875, 180.6060791015625], "flags": {}, "order": 4, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 12}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [6]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background"]}, {"id": 6, "type": "CLIPTextEncode", "pos": [415, 186], "size": [422.84503173828125, 164.31304931640625], "flags": {}, "order": 3, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 11}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [4]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["\nmasterpiece,best quality,amazing quality, rosmontis (arknights), 1girl, solo, long hair, blush, bow, animal ears, green eyes, grey hair, lying, hood, cat ears, on back, chibi, pillow, :3, cat girl, under covers, dreaming, thought bubble, pizza,"]}], "links": [[1, 4, 0, 3, 0, "MODEL"], [2, 5, 0, 3, 3, "LATENT"], [4, 6, 0, 3, 1, "CONDITIONING"], [6, 7, 0, 3, 2, "CONDITIONING"], [7, 3, 0, 8, 0, "LATENT"], [8, 4, 2, 8, 1, "VAE"], [9, 8, 0, 9, 0, "IMAGE"], [10, 4, 1, 10, 0, "CLIP"], [11, 10, 0, 6, 0, "CLIP"], [12, 10, 0, 7, 0, "CLIP"]], "groups": [], "config": {}, "extra": {"ue_links": [], "links_added_by_ue": [], "ds": {"scale": 1.0264894259818746, "offset": [18.98547760748014, 84.52661128132267]}, "frontendVersion": "1.21.7"}, "version": 0.4}}
Width
1024
Height
1536
Models
['illumiyume_35.safetensors']
Denoise
1
Modelids
[]
Scheduler
normal
Upscalers
[]
Versionids
[]
Controlnets
[]
Additionalresources
[]
+
+
+ + + + +
+ +
+
Prompt
(masterpiece,best quality,amazing quality), shinomiya kaguya, yor briar, 2girls, black dress, black gloves, black hair, black thighhighs, breasts, cleavage, closed mouth, color connection, crossover, dagger, dress, dual wielding, expressionless, fingerless gloves, flower, folded ponytail, gloves, gold hairband, gradient background, hair color connection, hair flower, hair ornament, hair ribbon, hairband, hairstyle connection, hand to own mouth, holding, holding dagger, holding knife, holding weapon, index finger raised, knife, large breasts, long hair, looking at viewer, medium hair, multiple girls, neck ribbon, pinafore dress, red background, red eyes, red ribbon, reverse grip, ribbon, rose, school uniform, season connection, short hair with long locks, shuuchiin academy school uniform, sidelocks, sleeveless, sleeveless dress, spikes, thighhighs, trait connection, two-sided dress, two-sided fabric, weapon + +
Negative prompt
bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background
Seed
706525802785343
Model
illumiyume_35
Sampler
DPM++ 2S a
Steps
30
CFG scale
5
+
+
+ + +
+
Vaes
[]
Comfy
{"prompt": {"3": {"inputs": {"seed": 706525802785343, "steps": 30, "cfg": 5.0, "sampler_name": "dpmpp_2s_ancestral", "scheduler": "normal", "denoise": 1.0, "model": ["4", 0], "positive": ["6", 0], "negative": ["7", 0], "latent_image": ["5", 0]}, "class_type": "KSampler", "_meta": {"title": "KSampler"}}, "4": {"inputs": {"ckpt_name": "illumiyume_35.safetensors"}, "class_type": "CheckpointLoaderSimple", "_meta": {"title": "Load Checkpoint"}}, "5": {"inputs": {"width": 1024, "height": 1536, "batch_size": 3}, "class_type": "EmptyLatentImage", "_meta": {"title": "Empty Latent Image"}}, "6": {"inputs": {"text": "(masterpiece,best quality,amazing quality), shinomiya kaguya, yor briar, 2girls, black dress, black gloves, black hair, black thighhighs, breasts, cleavage, closed mouth, color connection, crossover, dagger, dress, dual wielding, expressionless, fingerless gloves, flower, folded ponytail, gloves, gold hairband, gradient background, hair color connection, hair flower, hair ornament, hair ribbon, hairband, hairstyle connection, hand to own mouth, holding, holding dagger, holding knife, holding weapon, index finger raised, knife, large breasts, long hair, looking at viewer, medium hair, multiple girls, neck ribbon, pinafore dress, red background, red eyes, red ribbon, reverse grip, ribbon, rose, school uniform, season connection, short hair with long locks, shuuchiin academy school uniform, sidelocks, sleeveless, sleeveless dress, spikes, thighhighs, trait connection, two-sided dress, two-sided fabric, weapon\n\n", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "7": {"inputs": {"text": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "8": {"inputs": {"samples": ["3", 0], "vae": ["4", 2]}, "class_type": "VAEDecode", "_meta": {"title": "VAE Decode"}}, "9": {"inputs": {"filename_prefix": "ComfyUI", "images": ["8", 0]}, "class_type": "SaveImage", "_meta": {"title": "Save Image"}}, "10": {"inputs": {"stop_at_clip_layer": -2, "clip": ["4", 1]}, "class_type": "CLIPSetLastLayer", "_meta": {"title": "CLIP Set Last Layer"}}}, "workflow": {"id": "7f35ca53-6911-4aba-ade5-a7087933ca8d", "revision": 0, "last_node_id": 10, "last_link_id": 12, "nodes": [{"id": 8, "type": "VAEDecode", "pos": [1209, 188], "size": [210, 46], "flags": {}, "order": 6, "mode": 0, "inputs": [{"name": "samples", "type": "LATENT", "link": 7}, {"name": "vae", "type": "VAE", "link": 8}], "outputs": [{"name": "IMAGE", "type": "IMAGE", "slot_index": 0, "links": [9]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "VAEDecode", "widget_ue_connectable": {}}, "widgets_values": []}, {"id": 9, "type": "SaveImage", "pos": [1451, 189], "size": [210, 58], "flags": {}, "order": 7, "mode": 0, "inputs": [{"name": "images", "type": "IMAGE", "link": 9}], "outputs": [], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "widget_ue_connectable": {}}, "widgets_values": ["ComfyUI"]}, {"id": 4, "type": "CheckpointLoaderSimple", "pos": [-272.2301025390625, 457.4971923828125], "size": [315, 98], "flags": {}, "order": 0, "mode": 0, "inputs": [], "outputs": [{"name": "MODEL", "type": "MODEL", "slot_index": 0, "links": [1]}, {"name": "CLIP", "type": "CLIP", "slot_index": 1, "links": [10]}, {"name": "VAE", "type": "VAE", "slot_index": 2, "links": [8]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CheckpointLoaderSimple", "widget_ue_connectable": {}}, "widgets_values": ["illumiyume_35.safetensors"]}, {"id": 10, "type": "CLIPSetLastLayer", "pos": [41.93305206298828, 215.06910705566406], "size": [270, 58], "flags": {}, "order": 2, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 10}], "outputs": [{"name": "CLIP", "type": "CLIP", "links": [11, 12]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPSetLastLayer", "widget_ue_connectable": {}}, "widgets_values": [-2]}, {"id": 3, "type": "KSampler", "pos": [863, 186], "size": [315, 262], "flags": {}, "order": 5, "mode": 0, "inputs": [{"name": "model", "type": "MODEL", "link": 1}, {"name": "positive", "type": "CONDITIONING", "link": 4}, {"name": "negative", "type": "CONDITIONING", "link": 6}, {"name": "latent_image", "type": "LATENT", "link": 2}], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [7]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "KSampler", "widget_ue_connectable": {}}, "widgets_values": [706525802785343, "randomize", 30, 5, "dpmpp_2s_ancestral", "normal", 1]}, {"id": 5, "type": "EmptyLatentImage", "pos": [473, 609], "size": [315, 106], "flags": {}, "order": 1, "mode": 0, "inputs": [], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [2]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "EmptyLatentImage", "widget_ue_connectable": {}}, "widgets_values": [1024, 1536, 3]}, {"id": 7, "type": "CLIPTextEncode", "pos": [413, 389], "size": [425.27801513671875, 180.6060791015625], "flags": {}, "order": 4, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 12}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [6]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background"]}, {"id": 6, "type": "CLIPTextEncode", "pos": [415, 186], "size": [422.84503173828125, 164.31304931640625], "flags": {}, "order": 3, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 11}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [4]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["(masterpiece,best quality,amazing quality), shinomiya kaguya, yor briar, 2girls, black dress, black gloves, black hair, black thighhighs, breasts, cleavage, closed mouth, color connection, crossover, dagger, dress, dual wielding, expressionless, fingerless gloves, flower, folded ponytail, gloves, gold hairband, gradient background, hair color connection, hair flower, hair ornament, hair ribbon, hairband, hairstyle connection, hand to own mouth, holding, holding dagger, holding knife, holding weapon, index finger raised, knife, large breasts, long hair, looking at viewer, medium hair, multiple girls, neck ribbon, pinafore dress, red background, red eyes, red ribbon, reverse grip, ribbon, rose, school uniform, season connection, short hair with long locks, shuuchiin academy school uniform, sidelocks, sleeveless, sleeveless dress, spikes, thighhighs, trait connection, two-sided dress, two-sided fabric, weapon\n\n"]}], "links": [[1, 4, 0, 3, 0, "MODEL"], [2, 5, 0, 3, 3, "LATENT"], [4, 6, 0, 3, 1, "CONDITIONING"], [6, 7, 0, 3, 2, "CONDITIONING"], [7, 3, 0, 8, 0, "LATENT"], [8, 4, 2, 8, 1, "VAE"], [9, 8, 0, 9, 0, "IMAGE"], [10, 4, 1, 10, 0, "CLIP"], [11, 10, 0, 6, 0, "CLIP"], [12, 10, 0, 7, 0, "CLIP"]], "groups": [], "config": {}, "extra": {"ue_links": [], "links_added_by_ue": [], "ds": {"scale": 1.0264894259818746, "offset": [-187.54368346713517, -34.32507575218235]}, "frontendVersion": "1.21.7"}, "version": 0.4}}
Width
1024
Height
1536
Models
['illumiyume_35.safetensors']
Denoise
1
Modelids
[]
Scheduler
normal
Upscalers
[]
Versionids
[]
Controlnets
[]
Additionalresources
[]
+
+
+ + + + +
+ +
+
Prompt
(best quality,4k,8k,highres,masterpiece:1.2),ultra-detailed,suzukaze aoba,1girl,solo,upper body,victory hand gesture,looking at viewer,smiling,short brownSmooth Quality, 1girl, solo, cat girl, classroom, >_<, holding drawing of chibi cat girl with >_< expression, classroom, white hair,
Negative prompt
bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background
Seed
238613207477948
Model
illumiyume_35
Sampler
DPM++ 2S a
Steps
30
CFG scale
5
+
+
+ + +
+
Vaes
[]
Comfy
{"prompt": {"3": {"inputs": {"seed": 238613207477948, "steps": 30, "cfg": 5.0, "sampler_name": "dpmpp_2s_ancestral", "scheduler": "normal", "denoise": 1.0, "model": ["4", 0], "positive": ["6", 0], "negative": ["7", 0], "latent_image": ["5", 0]}, "class_type": "KSampler", "_meta": {"title": "KSampler"}}, "4": {"inputs": {"ckpt_name": "illumiyume_35.safetensors"}, "class_type": "CheckpointLoaderSimple", "_meta": {"title": "Load Checkpoint"}}, "5": {"inputs": {"width": 1024, "height": 1536, "batch_size": 3}, "class_type": "EmptyLatentImage", "_meta": {"title": "Empty Latent Image"}}, "6": {"inputs": {"text": "(best quality,4k,8k,highres,masterpiece:1.2),ultra-detailed,suzukaze aoba,1girl,solo,upper body,victory hand gesture,looking at viewer,smiling,short brownSmooth Quality, 1girl, solo, cat girl, classroom, >_<, holding drawing of chibi cat girl with >_< expression, classroom, white hair,", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "7": {"inputs": {"text": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "8": {"inputs": {"samples": ["3", 0], "vae": ["4", 2]}, "class_type": "VAEDecode", "_meta": {"title": "VAE Decode"}}, "9": {"inputs": {"filename_prefix": "ComfyUI", "images": ["8", 0]}, "class_type": "SaveImage", "_meta": {"title": "Save Image"}}, "10": {"inputs": {"stop_at_clip_layer": -2, "clip": ["4", 1]}, "class_type": "CLIPSetLastLayer", "_meta": {"title": "CLIP Set Last Layer"}}}, "workflow": {"id": "7f35ca53-6911-4aba-ade5-a7087933ca8d", "revision": 0, "last_node_id": 10, "last_link_id": 12, "nodes": [{"id": 8, "type": "VAEDecode", "pos": [1209, 188], "size": [210, 46], "flags": {}, "order": 6, "mode": 0, "inputs": [{"name": "samples", "type": "LATENT", "link": 7}, {"name": "vae", "type": "VAE", "link": 8}], "outputs": [{"name": "IMAGE", "type": "IMAGE", "slot_index": 0, "links": [9]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "VAEDecode", "widget_ue_connectable": {}}, "widgets_values": []}, {"id": 9, "type": "SaveImage", "pos": [1451, 189], "size": [210, 58], "flags": {}, "order": 7, "mode": 0, "inputs": [{"name": "images", "type": "IMAGE", "link": 9}], "outputs": [], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "widget_ue_connectable": {}}, "widgets_values": ["ComfyUI"]}, {"id": 4, "type": "CheckpointLoaderSimple", "pos": [-272.2301025390625, 457.4971923828125], "size": [315, 98], "flags": {}, "order": 0, "mode": 0, "inputs": [], "outputs": [{"name": "MODEL", "type": "MODEL", "slot_index": 0, "links": [1]}, {"name": "CLIP", "type": "CLIP", "slot_index": 1, "links": [10]}, {"name": "VAE", "type": "VAE", "slot_index": 2, "links": [8]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CheckpointLoaderSimple", "widget_ue_connectable": {}}, "widgets_values": ["illumiyume_35.safetensors"]}, {"id": 10, "type": "CLIPSetLastLayer", "pos": [41.93305206298828, 215.06910705566406], "size": [270, 58], "flags": {}, "order": 2, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 10}], "outputs": [{"name": "CLIP", "type": "CLIP", "links": [11, 12]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPSetLastLayer", "widget_ue_connectable": {}}, "widgets_values": [-2]}, {"id": 3, "type": "KSampler", "pos": [863, 186], "size": [315, 262], "flags": {}, "order": 5, "mode": 0, "inputs": [{"name": "model", "type": "MODEL", "link": 1}, {"name": "positive", "type": "CONDITIONING", "link": 4}, {"name": "negative", "type": "CONDITIONING", "link": 6}, {"name": "latent_image", "type": "LATENT", "link": 2}], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [7]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "KSampler", "widget_ue_connectable": {}}, "widgets_values": [238613207477948, "randomize", 30, 5, "dpmpp_2s_ancestral", "normal", 1]}, {"id": 5, "type": "EmptyLatentImage", "pos": [473, 609], "size": [315, 106], "flags": {}, "order": 1, "mode": 0, "inputs": [], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [2]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "EmptyLatentImage", "widget_ue_connectable": {}}, "widgets_values": [1024, 1536, 3]}, {"id": 7, "type": "CLIPTextEncode", "pos": [413, 389], "size": [425.27801513671875, 180.6060791015625], "flags": {}, "order": 4, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 12}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [6]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background"]}, {"id": 6, "type": "CLIPTextEncode", "pos": [415, 186], "size": [422.84503173828125, 164.31304931640625], "flags": {}, "order": 3, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 11}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [4]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["(best quality,4k,8k,highres,masterpiece:1.2),ultra-detailed,suzukaze aoba,1girl,solo,upper body,victory hand gesture,looking at viewer,smiling,short brownSmooth Quality, 1girl, solo, cat girl, classroom, >_<, holding drawing of chibi cat girl with >_< expression, classroom, white hair,"]}], "links": [[1, 4, 0, 3, 0, "MODEL"], [2, 5, 0, 3, 3, "LATENT"], [4, 6, 0, 3, 1, "CONDITIONING"], [6, 7, 0, 3, 2, "CONDITIONING"], [7, 3, 0, 8, 0, "LATENT"], [8, 4, 2, 8, 1, "VAE"], [9, 8, 0, 9, 0, "IMAGE"], [10, 4, 1, 10, 0, "CLIP"], [11, 10, 0, 6, 0, "CLIP"], [12, 10, 0, 7, 0, "CLIP"]], "groups": [], "config": {}, "extra": {"ue_links": [], "links_added_by_ue": [], "ds": {"scale": 1.0264894259818746, "offset": [-187.54368346713517, -34.32507575218235]}, "frontendVersion": "1.21.7"}, "version": 0.4}}
Width
1024
Height
1536
Models
['illumiyume_35.safetensors']
Denoise
1
Modelids
[]
Scheduler
normal
Upscalers
[]
Versionids
[]
Controlnets
[]
Additionalresources
[]
+
+
+ + + + +
+ +
+
Prompt
1girl,,furina \(genshin impact\),(dappled sunlight:1.40), (close-up:0.80), (shade:1.20), (grass:1.20), (forest:1.20), (nature:1.20), (sunlight:1.20), (tree:1.20),,from above, lens flare, closed mouth, lying, beachball, ring necklace, :3,,small breasts, hair between eyes, streaked hair, long hair, hat, gloves, blue hair, white hair, blue eyes,,hat ribbon, hat bow, blue bikini, ring, bow, cleavage, straw hat, jewelry, necklace, breasts, bikini, bracelet, towel, navel, swimsuit, collarbone, ribbon,,dappled sunlight, light rays, cloud, ball, day, beach, cloudy sky, beach towel, light particles, sand, sunlight, sun, sky, outdoors, shadow, bird, +,masterpiece,best quality,amazing quality,
Negative prompt
bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background
Seed
249788060692524
Model
illumiyume_35
Sampler
DPM++ 2S a
Steps
30
CFG scale
5
+
+
+ + +
+
Vaes
[]
Comfy
{"prompt": {"3": {"inputs": {"seed": 249788060692524, "steps": 30, "cfg": 5.0, "sampler_name": "dpmpp_2s_ancestral", "scheduler": "normal", "denoise": 1.0, "model": ["4", 0], "positive": ["6", 0], "negative": ["7", 0], "latent_image": ["5", 0]}, "class_type": "KSampler", "_meta": {"title": "KSampler"}}, "4": {"inputs": {"ckpt_name": "illumiyume_35.safetensors"}, "class_type": "CheckpointLoaderSimple", "_meta": {"title": "Load Checkpoint"}}, "5": {"inputs": {"width": 1024, "height": 1536, "batch_size": 3}, "class_type": "EmptyLatentImage", "_meta": {"title": "Empty Latent Image"}}, "6": {"inputs": {"text": "1girl,,furina \\(genshin impact\\),(dappled sunlight:1.40), (close-up:0.80), (shade:1.20), (grass:1.20), (forest:1.20), (nature:1.20), (sunlight:1.20), (tree:1.20),,from above, lens flare, closed mouth, lying, beachball, ring necklace, :3,,small breasts, hair between eyes, streaked hair, long hair, hat, gloves, blue hair, white hair, blue eyes,,hat ribbon, hat bow, blue bikini, ring, bow, cleavage, straw hat, jewelry, necklace, breasts, bikini, bracelet, towel, navel, swimsuit, collarbone, ribbon,,dappled sunlight, light rays, cloud, ball, day, beach, cloudy sky, beach towel, light particles, sand, sunlight, sun, sky, outdoors, shadow, bird,\n,masterpiece,best quality,amazing quality,", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "7": {"inputs": {"text": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "8": {"inputs": {"samples": ["3", 0], "vae": ["4", 2]}, "class_type": "VAEDecode", "_meta": {"title": "VAE Decode"}}, "9": {"inputs": {"filename_prefix": "ComfyUI", "images": ["8", 0]}, "class_type": "SaveImage", "_meta": {"title": "Save Image"}}, "10": {"inputs": {"stop_at_clip_layer": -2, "clip": ["4", 1]}, "class_type": "CLIPSetLastLayer", "_meta": {"title": "CLIP Set Last Layer"}}}, "workflow": {"id": "7f35ca53-6911-4aba-ade5-a7087933ca8d", "revision": 0, "last_node_id": 10, "last_link_id": 12, "nodes": [{"id": 8, "type": "VAEDecode", "pos": [1209, 188], "size": [210, 46], "flags": {}, "order": 6, "mode": 0, "inputs": [{"name": "samples", "type": "LATENT", "link": 7}, {"name": "vae", "type": "VAE", "link": 8}], "outputs": [{"name": "IMAGE", "type": "IMAGE", "slot_index": 0, "links": [9]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "VAEDecode", "widget_ue_connectable": {}}, "widgets_values": []}, {"id": 9, "type": "SaveImage", "pos": [1451, 189], "size": [210, 58], "flags": {}, "order": 7, "mode": 0, "inputs": [{"name": "images", "type": "IMAGE", "link": 9}], "outputs": [], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "widget_ue_connectable": {}}, "widgets_values": ["ComfyUI"]}, {"id": 4, "type": "CheckpointLoaderSimple", "pos": [-272.2301025390625, 457.4971923828125], "size": [315, 98], "flags": {}, "order": 0, "mode": 0, "inputs": [], "outputs": [{"name": "MODEL", "type": "MODEL", "slot_index": 0, "links": [1]}, {"name": "CLIP", "type": "CLIP", "slot_index": 1, "links": [10]}, {"name": "VAE", "type": "VAE", "slot_index": 2, "links": [8]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CheckpointLoaderSimple", "widget_ue_connectable": {}}, "widgets_values": ["illumiyume_35.safetensors"]}, {"id": 10, "type": "CLIPSetLastLayer", "pos": [41.93305206298828, 215.06910705566406], "size": [270, 58], "flags": {}, "order": 2, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 10}], "outputs": [{"name": "CLIP", "type": "CLIP", "links": [11, 12]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPSetLastLayer", "widget_ue_connectable": {}}, "widgets_values": [-2]}, {"id": 3, "type": "KSampler", "pos": [863, 186], "size": [315, 262], "flags": {}, "order": 5, "mode": 0, "inputs": [{"name": "model", "type": "MODEL", "link": 1}, {"name": "positive", "type": "CONDITIONING", "link": 4}, {"name": "negative", "type": "CONDITIONING", "link": 6}, {"name": "latent_image", "type": "LATENT", "link": 2}], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [7]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "KSampler", "widget_ue_connectable": {}}, "widgets_values": [249788060692524, "randomize", 30, 5, "dpmpp_2s_ancestral", "normal", 1]}, {"id": 5, "type": "EmptyLatentImage", "pos": [473, 609], "size": [315, 106], "flags": {}, "order": 1, "mode": 0, "inputs": [], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [2]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "EmptyLatentImage", "widget_ue_connectable": {}}, "widgets_values": [1024, 1536, 3]}, {"id": 7, "type": "CLIPTextEncode", "pos": [413, 389], "size": [425.27801513671875, 180.6060791015625], "flags": {}, "order": 4, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 12}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [6]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background"]}, {"id": 6, "type": "CLIPTextEncode", "pos": [415, 186], "size": [422.84503173828125, 164.31304931640625], "flags": {}, "order": 3, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 11}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [4]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["1girl,,furina \\(genshin impact\\),(dappled sunlight:1.40), (close-up:0.80), (shade:1.20), (grass:1.20), (forest:1.20), (nature:1.20), (sunlight:1.20), (tree:1.20),,from above, lens flare, closed mouth, lying, beachball, ring necklace, :3,,small breasts, hair between eyes, streaked hair, long hair, hat, gloves, blue hair, white hair, blue eyes,,hat ribbon, hat bow, blue bikini, ring, bow, cleavage, straw hat, jewelry, necklace, breasts, bikini, bracelet, towel, navel, swimsuit, collarbone, ribbon,,dappled sunlight, light rays, cloud, ball, day, beach, cloudy sky, beach towel, light particles, sand, sunlight, sun, sky, outdoors, shadow, bird,\n,masterpiece,best quality,amazing quality,"]}], "links": [[1, 4, 0, 3, 0, "MODEL"], [2, 5, 0, 3, 3, "LATENT"], [4, 6, 0, 3, 1, "CONDITIONING"], [6, 7, 0, 3, 2, "CONDITIONING"], [7, 3, 0, 8, 0, "LATENT"], [8, 4, 2, 8, 1, "VAE"], [9, 8, 0, 9, 0, "IMAGE"], [10, 4, 1, 10, 0, "CLIP"], [11, 10, 0, 6, 0, "CLIP"], [12, 10, 0, 7, 0, "CLIP"]], "groups": [], "config": {}, "extra": {"ue_links": [], "links_added_by_ue": [], "ds": {"scale": 1.0264894259818746, "offset": [-252.81469191996166, -3.1508627597875902]}, "frontendVersion": "1.21.7"}, "version": 0.4}}
Width
1024
Height
1536
Models
['illumiyume_35.safetensors']
Denoise
1
Modelids
[]
Scheduler
normal
Upscalers
[]
Versionids
[]
Controlnets
[]
Additionalresources
[]
+
+
+ + + + +
+ +
+
Prompt
masterpiece,best quality,amazing quality, anya (spy x family), twilight (spy x family), yor briar, 1boy, 2girls, black gloves, black hair, black headwear, black suit, blue eyes, brown hair, child, dress, eden academy school uniform, formal clothes, gloves, grey eyes, hat, headband, holding, holding hat, holding unworn clothes, long hair, looking at viewer, multiple girls, neck ribbon, necktie, red eyes, red necktie, red ribbon, red sweater, ribbed sweater, ribbon, school uniform, sidelocks, smile, suit, sweater, sweater dress, unworn hat, unworn headwear, upper body, white headband
Negative prompt
bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background
Seed
538624701520958
Model
illumiyume_35
Sampler
DPM++ 2S a
Steps
30
CFG scale
5
+
+
+ + +
+
Vaes
[]
Comfy
{"prompt": {"3": {"inputs": {"seed": 538624701520958, "steps": 30, "cfg": 5.0, "sampler_name": "dpmpp_2s_ancestral", "scheduler": "normal", "denoise": 1.0, "model": ["4", 0], "positive": ["6", 0], "negative": ["7", 0], "latent_image": ["5", 0]}, "class_type": "KSampler", "_meta": {"title": "KSampler"}}, "4": {"inputs": {"ckpt_name": "illumiyume_35.safetensors"}, "class_type": "CheckpointLoaderSimple", "_meta": {"title": "Load Checkpoint"}}, "5": {"inputs": {"width": 1024, "height": 1536, "batch_size": 3}, "class_type": "EmptyLatentImage", "_meta": {"title": "Empty Latent Image"}}, "6": {"inputs": {"text": "masterpiece,best quality,amazing quality, anya (spy x family), twilight (spy x family), yor briar, 1boy, 2girls, black gloves, black hair, black headwear, black suit, blue eyes, brown hair, child, dress, eden academy school uniform, formal clothes, gloves, grey eyes, hat, headband, holding, holding hat, holding unworn clothes, long hair, looking at viewer, multiple girls, neck ribbon, necktie, red eyes, red necktie, red ribbon, red sweater, ribbed sweater, ribbon, school uniform, sidelocks, smile, suit, sweater, sweater dress, unworn hat, unworn headwear, upper body, white headband", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "7": {"inputs": {"text": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "8": {"inputs": {"samples": ["3", 0], "vae": ["4", 2]}, "class_type": "VAEDecode", "_meta": {"title": "VAE Decode"}}, "9": {"inputs": {"filename_prefix": "ComfyUI", "images": ["8", 0]}, "class_type": "SaveImage", "_meta": {"title": "Save Image"}}, "10": {"inputs": {"stop_at_clip_layer": -2, "clip": ["4", 1]}, "class_type": "CLIPSetLastLayer", "_meta": {"title": "CLIP Set Last Layer"}}}, "workflow": {"id": "7f35ca53-6911-4aba-ade5-a7087933ca8d", "revision": 0, "last_node_id": 10, "last_link_id": 12, "nodes": [{"id": 8, "type": "VAEDecode", "pos": [1209, 188], "size": [210, 46], "flags": {}, "order": 6, "mode": 0, "inputs": [{"name": "samples", "type": "LATENT", "link": 7}, {"name": "vae", "type": "VAE", "link": 8}], "outputs": [{"name": "IMAGE", "type": "IMAGE", "slot_index": 0, "links": [9]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "VAEDecode", "widget_ue_connectable": {}}, "widgets_values": []}, {"id": 9, "type": "SaveImage", "pos": [1451, 189], "size": [210, 58], "flags": {}, "order": 7, "mode": 0, "inputs": [{"name": "images", "type": "IMAGE", "link": 9}], "outputs": [], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "widget_ue_connectable": {}}, "widgets_values": ["ComfyUI"]}, {"id": 4, "type": "CheckpointLoaderSimple", "pos": [-272.2301025390625, 457.4971923828125], "size": [315, 98], "flags": {}, "order": 0, "mode": 0, "inputs": [], "outputs": [{"name": "MODEL", "type": "MODEL", "slot_index": 0, "links": [1]}, {"name": "CLIP", "type": "CLIP", "slot_index": 1, "links": [10]}, {"name": "VAE", "type": "VAE", "slot_index": 2, "links": [8]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CheckpointLoaderSimple", "widget_ue_connectable": {}}, "widgets_values": ["illumiyume_35.safetensors"]}, {"id": 10, "type": "CLIPSetLastLayer", "pos": [41.93305206298828, 215.06910705566406], "size": [270, 58], "flags": {}, "order": 2, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 10}], "outputs": [{"name": "CLIP", "type": "CLIP", "links": [11, 12]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPSetLastLayer", "widget_ue_connectable": {}}, "widgets_values": [-2]}, {"id": 3, "type": "KSampler", "pos": [863, 186], "size": [315, 262], "flags": {}, "order": 5, "mode": 0, "inputs": [{"name": "model", "type": "MODEL", "link": 1}, {"name": "positive", "type": "CONDITIONING", "link": 4}, {"name": "negative", "type": "CONDITIONING", "link": 6}, {"name": "latent_image", "type": "LATENT", "link": 2}], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [7]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "KSampler", "widget_ue_connectable": {}}, "widgets_values": [538624701520958, "randomize", 30, 5, "dpmpp_2s_ancestral", "normal", 1]}, {"id": 7, "type": "CLIPTextEncode", "pos": [413, 389], "size": [425.27801513671875, 180.6060791015625], "flags": {}, "order": 4, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 12}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [6]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background"]}, {"id": 5, "type": "EmptyLatentImage", "pos": [473, 609], "size": [315, 106], "flags": {}, "order": 1, "mode": 0, "inputs": [], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [2]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "EmptyLatentImage", "widget_ue_connectable": {}}, "widgets_values": [1024, 1536, 3]}, {"id": 6, "type": "CLIPTextEncode", "pos": [415, 186], "size": [422.84503173828125, 164.31304931640625], "flags": {}, "order": 3, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 11}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [4]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["masterpiece,best quality,amazing quality, anya (spy x family), twilight (spy x family), yor briar, 1boy, 2girls, black gloves, black hair, black headwear, black suit, blue eyes, brown hair, child, dress, eden academy school uniform, formal clothes, gloves, grey eyes, hat, headband, holding, holding hat, holding unworn clothes, long hair, looking at viewer, multiple girls, neck ribbon, necktie, red eyes, red necktie, red ribbon, red sweater, ribbed sweater, ribbon, school uniform, sidelocks, smile, suit, sweater, sweater dress, unworn hat, unworn headwear, upper body, white headband"]}], "links": [[1, 4, 0, 3, 0, "MODEL"], [2, 5, 0, 3, 3, "LATENT"], [4, 6, 0, 3, 1, "CONDITIONING"], [6, 7, 0, 3, 2, "CONDITIONING"], [7, 3, 0, 8, 0, "LATENT"], [8, 4, 2, 8, 1, "VAE"], [9, 8, 0, 9, 0, "IMAGE"], [10, 4, 1, 10, 0, "CLIP"], [11, 10, 0, 6, 0, "CLIP"], [12, 10, 0, 7, 0, "CLIP"]], "groups": [], "config": {}, "extra": {"ue_links": [], "links_added_by_ue": [], "ds": {"scale": 1.1291383685800622, "offset": [-155.85485296077175, 25.93501757600727]}, "frontendVersion": "1.21.7"}, "version": 0.4}}
Width
1024
Height
1536
Models
['illumiyume_35.safetensors']
Denoise
1
Modelids
[]
Scheduler
normal
Upscalers
[]
Versionids
[]
Controlnets
[]
Additionalresources
[]
+
+
+ + + + +
+ +
+
Prompt
masterpiece, (zoom:1.1), +fujita kotone, ((whichs better?:1.3) spoken-text:1), thinking, hand up, open mouth
Negative prompt
bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background
Seed
800449096470161
Model
illumiyume_35
Sampler
DPM++ 2S a
Steps
30
CFG scale
5
+
+
+ + +
+
Vaes
[]
Comfy
{"prompt": {"3": {"inputs": {"seed": 800449096470161, "steps": 30, "cfg": 5.0, "sampler_name": "dpmpp_2s_ancestral", "scheduler": "normal", "denoise": 1.0, "model": ["4", 0], "positive": ["6", 0], "negative": ["7", 0], "latent_image": ["5", 0]}, "class_type": "KSampler", "_meta": {"title": "KSampler"}}, "4": {"inputs": {"ckpt_name": "illumiyume_35.safetensors"}, "class_type": "CheckpointLoaderSimple", "_meta": {"title": "Load Checkpoint"}}, "5": {"inputs": {"width": 1024, "height": 1536, "batch_size": 3}, "class_type": "EmptyLatentImage", "_meta": {"title": "Empty Latent Image"}}, "6": {"inputs": {"text": "masterpiece, (zoom:1.1), \nfujita kotone, ((whichs better?:1.3) spoken-text:1), thinking, hand up, open mouth", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "7": {"inputs": {"text": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background", "clip": ["10", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Prompt)"}}, "8": {"inputs": {"samples": ["3", 0], "vae": ["4", 2]}, "class_type": "VAEDecode", "_meta": {"title": "VAE Decode"}}, "9": {"inputs": {"filename_prefix": "ComfyUI", "images": ["8", 0]}, "class_type": "SaveImage", "_meta": {"title": "Save Image"}}, "10": {"inputs": {"stop_at_clip_layer": -2, "clip": ["4", 1]}, "class_type": "CLIPSetLastLayer", "_meta": {"title": "CLIP Set Last Layer"}}}, "workflow": {"id": "7f35ca53-6911-4aba-ade5-a7087933ca8d", "revision": 0, "last_node_id": 10, "last_link_id": 12, "nodes": [{"id": 8, "type": "VAEDecode", "pos": [1209, 188], "size": [210, 46], "flags": {}, "order": 6, "mode": 0, "inputs": [{"name": "samples", "type": "LATENT", "link": 7}, {"name": "vae", "type": "VAE", "link": 8}], "outputs": [{"name": "IMAGE", "type": "IMAGE", "slot_index": 0, "links": [9]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "VAEDecode", "widget_ue_connectable": {}}, "widgets_values": []}, {"id": 9, "type": "SaveImage", "pos": [1451, 189], "size": [210, 58], "flags": {}, "order": 7, "mode": 0, "inputs": [{"name": "images", "type": "IMAGE", "link": 9}], "outputs": [], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "widget_ue_connectable": {}}, "widgets_values": ["ComfyUI"]}, {"id": 4, "type": "CheckpointLoaderSimple", "pos": [-272.2301025390625, 457.4971923828125], "size": [315, 98], "flags": {}, "order": 0, "mode": 0, "inputs": [], "outputs": [{"name": "MODEL", "type": "MODEL", "slot_index": 0, "links": [1]}, {"name": "CLIP", "type": "CLIP", "slot_index": 1, "links": [10]}, {"name": "VAE", "type": "VAE", "slot_index": 2, "links": [8]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CheckpointLoaderSimple", "widget_ue_connectable": {}}, "widgets_values": ["illumiyume_35.safetensors"]}, {"id": 10, "type": "CLIPSetLastLayer", "pos": [41.93305206298828, 215.06910705566406], "size": [270, 58], "flags": {}, "order": 2, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 10}], "outputs": [{"name": "CLIP", "type": "CLIP", "links": [11, 12]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPSetLastLayer", "widget_ue_connectable": {}}, "widgets_values": [-2]}, {"id": 3, "type": "KSampler", "pos": [863, 186], "size": [315, 262], "flags": {}, "order": 5, "mode": 0, "inputs": [{"name": "model", "type": "MODEL", "link": 1}, {"name": "positive", "type": "CONDITIONING", "link": 4}, {"name": "negative", "type": "CONDITIONING", "link": 6}, {"name": "latent_image", "type": "LATENT", "link": 2}], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [7]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "KSampler", "widget_ue_connectable": {}}, "widgets_values": [800449096470161, "randomize", 30, 5, "dpmpp_2s_ancestral", "normal", 1]}, {"id": 7, "type": "CLIPTextEncode", "pos": [413, 389], "size": [425.27801513671875, 180.6060791015625], "flags": {}, "order": 4, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 12}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [6]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background"]}, {"id": 5, "type": "EmptyLatentImage", "pos": [473, 609], "size": [315, 106], "flags": {}, "order": 1, "mode": 0, "inputs": [], "outputs": [{"name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [2]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "EmptyLatentImage", "widget_ue_connectable": {}}, "widgets_values": [1024, 1536, 3]}, {"id": 6, "type": "CLIPTextEncode", "pos": [415, 186], "size": [422.84503173828125, 164.31304931640625], "flags": {}, "order": 3, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 11}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [4]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.40", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {}}, "widgets_values": ["masterpiece, (zoom:1.1), \nfujita kotone, ((whichs better?:1.3) spoken-text:1), thinking, hand up, open mouth"]}], "links": [[1, 4, 0, 3, 0, "MODEL"], [2, 5, 0, 3, 3, "LATENT"], [4, 6, 0, 3, 1, "CONDITIONING"], [6, 7, 0, 3, 2, "CONDITIONING"], [7, 3, 0, 8, 0, "LATENT"], [8, 4, 2, 8, 1, "VAE"], [9, 8, 0, 9, 0, "IMAGE"], [10, 4, 1, 10, 0, "CLIP"], [11, 10, 0, 6, 0, "CLIP"], [12, 10, 0, 7, 0, "CLIP"]], "groups": [], "config": {}, "extra": {"ue_links": [], "links_added_by_ue": [], "ds": {"scale": 0.8483383685800604, "offset": [195.00071225071235, 205.75356125356134]}, "frontendVersion": "1.21.7"}, "version": 0.4}}
Width
1024
Height
1536
Models
['illumiyume_35.safetensors']
Denoise
1
Modelids
[]
Scheduler
normal
Upscalers
[]
Versionids
[]
Controlnets
[]
Additionalresources
[]
+ \ No newline at end of file