eyeball / app.py
xerox-elf's picture
Update app.py
6c80d8c verified
import gradio as gr
import torch
import os
import random
import sys
from typing import Sequence, Mapping, Any, Union
from transformers import AutoModel, AutoTokenizer
# Example of loading a custom LoRA model
lora_model = AutoModel.from_pretrained("xerox-elf/eyeballlora")
# Example of loading a checkpoint
checkpoint = AutoModel.from_pretrained("stablediffusionapi/juggernaut-xl-v8")
import gradio as gr
from PIL import Image, ImageOps
def process_image(input_image):
"""
Convert the input image to grayscale.
Args:
- input_image: The input image as a PIL.Image object.
Returns:
- A PIL.Image object representing the processed image.
"""
# Convert the image to grayscale
grayscale_image = ImageOps.grayscale(input_image)
return grayscale_image
# Create a Gradio interface
interface = gr.Interface(fn=process_image,
inputs=gr.inputs.Image(shape=(224, 224)),
outputs=gr.outputs.Image(type="pil"),
title="Image Grayscale Converter",
description="Upload an image to convert it to grayscale.")
# Launch the interface
interface.launch()
def get_value_at_index(obj: Union[Sequence, Mapping], index: int) -> Any:
"""Returns the value at the given index of a sequence or mapping.
If the object is a sequence (like list or string), returns the value at the given index.
If the object is a mapping (like a dictionary), returns the value at the index-th key.
Some return a dictionary, in these cases, we look for the "results" key
Args:
obj (Union[Sequence, Mapping]): The object to retrieve the value from.
index (int): The index of the value to retrieve.
Returns:
Any: The value at the given index.
Raises:
IndexError: If the index is out of bounds for the object and the object is not a mapping.
"""
try:
return obj[index]
except KeyError:
return obj["result"][index]
def find_path(name: str, path: str = None) -> str:
"""
Recursively looks at parent folders starting from the given path until it finds the given name.
Returns the path as a Path object if found, or None otherwise.
"""
# If no path is given, use the current working directory
if path is None:
path = os.getcwd()
# Check if the current directory contains the name
if name in os.listdir(path):
path_name = os.path.join(path, name)
print(f"{name} found: {path_name}")
return path_name
# Get the parent directory
parent_directory = os.path.dirname(path)
# If the parent directory is the same as the current directory, we've reached the root and stop the search
if parent_directory == path:
return None
# Recursively call the function with the parent directory
return find_path(name, parent_directory)
def add_comfyui_directory_to_sys_path() -> None:
"""
Add 'ComfyUI' to the sys.path
"""
comfyui_path = find_path("ComfyUI")
if comfyui_path is not None and os.path.isdir(comfyui_path):
sys.path.append(comfyui_path)
print(f"'{comfyui_path}' added to sys.path")
def add_extra_model_paths() -> None:
"""
Parse the optional extra_model_paths.yaml file and add the parsed paths to the sys.path.
"""
from main import load_extra_path_config
extra_model_paths = find_path("extra_model_paths.yaml")
if extra_model_paths is not None:
load_extra_path_config(extra_model_paths)
else:
print("Could not find the extra_model_paths config file.")
add_comfyui_directory_to_sys_path()
add_extra_model_paths()
def import_custom_nodes() -> None:
"""Find all custom nodes in the custom_nodes folder and add those node objects to NODE_CLASS_MAPPINGS
This function sets up a new asyncio event loop, initializes the PromptServer,
creates a PromptQueue, and initializes the custom nodes.
"""
import asyncio
import execution
from nodes import init_custom_nodes
import server
# Creating a new event loop and setting it as the default loop
loop = asyncio.new_event_loop()
asyncio.set_event_loop(loop)
# Creating an instance of PromptServer with the loop
server_instance = server.PromptServer(loop)
execution.PromptQueue(server_instance)
# Initializing custom nodes
init_custom_nodes()
from nodes import (
LoadImage,
SaveImage,
LatentUpscale,
SetLatentNoiseMask,
CLIPTextEncode,
NODE_CLASS_MAPPINGS,
CheckpointLoaderSimple,
LoraLoader,
KSampler,
VAEDecode,
VAEEncode,
)
def main():
import_custom_nodes()
with torch.inference_mode():
checkpointloadersimple = CheckpointLoaderSimple()
checkpointloadersimple_38 = checkpointloadersimple.load_checkpoint(
ckpt_name="juggernautXL_v8Rundiffusion.safetensors"
)
loraloader = LoraLoader()
loraloader_51 = loraloader.load_lora(
lora_name="model/eyeballsXL-000025.safetensors",
strength_model=1,
strength_clip=1,
model=get_value_at_index(checkpointloadersimple_38, 0),
clip=get_value_at_index(checkpointloadersimple_38, 1),
)
cliptextencode = CLIPTextEncode()
cliptextencode_40 = cliptextencode.encode(
text="photograph of person with spherical round eyeballs, black pupils",
clip=get_value_at_index(loraloader_51, 1),
)
cliptextencode_41 = cliptextencode.encode(
text="", clip=get_value_at_index(loraloader_51, 1)
)
loadimage = LoadImage()
loadimage_167 = loadimage.load_image(image="IMG_2734_copy.jpg")
mediapipe_facemeshpreprocessor = NODE_CLASS_MAPPINGS[
"MediaPipe-FaceMeshPreprocessor"
]()
mediapipe_facemeshpreprocessor_25 = mediapipe_facemeshpreprocessor.detect(
max_faces=10,
min_confidence=0.5,
resolution=512,
image=get_value_at_index(loadimage_167, 0),
)
mediapipefacemeshtosegs = NODE_CLASS_MAPPINGS["MediaPipeFaceMeshToSEGS"]()
mediapipefacemeshtosegs_18 = mediapipefacemeshtosegs.doit(
crop_factor=3,
bbox_fill=False,
crop_min_size=50,
drop_size=1,
dilation=0,
face=False,
mouth=False,
left_eyebrow=False,
left_eye=True,
left_pupil=False,
right_eyebrow=False,
right_eye=True,
right_pupil=False,
image=get_value_at_index(mediapipe_facemeshpreprocessor_25, 0),
)
segstocombinedmask = NODE_CLASS_MAPPINGS["SegsToCombinedMask"]()
segstocombinedmask_23 = segstocombinedmask.doit(
segs=get_value_at_index(mediapipefacemeshtosegs_18, 0)
)
impactdilatemask = NODE_CLASS_MAPPINGS["ImpactDilateMask"]()
impactdilatemask_26 = impactdilatemask.doit(
dilation=55, mask=get_value_at_index(segstocombinedmask_23, 0)
)
impactgaussianblurmask = NODE_CLASS_MAPPINGS["ImpactGaussianBlurMask"]()
impactgaussianblurmask_27 = impactgaussianblurmask.doit(
kernel_size=20, sigma=10, mask=get_value_at_index(impactdilatemask_26, 0)
)
masktoimage = NODE_CLASS_MAPPINGS["MaskToImage"]()
masktoimage_21 = masktoimage.mask_to_image(
mask=get_value_at_index(impactgaussianblurmask_27, 0)
)
mask_to_region = NODE_CLASS_MAPPINGS["Mask To Region"]()
mask_to_region_29 = mask_to_region.get_region(
padding=0,
constraints="keep_ratio",
constraint_x=64,
constraint_y=64,
min_width=0,
min_height=0,
batch_behavior="match_ratio",
mask=get_value_at_index(masktoimage_21, 0),
)
cut_by_mask = NODE_CLASS_MAPPINGS["Cut By Mask"]()
cut_by_mask_30 = cut_by_mask.cut(
force_resize_width=0,
force_resize_height=0,
image=get_value_at_index(loadimage_167, 0),
mask=get_value_at_index(mask_to_region_29, 0),
)
vaeencode = VAEEncode()
vaeencode_48 = vaeencode.encode(
pixels=get_value_at_index(cut_by_mask_30, 0),
vae=get_value_at_index(checkpointloadersimple_38, 2),
)
latentupscale = LatentUpscale()
latentupscale_60 = latentupscale.upscale(
upscale_method="nearest-exact",
width=1024,
height=1024,
crop="disabled",
samples=get_value_at_index(vaeencode_48, 0),
)
ksampler = KSampler()
ksampler_39 = ksampler.sample(
seed=random.randint(1, 2**64),
steps=20,
cfg=8,
sampler_name="euler",
scheduler="normal",
denoise=0.65,
model=get_value_at_index(loraloader_51, 0),
positive=get_value_at_index(cliptextencode_40, 0),
negative=get_value_at_index(cliptextencode_41, 0),
latent_image=get_value_at_index(latentupscale_60, 0),
)
vaedecode = VAEDecode()
vaedecode_43 = vaedecode.decode(
samples=get_value_at_index(ksampler_39, 0),
vae=get_value_at_index(checkpointloadersimple_38, 2),
)
cut_by_mask_32 = cut_by_mask.cut(
force_resize_width=0,
force_resize_height=0,
image=get_value_at_index(masktoimage_21, 0),
mask=get_value_at_index(mask_to_region_29, 0),
)
cut_by_mask_62 = cut_by_mask.cut(
force_resize_width=0,
force_resize_height=0,
image=get_value_at_index(vaedecode_43, 0),
mask=get_value_at_index(cut_by_mask_32, 0),
)
paste_by_mask = NODE_CLASS_MAPPINGS["Paste By Mask"]()
paste_by_mask_45 = paste_by_mask.paste(
resize_behavior="resize",
image_base=get_value_at_index(loadimage_167, 0),
image_to_paste=get_value_at_index(cut_by_mask_62, 0),
mask=get_value_at_index(masktoimage_21, 0),
)
cut_by_mask_91 = cut_by_mask.cut(
force_resize_width=0,
force_resize_height=0,
image=get_value_at_index(paste_by_mask_45, 0),
mask=get_value_at_index(mask_to_region_29, 0),
)
vaeencode_93 = vaeencode.encode(
pixels=get_value_at_index(cut_by_mask_91, 0),
vae=get_value_at_index(checkpointloadersimple_38, 2),
)
load_image_batch = NODE_CLASS_MAPPINGS["Load Image Batch"]()
load_image_batch_166 = load_image_batch.load_batch_images(
mode="incremental_image",
index=0,
label="Batch 001",
path="/workspace/input",
pattern="*",
allow_RGBA_output="false",
filename_text_extension="true",
)
image_to_mask = NODE_CLASS_MAPPINGS["Image To Mask"]()
setlatentnoisemask = SetLatentNoiseMask()
saveimage = SaveImage()
for q in range(10):
image_to_mask_50 = image_to_mask.convert(
method="intensity", image=get_value_at_index(mask_to_region_29, 0)
)
setlatentnoisemask_49 = setlatentnoisemask.set_mask(
samples=get_value_at_index(latentupscale_60, 0),
mask=get_value_at_index(image_to_mask_50, 0),
)
latentupscale_111 = latentupscale.upscale(
upscale_method="nearest-exact",
width=1024,
height=1024,
crop="disabled",
samples=get_value_at_index(vaeencode_93, 0),
)
ksampler_95 = ksampler.sample(
seed=random.randint(1, 2**64),
steps=60,
cfg=8,
sampler_name="dpm_2",
scheduler="normal",
denoise=0.55,
model=get_value_at_index(loraloader_51, 0),
positive=get_value_at_index(cliptextencode_40, 0),
negative=get_value_at_index(cliptextencode_41, 0),
latent_image=get_value_at_index(latentupscale_111, 0),
)
vaedecode_96 = vaedecode.decode(
samples=get_value_at_index(ksampler_95, 0),
vae=get_value_at_index(checkpointloadersimple_38, 2),
)
cut_by_mask_124 = cut_by_mask.cut(
force_resize_width=0,
force_resize_height=0,
image=get_value_at_index(vaedecode_96, 0),
mask=get_value_at_index(cut_by_mask_32, 0),
)
paste_by_mask_127 = paste_by_mask.paste(
resize_behavior="resize",
image_base=get_value_at_index(paste_by_mask_45, 0),
image_to_paste=get_value_at_index(cut_by_mask_124, 0),
mask=get_value_at_index(masktoimage_21, 0),
)
saveimage_147 = saveimage.save_images(
filename_prefix="ComfyUI",
images=get_value_at_index(paste_by_mask_127, 0),
)
if __name__ == "__main__":
main()