import torch from skimage.filters import gaussian from skimage.util import compare_images import numpy as np import torch.nn.functional as F from PIL import Image from ..utils import tensor2pil, pil2tensor, tensor2np import torch import folder_paths from PIL.PngImagePlugin import PngInfo import json import os import math # try: # from cv2.ximgproc import guidedFilter # except ImportError: # log.warning("cv2.ximgproc.guidedFilter not found, use opencv-contrib-python") class ColorCorrect: """Various color correction methods""" @classmethod def INPUT_TYPES(cls): return { "required": { "image": ("IMAGE",), "clamp": ([True, False], {"default": True}), "gamma": ( "FLOAT", {"default": 1.0, "min": 0.0, "max": 5.0, "step": 0.01}, ), "contrast": ( "FLOAT", {"default": 1.0, "min": 0.0, "max": 5.0, "step": 0.01}, ), "exposure": ( "FLOAT", {"default": 0.0, "min": -5.0, "max": 5.0, "step": 0.01}, ), "offset": ( "FLOAT", {"default": 0.0, "min": -5.0, "max": 5.0, "step": 0.01}, ), "hue": ( "FLOAT", {"default": 0.0, "min": -0.5, "max": 0.5, "step": 0.01}, ), "saturation": ( "FLOAT", {"default": 1.0, "min": 0.0, "max": 5.0, "step": 0.01}, ), "value": ( "FLOAT", {"default": 1.0, "min": 0.0, "max": 5.0, "step": 0.01}, ), } } RETURN_TYPES = ("IMAGE",) FUNCTION = "correct" CATEGORY = "mtb/image processing" @staticmethod def gamma_correction_tensor(image, gamma): gamma_inv = 1.0 / gamma return image.pow(gamma_inv) @staticmethod def contrast_adjustment_tensor(image, contrast): contrasted = (image - 0.5) * contrast + 0.5 return torch.clamp(contrasted, 0.0, 1.0) @staticmethod def exposure_adjustment_tensor(image, exposure): return image * (2.0**exposure) @staticmethod def offset_adjustment_tensor(image, offset): return image + offset @staticmethod def hsv_adjustment(image: torch.Tensor, hue, saturation, value): images = tensor2pil(image) out = [] for img in images: hsv_image = img.convert("HSV") h, s, v = hsv_image.split() h = h.point(lambda x: (x + hue * 255) % 256) s = s.point(lambda x: int(x * saturation)) v = v.point(lambda x: int(x * value)) hsv_image = Image.merge("HSV", (h, s, v)) rgb_image = hsv_image.convert("RGB") out.append(rgb_image) return pil2tensor(out) @staticmethod def hsv_adjustment_tensor_not_working(image: torch.Tensor, hue, saturation, value): """Abandonning for now""" image = image.squeeze(0).permute(2, 0, 1) max_val, _ = image.max(dim=0, keepdim=True) min_val, _ = image.min(dim=0, keepdim=True) delta = max_val - min_val hue_image = torch.zeros_like(max_val) mask = delta != 0.0 r, g, b = image[0], image[1], image[2] hue_image[mask & (max_val == r)] = ((g - b) / delta)[ mask & (max_val == r) ] % 6.0 hue_image[mask & (max_val == g)] = ((b - r) / delta)[ mask & (max_val == g) ] + 2.0 hue_image[mask & (max_val == b)] = ((r - g) / delta)[ mask & (max_val == b) ] + 4.0 saturation_image = delta / (max_val + 1e-7) value_image = max_val hue_image = (hue_image + hue) % 1.0 saturation_image = torch.where( mask, saturation * saturation_image, saturation_image ) value_image = value * value_image c = value_image * saturation_image x = c * (1 - torch.abs((hue_image % 2) - 1)) m = value_image - c prime_image = torch.zeros_like(image) prime_image[0] = torch.where( max_val == r, c, torch.where(max_val == g, x, prime_image[0]) ) prime_image[1] = torch.where( max_val == r, x, torch.where(max_val == g, c, prime_image[1]) ) prime_image[2] = torch.where( max_val == g, x, torch.where(max_val == b, c, prime_image[2]) ) rgb_image = prime_image + m rgb_image = rgb_image.permute(1, 2, 0).unsqueeze(0) return rgb_image def correct( self, image: torch.Tensor, clamp: bool, gamma: float = 1.0, contrast: float = 1.0, exposure: float = 0.0, offset: float = 0.0, hue: float = 0.0, saturation: float = 1.0, value: float = 1.0, ): # Apply color correction operations image = self.gamma_correction_tensor(image, gamma) image = self.contrast_adjustment_tensor(image, contrast) image = self.exposure_adjustment_tensor(image, exposure) image = self.offset_adjustment_tensor(image, offset) image = self.hsv_adjustment(image, hue, saturation, value) if clamp: image = torch.clamp(image, 0.0, 1.0) return (image,) class ImageCompare: """Compare two images and return a difference image""" @classmethod def INPUT_TYPES(cls): return { "required": { "imageA": ("IMAGE",), "imageB": ("IMAGE",), "mode": ( ["checkerboard", "diff", "blend"], {"default": "checkerboard"}, ), } } RETURN_TYPES = ("IMAGE",) FUNCTION = "compare" CATEGORY = "mtb/image" def compare(self, imageA: torch.Tensor, imageB: torch.Tensor, mode): imageA = imageA.numpy() imageB = imageB.numpy() imageA = imageA.squeeze() imageB = imageB.squeeze() image = compare_images(imageA, imageB, method=mode) image = np.expand_dims(image, axis=0) return (torch.from_numpy(image),) import requests class LoadImageFromUrl: """Load an image from the given URL""" @classmethod def INPUT_TYPES(cls): return { "required": { "url": ( "STRING", { "default": "https://upload.wikimedia.org/wikipedia/commons/thumb/a/a7/Example.jpg/800px-Example.jpg" }, ), } } RETURN_TYPES = ("IMAGE",) FUNCTION = "load" CATEGORY = "mtb/IO" def load(self, url): # get the image from the url image = Image.open(requests.get(url, stream=True).raw) return (pil2tensor(image),) class Blur: """Blur an image using a Gaussian filter.""" @classmethod def INPUT_TYPES(cls): return { "required": { "image": ("IMAGE",), "sigmaX": ( "FLOAT", {"default": 3.0, "min": 0.0, "max": 10.0, "step": 0.01}, ), "sigmaY": ( "FLOAT", {"default": 3.0, "min": 0.0, "max": 10.0, "step": 0.01}, ), } } RETURN_TYPES = ("IMAGE",) FUNCTION = "blur" CATEGORY = "mtb/image processing" def blur(self, image: torch.Tensor, sigmaX, sigmaY): image = image.numpy() image = image.transpose(1, 2, 3, 0) image = gaussian(image, sigma=(sigmaX, sigmaY, 0, 0)) image = image.transpose(3, 0, 1, 2) return (torch.from_numpy(image),) # https://github.com/lllyasviel/AdverseCleaner/blob/main/clean.py # def deglaze_np_img(np_img): # y = np_img.copy() # for _ in range(64): # y = cv2.bilateralFilter(y, 5, 8, 8) # for _ in range(4): # y = guidedFilter(np_img, y, 4, 16) # return y # class DeglazeImage: # """Remove adversarial noise from images""" # @classmethod # def INPUT_TYPES(cls): # return {"required": {"image": ("IMAGE",)}} # CATEGORY = "mtb/image processing" # RETURN_TYPES = ("IMAGE",) # FUNCTION = "deglaze_image" # def deglaze_image(self, image): # return (np2tensor(deglaze_np_img(tensor2np(image))),) class MaskToImage: """Converts a mask (alpha) to an RGB image with a color and background""" @classmethod def INPUT_TYPES(cls): return { "required": { "mask": ("MASK",), "color": ("COLOR",), "background": ("COLOR", {"default": "#000000"}), } } CATEGORY = "mtb/generate" RETURN_TYPES = ("IMAGE",) FUNCTION = "render_mask" def render_mask(self, mask, color, background): mask = tensor2np(mask) mask = Image.fromarray(mask).convert("L") image = Image.new("RGBA", mask.size, color=color) # apply the mask image = Image.composite( image, Image.new("RGBA", mask.size, color=background), mask ) # image = ImageChops.multiply(image, mask) # apply over background # image = Image.alpha_composite(Image.new("RGBA", image.size, color=background), image) image = pil2tensor(image.convert("RGB")) return (image,) class ColoredImage: """Constant color image of given size""" def __init__(self) -> None: pass @classmethod def INPUT_TYPES(cls): return { "required": { "color": ("COLOR",), "width": ("INT", {"default": 512, "min": 16, "max": 8160}), "height": ("INT", {"default": 512, "min": 16, "max": 8160}), } } CATEGORY = "mtb/generate" RETURN_TYPES = ("IMAGE",) FUNCTION = "render_img" def render_img(self, color, width, height): image = Image.new("RGB", (width, height), color=color) image = pil2tensor(image) return (image,) class ImagePremultiply: """Premultiply image with mask""" @classmethod def INPUT_TYPES(cls): return { "required": { "image": ("IMAGE",), "mask": ("MASK",), "invert": ("BOOLEAN", {"default": False}), } } CATEGORY = "mtb/image" RETURN_TYPES = ("IMAGE",) FUNCTION = "premultiply" def premultiply(self, image, mask, invert): images = tensor2pil(image) if invert: masks = tensor2pil(mask) # .convert("L") else: masks = tensor2pil(1.0 - mask) single = False if len(mask) == 1: single = True masks = [x.convert("L") for x in masks] out = [] for i, img in enumerate(images): cur_mask = masks[0] if single else masks[i] img.putalpha(cur_mask) out.append(img) # if invert: # image = Image.composite(image,Image.new("RGBA", image.size, color=(0,0,0,0)), mask) # else: # image = Image.composite(Image.new("RGBA", image.size, color=(0,0,0,0)), image, mask) return (pil2tensor(out),) class ImageResizeFactor: """Extracted mostly from WAS Node Suite, with a few edits (most notably multiple image support) and less features.""" @classmethod def INPUT_TYPES(cls): return { "required": { "image": ("IMAGE",), "factor": ( "FLOAT", {"default": 2, "min": 0.01, "max": 16.0, "step": 0.01}, ), "supersample": ("BOOLEAN", {"default": True}), "resampling": ( [ "nearest", "linear", "bilinear", "bicubic", "trilinear", "area", "nearest-exact", ], {"default": "nearest"}, ), }, "optional": { "mask": ("MASK",), }, } CATEGORY = "mtb/image" RETURN_TYPES = ("IMAGE", "MASK") FUNCTION = "resize" def resize( self, image: torch.Tensor, factor: float, supersample: bool, resampling: str, mask=None, ): # Check if the tensor has the correct dimension if len(image.shape) not in [3, 4]: # HxWxC or BxHxWxC raise ValueError("Expected image tensor of shape (H, W, C) or (B, H, W, C)") # Transpose to CxHxW or BxCxHxW for PyTorch if len(image.shape) == 3: image = image.permute(2, 0, 1).unsqueeze(0) # CxHxW else: image = image.permute(0, 3, 1, 2) # BxCxHxW # Compute new dimensions B, C, H, W = image.shape new_H, new_W = int(H * factor), int(W * factor) align_corner_filters = ("linear", "bilinear", "bicubic", "trilinear") # Resize the image resized_image = F.interpolate( image, size=(new_H, new_W), mode=resampling, align_corners=resampling in align_corner_filters, ) # Optionally supersample if supersample: resized_image = F.interpolate( resized_image, scale_factor=2, mode=resampling, align_corners=resampling in align_corner_filters, ) # Transpose back to the original format: BxHxWxC or HxWxC if len(image.shape) == 4: resized_image = resized_image.permute(0, 2, 3, 1) else: resized_image = resized_image.squeeze(0).permute(1, 2, 0) # Apply mask if provided if mask is not None: if len(mask.shape) != len(resized_image.shape): raise ValueError( "Mask tensor should have the same dimensions as the image tensor" ) resized_image = resized_image * mask return (resized_image,) class SaveImageGrid: """Save all the images in the input batch as a grid of images.""" def __init__(self): self.output_dir = folder_paths.get_output_directory() self.type = "output" @classmethod def INPUT_TYPES(cls): return { "required": { "images": ("IMAGE",), "filename_prefix": ("STRING", {"default": "ComfyUI"}), "save_intermediate": ("BOOLEAN", {"default": False}), }, "hidden": {"prompt": "PROMPT", "extra_pnginfo": "EXTRA_PNGINFO"}, } RETURN_TYPES = () FUNCTION = "save_images" OUTPUT_NODE = True CATEGORY = "mtb/IO" def create_image_grid(self, image_list): total_images = len(image_list) # Calculate the grid size based on the square root of the total number of images grid_size = ( int(math.sqrt(total_images)), int(math.ceil(math.sqrt(total_images))), ) # Get the size of the first image to determine the grid size image_width, image_height = image_list[0].size # Create a new blank image to hold the grid grid_width = grid_size[0] * image_width grid_height = grid_size[1] * image_height grid_image = Image.new("RGB", (grid_width, grid_height)) # Iterate over the images and paste them onto the grid for i, image in enumerate(image_list): x = (i % grid_size[0]) * image_width y = (i // grid_size[0]) * image_height grid_image.paste(image, (x, y, x + image_width, y + image_height)) return grid_image def save_images( self, images, filename_prefix="Grid", save_intermediate=False, prompt=None, extra_pnginfo=None, ): ( full_output_folder, filename, counter, subfolder, filename_prefix, ) = folder_paths.get_save_image_path( filename_prefix, self.output_dir, images[0].shape[1], images[0].shape[0] ) image_list = [] batch_counter = counter metadata = PngInfo() if prompt is not None: metadata.add_text("prompt", json.dumps(prompt)) if extra_pnginfo is not None: for x in extra_pnginfo: metadata.add_text(x, json.dumps(extra_pnginfo[x])) for idx, image in enumerate(images): i = 255.0 * image.cpu().numpy() img = Image.fromarray(np.clip(i, 0, 255).astype(np.uint8)) image_list.append(img) if save_intermediate: file = f"{filename}_batch-{idx:03}_{batch_counter:05}_.png" img.save( os.path.join(full_output_folder, file), pnginfo=metadata, compress_level=4, ) batch_counter += 1 file = f"{filename}_{counter:05}_.png" grid = self.create_image_grid(image_list) grid.save( os.path.join(full_output_folder, file), pnginfo=metadata, compress_level=4 ) results = [{"filename": file, "subfolder": subfolder, "type": self.type}] return {"ui": {"images": results}} __nodes__ = [ ColorCorrect, ImageCompare, Blur, # DeglazeImage, MaskToImage, ColoredImage, ImagePremultiply, ImageResizeFactor, SaveImageGrid, LoadImageFromUrl, ]