|
|
import torch |
|
|
from skimage.filters import gaussian |
|
|
from skimage.util import compare_images |
|
|
import numpy as np |
|
|
import torch.nn.functional as F |
|
|
from PIL import Image |
|
|
from ..utils import tensor2pil, pil2tensor, tensor2np |
|
|
import torch |
|
|
import folder_paths |
|
|
from PIL.PngImagePlugin import PngInfo |
|
|
import json |
|
|
import os |
|
|
import math |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
class ColorCorrect: |
|
|
"""Various color correction methods""" |
|
|
|
|
|
@classmethod |
|
|
def INPUT_TYPES(cls): |
|
|
return { |
|
|
"required": { |
|
|
"image": ("IMAGE",), |
|
|
"clamp": ([True, False], {"default": True}), |
|
|
"gamma": ( |
|
|
"FLOAT", |
|
|
{"default": 1.0, "min": 0.0, "max": 5.0, "step": 0.01}, |
|
|
), |
|
|
"contrast": ( |
|
|
"FLOAT", |
|
|
{"default": 1.0, "min": 0.0, "max": 5.0, "step": 0.01}, |
|
|
), |
|
|
"exposure": ( |
|
|
"FLOAT", |
|
|
{"default": 0.0, "min": -5.0, "max": 5.0, "step": 0.01}, |
|
|
), |
|
|
"offset": ( |
|
|
"FLOAT", |
|
|
{"default": 0.0, "min": -5.0, "max": 5.0, "step": 0.01}, |
|
|
), |
|
|
"hue": ( |
|
|
"FLOAT", |
|
|
{"default": 0.0, "min": -0.5, "max": 0.5, "step": 0.01}, |
|
|
), |
|
|
"saturation": ( |
|
|
"FLOAT", |
|
|
{"default": 1.0, "min": 0.0, "max": 5.0, "step": 0.01}, |
|
|
), |
|
|
"value": ( |
|
|
"FLOAT", |
|
|
{"default": 1.0, "min": 0.0, "max": 5.0, "step": 0.01}, |
|
|
), |
|
|
} |
|
|
} |
|
|
|
|
|
RETURN_TYPES = ("IMAGE",) |
|
|
FUNCTION = "correct" |
|
|
CATEGORY = "mtb/image processing" |
|
|
|
|
|
@staticmethod |
|
|
def gamma_correction_tensor(image, gamma): |
|
|
gamma_inv = 1.0 / gamma |
|
|
return image.pow(gamma_inv) |
|
|
|
|
|
@staticmethod |
|
|
def contrast_adjustment_tensor(image, contrast): |
|
|
contrasted = (image - 0.5) * contrast + 0.5 |
|
|
return torch.clamp(contrasted, 0.0, 1.0) |
|
|
|
|
|
@staticmethod |
|
|
def exposure_adjustment_tensor(image, exposure): |
|
|
return image * (2.0**exposure) |
|
|
|
|
|
@staticmethod |
|
|
def offset_adjustment_tensor(image, offset): |
|
|
return image + offset |
|
|
|
|
|
@staticmethod |
|
|
def hsv_adjustment(image: torch.Tensor, hue, saturation, value): |
|
|
images = tensor2pil(image) |
|
|
out = [] |
|
|
for img in images: |
|
|
hsv_image = img.convert("HSV") |
|
|
|
|
|
h, s, v = hsv_image.split() |
|
|
|
|
|
h = h.point(lambda x: (x + hue * 255) % 256) |
|
|
s = s.point(lambda x: int(x * saturation)) |
|
|
v = v.point(lambda x: int(x * value)) |
|
|
|
|
|
hsv_image = Image.merge("HSV", (h, s, v)) |
|
|
rgb_image = hsv_image.convert("RGB") |
|
|
out.append(rgb_image) |
|
|
return pil2tensor(out) |
|
|
|
|
|
@staticmethod |
|
|
def hsv_adjustment_tensor_not_working(image: torch.Tensor, hue, saturation, value): |
|
|
"""Abandonning for now""" |
|
|
image = image.squeeze(0).permute(2, 0, 1) |
|
|
|
|
|
max_val, _ = image.max(dim=0, keepdim=True) |
|
|
min_val, _ = image.min(dim=0, keepdim=True) |
|
|
delta = max_val - min_val |
|
|
|
|
|
hue_image = torch.zeros_like(max_val) |
|
|
mask = delta != 0.0 |
|
|
|
|
|
r, g, b = image[0], image[1], image[2] |
|
|
hue_image[mask & (max_val == r)] = ((g - b) / delta)[ |
|
|
mask & (max_val == r) |
|
|
] % 6.0 |
|
|
hue_image[mask & (max_val == g)] = ((b - r) / delta)[ |
|
|
mask & (max_val == g) |
|
|
] + 2.0 |
|
|
hue_image[mask & (max_val == b)] = ((r - g) / delta)[ |
|
|
mask & (max_val == b) |
|
|
] + 4.0 |
|
|
|
|
|
saturation_image = delta / (max_val + 1e-7) |
|
|
value_image = max_val |
|
|
|
|
|
hue_image = (hue_image + hue) % 1.0 |
|
|
saturation_image = torch.where( |
|
|
mask, saturation * saturation_image, saturation_image |
|
|
) |
|
|
value_image = value * value_image |
|
|
|
|
|
c = value_image * saturation_image |
|
|
x = c * (1 - torch.abs((hue_image % 2) - 1)) |
|
|
m = value_image - c |
|
|
|
|
|
prime_image = torch.zeros_like(image) |
|
|
prime_image[0] = torch.where( |
|
|
max_val == r, c, torch.where(max_val == g, x, prime_image[0]) |
|
|
) |
|
|
prime_image[1] = torch.where( |
|
|
max_val == r, x, torch.where(max_val == g, c, prime_image[1]) |
|
|
) |
|
|
prime_image[2] = torch.where( |
|
|
max_val == g, x, torch.where(max_val == b, c, prime_image[2]) |
|
|
) |
|
|
|
|
|
rgb_image = prime_image + m |
|
|
|
|
|
rgb_image = rgb_image.permute(1, 2, 0).unsqueeze(0) |
|
|
|
|
|
return rgb_image |
|
|
|
|
|
def correct( |
|
|
self, |
|
|
image: torch.Tensor, |
|
|
clamp: bool, |
|
|
gamma: float = 1.0, |
|
|
contrast: float = 1.0, |
|
|
exposure: float = 0.0, |
|
|
offset: float = 0.0, |
|
|
hue: float = 0.0, |
|
|
saturation: float = 1.0, |
|
|
value: float = 1.0, |
|
|
): |
|
|
|
|
|
image = self.gamma_correction_tensor(image, gamma) |
|
|
image = self.contrast_adjustment_tensor(image, contrast) |
|
|
image = self.exposure_adjustment_tensor(image, exposure) |
|
|
image = self.offset_adjustment_tensor(image, offset) |
|
|
image = self.hsv_adjustment(image, hue, saturation, value) |
|
|
|
|
|
if clamp: |
|
|
image = torch.clamp(image, 0.0, 1.0) |
|
|
|
|
|
return (image,) |
|
|
|
|
|
|
|
|
class ImageCompare: |
|
|
"""Compare two images and return a difference image""" |
|
|
|
|
|
@classmethod |
|
|
def INPUT_TYPES(cls): |
|
|
return { |
|
|
"required": { |
|
|
"imageA": ("IMAGE",), |
|
|
"imageB": ("IMAGE",), |
|
|
"mode": ( |
|
|
["checkerboard", "diff", "blend"], |
|
|
{"default": "checkerboard"}, |
|
|
), |
|
|
} |
|
|
} |
|
|
|
|
|
RETURN_TYPES = ("IMAGE",) |
|
|
FUNCTION = "compare" |
|
|
CATEGORY = "mtb/image" |
|
|
|
|
|
def compare(self, imageA: torch.Tensor, imageB: torch.Tensor, mode): |
|
|
imageA = imageA.numpy() |
|
|
imageB = imageB.numpy() |
|
|
|
|
|
imageA = imageA.squeeze() |
|
|
imageB = imageB.squeeze() |
|
|
|
|
|
image = compare_images(imageA, imageB, method=mode) |
|
|
|
|
|
image = np.expand_dims(image, axis=0) |
|
|
return (torch.from_numpy(image),) |
|
|
|
|
|
|
|
|
import requests |
|
|
|
|
|
|
|
|
class LoadImageFromUrl: |
|
|
"""Load an image from the given URL""" |
|
|
|
|
|
@classmethod |
|
|
def INPUT_TYPES(cls): |
|
|
return { |
|
|
"required": { |
|
|
"url": ( |
|
|
"STRING", |
|
|
{ |
|
|
"default": "https://upload.wikimedia.org/wikipedia/commons/thumb/a/a7/Example.jpg/800px-Example.jpg" |
|
|
}, |
|
|
), |
|
|
} |
|
|
} |
|
|
|
|
|
RETURN_TYPES = ("IMAGE",) |
|
|
FUNCTION = "load" |
|
|
CATEGORY = "mtb/IO" |
|
|
|
|
|
def load(self, url): |
|
|
|
|
|
image = Image.open(requests.get(url, stream=True).raw) |
|
|
return (pil2tensor(image),) |
|
|
|
|
|
|
|
|
class Blur: |
|
|
"""Blur an image using a Gaussian filter.""" |
|
|
|
|
|
@classmethod |
|
|
def INPUT_TYPES(cls): |
|
|
return { |
|
|
"required": { |
|
|
"image": ("IMAGE",), |
|
|
"sigmaX": ( |
|
|
"FLOAT", |
|
|
{"default": 3.0, "min": 0.0, "max": 10.0, "step": 0.01}, |
|
|
), |
|
|
"sigmaY": ( |
|
|
"FLOAT", |
|
|
{"default": 3.0, "min": 0.0, "max": 10.0, "step": 0.01}, |
|
|
), |
|
|
} |
|
|
} |
|
|
|
|
|
RETURN_TYPES = ("IMAGE",) |
|
|
FUNCTION = "blur" |
|
|
CATEGORY = "mtb/image processing" |
|
|
|
|
|
def blur(self, image: torch.Tensor, sigmaX, sigmaY): |
|
|
image = image.numpy() |
|
|
image = image.transpose(1, 2, 3, 0) |
|
|
image = gaussian(image, sigma=(sigmaX, sigmaY, 0, 0)) |
|
|
image = image.transpose(3, 0, 1, 2) |
|
|
return (torch.from_numpy(image),) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
class MaskToImage: |
|
|
"""Converts a mask (alpha) to an RGB image with a color and background""" |
|
|
|
|
|
@classmethod |
|
|
def INPUT_TYPES(cls): |
|
|
return { |
|
|
"required": { |
|
|
"mask": ("MASK",), |
|
|
"color": ("COLOR",), |
|
|
"background": ("COLOR", {"default": "#000000"}), |
|
|
} |
|
|
} |
|
|
|
|
|
CATEGORY = "mtb/generate" |
|
|
|
|
|
RETURN_TYPES = ("IMAGE",) |
|
|
|
|
|
FUNCTION = "render_mask" |
|
|
|
|
|
def render_mask(self, mask, color, background): |
|
|
mask = tensor2np(mask) |
|
|
mask = Image.fromarray(mask).convert("L") |
|
|
|
|
|
image = Image.new("RGBA", mask.size, color=color) |
|
|
|
|
|
image = Image.composite( |
|
|
image, Image.new("RGBA", mask.size, color=background), mask |
|
|
) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
image = pil2tensor(image.convert("RGB")) |
|
|
|
|
|
return (image,) |
|
|
|
|
|
|
|
|
class ColoredImage: |
|
|
"""Constant color image of given size""" |
|
|
|
|
|
def __init__(self) -> None: |
|
|
pass |
|
|
|
|
|
@classmethod |
|
|
def INPUT_TYPES(cls): |
|
|
return { |
|
|
"required": { |
|
|
"color": ("COLOR",), |
|
|
"width": ("INT", {"default": 512, "min": 16, "max": 8160}), |
|
|
"height": ("INT", {"default": 512, "min": 16, "max": 8160}), |
|
|
} |
|
|
} |
|
|
|
|
|
CATEGORY = "mtb/generate" |
|
|
|
|
|
RETURN_TYPES = ("IMAGE",) |
|
|
|
|
|
FUNCTION = "render_img" |
|
|
|
|
|
def render_img(self, color, width, height): |
|
|
image = Image.new("RGB", (width, height), color=color) |
|
|
|
|
|
image = pil2tensor(image) |
|
|
|
|
|
return (image,) |
|
|
|
|
|
|
|
|
class ImagePremultiply: |
|
|
"""Premultiply image with mask""" |
|
|
|
|
|
@classmethod |
|
|
def INPUT_TYPES(cls): |
|
|
return { |
|
|
"required": { |
|
|
"image": ("IMAGE",), |
|
|
"mask": ("MASK",), |
|
|
"invert": ("BOOLEAN", {"default": False}), |
|
|
} |
|
|
} |
|
|
|
|
|
CATEGORY = "mtb/image" |
|
|
RETURN_TYPES = ("IMAGE",) |
|
|
FUNCTION = "premultiply" |
|
|
|
|
|
def premultiply(self, image, mask, invert): |
|
|
images = tensor2pil(image) |
|
|
if invert: |
|
|
masks = tensor2pil(mask) |
|
|
else: |
|
|
masks = tensor2pil(1.0 - mask) |
|
|
|
|
|
single = False |
|
|
if len(mask) == 1: |
|
|
single = True |
|
|
|
|
|
masks = [x.convert("L") for x in masks] |
|
|
|
|
|
out = [] |
|
|
for i, img in enumerate(images): |
|
|
cur_mask = masks[0] if single else masks[i] |
|
|
|
|
|
img.putalpha(cur_mask) |
|
|
out.append(img) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
return (pil2tensor(out),) |
|
|
|
|
|
|
|
|
class ImageResizeFactor: |
|
|
"""Extracted mostly from WAS Node Suite, with a few edits (most notably multiple image support) and less features.""" |
|
|
|
|
|
@classmethod |
|
|
def INPUT_TYPES(cls): |
|
|
return { |
|
|
"required": { |
|
|
"image": ("IMAGE",), |
|
|
"factor": ( |
|
|
"FLOAT", |
|
|
{"default": 2, "min": 0.01, "max": 16.0, "step": 0.01}, |
|
|
), |
|
|
"supersample": ("BOOLEAN", {"default": True}), |
|
|
"resampling": ( |
|
|
[ |
|
|
"nearest", |
|
|
"linear", |
|
|
"bilinear", |
|
|
"bicubic", |
|
|
"trilinear", |
|
|
"area", |
|
|
"nearest-exact", |
|
|
], |
|
|
{"default": "nearest"}, |
|
|
), |
|
|
}, |
|
|
"optional": { |
|
|
"mask": ("MASK",), |
|
|
}, |
|
|
} |
|
|
|
|
|
CATEGORY = "mtb/image" |
|
|
RETURN_TYPES = ("IMAGE", "MASK") |
|
|
FUNCTION = "resize" |
|
|
|
|
|
def resize( |
|
|
self, |
|
|
image: torch.Tensor, |
|
|
factor: float, |
|
|
supersample: bool, |
|
|
resampling: str, |
|
|
mask=None, |
|
|
): |
|
|
|
|
|
if len(image.shape) not in [3, 4]: |
|
|
raise ValueError("Expected image tensor of shape (H, W, C) or (B, H, W, C)") |
|
|
|
|
|
|
|
|
if len(image.shape) == 3: |
|
|
image = image.permute(2, 0, 1).unsqueeze(0) |
|
|
else: |
|
|
image = image.permute(0, 3, 1, 2) |
|
|
|
|
|
|
|
|
B, C, H, W = image.shape |
|
|
new_H, new_W = int(H * factor), int(W * factor) |
|
|
|
|
|
align_corner_filters = ("linear", "bilinear", "bicubic", "trilinear") |
|
|
|
|
|
resized_image = F.interpolate( |
|
|
image, |
|
|
size=(new_H, new_W), |
|
|
mode=resampling, |
|
|
align_corners=resampling in align_corner_filters, |
|
|
) |
|
|
|
|
|
|
|
|
if supersample: |
|
|
resized_image = F.interpolate( |
|
|
resized_image, |
|
|
scale_factor=2, |
|
|
mode=resampling, |
|
|
align_corners=resampling in align_corner_filters, |
|
|
) |
|
|
|
|
|
|
|
|
if len(image.shape) == 4: |
|
|
resized_image = resized_image.permute(0, 2, 3, 1) |
|
|
else: |
|
|
resized_image = resized_image.squeeze(0).permute(1, 2, 0) |
|
|
|
|
|
|
|
|
if mask is not None: |
|
|
if len(mask.shape) != len(resized_image.shape): |
|
|
raise ValueError( |
|
|
"Mask tensor should have the same dimensions as the image tensor" |
|
|
) |
|
|
resized_image = resized_image * mask |
|
|
|
|
|
return (resized_image,) |
|
|
|
|
|
|
|
|
class SaveImageGrid: |
|
|
"""Save all the images in the input batch as a grid of images.""" |
|
|
|
|
|
def __init__(self): |
|
|
self.output_dir = folder_paths.get_output_directory() |
|
|
self.type = "output" |
|
|
|
|
|
@classmethod |
|
|
def INPUT_TYPES(cls): |
|
|
return { |
|
|
"required": { |
|
|
"images": ("IMAGE",), |
|
|
"filename_prefix": ("STRING", {"default": "ComfyUI"}), |
|
|
"save_intermediate": ("BOOLEAN", {"default": False}), |
|
|
}, |
|
|
"hidden": {"prompt": "PROMPT", "extra_pnginfo": "EXTRA_PNGINFO"}, |
|
|
} |
|
|
|
|
|
RETURN_TYPES = () |
|
|
FUNCTION = "save_images" |
|
|
|
|
|
OUTPUT_NODE = True |
|
|
|
|
|
CATEGORY = "mtb/IO" |
|
|
|
|
|
def create_image_grid(self, image_list): |
|
|
total_images = len(image_list) |
|
|
|
|
|
|
|
|
grid_size = ( |
|
|
int(math.sqrt(total_images)), |
|
|
int(math.ceil(math.sqrt(total_images))), |
|
|
) |
|
|
|
|
|
|
|
|
image_width, image_height = image_list[0].size |
|
|
|
|
|
|
|
|
grid_width = grid_size[0] * image_width |
|
|
grid_height = grid_size[1] * image_height |
|
|
grid_image = Image.new("RGB", (grid_width, grid_height)) |
|
|
|
|
|
|
|
|
for i, image in enumerate(image_list): |
|
|
x = (i % grid_size[0]) * image_width |
|
|
y = (i // grid_size[0]) * image_height |
|
|
grid_image.paste(image, (x, y, x + image_width, y + image_height)) |
|
|
|
|
|
return grid_image |
|
|
|
|
|
def save_images( |
|
|
self, |
|
|
images, |
|
|
filename_prefix="Grid", |
|
|
save_intermediate=False, |
|
|
prompt=None, |
|
|
extra_pnginfo=None, |
|
|
): |
|
|
( |
|
|
full_output_folder, |
|
|
filename, |
|
|
counter, |
|
|
subfolder, |
|
|
filename_prefix, |
|
|
) = folder_paths.get_save_image_path( |
|
|
filename_prefix, self.output_dir, images[0].shape[1], images[0].shape[0] |
|
|
) |
|
|
image_list = [] |
|
|
batch_counter = counter |
|
|
|
|
|
metadata = PngInfo() |
|
|
if prompt is not None: |
|
|
metadata.add_text("prompt", json.dumps(prompt)) |
|
|
if extra_pnginfo is not None: |
|
|
for x in extra_pnginfo: |
|
|
metadata.add_text(x, json.dumps(extra_pnginfo[x])) |
|
|
|
|
|
for idx, image in enumerate(images): |
|
|
i = 255.0 * image.cpu().numpy() |
|
|
img = Image.fromarray(np.clip(i, 0, 255).astype(np.uint8)) |
|
|
image_list.append(img) |
|
|
|
|
|
if save_intermediate: |
|
|
file = f"{filename}_batch-{idx:03}_{batch_counter:05}_.png" |
|
|
img.save( |
|
|
os.path.join(full_output_folder, file), |
|
|
pnginfo=metadata, |
|
|
compress_level=4, |
|
|
) |
|
|
|
|
|
batch_counter += 1 |
|
|
|
|
|
file = f"{filename}_{counter:05}_.png" |
|
|
grid = self.create_image_grid(image_list) |
|
|
grid.save( |
|
|
os.path.join(full_output_folder, file), pnginfo=metadata, compress_level=4 |
|
|
) |
|
|
|
|
|
results = [{"filename": file, "subfolder": subfolder, "type": self.type}] |
|
|
return {"ui": {"images": results}} |
|
|
|
|
|
|
|
|
__nodes__ = [ |
|
|
ColorCorrect, |
|
|
ImageCompare, |
|
|
Blur, |
|
|
|
|
|
MaskToImage, |
|
|
ColoredImage, |
|
|
ImagePremultiply, |
|
|
ImageResizeFactor, |
|
|
SaveImageGrid, |
|
|
LoadImageFromUrl, |
|
|
] |
|
|
|