| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | from typing import Tuple |
| |
|
| | from PIL import Image |
| | from torchvision import transforms |
| | from transformers import Siglip2ImageProcessorFast |
| |
|
| | from .tokenizer_wrapper import ImageInfo, JointImageInfo, ResolutionGroup |
| |
|
| |
|
| | def resize_and_crop(image: Image.Image, target_size: Tuple[int, int]) -> Image.Image: |
| | tw, th = target_size |
| | w, h = image.size |
| |
|
| | tr = th / tw |
| | r = h / w |
| |
|
| | |
| | if r < tr: |
| | resize_height = th |
| | resize_width = int(round(th / h * w)) |
| | else: |
| | resize_width = tw |
| | resize_height = int(round(tw / w * h)) |
| |
|
| | image = image.resize((resize_width, resize_height), resample=Image.Resampling.LANCZOS) |
| |
|
| | |
| | crop_top = int(round((resize_height - th) / 2.0)) |
| | crop_left = int(round((resize_width - tw) / 2.0)) |
| |
|
| | image = image.crop((crop_left, crop_top, crop_left + tw, crop_top + th)) |
| | return image |
| |
|
| |
|
| | class HunyuanImage3ImageProcessor(object): |
| | def __init__(self, config): |
| | self.config = config |
| |
|
| | min_multiple = getattr(config, "image_min_multiple", 0.5) |
| | max_multiple = getattr(config, "image_max_multiple", 2.0) |
| | step = getattr(config, "image_resolution_step", None) |
| | align = getattr(config, "image_resolution_align", 1) |
| | max_entries = getattr(config, "image_resolution_count", 33) |
| | presets = getattr(config, "image_resolution_presets", None) |
| | self.reso_group = ResolutionGroup( |
| | base_size=config.image_base_size, |
| | step=step, |
| | align=align, |
| | min_multiple=min_multiple, |
| | max_multiple=max_multiple, |
| | max_entries=max_entries, |
| | presets=presets, |
| | ) |
| | self.vae_processor = transforms.Compose([ |
| | transforms.ToTensor(), |
| | transforms.Normalize([0.5], [0.5]), |
| | ]) |
| | self.vision_encoder_processor = Siglip2ImageProcessorFast.from_dict(config.vit_processor) |
| |
|
| | def build_image_info(self, image_size): |
| | |
| | if isinstance(image_size, str): |
| | if image_size.startswith("<img_ratio_"): |
| | ratio_index = int(image_size.split("_")[-1].rstrip(">")) |
| | reso = self.reso_group[ratio_index] |
| | image_size = reso.height, reso.width |
| | elif 'x' in image_size: |
| | image_size = [int(s) for s in image_size.split('x')] |
| | elif ':' in image_size: |
| | image_size = [int(s) for s in image_size.split(':')] |
| | else: |
| | raise ValueError( |
| | f"`image_size` should be in the format of 'HxW', 'H:W' or <img_ratio_i>, got {image_size}.") |
| | assert len(image_size) == 2, f"`image_size` should be in the format of 'HxW', got {image_size}." |
| | elif isinstance(image_size, (list, tuple)): |
| | assert len(image_size) == 2 and all(isinstance(s, int) for s in image_size), \ |
| | f"`image_size` should be a tuple of two integers or a string in the format of 'HxW', got {image_size}." |
| | else: |
| | raise ValueError(f"`image_size` should be a tuple of two integers or a string in the format of 'WxH', " |
| | f"got {image_size}.") |
| | image_width, image_height = self.reso_group.get_target_size(image_size[1], image_size[0]) |
| | token_height = image_height // (self.config.vae_downsample_factor[0] * self.config.patch_size) |
| | token_width = image_width // (self.config.vae_downsample_factor[1] * self.config.patch_size) |
| | base_size, ratio_idx = self.reso_group.get_base_size_and_ratio_index(image_size[1], image_size[0]) |
| | image_info = ImageInfo( |
| | image_type="gen_image", image_width=image_width, image_height=image_height, |
| | token_width=token_width, token_height=token_height, base_size=base_size, ratio_index=ratio_idx, |
| | ) |
| | return image_info |
| |
|
| | def preprocess(self, image: Image.Image): |
| | |
| | image_width, image_height = self.reso_group.get_target_size(image.width, image.height) |
| | resized_image = resize_and_crop(image, (image_width, image_height)) |
| | image_tensor = self.vae_processor(resized_image) |
| | token_height = image_height // (self.config.vae_downsample_factor[0] * self.config.patch_size) |
| | token_width = image_width // (self.config.vae_downsample_factor[1] * self.config.patch_size) |
| | base_size, ratio_index = self.reso_group.get_base_size_and_ratio_index(width=image_width, height=image_height) |
| | vae_image_info = ImageInfo( |
| | image_type="vae", |
| | image_tensor=image_tensor.unsqueeze(0), |
| | image_width=image_width, image_height=image_height, |
| | token_width=token_width, token_height=token_height, |
| | base_size=base_size, ratio_index=ratio_index, |
| | ) |
| |
|
| | |
| | inputs = self.vision_encoder_processor(image) |
| | image = inputs["pixel_values"].squeeze(0) |
| | pixel_attention_mask = inputs["pixel_attention_mask"].squeeze(0) |
| | spatial_shapes = inputs["spatial_shapes"].squeeze(0) |
| | vision_encoder_kwargs = dict( |
| | pixel_attention_mask=pixel_attention_mask, |
| | spatial_shapes=spatial_shapes, |
| | ) |
| | vision_image_info = ImageInfo( |
| | image_type="vit", |
| | image_tensor=image.unsqueeze(0), |
| | image_width=spatial_shapes[1].item() * self.config.vit_processor["patch_size"], |
| | image_height=spatial_shapes[0].item() * self.config.vit_processor["patch_size"], |
| | token_width=spatial_shapes[1].item(), |
| | token_height=spatial_shapes[0].item(), |
| | image_token_length=self.config.vit_processor["max_num_patches"], |
| | |
| | ) |
| | return JointImageInfo(vae_image_info, vision_image_info, vision_encoder_kwargs) |
| |
|