| |
| |
| |
| |
| |
|
|
| from dataclasses import dataclass |
|
|
| import numpy as np |
| import torch |
| import torch.nn.functional as F |
| from PIL import Image |
|
|
|
|
| def is_valid_alpha( |
| alpha: np.ndarray, min_ratio: float = 0.01, threshold: int = 127 |
| ) -> bool: |
| """Check if alpha channel has sufficient foreground and background pixels.""" |
| total_pixels = alpha.size |
| min_count = int(total_pixels * min_ratio) |
| fg_count = np.count_nonzero(alpha > threshold) |
| bg_count = total_pixels - fg_count |
| return bg_count >= min_count and fg_count >= min_count |
|
|
|
|
| def load_image( |
| image: Image.Image, |
| bg_color: np.ndarray, |
| ) -> tuple[torch.Tensor, tuple[int, int, int, int]]: |
| """Load and process an RGBA image, applying background color.""" |
| |
| if image.mode != "RGBA": |
| image = image.convert("RGBA") |
|
|
| |
| img_array = np.ascontiguousarray(image) |
| rgb = img_array[..., :3] |
| alpha = img_array[..., 3] |
|
|
| if not is_valid_alpha(alpha): |
| raise ValueError("Invalid alpha channel: insufficient foreground/background") |
|
|
| |
| alpha_norm = alpha.astype(np.float32) * (1.0 / 255.0) |
|
|
| |
| |
| alpha_3ch = alpha_norm[..., np.newaxis] |
| bg_color_f32 = bg_color.astype(np.float32) |
| rgb_composite = rgb.astype(np.float32) * ( |
| 1.0 / 255.0 |
| ) * alpha_3ch + bg_color_f32 * (1.0 - alpha_3ch) |
|
|
| |
| rgb_pt = torch.from_numpy(np.ascontiguousarray(rgb_composite.transpose(2, 0, 1))) |
|
|
| |
| alpha_mask = alpha > 0 |
| rows_any = np.any(alpha_mask, axis=1) |
| cols_any = np.any(alpha_mask, axis=0) |
| row_indices = np.nonzero(rows_any)[0] |
| col_indices = np.nonzero(cols_any)[0] |
| y, y_max = row_indices[0], row_indices[-1] |
| x, x_max = col_indices[0], col_indices[-1] |
| w, h = x_max - x + 1, y_max - y + 1 |
|
|
| return rgb_pt, (x, y, w, h) |
|
|
|
|
| def aggregate_bboxes( |
| bboxes: list[tuple[int, int, int, int]], |
| ) -> tuple[int, int, int, int]: |
| """Compute the bounding box that encompasses all input bboxes.""" |
| x_min = min(b[0] for b in bboxes) |
| y_min = min(b[1] for b in bboxes) |
| x_max = max(b[0] + b[2] for b in bboxes) |
| y_max = max(b[1] + b[3] for b in bboxes) |
| return x_min, y_min, x_max - x_min, y_max - y_min |
|
|
|
|
| def apply_padding( |
| rgb_image: torch.Tensor, |
| bbox: tuple[int, int, int, int], |
| padding_ratio: float = 0.1, |
| padding_value: float = 1.0, |
| ) -> torch.Tensor: |
| """Crop to bounding box and pad to make square with additional margin.""" |
| x, y, w, h = bbox |
|
|
| |
| cropped = rgb_image[:, y : y + h, x : x + w] |
|
|
| |
| max_dim = max(w, h) |
| pad_base = int(max_dim * padding_ratio) |
| pad_x = pad_base + (max_dim - w) // 2 |
| pad_y = pad_base + (max_dim - h) // 2 |
|
|
| return F.pad( |
| cropped, (pad_x, pad_x, pad_y, pad_y), mode="constant", value=padding_value |
| ) |
|
|
|
|
| @dataclass(eq=False) |
| class ImagePreprocessor: |
| """Preprocessor for ActionMesh input images. |
| |
| Handles RGBA images by compositing onto a white background, cropping to |
| the foreground region, and padding to create square outputs. |
| |
| Attributes: |
| independent_cropping: If True, crop each frame independently. If False, |
| use a shared bounding box across all frames for consistent framing. |
| padding_ratio: Padding to add around the cropped region as a fraction |
| of the largest dimension. |
| """ |
|
|
| independent_cropping: bool = False |
| padding_ratio: float = 0.1 |
|
|
| def __post_init__(self): |
| self.bg_color = np.array([1.0, 1.0, 1.0]) |
|
|
| def process_images(self, frames: list[Image.Image]) -> list[Image.Image]: |
| """Process frames: load, crop, pad, and return as PIL images.""" |
|
|
| results = [load_image(frame, self.bg_color) for frame in frames] |
| images = [r[0] for r in results] |
| bboxes = [r[1] for r in results] |
|
|
| |
| if not self.independent_cropping: |
| agg_bbox = aggregate_bboxes(bboxes) |
| bboxes = [agg_bbox] * len(bboxes) |
|
|
| |
| processed = [ |
| apply_padding(img, bbox, self.padding_ratio, self.bg_color[0]) |
| for img, bbox in zip(images, bboxes) |
| ] |
|
|
| |
| return [ |
| Image.fromarray((img.permute(1, 2, 0).cpu().numpy() * 255).astype(np.uint8)) |
| for img in processed |
| ] |
|
|