ActionMesh / data /actionmesh /preprocessing /image_processor.py
introvoyz041's picture
Migrated from GitHub
d0e86f6 verified
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
from dataclasses import dataclass
import numpy as np
import torch
import torch.nn.functional as F
from PIL import Image
def is_valid_alpha(
alpha: np.ndarray, min_ratio: float = 0.01, threshold: int = 127
) -> bool:
"""Check if alpha channel has sufficient foreground and background pixels."""
total_pixels = alpha.size
min_count = int(total_pixels * min_ratio)
fg_count = np.count_nonzero(alpha > threshold)
bg_count = total_pixels - fg_count
return bg_count >= min_count and fg_count >= min_count
def load_image(
image: Image.Image,
bg_color: np.ndarray,
) -> tuple[torch.Tensor, tuple[int, int, int, int]]:
"""Load and process an RGBA image, applying background color."""
# Ensure RGBA format
if image.mode != "RGBA":
image = image.convert("RGBA")
# Get image as contiguous numpy array
img_array = np.ascontiguousarray(image)
rgb = img_array[..., :3]
alpha = img_array[..., 3]
if not is_valid_alpha(alpha):
raise ValueError("Invalid alpha channel: insufficient foreground/background")
# Normalize alpha to [0, 1] as float32 for compositing
alpha_norm = alpha.astype(np.float32) * (1.0 / 255.0)
# Composite RGB with background
# rgb_out = rgb * alpha + bg * (1 - alpha)
alpha_3ch = alpha_norm[..., np.newaxis] # (H, W, 1)
bg_color_f32 = bg_color.astype(np.float32)
rgb_composite = rgb.astype(np.float32) * (
1.0 / 255.0
) * alpha_3ch + bg_color_f32 * (1.0 - alpha_3ch)
# Convert to tensor (HWC -> CHW)
rgb_pt = torch.from_numpy(np.ascontiguousarray(rgb_composite.transpose(2, 0, 1)))
# Compute bounding box from alpha mask
alpha_mask = alpha > 0
rows_any = np.any(alpha_mask, axis=1)
cols_any = np.any(alpha_mask, axis=0)
row_indices = np.nonzero(rows_any)[0]
col_indices = np.nonzero(cols_any)[0]
y, y_max = row_indices[0], row_indices[-1]
x, x_max = col_indices[0], col_indices[-1]
w, h = x_max - x + 1, y_max - y + 1
return rgb_pt, (x, y, w, h)
def aggregate_bboxes(
bboxes: list[tuple[int, int, int, int]],
) -> tuple[int, int, int, int]:
"""Compute the bounding box that encompasses all input bboxes."""
x_min = min(b[0] for b in bboxes)
y_min = min(b[1] for b in bboxes)
x_max = max(b[0] + b[2] for b in bboxes)
y_max = max(b[1] + b[3] for b in bboxes)
return x_min, y_min, x_max - x_min, y_max - y_min
def apply_padding(
rgb_image: torch.Tensor,
bbox: tuple[int, int, int, int],
padding_ratio: float = 0.1,
padding_value: float = 1.0,
) -> torch.Tensor:
"""Crop to bounding box and pad to make square with additional margin."""
x, y, w, h = bbox
# Crop to bounding box
cropped = rgb_image[:, y : y + h, x : x + w]
# Calculate padding to make square with margin
max_dim = max(w, h)
pad_base = int(max_dim * padding_ratio)
pad_x = pad_base + (max_dim - w) // 2
pad_y = pad_base + (max_dim - h) // 2
return F.pad(
cropped, (pad_x, pad_x, pad_y, pad_y), mode="constant", value=padding_value
)
@dataclass(eq=False)
class ImagePreprocessor:
"""Preprocessor for ActionMesh input images.
Handles RGBA images by compositing onto a white background, cropping to
the foreground region, and padding to create square outputs.
Attributes:
independent_cropping: If True, crop each frame independently. If False,
use a shared bounding box across all frames for consistent framing.
padding_ratio: Padding to add around the cropped region as a fraction
of the largest dimension.
"""
independent_cropping: bool = False
padding_ratio: float = 0.1
def __post_init__(self):
self.bg_color = np.array([1.0, 1.0, 1.0])
def process_images(self, frames: list[Image.Image]) -> list[Image.Image]:
"""Process frames: load, crop, pad, and return as PIL images."""
results = [load_image(frame, self.bg_color) for frame in frames]
images = [r[0] for r in results]
bboxes = [r[1] for r in results]
# Use aggregated bbox if not cropping independently
if not self.independent_cropping:
agg_bbox = aggregate_bboxes(bboxes)
bboxes = [agg_bbox] * len(bboxes)
# Crop and pad each image
processed = [
apply_padding(img, bbox, self.padding_ratio, self.bg_color[0])
for img, bbox in zip(images, bboxes)
]
# Convert back to PIL images
return [
Image.fromarray((img.permute(1, 2, 0).cpu().numpy() * 255).astype(np.uint8))
for img in processed
]