File size: 4,859 Bytes
d0e86f6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.

from dataclasses import dataclass

import numpy as np
import torch
import torch.nn.functional as F
from PIL import Image


def is_valid_alpha(
    alpha: np.ndarray, min_ratio: float = 0.01, threshold: int = 127
) -> bool:
    """Check if alpha channel has sufficient foreground and background pixels."""
    total_pixels = alpha.size
    min_count = int(total_pixels * min_ratio)
    fg_count = np.count_nonzero(alpha > threshold)
    bg_count = total_pixels - fg_count
    return bg_count >= min_count and fg_count >= min_count


def load_image(
    image: Image.Image,
    bg_color: np.ndarray,
) -> tuple[torch.Tensor, tuple[int, int, int, int]]:
    """Load and process an RGBA image, applying background color."""
    # Ensure RGBA format
    if image.mode != "RGBA":
        image = image.convert("RGBA")

    # Get image as contiguous numpy array
    img_array = np.ascontiguousarray(image)
    rgb = img_array[..., :3]
    alpha = img_array[..., 3]

    if not is_valid_alpha(alpha):
        raise ValueError("Invalid alpha channel: insufficient foreground/background")

    # Normalize alpha to [0, 1] as float32 for compositing
    alpha_norm = alpha.astype(np.float32) * (1.0 / 255.0)

    # Composite RGB with background
    # rgb_out = rgb * alpha + bg * (1 - alpha)
    alpha_3ch = alpha_norm[..., np.newaxis]  # (H, W, 1)
    bg_color_f32 = bg_color.astype(np.float32)
    rgb_composite = rgb.astype(np.float32) * (
        1.0 / 255.0
    ) * alpha_3ch + bg_color_f32 * (1.0 - alpha_3ch)

    # Convert to tensor (HWC -> CHW)
    rgb_pt = torch.from_numpy(np.ascontiguousarray(rgb_composite.transpose(2, 0, 1)))

    # Compute bounding box from alpha mask
    alpha_mask = alpha > 0
    rows_any = np.any(alpha_mask, axis=1)
    cols_any = np.any(alpha_mask, axis=0)
    row_indices = np.nonzero(rows_any)[0]
    col_indices = np.nonzero(cols_any)[0]
    y, y_max = row_indices[0], row_indices[-1]
    x, x_max = col_indices[0], col_indices[-1]
    w, h = x_max - x + 1, y_max - y + 1

    return rgb_pt, (x, y, w, h)


def aggregate_bboxes(
    bboxes: list[tuple[int, int, int, int]],
) -> tuple[int, int, int, int]:
    """Compute the bounding box that encompasses all input bboxes."""
    x_min = min(b[0] for b in bboxes)
    y_min = min(b[1] for b in bboxes)
    x_max = max(b[0] + b[2] for b in bboxes)
    y_max = max(b[1] + b[3] for b in bboxes)
    return x_min, y_min, x_max - x_min, y_max - y_min


def apply_padding(
    rgb_image: torch.Tensor,
    bbox: tuple[int, int, int, int],
    padding_ratio: float = 0.1,
    padding_value: float = 1.0,
) -> torch.Tensor:
    """Crop to bounding box and pad to make square with additional margin."""
    x, y, w, h = bbox

    # Crop to bounding box
    cropped = rgb_image[:, y : y + h, x : x + w]

    # Calculate padding to make square with margin
    max_dim = max(w, h)
    pad_base = int(max_dim * padding_ratio)
    pad_x = pad_base + (max_dim - w) // 2
    pad_y = pad_base + (max_dim - h) // 2

    return F.pad(
        cropped, (pad_x, pad_x, pad_y, pad_y), mode="constant", value=padding_value
    )


@dataclass(eq=False)
class ImagePreprocessor:
    """Preprocessor for ActionMesh input images.

    Handles RGBA images by compositing onto a white background, cropping to
    the foreground region, and padding to create square outputs.

    Attributes:
        independent_cropping: If True, crop each frame independently. If False,
            use a shared bounding box across all frames for consistent framing.
        padding_ratio: Padding to add around the cropped region as a fraction
            of the largest dimension.
    """

    independent_cropping: bool = False
    padding_ratio: float = 0.1

    def __post_init__(self):
        self.bg_color = np.array([1.0, 1.0, 1.0])

    def process_images(self, frames: list[Image.Image]) -> list[Image.Image]:
        """Process frames: load, crop, pad, and return as PIL images."""

        results = [load_image(frame, self.bg_color) for frame in frames]
        images = [r[0] for r in results]
        bboxes = [r[1] for r in results]

        # Use aggregated bbox if not cropping independently
        if not self.independent_cropping:
            agg_bbox = aggregate_bboxes(bboxes)
            bboxes = [agg_bbox] * len(bboxes)

        # Crop and pad each image
        processed = [
            apply_padding(img, bbox, self.padding_ratio, self.bg_color[0])
            for img, bbox in zip(images, bboxes)
        ]

        # Convert back to PIL images
        return [
            Image.fromarray((img.permute(1, 2, 0).cpu().numpy() * 255).astype(np.uint8))
            for img in processed
        ]