|
|
import math |
|
|
from PIL import Image |
|
|
from torchvision.transforms import ToTensor, ToPILImage |
|
|
import torch |
|
|
|
|
|
|
|
|
import random |
|
|
from imgaug import augmenters as iaa |
|
|
import numpy as np |
|
|
|
|
|
NEWLINE_TOKEN = 13 |
|
|
DOT_TOKEN = 29892 |
|
|
|
|
|
def split_to_patches(image, grid): |
|
|
patches = [] |
|
|
width, height = image.size |
|
|
grid_x = int(width / grid[0]) |
|
|
grid_y = int(height / grid[1]) |
|
|
|
|
|
for i in range(0, height, grid_y): |
|
|
images = [] |
|
|
for j in range(0, width, grid_x): |
|
|
box = (j, i, j + grid_x, i + grid_y) |
|
|
patch = image.crop(box) |
|
|
images.append(patch) |
|
|
patches.append(images) |
|
|
|
|
|
return patches |
|
|
|
|
|
def get_refine_size( |
|
|
original_size, grid, scale_resolution, patch_size, allow_upscale=False |
|
|
): |
|
|
width, height = original_size |
|
|
grid_x, grid_y = grid |
|
|
|
|
|
refine_width = ensure_divide(width, grid_x) |
|
|
refine_height = ensure_divide(height, grid_y) |
|
|
|
|
|
grid_width = refine_width / grid_x |
|
|
grid_height = refine_height / grid_y |
|
|
|
|
|
best_grid_size = find_best_resize( |
|
|
(grid_width, grid_height), |
|
|
scale_resolution, |
|
|
patch_size, |
|
|
allow_upscale=allow_upscale, |
|
|
) |
|
|
|
|
|
refine_size = (best_grid_size[0] * grid_x, best_grid_size[1] * grid_y) |
|
|
|
|
|
return refine_size |
|
|
|
|
|
def ensure_divide(length, patch_size): |
|
|
|
|
|
return max(math.floor(length / patch_size) * patch_size, patch_size) |
|
|
|
|
|
def find_best_resize(original_size, scale_resolution, patch_size, allow_upscale=False, any_res=False): |
|
|
width, height = original_size |
|
|
if any_res: |
|
|
r = width / height |
|
|
if (width * height > scale_resolution * scale_resolution): |
|
|
height = int(scale_resolution / math.sqrt(r)) |
|
|
width = int(height * r) |
|
|
elif (width * height < 256 * 256): |
|
|
height = int(256 / math.sqrt(r)) |
|
|
width = int(height * r) |
|
|
else: |
|
|
if (width * height > scale_resolution * scale_resolution) or allow_upscale: |
|
|
r = width / height |
|
|
height = int(scale_resolution / math.sqrt(r)) |
|
|
width = int(height * r) |
|
|
best_width = ensure_divide(width, patch_size) |
|
|
best_height = ensure_divide(height, patch_size) |
|
|
return (best_width, best_height) |
|
|
|
|
|
def slice_image_minicpm( |
|
|
image, max_slice_nums=9, scale_resolution=448, patch_size=14, never_split=False, any_res=False |
|
|
): |
|
|
original_size = image.size |
|
|
original_width, original_height = original_size |
|
|
log_ratio = math.log(original_width / original_height) |
|
|
ratio = original_width * original_height / (scale_resolution * scale_resolution) |
|
|
multiple = min(math.ceil(ratio), max_slice_nums) |
|
|
|
|
|
source_image = None |
|
|
best_grid = None |
|
|
patches = [] |
|
|
|
|
|
if multiple <= 1 or never_split: |
|
|
|
|
|
best_size = find_best_resize( |
|
|
original_size, scale_resolution, patch_size, allow_upscale=True, any_res=any_res |
|
|
) |
|
|
source_image = image.resize(best_size, Image.Resampling.BICUBIC) |
|
|
else: |
|
|
candidate_split_grids_nums = [] |
|
|
for i in [multiple - 1, multiple, multiple + 1]: |
|
|
if i == 1 or i > max_slice_nums: |
|
|
continue |
|
|
candidate_split_grids_nums.append(i) |
|
|
|
|
|
|
|
|
best_resize = find_best_resize(original_size, scale_resolution, patch_size, any_res=any_res) |
|
|
source_image = image.copy().resize(best_resize, Image.Resampling.BICUBIC) |
|
|
candidate_grids = [] |
|
|
|
|
|
|
|
|
for split_grids_nums in candidate_split_grids_nums: |
|
|
m = 1 |
|
|
while m <= split_grids_nums: |
|
|
if split_grids_nums % m == 0: |
|
|
candidate_grids.append([m, split_grids_nums // m]) |
|
|
m += 1 |
|
|
|
|
|
best_grid = [1, 1] |
|
|
min_error = float("inf") |
|
|
for grid in candidate_grids: |
|
|
error = abs(log_ratio - math.log(grid[0] / grid[1])) |
|
|
if error < min_error: |
|
|
best_grid = grid |
|
|
min_error = error |
|
|
|
|
|
refine_size = get_refine_size( |
|
|
original_size, best_grid, scale_resolution, patch_size, allow_upscale=True |
|
|
) |
|
|
|
|
|
refine_image = image.resize(refine_size, Image.Resampling.BICUBIC) |
|
|
patches = split_to_patches(refine_image, best_grid) |
|
|
|
|
|
ind_tokens = [] |
|
|
if best_grid is None: |
|
|
return source_image, patches, best_grid, ind_tokens |
|
|
else: |
|
|
|
|
|
patches = [item for sublist in patches for item in sublist] |
|
|
|
|
|
for j in range(best_grid[1]): |
|
|
for i in range(best_grid[0]): |
|
|
if i != best_grid[0] - 1: |
|
|
ind_tokens.append(DOT_TOKEN) |
|
|
else: |
|
|
ind_tokens.append(NEWLINE_TOKEN) |
|
|
|
|
|
return source_image, patches, best_grid, ind_tokens |
|
|
|
|
|
|
|
|
|
|
|
def split_image(image, scale=672, grid=(2, 2)): |
|
|
resized_image = image.resize((scale, scale)) |
|
|
width, height = resized_image.size |
|
|
grid_width = width // grid[0] |
|
|
grid_height = height // grid[1] |
|
|
|
|
|
sub_images = [] |
|
|
|
|
|
for i in range(grid[0]): |
|
|
for j in range(grid[1]): |
|
|
left = i * grid_width |
|
|
upper = j * grid_height |
|
|
right = left + grid_width |
|
|
lower = upper + grid_height |
|
|
sub_image = resized_image.crop((left, upper, right, lower)) |
|
|
sub_images.append(sub_image) |
|
|
|
|
|
return sub_images |
|
|
|
|
|
|
|
|
def generate_subimage_coordinates(H, W, h, w, num_windows): |
|
|
""" |
|
|
生成子图的左上角和右下角坐标,并返回一个形状为 (n, 4) 的 PyTorch tensor。 |
|
|
|
|
|
参数: |
|
|
H (int): 原始图像的高度 |
|
|
W (int): 原始图像的宽度 |
|
|
h (int): 子图的高度 |
|
|
w (int): 子图的宽度 |
|
|
|
|
|
返回: |
|
|
torch.Tensor: 形状为 (n, 4) 的张量,包含所有子图的左上角和右下角坐标 |
|
|
""" |
|
|
|
|
|
|
|
|
rows = int(round(H / h)) |
|
|
cols = int(round(W / w)) |
|
|
assert rows * cols == num_windows, f'H:{H}, W:{W}, h:{h}, w:{w}, rows:{H/h}, cols:{W/w}' |
|
|
coordinates = [] |
|
|
for i in range(rows): |
|
|
for j in range(cols): |
|
|
x1 = j * w |
|
|
y1 = i * h |
|
|
x2 = x1 + w |
|
|
y2 = y1 + h |
|
|
coordinates.append([x1, y1, x2, y2]) |
|
|
|
|
|
return torch.tensor(coordinates, dtype=torch.float32) |
|
|
|
|
|
def slice_image_feature_minicpm( |
|
|
image_feature, num_windows=144, max_slice_nums=1000, num_ratio=1): |
|
|
|
|
|
|
|
|
|
|
|
bs = image_feature.shape[0] |
|
|
dtype, device = image_feature.dtype, image_feature.device |
|
|
feature_size = image_feature.shape[-2:] |
|
|
feature_height, feature_width = feature_size |
|
|
log_ratio = math.log(feature_width / feature_height) |
|
|
ratio = feature_height * feature_width / num_windows |
|
|
multiple = min(math.ceil(ratio), max_slice_nums) |
|
|
|
|
|
candidate_split_grids_nums = [] |
|
|
for i in [multiple - 1, multiple, multiple + 1]: |
|
|
if i == 1 or i > max_slice_nums: |
|
|
continue |
|
|
candidate_split_grids_nums.append(i) |
|
|
|
|
|
candidate_grids = [] |
|
|
|
|
|
for split_grids_nums in candidate_split_grids_nums: |
|
|
m = 1 |
|
|
while m <= split_grids_nums: |
|
|
if split_grids_nums % m == 0: |
|
|
candidate_grids.append([m, split_grids_nums // m]) |
|
|
m += 1 |
|
|
|
|
|
best_grid = [1, 1] |
|
|
min_error = float("inf") |
|
|
for grid in candidate_grids: |
|
|
error = abs(log_ratio - math.log(grid[0] / grid[1])) |
|
|
if error < min_error: |
|
|
best_grid = grid |
|
|
min_error = error |
|
|
|
|
|
|
|
|
float_crop_height = math.sqrt(ratio / (feature_width / feature_height)) |
|
|
float_crop_width = float_crop_height * (feature_width / feature_height) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
region_boxes = generate_subimage_coordinates(feature_height, feature_width, |
|
|
float_crop_height, float_crop_width, num_windows) |
|
|
|
|
|
region_boxes = region_boxes.to(dtype=dtype, device=device).detach() |
|
|
batch_region_boxes = [] |
|
|
for i in range(bs): |
|
|
batch_id = torch.ones_like(region_boxes)[:, :1] * i |
|
|
batch_region_boxes.append(torch.cat([batch_id, region_boxes], dim=1)) |
|
|
batch_region_boxes = torch.cat(batch_region_boxes) |
|
|
|
|
|
return batch_region_boxes, best_grid, feature_width / feature_height |
|
|
|
|
|
|
|
|
def resize_image_keep_ratio(image, max_size=1024): |
|
|
original_width, original_height = image.size |
|
|
if original_width > original_height: |
|
|
new_width = max_size |
|
|
new_height = int((max_size / original_width) * original_height) |
|
|
else: |
|
|
new_height = max_size |
|
|
new_width = int((max_size / original_height) * original_width) |
|
|
resized_image = image.resize((new_width, new_height), Image.Resampling.BICUBIC) |
|
|
return resized_image |
|
|
|
|
|
|
|
|
def aug_image(image): |
|
|
if random.random() < 0.5: |
|
|
image = resize_image_keep_ratio(image, max_size=1024) |
|
|
if random.random() < 0.1: |
|
|
aug = iaa.contrast.LinearContrast((0.5, 2.0), per_channel=False) |
|
|
image = Image.fromarray(aug(image=np.array(image))) |
|
|
if random.random() < 0.1: |
|
|
aug = iaa.Sharpen(alpha=(0.0, 0.5), lightness=(0.75, 1.5)) |
|
|
image = Image.fromarray(aug(image=np.array(image))) |
|
|
if random.random() < 0.2: |
|
|
aug = iaa.AddToHue((-50, 50)) |
|
|
image = Image.fromarray(aug(image=np.array(image))) |
|
|
if random.random() < 0.1: |
|
|
aug = iaa.JpegCompression(compression=(75, 95)) |
|
|
image = Image.fromarray(aug(image=np.array(image))) |
|
|
return image |
|
|
|
|
|
|