BryanW's picture
Upload folder using huggingface_hub
3d1c0e1 verified
# Copyright (c) 2025 FoundationVision
# SPDX-License-Identifier: MIT
import os
import os.path as osp
import cv2
import torch
import torch.nn.functional as F
import numpy as np
from infinity.schedules.dynamic_resolution import get_first_full_spatial_size_scale_index
def labels2image(all_indices, label_type='int_label', scale_schedule=None):
summed_codes, recons_imgs = self.vae.decode_from_indices(all_indices, scale_schedule, label_type)
recons_img = recons_imgs[0]
recons_img = (recons_img + 1) / 2
recons_img = recons_img.permute(1, 2, 0).mul_(255).cpu().numpy().astype(np.uint8)[:,:,::-1]
return recons_img
def features2image(raw_features):
recons_imgs = self.vae.decode(raw_features.squeeze(-3))
recons_img = recons_imgs[0]
recons_img = (recons_img + 1) / 2
recons_img = recons_img.permute(1, 2, 0).mul_(255).cpu().numpy().astype(np.uint8)[:,:,::-1]
return recons_img
class SelfCorrection(object):
def __init__(self, vae, args):
self.noise_apply_layers = args.noise_apply_layers
self.noise_apply_requant = args.noise_apply_requant
self.noise_apply_strength = args.noise_apply_strength
if not isinstance(self.noise_apply_strength, list):
self.noise_apply_strength = str(self.noise_apply_strength)
self.noise_apply_strength = list(map(float, self.noise_apply_strength.split(',')))
if len(self.noise_apply_strength) == 1:
self.noise_apply_strength = self.noise_apply_strength[0]
self.apply_spatial_patchify = args.apply_spatial_patchify
self.vae = vae
print(f'self.noise_apply_strength: {self.noise_apply_strength}')
def apply_noise_requant(self, bit_indices, quantized, args, device, si, lfq=None, noise_apply_strength=None):
if lfq is None:
lfq = self.vae.quantizer.lfq
if noise_apply_strength is None:
noise_apply_strength = self.noise_apply_strength
if isinstance(noise_apply_strength, list):
noise_apply_strength = np.random.randint(0, max(1, 100 * noise_apply_strength[si]+1)) * 0.01
else:
noise_apply_strength = np.random.randint(0, max(1, 100 * noise_apply_strength+1)) * 0.01
mask = torch.rand(*bit_indices.shape, device=device) < noise_apply_strength
pred_bit_indices = bit_indices.clone()
if args.num_of_label_value == 2:
pred_bit_indices[mask] = 1 - pred_bit_indices[mask]
else:
noise = torch.randint(0, args.num_of_label_value, bit_indices.shape, dtype=bit_indices.dtype, device=device)
pred_bit_indices[mask] = noise[mask]
if self.noise_apply_requant:
quantized = lfq.indices_to_codes(pred_bit_indices, label_type = 'bit_label')
return pred_bit_indices, quantized
def visualize(self, vae_scale_schedule, inp_B3HW, gt_all_bit_indices, pred_all_bit_indices):
gt_img = (inp_B3HW.squeeze(-3) + 1) / 2 * 255
gt_img = gt_img[0].permute(1,2,0).cpu().numpy().astype(np.uint8)[:,:,::-1]
recons_img_2 = labels2image(gt_all_bit_indices, label_type='bit_label', scale_schedule=vae_scale_schedule)
recons_img_3 = labels2image(pred_all_bit_indices, label_type='bit_label', scale_schedule=vae_scale_schedule)
cat_image = np.concatenate([gt_img, recons_img_2, recons_img_3], axis=1)
save_path = osp.abspath('non_teacher_force.jpg')
cv2.imwrite(save_path, cat_image)
print(f'Save to {save_path}')
import pdb; pdb.set_trace()
print(cat_image.shape)