import os import torch import safetensors.torch import matplotlib import numpy as np import torch.nn.functional as F from torchvision import transforms as tt from omegaconf import OmegaConf from PIL import Image from transformers import Blip2Processor, Blip2ForConditionalGeneration from PrimeDepth_inference.ldm.util import instantiate_from_config def get_state_dict(d): return d.get('state_dict', d) def load_state_dict(ckpt_path, location='cpu'): _, extension = os.path.splitext(ckpt_path) if extension.lower() == ".safetensors": state_dict = safetensors.torch.load_file(ckpt_path, device=location) else: state_dict = get_state_dict( torch.load( ckpt_path, map_location=torch.device(location) ) ) state_dict = get_state_dict(state_dict) print(f'Loaded state_dict from [{ckpt_path}]') return state_dict def create_model(config_path, device="cpu"): config = OmegaConf.load(config_path) model = instantiate_from_config(config.model).to(device) print(f'Loaded model config from [{config_path}]') return model def get_image(path): image = Image.open(path) if not image.mode == "RGB": image = image.convert("RGB") return image class InferenceEngine: """ Utility class for obtaining PrimeDepth predictions """ def __init__(self, pd_config_path, blip2_cache_dir=None, cmap="Spectral", device="cuda"): """ pd_config_path : str Path to the model configuration file blip2_cache_dir : str, optional Path to the cache directory for the BLIP2 model, by default None cmap: str, optional Matplotlib colormap name, by default "Spectral" device : str, optional Device to run the model on, by default "cuda" """ self.pd = create_model(pd_config_path, device) self.processor, self.model = self.load_BLIP2(blip2_cache_dir, device) self.cm = matplotlib.colormaps[cmap] ''' def load_BLIP2(self, cache_dir=None, device="cuda"): processor = Blip2Processor.from_pretrained("Salesforce/blip2-opt-2.7b", cache_dir=cache_dir) model = Blip2ForConditionalGeneration.from_pretrained( "Salesforce/blip2-opt-2.7b", torch_dtype=torch.float16, cache_dir=cache_dir ) model = model.to(device) return processor, model ''' def load_BLIP2(self, cache_dir=None, device="cuda"): import os from transformers import AutoProcessor, AutoModelForCausalLM import torch # 使用 git-base 作为替代 model_name = "microsoft/git-base" try: print(f"Loading alternative model: {model_name}") processor = AutoProcessor.from_pretrained(model_name) model = AutoModelForCausalLM.from_pretrained(model_name) model = model.to(device) print("Model loaded successfully!") return processor, model except Exception as e: print(f"Error loading model: {e}") raise def captionize(self, image): # 1. 首先转换输入 inputs = self.processor(images=image, return_tensors="pt") # 2. 然后将所有张量移到GPU并转换为float16 inputs = {k: v.to(device='cuda', dtype=torch.float16) if torch.is_tensor(v) else v for k, v in inputs.items()} # 3. 生成caption generated_ids = self.model.generate( **inputs, max_length=30, num_beams=4, num_return_sequences=1 ) # 4. 解码生成的文本 caption = self.processor.batch_decode(generated_ids, skip_special_tokens=True)[0] return caption ''' def captionize(self, image): inputs = self.processor(images=image, return_tensors="pt").to('cuda', torch.float16) generated_ids = self.model.generate(**inputs) caption = self.processor.batch_decode(generated_ids, skip_special_tokens=True)[0].strip() return caption ''' def predict(self, image_path, max_size=1024): """ Parameters ---------- image_path : str Path to the image max_size : int, optional Maximal processing size of the longer image edge, by default 1024 Returns ------- depth_ssi : np.ndarray Scale and shift invariant depth map prediction depth_color : PIL.Image Colorized depth map prediction """ image = get_image(image_path) caption = self.captionize(image) w, h = image.size if max_size is not None and max(h, w) > max_size: if h == w: image = tt.Resize(max_size)(image) else: image = tt.Resize(max_size-1, max_size=max_size)(image) with torch.no_grad(): labels = self.pd.get_label_from_image(image, prompt=caption) depth_ssi = labels['depth'].clone().mean(dim=1) if max_size is not None and max(h, w) > max_size: depth_ssi = F.interpolate(depth_ssi[None], (h, w), mode='bilinear', align_corners=False)[0, 0] else: depth_ssi = depth_ssi[0] depth_ssi = depth_ssi.cpu().numpy().astype(np.float32) depth_color = depth_ssi.copy() depth_color = (depth_color - depth_color.min()) / (depth_color.max() - depth_color.min()) depth_color = self.cm(depth_color)[:, :, :3] depth_color = (depth_color * 255).astype(np.uint8) depth_color = Image.fromarray(depth_color) return depth_ssi, depth_color