Qnancy's picture
Upload folder using huggingface_hub
ad44ad4 verified
import os
import cv2
import numpy as np
import torch
from PIL import Image
from transformers import AutoProcessor, AutoModelForZeroShotObjectDetection
from sam2.sam2_image_predictor import SAM2ImagePredictor
from sam2.build_sam import build_sam2
def segment_single_image(image_path, text_prompt,
sam2_checkpoint="/mnt/prev_nas/qhy/MagicMotion/trajectory_construction/Grounded_SAM2/checkpoints/sam2_hiera_large.pt",
model_cfg="sam2_hiera_l.yaml"):
"""
对单张图片进行文本引导分割,返回二值掩码。
Args:
image_path (str): 输入图片路径
text_prompt (str): 文本提示(如 "car")
sam2_checkpoint (str): SAM2 模型路径
model_cfg (str): SAM2 配置文件名
Returns:
mask (np.ndarray): 二值掩码 (H, W), dtype=bool
success (bool): 是否成功检测到物体
"""
# === 1. 加载模型 ===
device = "cuda" if torch.cuda.is_available() else "cpu"
# Grounding DINO
processor = AutoProcessor.from_pretrained("/mnt/prev_nas/qhy/MagicMotion/trajectory_construction/Grounded_SAM2/checkpoints/grounding-dino-tiny")
grounding_model = AutoModelForZeroShotObjectDetection.from_pretrained(
"/mnt/prev_nas/qhy/MagicMotion/trajectory_construction/Grounded_SAM2/checkpoints/grounding-dino-tiny"
).to(device)
# SAM2 Image Predictor
sam2_model = build_sam2(model_cfg, sam2_checkpoint)
predictor = SAM2ImagePredictor(sam2_model)
# === 2. 读取图像 ===
image_pil = Image.open(image_path).convert("RGB")
image_np = np.array(image_pil)
# === 3. 文本预处理(Grounding DINO 要求小写 + 句号)===
text = text_prompt.strip().lower()
if not text.endswith("."):
text += "."
# === 4. Grounding DINO 检测边界框 ===
inputs = processor(images=image_pil, text=text, return_tensors="pt").to(device)
with torch.no_grad():
outputs = grounding_model(**inputs)
results = processor.post_process_grounded_object_detection(
outputs,
inputs.input_ids,
box_threshold=0.25,
text_threshold=0.3,
target_sizes=[image_pil.size[::-1]] # (H, W)
)
boxes = results[0]["boxes"].cpu().numpy()
if len(boxes) == 0:
print(f"❌ 未检测到 '{text_prompt}'")
return None, False
print(f"✅ 检测到 {len(boxes)} 个 '{text_prompt}'")
# === 5. SAM2 生成掩码 ===
predictor.set_image(image_np)
masks, _, _ = predictor.predict(
point_coords=None,
point_labels=None,
box=boxes, # 可一次处理多个框
multimask_output=False
)
# 合并所有检测到的物体(逻辑或)
if masks.ndim == 4:
masks = masks.squeeze(1) # (N, H, W)
final_mask = np.any(masks, axis=0) # (H, W), bool
return final_mask, True
def save_mask(mask, output_path):
"""保存二值掩码为 PNG(白色=前景,黑色=背景)"""
mask_uint8 = (mask * 255).astype(np.uint8)
cv2.imwrite(output_path, mask_uint8)
if __name__ == "__main__":
import argparse
parser = argparse.ArgumentParser(description="单图文本引导分割")
parser.add_argument("--image", required=True, help="输入图片路径")
parser.add_argument("--text", required=True, help="文本提示(如 'car')")
parser.add_argument("--output", default="mask.png", help="输出掩码路径")
args = parser.parse_args()
mask, success = segment_single_image(args.image, args.text)
if success:
save_mask(mask, args.output)
print(f"✅ 掩码已保存至: {args.output}")
else:
print("❌ 分割失败")