import gradio as gr import cv2 import numpy as np import matplotlib.pyplot as plt import json import math # === Helper: Rotated rectangle corners === def get_rotated_rect_corners(x, y, w, h, rotation_deg): rot_rad = np.deg2rad(rotation_deg) cos_r = np.cos(rot_rad) sin_r = np.sin(rot_rad) cx = x + w/2 cy = y + h/2 local_corners = np.array([ [-w/2, -h/2], [ w/2, -h/2], [ w/2, h/2], [-w/2, h/2] ]) R = np.array([[cos_r, -sin_r], [sin_r, cos_r]]) rotated_corners = np.dot(local_corners, R.T) corners = rotated_corners + np.array([cx, cy]) return corners.astype(np.float32) # === Preprocessing === def preprocess_gray_clahe(img): gray = cv2.cvtColor(img, cv2.COLOR_RGB2GRAY) clahe = cv2.createCLAHE(clipLimit=3.0, tileGridSize=(8, 8)) return clahe.apply(gray) # === Feature detectors === def get_detector(detector_name): if detector_name == "SIFT": return cv2.SIFT_create(nfeatures=5000) elif detector_name == "ORB": return cv2.ORB_create(5000) elif detector_name == "BRISK": return cv2.BRISK_create() elif detector_name == "AKAZE": return cv2.AKAZE_create() elif detector_name == "KAZE": return cv2.KAZE_create() else: return None def detect_and_match(img1_gray, img2_gray, detector_name, ratio_thresh=0.78): detector = get_detector(detector_name) kp1, des1 = detector.detectAndCompute(img1_gray, None) kp2, des2 = detector.detectAndCompute(img2_gray, None) if detector_name in ["SIFT", "KAZE"]: matcher = cv2.BFMatcher(cv2.NORM_L2) else: matcher = cv2.BFMatcher(cv2.NORM_HAMMING) matches = matcher.knnMatch(des1, des2, k=2) good = [] for m, n in matches: if m.distance < ratio_thresh * n.distance: good.append(m) return kp1, kp2, good # === Main processing === def process_images(flat_img, persp_img, json_file): if flat_img is None or persp_img is None or json_file is None: return [None]*6 # Load JSON try: data = json.load(open(json_file.name)) except Exception as e: print("JSON read error:", e) return [None]*6 roi = data["printAreas"][0] roi_x = roi["position"]["x"] roi_y = roi["position"]["y"] roi_w = roi["width"] roi_h = roi["height"] roi_rot_deg = roi["rotation"] # Preprocess images flat_gray = preprocess_gray_clahe(flat_img) persp_gray = preprocess_gray_clahe(persp_img) detectors = ["SIFT", "ORB", "BRISK", "AKAZE", "KAZE"] results = [] for det in detectors: kp1, kp2, good_matches = detect_and_match(flat_gray, persp_gray, det) if len(good_matches) < 4: print(f"Not enough matches for {det}") results.append(None) continue src_pts = np.float32([kp1[m.queryIdx].pt for m in good_matches]).reshape(-1,1,2) dst_pts = np.float32([kp2[m.trainIdx].pt for m in good_matches]).reshape(-1,1,2) H, mask = cv2.findHomography(src_pts, dst_pts, cv2.RANSAC, 5.0) # ROI corners roi_corners_flat = get_rotated_rect_corners(roi_x, roi_y, roi_w, roi_h, roi_rot_deg) roi_corners_persp = cv2.perspectiveTransform(roi_corners_flat.reshape(-1,1,2), H).reshape(-1,2) # Draw ROI flat_out = flat_img.copy() persp_out = persp_img.copy() cv2.polylines(flat_out, [roi_corners_flat.astype(int)], True, (255,0,0), 3) cv2.polylines(persp_out, [roi_corners_persp.astype(int)], True, (0,255,0), 3) results.append([flat_out, persp_out]) return results # List of [flat_out, persp_out] for each detector # === Gradio Interface === def wrap_gradio(flat_img, persp_img, json_file): outputs = process_images(flat_img, persp_img, json_file) # Flatten the outputs for Gallery display gallery_images = [] for item in outputs: if item is not None: gallery_images.extend([item[0], item[1]]) return gallery_images iface = gr.Interface( fn=wrap_gradio, inputs=[ gr.Image(type="numpy", label="Flat Image"), gr.Image(type="numpy", label="Perspective Image"), gr.File(type="filepath", label="JSON File") ], outputs=[ gr.Gallery(label="Results (Flat + Perspective per Detector)") ], title="Feature Detection with ROI Projection", description="Shows SIFT, ORB, BRISK, AKAZE, KAZE feature-based ROI projections. Each detector outputs Flat and Perspective images." ) iface.launch()