Ayesha352's picture
Update app.py
b4ecde4 verified
raw
history blame
4.57 kB
import gradio as gr
import cv2
import numpy as np
import matplotlib.pyplot as plt
import json
import math
# === Helper: Rotated rectangle corners ===
def get_rotated_rect_corners(x, y, w, h, rotation_deg):
rot_rad = np.deg2rad(rotation_deg)
cos_r = np.cos(rot_rad)
sin_r = np.sin(rot_rad)
cx = x + w/2
cy = y + h/2
local_corners = np.array([
[-w/2, -h/2],
[ w/2, -h/2],
[ w/2, h/2],
[-w/2, h/2]
])
R = np.array([[cos_r, -sin_r],
[sin_r, cos_r]])
rotated_corners = np.dot(local_corners, R.T)
corners = rotated_corners + np.array([cx, cy])
return corners.astype(np.float32)
# === Preprocessing ===
def preprocess_gray_clahe(img):
gray = cv2.cvtColor(img, cv2.COLOR_RGB2GRAY)
clahe = cv2.createCLAHE(clipLimit=3.0, tileGridSize=(8, 8))
return clahe.apply(gray)
# === Feature detectors ===
def get_detector(detector_name):
if detector_name == "SIFT":
return cv2.SIFT_create(nfeatures=5000)
elif detector_name == "ORB":
return cv2.ORB_create(5000)
elif detector_name == "BRISK":
return cv2.BRISK_create()
elif detector_name == "AKAZE":
return cv2.AKAZE_create()
elif detector_name == "KAZE":
return cv2.KAZE_create()
else:
return None
def detect_and_match(img1_gray, img2_gray, detector_name, ratio_thresh=0.78):
detector = get_detector(detector_name)
kp1, des1 = detector.detectAndCompute(img1_gray, None)
kp2, des2 = detector.detectAndCompute(img2_gray, None)
if detector_name in ["SIFT", "KAZE"]:
matcher = cv2.BFMatcher(cv2.NORM_L2)
else:
matcher = cv2.BFMatcher(cv2.NORM_HAMMING)
matches = matcher.knnMatch(des1, des2, k=2)
good = []
for m, n in matches:
if m.distance < ratio_thresh * n.distance:
good.append(m)
return kp1, kp2, good
# === Main processing ===
def process_images(flat_img, persp_img, json_file):
if flat_img is None or persp_img is None or json_file is None:
return [None]*6
# Load JSON
try:
data = json.load(open(json_file.name))
except Exception as e:
print("JSON read error:", e)
return [None]*6
roi = data["printAreas"][0]
roi_x = roi["position"]["x"]
roi_y = roi["position"]["y"]
roi_w = roi["width"]
roi_h = roi["height"]
roi_rot_deg = roi["rotation"]
# Preprocess images
flat_gray = preprocess_gray_clahe(flat_img)
persp_gray = preprocess_gray_clahe(persp_img)
detectors = ["SIFT", "ORB", "BRISK", "AKAZE", "KAZE"]
results = []
for det in detectors:
kp1, kp2, good_matches = detect_and_match(flat_gray, persp_gray, det)
if len(good_matches) < 4:
print(f"Not enough matches for {det}")
results.append(None)
continue
src_pts = np.float32([kp1[m.queryIdx].pt for m in good_matches]).reshape(-1,1,2)
dst_pts = np.float32([kp2[m.trainIdx].pt for m in good_matches]).reshape(-1,1,2)
H, mask = cv2.findHomography(src_pts, dst_pts, cv2.RANSAC, 5.0)
# ROI corners
roi_corners_flat = get_rotated_rect_corners(roi_x, roi_y, roi_w, roi_h, roi_rot_deg)
roi_corners_persp = cv2.perspectiveTransform(roi_corners_flat.reshape(-1,1,2), H).reshape(-1,2)
# Draw ROI
flat_out = flat_img.copy()
persp_out = persp_img.copy()
cv2.polylines(flat_out, [roi_corners_flat.astype(int)], True, (255,0,0), 3)
cv2.polylines(persp_out, [roi_corners_persp.astype(int)], True, (0,255,0), 3)
results.append([flat_out, persp_out])
return results # List of [flat_out, persp_out] for each detector
# === Gradio Interface ===
def wrap_gradio(flat_img, persp_img, json_file):
outputs = process_images(flat_img, persp_img, json_file)
# Flatten the outputs for Gallery display
gallery_images = []
for item in outputs:
if item is not None:
gallery_images.extend([item[0], item[1]])
return gallery_images
iface = gr.Interface(
fn=wrap_gradio,
inputs=[
gr.Image(type="numpy", label="Flat Image"),
gr.Image(type="numpy", label="Perspective Image"),
gr.File(type="filepath", label="JSON File")
],
outputs=[
gr.Gallery(label="Results (Flat + Perspective per Detector)")
],
title="Feature Detection with ROI Projection",
description="Shows SIFT, ORB, BRISK, AKAZE, KAZE feature-based ROI projections. Each detector outputs Flat and Perspective images."
)
iface.launch()