File size: 25,817 Bytes
26c7cba f4e5bc9 3b9b54d 26c7cba f4e5bc9 4742baa 26c7cba c6fefb4 f5fad1d f4e5bc9 26c7cba f4e5bc9 26c7cba c6e5d3f cb3b2ec 26c7cba f4e5bc9 3b9b54d f4e5bc9 c6e5d3f bd8d470 26c7cba bd8d470 26c7cba dbbb9cd 26c7cba c6e5d3f cb3b2ec c6e5d3f cb3b2ec 26c7cba ee1348a 26c7cba ee1348a cb3b2ec ee1348a 26c7cba ee1348a 26c7cba ee1348a 26c7cba ee1348a 26c7cba ee1348a 26c7cba bd8d470 26c7cba bd8d470 26c7cba b5648b7 26c7cba b5648b7 26c7cba e5c33cd 26c7cba b5648b7 26c7cba b5648b7 c6e5d3f cb3b2ec c6e5d3f cb3b2ec b5648b7 ee1348a b5648b7 ee1348a cb3b2ec ee1348a b5648b7 ee1348a b5648b7 c6fefb4 6d41c34 c6fefb4 b5648b7 26c7cba ee1348a 26c7cba b5648b7 ee1348a b5648b7 ee1348a f4e5bc9 b5648b7 f4e5bc9 4742baa f4e5bc9 4742baa f4e5bc9 4742baa f4e5bc9 4742baa f4e5bc9 4742baa f4e5bc9 4742baa f4e5bc9 4742baa f4e5bc9 4742baa f4e5bc9 4742baa f4e5bc9 1c68fe6 3b9b54d f4e5bc9 1c68fe6 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 | from fastapi import FastAPI, File, UploadFile, HTTPException, Body
from fastapi.responses import JSONResponse
from fastapi.middleware.cors import CORSMiddleware
from fastapi.staticfiles import StaticFiles
from pydantic import BaseModel
import cv2
import numpy as np
import tempfile
import os
from io import BytesIO
from PIL import Image
import uvicorn
import traceback
import json
from typing import List, Dict, Optional
import re
# Load .env file for local development.
# Search from this file's directory upward so it works whether the server
# is launched from project root (uvicorn backend.app:app) or from
# inside the backend/ folder (python app.py).
try:
from dotenv import load_dotenv
_here = os.path.dirname(os.path.abspath(__file__))
# Try backend/.env first, then project root .env
for _env_path in [
os.path.join(_here, ".env"),
os.path.join(_here, "..", ".env"),
]:
if os.path.isfile(_env_path):
load_dotenv(_env_path)
print(f"β
Loaded .env from: {os.path.abspath(_env_path)}")
break
else:
print("β οΈ No .env file found. Set GEMINI_API_KEY in your environment.")
except ImportError:
pass
try:
from .inference import infer_aw_contour, analyze_frame, analyze_video_frame, infer_cervix_bbox
except ImportError:
from inference import infer_aw_contour, analyze_frame, analyze_video_frame, infer_cervix_bbox
# Import Google Gemini (optional - graceful degradation if not installed)
try:
import google.generativeai as genai
GEMINI_AVAILABLE = True
except ImportError:
GEMINI_AVAILABLE = False
print("β οΈ google-generativeai not installed. LLM endpoints will be unavailable.")
app = FastAPI(title="Pathora Colposcopy API", version="1.0.0")
# Add CORS middleware to allow requests from frontend
app.add_middleware(
CORSMiddleware,
allow_origins=["*"],
allow_credentials=True,
allow_methods=["*"],
allow_headers=["*"],
)
# Initialize Gemini if available
GEMINI_API_KEY = os.getenv("GEMINI_API_KEY") or os.getenv("VITE_GEMINI_API_KEY")
if GEMINI_AVAILABLE and GEMINI_API_KEY:
try:
genai.configure(api_key=GEMINI_API_KEY)
print("β
Gemini AI configured successfully")
except Exception as e:
print(f"β οΈ Failed to configure Gemini: {e}")
GEMINI_AVAILABLE = False
elif GEMINI_AVAILABLE:
print("β οΈ GEMINI_API_KEY not found in environment variables")
def get_supported_gemini_models() -> List[str]:
"""Return model names that support generateContent for this API key."""
if not GEMINI_AVAILABLE or not GEMINI_API_KEY:
return []
discovered: List[str] = []
try:
for model in genai.list_models():
methods = getattr(model, "supported_generation_methods", []) or []
if "generateContent" not in methods:
continue
raw_name = getattr(model, "name", "")
if not raw_name:
continue
discovered.append(raw_name)
# Some SDK calls accept short names while discovery returns models/<name>.
if raw_name.startswith("models/"):
discovered.append(raw_name[len("models/"):])
except Exception as e:
print(f"β οΈ Could not list Gemini models: {e}")
return []
# De-duplicate while preserving order.
unique_models: List[str] = []
seen = set()
for name in discovered:
if name not in seen:
unique_models.append(name)
seen.add(name)
return unique_models
# Cache models that fail due to quota so we skip them on subsequent requests.
QUOTA_BLOCKED_MODELS: set[str] = set()
def get_ordered_model_candidates(available_models: List[str]) -> List[str]:
"""Order models by preference and exclude quota-blocked models."""
preferred_models = [
# Put models that are usually available on free keys first.
"models/gemini-2.5-flash",
"gemini-2.5-flash",
"models/gemini-flash-latest",
"gemini-flash-latest",
"models/gemini-2.5-flash-lite",
"gemini-2.5-flash-lite",
"models/gemini-flash-lite-latest",
"gemini-flash-lite-latest",
# Keep older families as fallback.
"models/gemini-2.0-flash",
"gemini-2.0-flash",
"models/gemini-2.0-flash-lite",
"gemini-2.0-flash-lite",
"models/gemini-1.5-flash",
"gemini-1.5-flash",
"models/gemini-1.5-pro",
"gemini-1.5-pro",
"models/gemini-pro-latest",
"gemini-pro-latest",
"models/gemini-pro",
"gemini-pro",
]
available = [m for m in available_models if m not in QUOTA_BLOCKED_MODELS]
ordered = [m for m in preferred_models if m in available]
ordered.extend(m for m in available if m not in ordered)
return ordered
# Pydantic models for LLM endpoints
class ChatMessage(BaseModel):
role: str
text: str
class ChatRequest(BaseModel):
message: str
history: List[ChatMessage] = []
system_prompt: Optional[str] = None
class ReportGenerationRequest(BaseModel):
patient_data: Dict
exam_findings: Dict
images: Optional[List[str]] = [] # base64 encoded images
system_prompt: Optional[str] = None
class SPAStaticFiles(StaticFiles):
async def get_response(self, path: str, scope):
response = await super().get_response(path, scope)
if response.status_code == 404:
return await super().get_response("index.html", scope)
return response
@app.get("/health")
async def health_check():
"""Health check endpoint"""
available_models = get_supported_gemini_models()
return {
"status": "healthy",
"service": "Pathora Colposcopy API",
"ai_models": {
"acetowhite_model": "loaded",
"cervix_model": "loaded"
},
"llm": {
"gemini_available": GEMINI_AVAILABLE,
"api_key_configured": bool(GEMINI_API_KEY),
"available_models": available_models
}
}
@app.get("/api/health")
async def api_health_check():
"""Health check endpoint under /api for HF Spaces compatibility."""
return await health_check()
@app.post("/api/chat")
async def chat_endpoint(request: ChatRequest):
"""
LLM Chat endpoint for conversational AI assistant
Args:
request: ChatRequest with message, history, and optional system_prompt
Returns:
JSON with AI response
"""
if not GEMINI_AVAILABLE:
raise HTTPException(
status_code=503,
detail="Gemini AI is not available. Install google-generativeai package."
)
if not GEMINI_API_KEY:
raise HTTPException(
status_code=503,
detail="GEMINI_API_KEY not configured in environment variables"
)
try:
# Use system prompt or default
system_prompt = request.system_prompt or """You are Pathora AI β a specialist colposcopy assistant. \
Provide expert guidance on examination techniques, findings interpretation, and management guidelines. \
Be professional, evidence-based, and concise."""
# Prefer modern fast models, then fall back to any model exposed by this key.
available_models = get_supported_gemini_models()
if not available_models:
raise Exception(
"No Gemini models with generateContent are available for this API key. "
"Check API key permissions and Gemini API enablement."
)
model_names = get_ordered_model_candidates(available_models)
print(f"β
Chat available models: {available_models}")
print(f"β
Chat candidate models: {model_names}")
response_text = None
used_model = None
for model_name in model_names:
try:
print(f"π Trying chat model: {model_name}")
# Initialize Gemini model
model = genai.GenerativeModel(
model_name=model_name,
system_instruction=system_prompt
)
# Build conversation history
chat_history = []
for msg in request.history:
role = "model" if msg.role == "bot" else "user"
chat_history.append({
"role": role,
"parts": [msg.text]
})
# Start chat with history
chat = model.start_chat(history=chat_history)
# Send message and get response
response = chat.send_message(request.message)
response_text = response.text
used_model = model_name
print(f"β
Successfully used chat model: {model_name}")
break
except Exception as model_err:
err_str = str(model_err)
if "429" in err_str or "quota exceeded" in err_str.lower():
QUOTA_BLOCKED_MODELS.add(model_name)
print(f"βοΈ Skipping quota-blocked chat model: {model_name}")
print(f"β οΈ Chat model {model_name} failed: {err_str}")
continue
if not response_text:
raise Exception("All model attempts failed. Please check API key and model availability.")
return JSONResponse({
"status": "success",
"response": response_text,
"model": used_model
})
except Exception as e:
error_msg = str(e)
print(f"β Chat error: {error_msg}")
traceback.print_exc()
# Provide more helpful error messages
if "API key" in error_msg or "authentication" in error_msg.lower():
detail = "API key authentication failed. Please add GEMINI_API_KEY to HF Space secrets."
elif "not found" in error_msg.lower() or "404" in error_msg:
detail = f"Gemini model not available. Error: {error_msg}. Please verify API key."
else:
detail = f"Chat error: {error_msg}"
raise HTTPException(status_code=500, detail=detail)
@app.post("/api/generate-report")
async def generate_report_endpoint(request: ReportGenerationRequest):
"""
Generate colposcopy report using LLM based on patient data and exam findings
Args:
request: ReportGenerationRequest with patient data, exam findings, and images
Returns:
JSON with generated report
"""
if not GEMINI_AVAILABLE:
raise HTTPException(
status_code=503,
detail="Gemini AI is not available. Install google-generativeai package."
)
if not GEMINI_API_KEY:
raise HTTPException(
status_code=503,
detail="GEMINI_API_KEY not configured in environment variables"
)
try:
# Use system prompt from frontend if provided, otherwise use a strict JSON-forcing default
system_prompt = request.system_prompt or """You are an expert colposcopy AI assistant acting as a specialist gynaecologist.
Analyse ALL the clinical data provided and return ONLY a valid JSON object β no markdown, no extra text, no code fences.
The JSON must have EXACTLY these 10 keys and no others:
{
"examQuality": "<Adequate or Inadequate>",
"transformationZone": "<I, II, or III>",
"acetowL": "<Present or Absent>",
"nativeFindings": "<2-3 sentence summary of native view findings>",
"aceticFindings": "<2-3 sentence summary of acetic acid findings>",
"biopsySites": "<recommended biopsy sites by clock position, or None>",
"biopsyNotes": "<brief biopsy notes: lesion grade, type, number of samples>",
"colposcopicFindings": "<professional colposcopic findings: 3-4 sentences including Swede score if available>",
"treatmentPlan": "<evidence-based treatment plan: 2-3 sentences>",
"followUp": "<follow-up schedule with specific timeframes>"
}"""
# Build a clean data prompt β just present the clinical data,
# the system_instruction above enforces the output format.
prompt_parts = []
prompt_parts.append("PATIENT DATA:")
prompt_parts.append(json.dumps(request.patient_data, indent=2))
prompt_parts.append("\n\nEXAMINATION FINDINGS & OBSERVATIONS:")
prompt_parts.append(json.dumps(request.exam_findings, indent=2))
prompt_parts.append("""
Based on all the above clinical data, return ONLY the JSON object with exactly these 10 keys:
examQuality, transformationZone, acetowL, nativeFindings, aceticFindings,
biopsySites, biopsyNotes, colposcopicFindings, treatmentPlan, followUp
Do NOT include any other keys. Do NOT wrap in markdown. Return raw JSON only.""")
full_prompt = "\n".join(prompt_parts)
# Prefer modern fast models, then fall back to any model exposed by this key.
available_models = get_supported_gemini_models()
if not available_models:
raise Exception(
"No Gemini models with generateContent are available for this API key. "
"Check API key permissions and Gemini API enablement."
)
model_names = get_ordered_model_candidates(available_models)
print(f"β
Report available models: {available_models}")
print(f"β
Report candidate models: {model_names}")
response_text = None
used_model = None
for model_name in model_names:
try:
print(f"π Trying model: {model_name}")
model = genai.GenerativeModel(
model_name=model_name,
system_instruction=system_prompt
)
response = model.generate_content(full_prompt)
response_text = response.text
used_model = model_name
print(f"β
Successfully used model: {model_name}")
break
except Exception as model_err:
err_str = str(model_err)
if "429" in err_str or "quota exceeded" in err_str.lower():
QUOTA_BLOCKED_MODELS.add(model_name)
print(f"βοΈ Skipping quota-blocked report model: {model_name}")
print(f"β οΈ Model {model_name} failed: {err_str}")
continue
if not response_text:
raise Exception("All model attempts failed. Please check API key and model availability.")
# Ensure response_text is valid JSON before returning
try:
# Strip markdown if present
cleaned_text = response_text.strip()
if cleaned_text.startswith('```'):
cleaned_text = re.sub(r'^```[a-z]*\n?', '', cleaned_text, flags=re.IGNORECASE)
cleaned_text = re.sub(r'\n?```\s*$', '', cleaned_text)
cleaned_text = cleaned_text.strip()
# Parse to verify it's valid JSON
parsed_json = json.loads(cleaned_text)
print(f"β
Report is valid JSON with keys: {list(parsed_json.keys())}")
# Return as JSON object (not string) so it's properly encoded by FastAPI
return JSONResponse({
"status": "success",
"report": cleaned_text, # Backward-compatible JSON string
"report_json": parsed_json, # Structured payload for robust frontend mapping
"model": used_model
})
except json.JSONDecodeError as je:
print(f"β οΈ Response is not valid JSON: {je}")
print(f"Response text: {response_text[:500]}")
raise Exception(f"Gemini returned invalid JSON: {str(je)}")
except Exception as e:
error_msg = str(e)
print(f"β Report generation error: {error_msg}")
traceback.print_exc()
if "API key" in error_msg or "authentication" in error_msg.lower():
detail = "API key authentication failed. Please check GEMINI_API_KEY in HF Space secrets."
elif "not found" in error_msg.lower() or "404" in error_msg:
detail = f"Gemini model not available. Error: {error_msg}. Please verify API key has access to Gemini models."
else:
detail = f"Report generation error: {error_msg}"
raise HTTPException(status_code=500, detail=detail)
@app.post("/api/infer-aw-contour")
async def infer_aw_contour_endpoint(file: UploadFile = File(...), conf_threshold: float = 0.4):
"""
Inference endpoint for Acetowhite contour detection
Args:
file: Image file (jpg, png, etc.)
conf_threshold: Confidence threshold for YOLO model (0.0-1.0)
Returns:
JSON with base64 encoded result image
"""
try:
# Read image file
image_data = await file.read()
print(f"β
File received, size: {len(image_data)} bytes")
# Try to open image - this will work regardless of content type
try:
image = Image.open(BytesIO(image_data))
print(f"β
Image opened, mode: {image.mode}, size: {image.size}")
except Exception as e:
print(f"β Image open error: {e}")
traceback.print_exc()
raise HTTPException(status_code=400, detail=f"Invalid image file: {str(e)}")
# Convert to numpy array and BGR format (OpenCV uses BGR)
# Handle different image modes
if image.mode == 'RGBA':
# Convert RGBA to RGB
image = image.convert('RGB')
elif image.mode != 'RGB':
# Convert other modes to RGB
image = image.convert('RGB')
frame = cv2.cvtColor(np.array(image), cv2.COLOR_RGB2BGR)
print(f"β
Frame converted, shape: {frame.shape}")
# Run inference - returns dict with 'overlay', 'contours', 'detections', etc.
print(f"π Running infer_aw_contour with conf_threshold={conf_threshold}")
result = infer_aw_contour(frame, conf_threshold=conf_threshold)
print(f"β
Inference complete, detections: {result['detections']}")
# Convert result overlay back to RGB for JSON serialization
if result["overlay"] is not None:
result_rgb = cv2.cvtColor(result["overlay"], cv2.COLOR_BGR2RGB)
result_image = Image.fromarray(result_rgb)
# Encode to base64
buffer = BytesIO()
result_image.save(buffer, format="PNG")
buffer.seek(0)
import base64
image_base64 = base64.b64encode(buffer.getvalue()).decode()
print(f"β
Image encoded to base64, size: {len(image_base64)} chars")
else:
image_base64 = None
print("β οΈ No overlay returned from inference")
return JSONResponse({
"status": "success",
"message": "Inference completed successfully",
"result_image": image_base64,
"contours": result["contours"],
"detections": result["detections"],
"confidence_threshold": conf_threshold
})
except Exception as e:
print(f"β EXCEPTION in infer_aw_contour:")
traceback.print_exc()
raise HTTPException(status_code=500, detail=f"Error during inference: {str(e)}")
@app.post("/api/batch-infer")
async def batch_infer(files: list[UploadFile] = File(...), conf_threshold: float = 0.4):
"""
Batch inference endpoint for multiple images
Args:
files: List of image files
conf_threshold: Confidence threshold for YOLO model
Returns:
JSON with results for all images
"""
results = []
for file in files:
try:
image_data = await file.read()
image = Image.open(BytesIO(image_data))
# Handle different image modes
if image.mode == 'RGBA':
image = image.convert('RGB')
elif image.mode != 'RGB':
image = image.convert('RGB')
frame = cv2.cvtColor(np.array(image), cv2.COLOR_RGB2BGR)
# Run inference - returns dict with 'overlay', 'contours', 'detections', etc.
result = infer_aw_contour(frame, conf_threshold=conf_threshold)
if result["overlay"] is not None:
result_rgb = cv2.cvtColor(result["overlay"], cv2.COLOR_BGR2RGB)
result_image = Image.fromarray(result_rgb)
buffer = BytesIO()
result_image.save(buffer, format="PNG")
buffer.seek(0)
import base64
image_base64 = base64.b64encode(buffer.getvalue()).decode()
else:
image_base64 = None
results.append({
"filename": file.filename,
"status": "success",
"result_image": image_base64,
"contours": result["contours"],
"detections": result["detections"]
})
except Exception as e:
results.append({
"filename": file.filename,
"status": "error",
"error": str(e)
})
return JSONResponse({
"status": "completed",
"total_files": len(results),
"results": results
})
@app.post("/infer/image")
async def infer_image(file: UploadFile = File(...)):
"""
Single image inference endpoint for cervix detection/quality.
"""
try:
contents = await file.read()
nparr = np.frombuffer(contents, np.uint8)
frame = cv2.imdecode(nparr, cv2.IMREAD_COLOR)
result = analyze_frame(frame)
return JSONResponse(content=result)
except Exception as e:
raise HTTPException(status_code=500, detail=str(e))
@app.post("/infer/video")
async def infer_video(file: UploadFile = File(...)):
"""
Video inference endpoint for cervix detection/quality (frame-by-frame).
"""
try:
with tempfile.NamedTemporaryFile(delete=False) as tmp:
tmp.write(await file.read())
temp_path = tmp.name
cap = cv2.VideoCapture(temp_path)
responses = []
frame_count = 0
while True:
ret, frame = cap.read()
if not ret:
break
result = analyze_video_frame(frame)
responses.append({
"frame": frame_count,
"status": result["status"],
"quality_percent": result["quality_percent"]
})
frame_count += 1
cap.release()
os.remove(temp_path)
return JSONResponse(content={
"total_frames": frame_count,
"results": responses
})
except Exception as e:
raise HTTPException(status_code=500, detail=str(e))
@app.post("/api/infer-cervix-bbox")
async def infer_cervix_bbox_endpoint(file: UploadFile = File(...), conf_threshold: float = 0.4):
"""
Cervix bounding box detection endpoint for annotation.
Detects cervix location and returns bounding boxes.
Args:
file: Image file (jpg, png, etc.)
conf_threshold: Confidence threshold for YOLO model (0.0-1.0)
Returns:
JSON with base64 encoded annotated image and bounding box coordinates
"""
try:
# Read image file
image_data = await file.read()
# Try to open image
try:
image = Image.open(BytesIO(image_data))
except Exception as e:
raise HTTPException(status_code=400, detail=f"Invalid image file: {str(e)}")
# Convert to numpy array and BGR format (OpenCV uses BGR)
if image.mode == 'RGBA':
image = image.convert('RGB')
elif image.mode != 'RGB':
image = image.convert('RGB')
frame = cv2.cvtColor(np.array(image), cv2.COLOR_RGB2BGR)
# Run inference
result = infer_cervix_bbox(frame, conf_threshold=conf_threshold)
# Convert result overlay back to RGB for JSON serialization
if result["overlay"] is not None:
result_rgb = cv2.cvtColor(result["overlay"], cv2.COLOR_BGR2RGB)
result_image = Image.fromarray(result_rgb)
# Encode to base64
buffer = BytesIO()
result_image.save(buffer, format="PNG")
buffer.seek(0)
import base64
image_base64 = base64.b64encode(buffer.getvalue()).decode()
else:
image_base64 = None
return JSONResponse({
"status": "success",
"message": "Cervix bounding box detection completed",
"result_image": image_base64,
"bounding_boxes": result["bounding_boxes"],
"detections": result["detections"],
"frame_width": result["frame_width"],
"frame_height": result["frame_height"],
"confidence_threshold": conf_threshold
})
except Exception as e:
raise HTTPException(status_code=500, detail=f"Error during cervix bbox inference: {str(e)}")
# Serve the built frontend if present (Space/Docker runtime)
frontend_dist = os.path.join(os.path.dirname(__file__), "..", "dist")
if os.path.isdir(frontend_dist):
app.mount("/", SPAStaticFiles(directory=frontend_dist, html=True), name="frontend")
if __name__ == "__main__":
uvicorn.run(app, host="0.0.0.0", port=8000) |