cardio-deploy
Deploy CardioScan inference 2026-04-24T11:06:40Z
560f62e
"""FastAPI inference server for the Cardiomegaly classifier.
Loads the multi-seed ensemble trained in ``model_training/`` and exposes a
single ``POST /predict`` endpoint that the frontend (`src/services/predict.ts`)
already knows how to consume.
Nothing inside ``model_training/`` is modified — we only *import* the model
factory (``src.model.build_model``) to rebuild the exact architecture that was
saved to disk, then load the weights on top.
Run locally
-----------
cd inference_server
pip install -r requirements.txt
uvicorn server:app --host 0.0.0.0 --port 8000
Environment overrides (optional)
--------------------------------
MODEL_BACKBONE default: CFG.backbone (e.g. "efficientnet_b0")
MODEL_IMG_SIZE default: CFG.img_size (e.g. 224)
MODEL_THRESHOLD default: 0.5 (binary cut-off for the label)
MODEL_USE_TTA default: "false" ("true" → 6-pass TTA per image)
ALLOWED_ORIGINS comma-separated CORS origins (exact match)
ALLOWED_ORIGIN_REGEX regex origin whitelist (e.g. Lovable preview URLs:
"https://.*\\.lovable\\.app")
LOG_LEVEL default: INFO
"""
from __future__ import annotations
import io
import json
import logging
import os
import sys
from pathlib import Path
from typing import List, Literal
import numpy as np
import pandas as pd
import torch
import torch.nn as nn
import torchvision.transforms as T
from fastapi import FastAPI, File, HTTPException, Query, UploadFile
from fastapi.middleware.cors import CORSMiddleware
from huggingface_hub import hf_hub_download
from PIL import Image
# ---------------------------------------------------------------------------
# Paths — make `from src.model import ...` resolvable without touching
# `model_training/`. We prepend the training directory to sys.path so its
# internal `from src.config import CFG` style imports keep working.
# ---------------------------------------------------------------------------
REPO_ROOT = Path(__file__).resolve().parent.parent
TRAINING_DIR = REPO_ROOT / "model_training"
NOTEBOOKS_DIR = TRAINING_DIR / "notebooks"
RESULTS_DIR = NOTEBOOKS_DIR / "results"
HF_MODEL_REPO_ID = os.environ.get("HF_MODEL_REPO_ID", "").strip()
HF_MODEL_REVISION = os.environ.get("HF_MODEL_REVISION", "main")
HF_HUB_TOKEN = os.environ.get("HF_TOKEN") or os.environ.get("HUGGINGFACE_HUB_TOKEN")
HF_MODEL_CACHE_DIR = os.environ.get("HF_MODEL_CACHE_DIR", str(REPO_ROOT / ".hf-model-cache"))
ACCURATE_MANIFEST = os.environ.get("ACCURATE_MANIFEST_FILE", "ensemble_manifest.csv")
FAST_MANIFEST = os.environ.get("FAST_MANIFEST_FILE", "fast/fast_ensemble_manifest.csv")
ACCURATE_METRICS = os.environ.get("ACCURATE_METRICS_FILE", "val_metrics_final.json")
FAST_METRICS = os.environ.get("FAST_METRICS_FILE", "fast/fast_val_metrics_final.json")
if str(TRAINING_DIR) not in sys.path:
sys.path.insert(0, str(TRAINING_DIR))
# Point torch's hub cache to a writable in-project location so the server
# works in sandboxed environments where ``~/.cache`` is read-only. Setting
# this BEFORE importing torchvision is critical.
os.environ.setdefault("TORCH_HOME", str(REPO_ROOT / ".torch-cache"))
# `build_model` in ``model_training/src/model.py`` constructs torchvision or
# torchxrayvision backbones WITH their pretrained weights. Those weights are
# irrelevant at inference time because we immediately overwrite them with the
# trained checkpoint from ``model_training/notebooks/results/``. We monkey-
# patch the constructors so the server skips every pretrained-weight
# download. This avoids needless bandwidth AND cache-dir permission errors
# when running in sandboxed environments.
import torchvision.models as _tvm # noqa: E402 pylint: disable=wrong-import-position
import torchxrayvision as _xrv # noqa: E402 pylint: disable=wrong-import-position
for _fn_name in ("efficientnet_b0", "efficientnet_b3", "mobilenet_v3_large"):
_orig = getattr(_tvm, _fn_name, None)
if _orig is None:
continue
def _no_download_builder(*args, __orig=_orig, **kwargs):
kwargs["weights"] = None
return __orig(*args, **kwargs)
setattr(_tvm, _fn_name, _no_download_builder)
# torchxrayvision DenseNet also attempts a download when weights="..." is set.
# We wrap its __init__ so the caller's weights argument is remembered, but
# the actual download is skipped. We still restore the canonical label list
# (``self.pathologies`` / ``self.targets``) that downstream code in
# ``model_training/src/model.py::cardio_logit`` relies on to locate the
# Cardiomegaly output index.
_orig_xrv_densenet_init = _xrv.models.DenseNet.__init__
def _xrv_densenet_init_no_download(self, *args, **kwargs):
requested_weights = kwargs.get("weights")
kwargs["weights"] = None
_orig_xrv_densenet_init(self, *args, **kwargs)
if requested_weights and requested_weights in _xrv.models.model_urls:
labels = _xrv.models.model_urls[requested_weights]["labels"]
self.targets = labels
self.pathologies = labels
_xrv.models.DenseNet.__init__ = _xrv_densenet_init_no_download
from src.config import CFG # noqa: E402 pylint: disable=wrong-import-position
from src.model import build_model, cardio_logit # noqa: E402 pylint: disable=wrong-import-position
from src.dataset import get_normalize_fn # noqa: E402 pylint: disable=wrong-import-position
def _detect_backbone_from_checkpoint(ckpt_path: Path) -> str:
"""Inspect a saved state_dict and guess which backbone produced it.
Rules:
* torchxrayvision DenseNet-121 → has ``features.denseblockN.*`` keys
* torchvision EfficientNet → top-level ``features.0.0.weight`` (stem conv)
and depth ≥ 9 feature groups
* torchvision MobileNetV3-Large → ``features.0.0.weight`` with depth ~17
* microsoft/rad-dino → keys under ``features.embeddings`` /
``features.encoder.layer.``
Defaults to ``CFG.backbone`` if no signature matches.
"""
state = torch.load(ckpt_path, map_location="cpu", weights_only=True)
if isinstance(state, dict) and "state_dict" in state:
state = state["state_dict"]
keys = list(state.keys())
if any("denseblock" in k for k in keys):
return "densenet121"
if any(k.startswith("features.embeddings.") for k in keys) or any(
k.startswith("features.encoder.layer.") for k in keys
):
return "rad-dino"
# torchvision feature indices
feature_indices = {
int(k.split(".")[1])
for k in keys
if k.startswith("features.") and k.split(".")[1].isdigit()
}
if feature_indices:
# EfficientNet-B0 has 9 groups (features.0 … features.8)
# MobileNetV3-Large has 17 groups (features.0 … features.16)
if max(feature_indices) >= 12:
return "mobilenet_v3_large"
if max(feature_indices) >= 7:
return "efficientnet_b0"
return CFG.backbone
def _hf_download(filename: str) -> Path:
"""Download a file from HF model repo and return local cached path."""
if not HF_MODEL_REPO_ID:
raise FileNotFoundError(
f"File {filename!r} not found locally and HF_MODEL_REPO_ID is not set."
)
path = hf_hub_download(
repo_id=HF_MODEL_REPO_ID,
filename=filename,
revision=HF_MODEL_REVISION,
token=HF_HUB_TOKEN,
cache_dir=HF_MODEL_CACHE_DIR,
)
return Path(path)
def _manifest_candidates(mode: Literal["accurate", "fast"]) -> tuple[list[Path], list[str]]:
if mode == "fast":
local = [RESULTS_DIR / "fast_model" / "fast_ensemble_manifest.csv", RESULTS_DIR / "fast_ensemble_manifest.csv"]
remote = [FAST_MANIFEST, "fast_ensemble_manifest.csv"]
else:
local = [RESULTS_DIR / "ensemble_manifest.csv"]
remote = [ACCURATE_MANIFEST, "ensemble_manifest.csv"]
return local, remote
def _metrics_candidates(mode: Literal["accurate", "fast"]) -> tuple[list[Path], list[str]]:
if mode == "fast":
local = [RESULTS_DIR / "fast_model" / "fast_val_metrics_final.json", RESULTS_DIR / "fast_val_metrics_final.json"]
remote = [FAST_METRICS, "fast_val_metrics_final.json"]
else:
local = [RESULTS_DIR / "val_metrics_final.json"]
remote = [ACCURATE_METRICS, "val_metrics_final.json"]
return local, remote
def _resolve_manifest_path(mode: Literal["accurate", "fast"] = "accurate") -> Path:
"""Find mode-specific manifest locally first, else download from HF."""
local_candidates, remote_candidates = _manifest_candidates(mode)
for local in local_candidates:
if local.exists():
return local
log.info(
"Local %s manifest not found under %s; downloading from HF repo %s",
mode,
RESULTS_DIR,
HF_MODEL_REPO_ID or "<unset>",
)
for filename in remote_candidates:
try:
return _hf_download(filename)
except Exception: # noqa: BLE001
continue
raise FileNotFoundError(f"Could not resolve {mode} manifest from local paths or HF")
def _resolve_optional_support_file(
name: str,
mode: Literal["accurate", "fast"] = "accurate",
) -> Path | None:
"""Find optional support file locally; if missing try HF model repo."""
if name == "val_metrics_final.json":
local_candidates, remote_candidates = _metrics_candidates(mode)
else:
local_candidates = [RESULTS_DIR / name]
remote_candidates = [name]
for local in local_candidates:
if local.exists():
return local
for remote in remote_candidates:
try:
return _hf_download(remote)
except Exception: # noqa: BLE001
continue
return None
# ---------------------------------------------------------------------------
# Backbone + image size: auto-detected from the checkpoint so the server never
# runs with a mismatched architecture. Can still be forced via env vars.
# ---------------------------------------------------------------------------
def _first_checkpoint_path(mode: Literal["accurate", "fast"] = "accurate") -> Path:
try:
manifest = _resolve_manifest_path(mode)
df = pd.read_csv(manifest)
first = df["checkpoint"].iloc[0]
p = Path(first)
if p.is_absolute() and p.exists():
return p
# Local resolution first.
for candidate in (
NOTEBOOKS_DIR / first,
RESULTS_DIR / Path(first).name,
RESULTS_DIR / "fast_model" / Path(first).name,
):
if candidate.exists():
return candidate
# `_first_checkpoint_path` is executed during module import (before
# `_resolve_checkpoint` is defined), so we do HF download inline here.
for name in (first, Path(first).name):
try:
return _hf_download(name)
except Exception: # noqa: BLE001
continue
raise FileNotFoundError(f"Could not resolve first checkpoint from manifest entry: {first!r}")
except Exception: # noqa: BLE001
pass
fallback = RESULTS_DIR / ("fast_best_model.pth" if mode == "fast" else "best_model.pth")
if fallback.exists():
return fallback
try:
return _hf_download("fast_best_model.pth" if mode == "fast" else "best_model.pth")
except Exception as exc: # noqa: BLE001
raise FileNotFoundError(
"No checkpoints found locally and could not download from HF. "
"Set HF_MODEL_REPO_ID and upload ensemble_manifest.csv + *.pth."
) from exc
_DETECTED_BACKBONE = _detect_backbone_from_checkpoint(_first_checkpoint_path("accurate"))
try:
_DETECTED_FAST_BACKBONE = _detect_backbone_from_checkpoint(_first_checkpoint_path("fast"))
except Exception as exc: # noqa: BLE001
# Do not crash the app during import if fast artefacts are temporarily unavailable.
# Fast mode will attempt loading again at startup and can still fallback gracefully.
_DETECTED_FAST_BACKBONE = _DETECTED_BACKBONE
log = logging.getLogger("inference")
log.warning(
"Fast backbone auto-detect failed at import (%s). Falling back to accurate backbone %s.",
exc,
_DETECTED_BACKBONE,
)
# DenseNet-121 (torchxrayvision) is trained on 224x224; ViT-B/14 needs 518.
_DEFAULT_IMG_SIZE = 518 if _DETECTED_BACKBONE == "rad-dino" else 224
_DEFAULT_FAST_IMG_SIZE = 518 if _DETECTED_FAST_BACKBONE == "rad-dino" else 224
BACKBONE: str = os.environ.get("MODEL_BACKBONE", _DETECTED_BACKBONE)
IMG_SIZE: int = int(os.environ.get("MODEL_IMG_SIZE", str(_DEFAULT_IMG_SIZE)))
FAST_BACKBONE: str = os.environ.get("MODEL_FAST_BACKBONE", _DETECTED_FAST_BACKBONE)
FAST_IMG_SIZE: int = int(os.environ.get("MODEL_FAST_IMG_SIZE", str(_DEFAULT_FAST_IMG_SIZE)))
USE_TTA: bool = os.environ.get("MODEL_USE_TTA", "true").lower() in {"1", "true", "yes"}
def _default_threshold() -> float:
"""Use the training-selected threshold when available."""
metrics_path = _resolve_optional_support_file("val_metrics_final.json", mode="accurate")
if metrics_path is not None:
try:
with open(metrics_path, "r", encoding="utf-8") as f:
data = json.load(f)
thr = float(data.get("threshold", 0.5))
if 0.0 <= thr <= 1.0:
return thr
except Exception: # noqa: BLE001
pass
return 0.5
DECISION_THRESHOLD: float = float(os.environ.get("MODEL_THRESHOLD", str(_default_threshold())))
def _default_fast_threshold() -> float:
metrics_path = _resolve_optional_support_file("val_metrics_final.json", mode="fast")
if metrics_path is not None:
try:
with open(metrics_path, "r", encoding="utf-8") as f:
data = json.load(f)
thr = float(data.get("threshold", DECISION_THRESHOLD))
if 0.0 <= thr <= 1.0:
return thr
except Exception: # noqa: BLE001
pass
return DECISION_THRESHOLD
FAST_DECISION_THRESHOLD: float = float(
os.environ.get("MODEL_FAST_THRESHOLD", str(_default_fast_threshold()))
)
_DEFAULT_ORIGINS = (
"http://localhost:3000,"
"http://localhost:5173,"
"http://localhost:8080,"
"http://127.0.0.1:3000,"
"http://127.0.0.1:5173,"
"http://127.0.0.1:8080"
)
ALLOWED_ORIGINS: list[str] = [
o.strip()
for o in os.environ.get("ALLOWED_ORIGINS", _DEFAULT_ORIGINS).split(",")
if o.strip()
]
# Optional regex list — useful when the production frontend is served from a
# hash-based preview URL (e.g. Lovable / Vercel preview deployments).
# By default we allow:
# * any *.lovable.app and *.lovableproject.com subdomain (deployed Lovable apps)
# * any *.ngrok-free.app / *.ngrok.app / *.ngrok.io subdomain (when the user
# forwards the dev server through ngrok and previews the app from anywhere)
# Override with `ALLOWED_ORIGIN_REGEX` to lock things down in production.
# Include common private LAN dev URLs (Vite "Network" URL is often
# `http://192.168.x.x:8080` — the Origin header is not localhost, so
# it must be accepted here or the browser will block with "Network Error").
_DEFAULT_ORIGIN_REGEX = (
r"https://([a-z0-9-]+\.)*lovable\.app"
r"|https://([a-z0-9-]+\.)*lovableproject\.com"
r"|https://([a-z0-9-]+\.)*ngrok-free\.app"
r"|https://([a-z0-9-]+\.)*ngrok\.app"
r"|https://([a-z0-9-]+\.)*ngrok\.io"
r"|http://(192\.168\.\d{1,3}\.\d{1,3}|10\.\d{1,3}\.\d{1,3}\.\d{1,3}):\d+"
)
_ORIGIN_REGEX: str | None = os.environ.get("ALLOWED_ORIGIN_REGEX", _DEFAULT_ORIGIN_REGEX) or None
DEVICE: torch.device = torch.device(
"cuda" if torch.cuda.is_available()
else "mps" if torch.backends.mps.is_available()
else "cpu"
)
POSITIVE_LABEL = "Cardiomegaly"
NEGATIVE_LABEL = "No Cardiomegaly indication"
# ---------------------------------------------------------------------------
# Logging
# ---------------------------------------------------------------------------
logging.basicConfig(
level=os.environ.get("LOG_LEVEL", "INFO"),
format="%(asctime)s %(levelname)-5s %(message)s",
)
log = logging.getLogger("inference")
# ---------------------------------------------------------------------------
# Preprocessing — delegate to the SAME normalization functions the training
# dataset uses (`xrv_normalize_np` for densenet121, `imagenet_normalize_np`
# for every other backbone). This guarantees byte-for-byte identical
# preprocessing between training and inference.
# ---------------------------------------------------------------------------
_normalize_fn = get_normalize_fn(BACKBONE)
_fast_normalize_fn = get_normalize_fn(FAST_BACKBONE)
def _pil_hflip(img: Image.Image) -> Image.Image:
return img.transpose(Image.FLIP_LEFT_RIGHT)
def _tta_pipelines(size: int) -> List[T.Compose]:
"""Match `src.transforms.make_tta_transforms` (6 deterministic passes)."""
s = (size, size)
return [
T.Compose([T.Resize(s)]),
T.Compose([T.Resize(s), T.Lambda(_pil_hflip)]),
T.Compose([T.Resize((size + 20, size + 20)), T.CenterCrop(s)]),
T.Compose([T.Resize((size - 20, size - 20)), T.Pad(10, fill=0), T.CenterCrop(s)]),
T.Compose([T.Resize(s), T.RandomAffine(degrees=(6, 6), fill=0)]),
T.Compose([T.Resize(s), T.RandomAffine(degrees=(-6, -6), fill=0)]),
]
def _single_eval_pipeline(size: int) -> T.Compose:
return T.Compose([T.Resize((size, size))])
# ---------------------------------------------------------------------------
# Ensemble loading
# ---------------------------------------------------------------------------
def _resolve_checkpoint(p: str, mode: Literal["accurate", "fast"] = "accurate") -> Path:
"""Resolve checkpoint locally first, else download from HF model repo."""
path = Path(p)
if path.is_absolute() and path.exists():
return path
for candidate in (
NOTEBOOKS_DIR / p,
RESULTS_DIR / Path(p).name,
RESULTS_DIR / "fast_model" / Path(p).name,
):
if candidate.exists():
return candidate
# In model repos we usually store files flat, so try both raw entry and basename.
tried = [p]
if Path(p).name != p:
tried.append(Path(p).name)
if mode == "fast":
tried.extend([f"fast/{Path(p).name}", f"fast_model/{Path(p).name}"])
for name in tried:
try:
downloaded = _hf_download(name)
log.info(" → downloaded %s from HF repo %s", name, HF_MODEL_REPO_ID)
return downloaded
except Exception: # noqa: BLE001
continue
raise FileNotFoundError(
f"Checkpoint not found locally and not downloadable from HF repo: {p!r}"
)
def _load_ensemble(mode: Literal["accurate", "fast"] = "accurate") -> tuple[List[nn.Module], list[str]]:
# Align CFG so build_model() reads the right backbone/size internally.
mode_backbone = FAST_BACKBONE if mode == "fast" else BACKBONE
mode_img_size = FAST_IMG_SIZE if mode == "fast" else IMG_SIZE
mode_norm_name = _fast_normalize_fn.__name__ if mode == "fast" else _normalize_fn.__name__
CFG.backbone = mode_backbone
CFG.img_size = mode_img_size
try:
manifest = _resolve_manifest_path(mode)
df = pd.read_csv(manifest)
checkpoint_paths = [_resolve_checkpoint(p, mode=mode) for p in df["checkpoint"].tolist()]
log.info(
"Loading %s ensemble of %d models from %s",
mode,
len(checkpoint_paths),
manifest,
)
except Exception:
fallback_name = "fast_best_model.pth" if mode == "fast" else "best_model.pth"
best = RESULTS_DIR / fallback_name
if best.exists():
checkpoint_paths = [best]
log.info("No %s manifest found, falling back to local checkpoint: %s", mode, best.name)
else:
checkpoint_paths = [_resolve_checkpoint(fallback_name, mode=mode)]
log.info("No %s manifest found, falling back to HF checkpoint: %s", mode, fallback_name)
models: list[nn.Module] = []
for ckpt_path in checkpoint_paths:
log.info(" → loading %s (%s)", ckpt_path.name, ckpt_path.resolve())
model = build_model(mode_backbone)
state = torch.load(ckpt_path, map_location=DEVICE)
if isinstance(state, dict) and "state_dict" in state:
state = state["state_dict"]
missing, unexpected = model.load_state_dict(state, strict=False)
if missing or unexpected:
raise RuntimeError(
"Checkpoint architecture mismatch. "
f"backbone={mode_backbone!r}, checkpoint={ckpt_path.name!r}, "
f"missing_keys={len(missing)}, unexpected_keys={len(unexpected)}. "
"Use the correct mode-specific backbone / img_size and ensure "
"ensemble_manifest.csv points to checkpoints from that training run."
)
model.to(DEVICE).eval()
models.append(model)
loaded_checkpoints = [p.name for p in checkpoint_paths]
mode_thr = FAST_DECISION_THRESHOLD if mode == "fast" else DECISION_THRESHOLD
log.info(
"%s ensemble ready — %d model(s) · device=%s · backbone=%s · "
"normalize=%s · img_size=%d · tta=%s · threshold=%.4f",
mode, len(models), DEVICE, mode_backbone,
mode_norm_name, mode_img_size, USE_TTA, mode_thr,
)
return models, loaded_checkpoints
# ---------------------------------------------------------------------------
# FastAPI app
# ---------------------------------------------------------------------------
app = FastAPI(title="CardioScan inference", version="1.0")
app.add_middleware(
CORSMiddleware,
allow_origins=ALLOWED_ORIGINS,
allow_origin_regex=_ORIGIN_REGEX,
allow_credentials=True,
allow_methods=["*"],
allow_headers=["*"],
)
_ensemble: list[nn.Module] = []
_loaded_checkpoints: list[str] = []
_fast_ensemble: list[nn.Module] = []
_fast_loaded_checkpoints: list[str] = []
@app.on_event("startup")
def _startup() -> None:
global _ensemble, _loaded_checkpoints, _fast_ensemble, _fast_loaded_checkpoints
_ensemble, _loaded_checkpoints = _load_ensemble("accurate")
try:
_fast_ensemble, _fast_loaded_checkpoints = _load_ensemble("fast")
except Exception: # noqa: BLE001
log.warning("Fast ensemble unavailable; using accurate ensemble for fast mode fallback.")
_fast_ensemble, _fast_loaded_checkpoints = _ensemble, _loaded_checkpoints
@app.get("/health")
def health() -> dict:
return {
"ok": bool(_ensemble),
"models": len(_ensemble),
"checkpoints": _loaded_checkpoints,
"backbone": BACKBONE,
"detected_backbone": _DETECTED_BACKBONE,
"normalization": _normalize_fn.__name__,
"img_size": IMG_SIZE,
"device": str(DEVICE),
"use_tta": USE_TTA,
"threshold": DECISION_THRESHOLD,
"fast_backbone": FAST_BACKBONE,
"fast_detected_backbone": _DETECTED_FAST_BACKBONE,
"fast_normalization": _fast_normalize_fn.__name__,
"fast_img_size": FAST_IMG_SIZE,
"fast_models": len(_fast_ensemble),
"fast_checkpoints": _fast_loaded_checkpoints,
"fast_threshold": FAST_DECISION_THRESHOLD,
}
@torch.no_grad()
def _predict_probability_detailed(
pil_gray: Image.Image,
use_tta: bool,
ensemble: list[nn.Module],
checkpoints: list[str],
img_size: int,
normalize_fn,
max_models: int | None = None,
) -> dict:
"""Run ensemble (+ optional TTA) on a single PIL image.
Returns a dict with per-model / per-TTA logits for transparency.
Matches `tta_predict` / `tta_predict_ensemble` in ``src.train`` exactly:
average logits across TTA (per model), then average across models,
then sigmoid.
"""
pipelines = _tta_pipelines(img_size) if use_tta else [_single_eval_pipeline(img_size)]
tensors = [normalize_fn(pipeline(pil_gray)) for pipeline in pipelines]
batch = torch.stack(tensors, dim=0).to(DEVICE) # (num_tta, 3, H, W)
active_model_count = len(ensemble) if max_models is None else max(1, min(max_models, len(ensemble)))
active_models = ensemble[:active_model_count]
active_checkpoints = checkpoints[:active_model_count]
per_model_tta_logits: list[np.ndarray] = []
per_model_mean_logit: list[float] = []
for model in active_models:
logit_vec = cardio_logit(model, batch).float().cpu().numpy() # (num_tta,)
per_model_tta_logits.append(logit_vec)
per_model_mean_logit.append(float(np.mean(logit_vec)))
ensemble_mean_logit = float(np.mean(per_model_mean_logit))
probability = float(1.0 / (1.0 + np.exp(-ensemble_mean_logit)))
return {
"probability": probability,
"ensemble_mean_logit": ensemble_mean_logit,
"per_model_mean_logit": {
name: lg for name, lg in zip(active_checkpoints, per_model_mean_logit)
},
"per_model_tta_logits": {
name: lg.tolist() for name, lg in zip(active_checkpoints, per_model_tta_logits)
},
"num_tta_passes": batch.shape[0],
"models_used": active_model_count,
"checkpoints_used": active_checkpoints,
}
@app.post("/predict")
async def predict(
image: UploadFile = File(...),
mode: Literal["accurate", "fast"] = Query(default="accurate", description="Inference mode"),
use_tta: bool | None = Query(default=None, description="Override TTA for this request."),
max_models: int | None = Query(default=None, ge=1, description="Use only first N models for speed."),
) -> dict:
if not _ensemble:
raise HTTPException(status_code=503, detail="Model not ready")
raw = await image.read()
if not raw:
raise HTTPException(status_code=400, detail="Empty upload")
try:
pil = Image.open(io.BytesIO(raw)).convert("L")
except Exception as exc: # noqa: BLE001
raise HTTPException(status_code=400, detail=f"Could not decode image: {exc}") from exc
selected_ensemble = _fast_ensemble if mode == "fast" else _ensemble
selected_checkpoints = _fast_loaded_checkpoints if mode == "fast" else _loaded_checkpoints
selected_img_size = FAST_IMG_SIZE if mode == "fast" else IMG_SIZE
selected_normalize_fn = _fast_normalize_fn if mode == "fast" else _normalize_fn
if not selected_ensemble:
raise HTTPException(status_code=503, detail=f"{mode} model not ready")
effective_use_tta = (False if mode == "fast" else USE_TTA) if use_tta is None else use_tta
try:
details = _predict_probability_detailed(
pil,
use_tta=effective_use_tta,
ensemble=selected_ensemble,
checkpoints=selected_checkpoints,
img_size=selected_img_size,
normalize_fn=selected_normalize_fn,
max_models=max_models,
)
except Exception as exc: # noqa: BLE001
log.exception("Inference failed")
raise HTTPException(status_code=500, detail=f"Inference error: {exc}") from exc
probability = details["probability"]
active_threshold = FAST_DECISION_THRESHOLD if mode == "fast" else DECISION_THRESHOLD
is_positive = probability >= active_threshold
log.info(
"/predict file=%s size=%d prob=%.4f thr=%.4f -> %s (per-model=%s, tta=%d)",
image.filename,
len(raw),
probability,
active_threshold,
"Cardiomegaly" if is_positive else "Negative",
{k: round(v, 4) for k, v in details["per_model_mean_logit"].items()},
details["num_tta_passes"],
)
return {
"prediction": POSITIVE_LABEL if is_positive else NEGATIVE_LABEL,
"prediction_binary": 1 if is_positive else 0,
"confidence": probability,
"heatmap_url": None,
"source": "model",
"threshold": active_threshold,
"ensemble_size": details["models_used"],
"use_tta": effective_use_tta,
"checkpoints": details["checkpoints_used"],
"mode": mode,
}
@app.post("/debug/predict")
async def debug_predict(image: UploadFile = File(...)) -> dict:
"""Same as /predict but returns per-model and per-TTA raw logits for
verification against the training notebook's val/test CSVs."""
if not _ensemble:
raise HTTPException(status_code=503, detail="Model not ready")
raw = await image.read()
if not raw:
raise HTTPException(status_code=400, detail="Empty upload")
try:
pil = Image.open(io.BytesIO(raw)).convert("L")
except Exception as exc: # noqa: BLE001
raise HTTPException(status_code=400, detail=f"Could not decode image: {exc}") from exc
details = _predict_probability_detailed(
pil,
use_tta=USE_TTA,
ensemble=_ensemble,
checkpoints=_loaded_checkpoints,
img_size=IMG_SIZE,
normalize_fn=_normalize_fn,
)
details["prediction"] = (
POSITIVE_LABEL if details["probability"] >= DECISION_THRESHOLD else NEGATIVE_LABEL
)
details["prediction_binary"] = 1 if details["probability"] >= DECISION_THRESHOLD else 0
details["threshold"] = DECISION_THRESHOLD
details["use_tta"] = USE_TTA
details["checkpoints"] = _loaded_checkpoints
return details