Spaces:
Running
Running
File size: 16,794 Bytes
59fa244 242f528 59fa244 242f528 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 | """Gradio demo: P&ID graph extraction with Claude VLM + evaluation.
Usage (local):
ANTHROPIC_API_KEY=sk-ant-... python app.py
Or put the key in a `.env` next to this file. The app:
1. Takes a P&ID image (preset or upload)
2. Runs extraction (optionally tiled 2x2) via Claude Opus 4.6
3. If a ground-truth graphml is provided, collapses it to semantic-only
form and computes node/edge P/R/F1 via `pid2graph_eval.metrics`
4. Draws both the prediction and the ground truth as NetworkX graphs
using bbox-based layouts so the topology matches the source image
"""
from __future__ import annotations
import json
import os
import time
from pathlib import Path
from typing import Optional
# Matplotlib backend must be set before pyplot import for headless use;
# the `matplotlib.use()` call below taints every subsequent import with
# E402 ("module level import not at top of file"), which is expected here.
import matplotlib
matplotlib.use("Agg")
import matplotlib.patches as mpatches # noqa: E402
import matplotlib.pyplot as plt # noqa: E402
import networkx as nx # noqa: E402
import anthropic # noqa: E402
# ---------------------------------------------------------------------------
# Gradio 4.44 / gradio_client 1.3.0 bug workaround
# ---------------------------------------------------------------------------
# At `/info` boot, Gradio walks every component's JSON schema via
# `gradio_client.utils._json_schema_to_python_type`. That function does not
# handle bool schemas (`additionalProperties: false` or `true`, both of which
# are valid JSON Schema) — it recurses with the bool and then `if "const" in
# schema:` on line 863 raises `TypeError: argument of type 'bool' is not
# iterable`. Patch the function here before importing gradio so the crash is
# avoided regardless of which component triggers it. (Fixed upstream in later
# gradio_client releases; we stay on 4.44 because Python 3.9 can't run
# gradio 5.) Harmless on versions where the bug is already fixed.
import gradio_client.utils as _gc_utils # noqa: E402
_orig_json_schema_to_python_type = _gc_utils._json_schema_to_python_type
def _patched_json_schema_to_python_type(schema, defs=None): # type: ignore[override]
if isinstance(schema, bool):
# `True` means "any value is allowed"; `False` means "no value".
return "Any" if schema else "None"
return _orig_json_schema_to_python_type(schema, defs)
_gc_utils._json_schema_to_python_type = _patched_json_schema_to_python_type
import gradio as gr # noqa: E402 (must come after the monkey-patch)
from dotenv import load_dotenv # noqa: E402
from pid2graph_eval.extractor import ( # noqa: E402
DEFAULT_MODEL,
extract_graph,
extract_graph_tiled,
)
from pid2graph_eval.gt_loader import ( # noqa: E402
SEMANTIC_EQUIPMENT_TYPES,
collapse_through_primitives,
filter_by_types,
load_graphml,
)
from pid2graph_eval.metrics import evaluate # noqa: E402
# ---------------------------------------------------------------------------
# Config
# ---------------------------------------------------------------------------
APP_ROOT = Path(__file__).parent
SAMPLES_DIR = APP_ROOT / "samples"
load_dotenv(APP_ROOT / ".env")
# Presets: (display name) -> (image path, graphml path)
PRESETS: dict[str, tuple[Path, Path]] = {
"OPEN100 #1 — small (27 semantic nodes)": (
SAMPLES_DIR / "open100_01_small.png",
SAMPLES_DIR / "open100_01_small.graphml",
),
"OPEN100 #3 — medium (53 semantic nodes)": (
SAMPLES_DIR / "open100_03_medium.png",
SAMPLES_DIR / "open100_03_medium.graphml",
),
"OPEN100 #0 — large (82 semantic nodes)": (
SAMPLES_DIR / "open100_00_large.png",
SAMPLES_DIR / "open100_00_large.graphml",
),
}
NONE_LABEL = "(none — upload your own)"
# Fixed palette so pred/GT visualizations use matching colors.
TYPE_COLORS: dict[str, str] = {
"valve": "#ff6b6b",
"pump": "#4ecdc4",
"tank": "#ffd93d",
"instrumentation": "#6bcfff",
"inlet/outlet": "#c47bff",
}
LEGEND_HANDLES = [
mpatches.Patch(color=c, label=t) for t, c in TYPE_COLORS.items()
]
# ---------------------------------------------------------------------------
# Visualization
# ---------------------------------------------------------------------------
def _bbox_to_xyxy(bbox) -> Optional[tuple[float, float, float, float]]:
"""Normalize a bbox to `(xmin, ymin, xmax, ymax)` floats.
Accepts both shapes that flow through the pipeline:
* **list / tuple** `[x1, y1, x2, y2]` — produced by `gt_loader._bbox`
and by `tile.merge_tile_graphs` for the tiled pred path.
* **dict** `{"xmin": ..., "ymin": ..., "xmax": ..., "ymax": ...}` —
produced by `GraphOut.to_dict()` in single-shot mode, because the
Pydantic `BBox` model round-trips through `model_dump()`.
Returns `None` if the bbox is missing or malformed.
"""
if bbox is None:
return None
if isinstance(bbox, dict):
try:
return (
float(bbox["xmin"]),
float(bbox["ymin"]),
float(bbox["xmax"]),
float(bbox["ymax"]),
)
except (KeyError, TypeError, ValueError):
return None
if isinstance(bbox, (list, tuple)) and len(bbox) == 4:
try:
return (
float(bbox[0]),
float(bbox[1]),
float(bbox[2]),
float(bbox[3]),
)
except (TypeError, ValueError):
return None
return None
def draw_graph(graph_dict: dict, title: str, figsize=(8, 6)) -> plt.Figure:
"""Render a graph as a matplotlib figure.
Node positions come from bbox centers when available (so the drawing
preserves the spatial layout of the original P&ID); nodes without a
bbox fall back to networkx spring layout.
"""
fig, ax = plt.subplots(figsize=figsize, dpi=110)
G = nx.Graph()
pos: dict[str, tuple[float, float]] = {}
colors: list[str] = []
node_list: list[str] = []
for n in graph_dict.get("nodes", []):
nid = n["id"]
G.add_node(nid)
node_list.append(nid)
colors.append(TYPE_COLORS.get(n.get("type", ""), "#cccccc"))
coords = _bbox_to_xyxy(n.get("bbox"))
if coords is not None:
x1, y1, x2, y2 = coords
cx = (x1 + x2) / 2.0
cy = (y1 + y2) / 2.0
pos[nid] = (cx, -cy) # flip y so the image is right-side up
for e in graph_dict.get("edges", []):
s, t = e.get("source"), e.get("target")
if s in G.nodes and t in G.nodes:
G.add_edge(s, t)
# Fall back to spring layout for any nodes that lack a bbox.
missing = [nid for nid in G.nodes if nid not in pos]
if missing:
if not pos:
pos = nx.spring_layout(G, seed=42)
else:
# Place missing nodes near the existing bbox cloud center.
xs = [p[0] for p in pos.values()]
ys = [p[1] for p in pos.values()]
cx0 = sum(xs) / len(xs)
cy0 = sum(ys) / len(ys)
for nid in missing:
pos[nid] = (cx0, cy0)
nx.draw_networkx_edges(G, pos, alpha=0.35, width=0.6, ax=ax)
nx.draw_networkx_nodes(
G, pos,
nodelist=node_list,
node_color=colors,
node_size=55,
linewidths=0.5,
edgecolors="#222",
ax=ax,
)
ax.set_title(title, fontsize=11)
ax.set_aspect("equal")
ax.axis("off")
ax.legend(handles=LEGEND_HANDLES, loc="lower right", fontsize=7, framealpha=0.9)
fig.tight_layout()
return fig
# ---------------------------------------------------------------------------
# Pipeline
# ---------------------------------------------------------------------------
def _preset_paths(preset_name: str) -> tuple[Optional[str], Optional[str]]:
"""Resolve a preset dropdown selection to (image_path, graphml_path)."""
if preset_name == NONE_LABEL or preset_name not in PRESETS:
return None, None
img, gt = PRESETS[preset_name]
return (str(img) if img.exists() else None,
str(gt) if gt.exists() else None)
def _format_metrics(metrics: dict, latency_s: float, mode: str) -> str:
nm = metrics["nodes"]
em = metrics["edges"]
return f"""
### Metrics
| | Precision | Recall | F1 | TP | FP | FN |
|---|---:|---:|---:|---:|---:|---:|
| **Nodes** | {nm['precision']:.3f} | {nm['recall']:.3f} | **{nm['f1']:.3f}** | {nm['tp']} | {nm['fp']} | {nm['fn']} |
| **Edges** | {em['precision']:.3f} | {em['recall']:.3f} | **{em['f1']:.3f}** | {em['tp']} | {em['fp']} | {em['fn']} |
- Pred: **{metrics['n_pred_nodes']}** ノード / **{metrics['n_pred_edges']}** エッジ
- GT (semantic-collapsed): **{metrics['n_gt_nodes']}** ノード / **{metrics['n_gt_edges']}** エッジ
- Mode: `{mode}` · Latency: **{latency_s:.1f}s**
"""
def _format_pred_only(pred_dict: dict, latency_s: float, mode: str) -> str:
return f"""
### Prediction
- **{len(pred_dict['nodes'])}** ノード / **{len(pred_dict['edges'])}** エッジ
- Mode: `{mode}` · Latency: **{latency_s:.1f}s**
- (正解 graphml 未指定のため評価スキップ)
"""
def run_extraction(
preset_name: str,
image_path: Optional[str],
gt_path: Optional[str],
use_tiling: bool,
progress: gr.Progress = gr.Progress(),
) -> tuple[str, Optional[plt.Figure], Optional[plt.Figure], str]:
"""Entry point wired to the Run button."""
# Preset overrides manual upload so the demo is reproducible.
preset_img, preset_gt = _preset_paths(preset_name)
if preset_img:
image_path = preset_img
if preset_gt:
gt_path = preset_gt
if not image_path:
return (
"⚠️ 画像をアップロードするか、プリセットを選択してください。",
None, None, "",
)
if not os.environ.get("ANTHROPIC_API_KEY"):
return (
"⚠️ `ANTHROPIC_API_KEY` が設定されていません。`.env` に追記して再起動してください。",
None, None, "",
)
client = anthropic.Anthropic()
mode = "tiled 2x2 + seam filter" if use_tiling else "single-shot"
try:
progress(0.05, desc=f"VLM 抽出開始 ({mode})…")
t0 = time.time()
if use_tiling:
pred_dict = extract_graph_tiled(
Path(image_path),
client=client,
rows=2,
cols=2,
overlap=0.1,
dedup_px=40.0,
)
else:
pred = extract_graph(Path(image_path), client=client)
pred_dict = pred.to_dict()
latency = time.time() - t0
progress(0.55, desc="予測を semantic types に絞り込み…")
# Defensive: drop anything non-semantic the VLM may have emitted.
pred_dict = filter_by_types(pred_dict, SEMANTIC_EQUIPMENT_TYPES)
except Exception as e:
return (f"❌ VLM 抽出中にエラー: `{e}`", None, None, "")
progress(0.65, desc="予測グラフを描画…")
pred_fig = draw_graph(
pred_dict,
title=f"Prediction — {len(pred_dict['nodes'])} nodes, {len(pred_dict['edges'])} edges",
)
gt_fig = None
metrics_md = _format_pred_only(pred_dict, latency, mode)
if gt_path and Path(gt_path).exists():
try:
progress(0.75, desc="GT graphml をロード & 縮約…")
gt_raw = load_graphml(Path(gt_path))
gt_dict = collapse_through_primitives(gt_raw, SEMANTIC_EQUIPMENT_TYPES)
progress(0.85, desc="P/R/F1 を評価…")
metrics = evaluate(
pred_dict,
gt_dict,
directed=False,
match_threshold=0.5,
)
metrics_md = _format_metrics(metrics, latency, mode)
progress(0.95, desc="GT グラフを描画…")
gt_fig = draw_graph(
gt_dict,
title=f"Ground Truth — {len(gt_dict['nodes'])} nodes, {len(gt_dict['edges'])} edges",
)
except Exception as e:
metrics_md += f"\n\n⚠️ GT 処理でエラー: `{e}`"
# Strip heavy-ish keys before JSON display.
display_dict = {
"nodes": pred_dict["nodes"],
"edges": pred_dict["edges"],
}
pred_json = json.dumps(display_dict, indent=2, ensure_ascii=False)
progress(1.0, desc="完了")
return metrics_md, pred_fig, gt_fig, pred_json
def on_preset_change(preset_name: str):
"""When a preset is picked, auto-fill the image and graphml fields."""
img, gt = _preset_paths(preset_name)
return img, gt
# ---------------------------------------------------------------------------
# UI
# ---------------------------------------------------------------------------
DESCRIPTION = """
# PID2Graph × Claude VLM Demo
P&ID (配管計装図) を Claude Opus 4.6 のビジョンで読み取り、シンボル(valve / pump /
tank / instrumentation / inlet・outlet)とその接続関係を JSON グラフに変換します。
正解 graphml を指定すると、ノード/エッジ単位の Precision / Recall / F1 を算出します。
- **タイル分割 (2x2)**: 大きな図面では 1 枚を 4 タイルに分割してから抽出し、マージ時に
bbox 距離で重複排除 + タイル境界の inlet/outlet FP を後処理で除去します。
- **評価ルール**: GT 側は semantic equipment のみを残し、配管プリミティブ (connector /
crossing / arrow / background) を経由する接続を 1 エッジに縮約します。
- **VLM 設定**: `temperature=0` で決定論的サンプリング、構造化出力で JSON スキーマを強制。
"""
def build_ui() -> gr.Blocks:
with gr.Blocks(title="PID2Graph × Claude VLM Demo") as demo:
gr.Markdown(DESCRIPTION)
with gr.Row():
with gr.Column(scale=1):
preset = gr.Dropdown(
choices=[NONE_LABEL] + list(PRESETS.keys()),
value=NONE_LABEL,
label="プリセット (OPEN100 より)",
)
image_in = gr.Image(
type="filepath",
label="P&ID 画像",
height=260,
)
gt_in = gr.File(
label="正解 graphml (任意)",
file_types=[".graphml", ".xml"],
type="filepath",
)
tiling = gr.Checkbox(
value=True,
label="タイル分割 (2x2) で抽出 — 高精度だがコスト・時間 4 倍",
)
run_btn = gr.Button("抽出実行", variant="primary")
gr.Markdown(
"モデル: `" + DEFAULT_MODEL + "` · 所要時間目安: single ~20s / tiled ~60-80s"
)
with gr.Column(scale=2):
metrics_md = gr.Markdown()
with gr.Row():
pred_plot = gr.Plot(label="Prediction")
gt_plot = gr.Plot(label="Ground Truth")
with gr.Accordion("予測 JSON (nodes / edges)", open=False):
# NOTE: using Textbox rather than `gr.Code(language="json")`
# because the latter's schema has tripped the gradio_client
# `additionalProperties: false` bug on 4.44.1 in the past.
# Textbox is a plain string component — zero schema surface.
pred_json = gr.Textbox(
label="",
lines=20,
max_lines=30,
show_copy_button=True,
interactive=False,
)
preset.change(on_preset_change, inputs=[preset], outputs=[image_in, gt_in])
run_btn.click(
run_extraction,
inputs=[preset, image_in, gt_in, tiling],
outputs=[metrics_md, pred_plot, gt_plot, pred_json],
)
return demo
if __name__ == "__main__":
# HF Spaces runs this inside a container where localhost is not
# reachable from outside — binding to 0.0.0.0 is required, otherwise
# Gradio raises "When localhost is not accessible, a shareable link
# must be created". Locally this is harmless.
#
# `show_api=False` hides the docs panel in the UI; the monkey-patch
# at the top of this file is what actually prevents the 4.44 schema
# crash (now kept as defensive dead-code since we pin to 4.31.5).
build_ui().launch(show_api=False, server_name="0.0.0.0")
|