| """ |
| SSIM + Bootstrap CI computation for all benchmark methods. |
| Works with existing rendered screenshots and per-sample CLIP scores. |
| |
| Usage: |
| python scripts/step_ssim_bootstrap.py --benchmark_dir results/benchmark --ref_dir data/ref_screenshots |
| python scripts/step_ssim_bootstrap.py --benchmark_dir results/benchmark_websight --ref_dir data/ref_screenshots_websight |
| """ |
|
|
| import argparse |
| import json |
| import os |
| import sys |
| from pathlib import Path |
|
|
| import numpy as np |
| from PIL import Image |
|
|
| PROJECT_ROOT = Path(__file__).parent.parent |
|
|
|
|
| def compute_ssim_pil(img1, img2, win_size=7): |
| """Compute SSIM between two PIL images using numpy (no skimage dependency).""" |
| target_size = (min(img1.width, img2.width, 512), min(img1.height, img2.height, 512)) |
| a = np.array(img1.resize(target_size).convert("RGB"), dtype=np.float64) |
| b = np.array(img2.resize(target_size).convert("RGB"), dtype=np.float64) |
|
|
| C1 = (0.01 * 255) ** 2 |
| C2 = (0.03 * 255) ** 2 |
|
|
| ssims = [] |
| for ch in range(3): |
| mu1 = uniform_filter(a[:, :, ch], win_size) |
| mu2 = uniform_filter(b[:, :, ch], win_size) |
| mu1_sq = mu1 ** 2 |
| mu2_sq = mu2 ** 2 |
| mu1_mu2 = mu1 * mu2 |
| sigma1_sq = uniform_filter(a[:, :, ch] ** 2, win_size) - mu1_sq |
| sigma2_sq = uniform_filter(b[:, :, ch] ** 2, win_size) - mu2_sq |
| sigma12 = uniform_filter(a[:, :, ch] * b[:, :, ch], win_size) - mu1_mu2 |
|
|
| ssim_map = ((2 * mu1_mu2 + C1) * (2 * sigma12 + C2)) / \ |
| ((mu1_sq + mu2_sq + C1) * (sigma1_sq + sigma2_sq + C2)) |
| ssims.append(ssim_map.mean()) |
|
|
| return float(np.mean(ssims)) |
|
|
|
|
| def uniform_filter(arr, size): |
| """Simple uniform (box) filter.""" |
| from scipy.ndimage import uniform_filter as _uf |
| return _uf(arr, size=size, mode='reflect') |
|
|
|
|
| def render_html_to_screenshot(html_path, out_path, width=1280, height=1024, timeout=15000): |
| """Render HTML file to PNG screenshot using Playwright.""" |
| try: |
| from playwright.sync_api import sync_playwright |
| with sync_playwright() as p: |
| browser = p.chromium.launch(headless=True, args=['--no-sandbox', '--disable-gpu']) |
| page = browser.new_page(viewport={"width": width, "height": height}) |
| page.goto(f"file://{html_path}", wait_until="networkidle", timeout=timeout) |
| page.wait_for_timeout(1000) |
| page.screenshot(path=str(out_path), full_page=True) |
| browser.close() |
| return True |
| except Exception as e: |
| print(f" Render failed for {html_path}: {e}") |
| return False |
|
|
|
|
| def compute_ssim_for_method(method_dir, ref_dir, render_cache_dir): |
| """Compute per-sample SSIM for a method.""" |
| html_dir = Path(method_dir) / "html_predictions" |
| if not html_dir.exists(): |
| return None |
|
|
| ref_dir = Path(ref_dir) |
| render_dir = Path(render_cache_dir) / Path(method_dir).name |
| render_dir.mkdir(parents=True, exist_ok=True) |
|
|
| html_files = sorted(html_dir.glob("*.html")) |
| per_sample = {} |
|
|
| for html_path in html_files: |
| sample_id = html_path.stem |
| ref_path = ref_dir / f"{sample_id}.png" |
| if not ref_path.exists(): |
| continue |
|
|
| rendered_path = render_dir / f"{sample_id}.png" |
| if not rendered_path.exists(): |
| ok = render_html_to_screenshot(str(html_path.resolve()), str(rendered_path)) |
| if not ok: |
| continue |
|
|
| try: |
| ref_img = Image.open(ref_path).convert("RGB") |
| rendered_img = Image.open(rendered_path).convert("RGB") |
| ssim = compute_ssim_pil(ref_img, rendered_img) |
| per_sample[sample_id] = ssim |
| except Exception as e: |
| print(f" SSIM error for {sample_id}: {e}") |
| continue |
|
|
| if not per_sample: |
| return None |
|
|
| return { |
| "n_samples": len(per_sample), |
| "avg_ssim": round(float(np.mean(list(per_sample.values()))), 4), |
| "std_ssim": round(float(np.std(list(per_sample.values()))), 4), |
| "per_sample": per_sample, |
| } |
|
|
|
|
| def bootstrap_ci(scores, n_bootstrap=10000, ci=0.95, seed=42): |
| """Compute bootstrap confidence interval.""" |
| rng = np.random.RandomState(seed) |
| scores = np.array(scores) |
| n = len(scores) |
| boot_means = np.array([ |
| rng.choice(scores, size=n, replace=True).mean() |
| for _ in range(n_bootstrap) |
| ]) |
| alpha = (1 - ci) / 2 |
| lo = float(np.percentile(boot_means, 100 * alpha)) |
| hi = float(np.percentile(boot_means, 100 * (1 - alpha))) |
| return { |
| "mean": float(scores.mean()), |
| "ci_lower": round(lo, 4), |
| "ci_upper": round(hi, 4), |
| "ci_width": round(hi - lo, 4), |
| "std": round(float(scores.std()), 4), |
| "n": n, |
| } |
|
|
|
|
| def compute_bootstrap_for_all(benchmark_dir): |
| """Compute bootstrap CI for all methods from per-sample CLIP scores.""" |
| benchmark_dir = Path(benchmark_dir) |
| results = {} |
|
|
| for method_dir in sorted(benchmark_dir.iterdir()): |
| if not method_dir.is_dir(): |
| continue |
| clip_file = method_dir / "clip_scores.json" |
| if not clip_file.exists(): |
| continue |
|
|
| with open(clip_file) as f: |
| clip_data = json.load(f) |
|
|
| per_sample = clip_data.get("per_sample", {}) |
| if not per_sample: |
| continue |
|
|
| scores = [] |
| for k, v in per_sample.items(): |
| if isinstance(v, dict): |
| scores.append(v.get("clip_score", 0)) |
| else: |
| scores.append(float(v)) |
|
|
| if not scores: |
| continue |
|
|
| ci_result = bootstrap_ci(scores) |
| results[method_dir.name] = ci_result |
| print(f" {method_dir.name}: CLIP={ci_result['mean']:.4f} [{ci_result['ci_lower']:.4f}, {ci_result['ci_upper']:.4f}]") |
|
|
| return results |
|
|
|
|
| def main(): |
| parser = argparse.ArgumentParser() |
| parser.add_argument("--benchmark_dir", type=str, default=str(PROJECT_ROOT / "results" / "benchmark")) |
| parser.add_argument("--ref_dir", type=str, default=str(PROJECT_ROOT / "data" / "ref_screenshots")) |
| parser.add_argument("--render_cache", type=str, default=str(PROJECT_ROOT / "results" / "rendered_screenshots")) |
| parser.add_argument("--skip_ssim", action="store_true") |
| parser.add_argument("--skip_bootstrap", action="store_true") |
| args = parser.parse_args() |
|
|
| benchmark_dir = Path(args.benchmark_dir) |
| output = {} |
|
|
| if not args.skip_bootstrap: |
| print("=" * 60) |
| print("Computing Bootstrap CI for CLIP scores...") |
| print("=" * 60) |
| bootstrap_results = compute_bootstrap_for_all(args.benchmark_dir) |
| output["bootstrap_ci"] = bootstrap_results |
|
|
| ci_file = benchmark_dir / "bootstrap_ci.json" |
| with open(ci_file, "w") as f: |
| json.dump(bootstrap_results, f, indent=2) |
| print(f"\nSaved to {ci_file}") |
|
|
| if not args.skip_ssim: |
| print("\n" + "=" * 60) |
| print("Computing SSIM scores...") |
| print("=" * 60) |
| ssim_results = {} |
| for method_dir in sorted(benchmark_dir.iterdir()): |
| if not method_dir.is_dir(): |
| continue |
| html_dir = method_dir / "html_predictions" |
| if not html_dir.exists(): |
| continue |
| print(f"\n Processing {method_dir.name}...") |
| result = compute_ssim_for_method(str(method_dir), args.ref_dir, args.render_cache) |
| if result: |
| ssim_results[method_dir.name] = { |
| "n_samples": result["n_samples"], |
| "avg_ssim": result["avg_ssim"], |
| "std_ssim": result["std_ssim"], |
| } |
| print(f" SSIM={result['avg_ssim']:.4f} ± {result['std_ssim']:.4f} (n={result['n_samples']})") |
| output[f"ssim_{method_dir.name}"] = result |
|
|
| ssim_file = benchmark_dir / "ssim_scores.json" |
| with open(ssim_file, "w") as f: |
| json.dump(ssim_results, f, indent=2) |
| print(f"\nSaved to {ssim_file}") |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|