UIPress / scripts /step_ssim_bootstrap.py
DesonDai's picture
Add files using upload-large-folder tool
bac741f verified
"""
SSIM + Bootstrap CI computation for all benchmark methods.
Works with existing rendered screenshots and per-sample CLIP scores.
Usage:
python scripts/step_ssim_bootstrap.py --benchmark_dir results/benchmark --ref_dir data/ref_screenshots
python scripts/step_ssim_bootstrap.py --benchmark_dir results/benchmark_websight --ref_dir data/ref_screenshots_websight
"""
import argparse
import json
import os
import sys
from pathlib import Path
import numpy as np
from PIL import Image
PROJECT_ROOT = Path(__file__).parent.parent
def compute_ssim_pil(img1, img2, win_size=7):
"""Compute SSIM between two PIL images using numpy (no skimage dependency)."""
target_size = (min(img1.width, img2.width, 512), min(img1.height, img2.height, 512))
a = np.array(img1.resize(target_size).convert("RGB"), dtype=np.float64)
b = np.array(img2.resize(target_size).convert("RGB"), dtype=np.float64)
C1 = (0.01 * 255) ** 2
C2 = (0.03 * 255) ** 2
ssims = []
for ch in range(3):
mu1 = uniform_filter(a[:, :, ch], win_size)
mu2 = uniform_filter(b[:, :, ch], win_size)
mu1_sq = mu1 ** 2
mu2_sq = mu2 ** 2
mu1_mu2 = mu1 * mu2
sigma1_sq = uniform_filter(a[:, :, ch] ** 2, win_size) - mu1_sq
sigma2_sq = uniform_filter(b[:, :, ch] ** 2, win_size) - mu2_sq
sigma12 = uniform_filter(a[:, :, ch] * b[:, :, ch], win_size) - mu1_mu2
ssim_map = ((2 * mu1_mu2 + C1) * (2 * sigma12 + C2)) / \
((mu1_sq + mu2_sq + C1) * (sigma1_sq + sigma2_sq + C2))
ssims.append(ssim_map.mean())
return float(np.mean(ssims))
def uniform_filter(arr, size):
"""Simple uniform (box) filter."""
from scipy.ndimage import uniform_filter as _uf
return _uf(arr, size=size, mode='reflect')
def render_html_to_screenshot(html_path, out_path, width=1280, height=1024, timeout=15000):
"""Render HTML file to PNG screenshot using Playwright."""
try:
from playwright.sync_api import sync_playwright
with sync_playwright() as p:
browser = p.chromium.launch(headless=True, args=['--no-sandbox', '--disable-gpu'])
page = browser.new_page(viewport={"width": width, "height": height})
page.goto(f"file://{html_path}", wait_until="networkidle", timeout=timeout)
page.wait_for_timeout(1000)
page.screenshot(path=str(out_path), full_page=True)
browser.close()
return True
except Exception as e:
print(f" Render failed for {html_path}: {e}")
return False
def compute_ssim_for_method(method_dir, ref_dir, render_cache_dir):
"""Compute per-sample SSIM for a method."""
html_dir = Path(method_dir) / "html_predictions"
if not html_dir.exists():
return None
ref_dir = Path(ref_dir)
render_dir = Path(render_cache_dir) / Path(method_dir).name
render_dir.mkdir(parents=True, exist_ok=True)
html_files = sorted(html_dir.glob("*.html"))
per_sample = {}
for html_path in html_files:
sample_id = html_path.stem
ref_path = ref_dir / f"{sample_id}.png"
if not ref_path.exists():
continue
rendered_path = render_dir / f"{sample_id}.png"
if not rendered_path.exists():
ok = render_html_to_screenshot(str(html_path.resolve()), str(rendered_path))
if not ok:
continue
try:
ref_img = Image.open(ref_path).convert("RGB")
rendered_img = Image.open(rendered_path).convert("RGB")
ssim = compute_ssim_pil(ref_img, rendered_img)
per_sample[sample_id] = ssim
except Exception as e:
print(f" SSIM error for {sample_id}: {e}")
continue
if not per_sample:
return None
return {
"n_samples": len(per_sample),
"avg_ssim": round(float(np.mean(list(per_sample.values()))), 4),
"std_ssim": round(float(np.std(list(per_sample.values()))), 4),
"per_sample": per_sample,
}
def bootstrap_ci(scores, n_bootstrap=10000, ci=0.95, seed=42):
"""Compute bootstrap confidence interval."""
rng = np.random.RandomState(seed)
scores = np.array(scores)
n = len(scores)
boot_means = np.array([
rng.choice(scores, size=n, replace=True).mean()
for _ in range(n_bootstrap)
])
alpha = (1 - ci) / 2
lo = float(np.percentile(boot_means, 100 * alpha))
hi = float(np.percentile(boot_means, 100 * (1 - alpha)))
return {
"mean": float(scores.mean()),
"ci_lower": round(lo, 4),
"ci_upper": round(hi, 4),
"ci_width": round(hi - lo, 4),
"std": round(float(scores.std()), 4),
"n": n,
}
def compute_bootstrap_for_all(benchmark_dir):
"""Compute bootstrap CI for all methods from per-sample CLIP scores."""
benchmark_dir = Path(benchmark_dir)
results = {}
for method_dir in sorted(benchmark_dir.iterdir()):
if not method_dir.is_dir():
continue
clip_file = method_dir / "clip_scores.json"
if not clip_file.exists():
continue
with open(clip_file) as f:
clip_data = json.load(f)
per_sample = clip_data.get("per_sample", {})
if not per_sample:
continue
scores = []
for k, v in per_sample.items():
if isinstance(v, dict):
scores.append(v.get("clip_score", 0))
else:
scores.append(float(v))
if not scores:
continue
ci_result = bootstrap_ci(scores)
results[method_dir.name] = ci_result
print(f" {method_dir.name}: CLIP={ci_result['mean']:.4f} [{ci_result['ci_lower']:.4f}, {ci_result['ci_upper']:.4f}]")
return results
def main():
parser = argparse.ArgumentParser()
parser.add_argument("--benchmark_dir", type=str, default=str(PROJECT_ROOT / "results" / "benchmark"))
parser.add_argument("--ref_dir", type=str, default=str(PROJECT_ROOT / "data" / "ref_screenshots"))
parser.add_argument("--render_cache", type=str, default=str(PROJECT_ROOT / "results" / "rendered_screenshots"))
parser.add_argument("--skip_ssim", action="store_true")
parser.add_argument("--skip_bootstrap", action="store_true")
args = parser.parse_args()
benchmark_dir = Path(args.benchmark_dir)
output = {}
if not args.skip_bootstrap:
print("=" * 60)
print("Computing Bootstrap CI for CLIP scores...")
print("=" * 60)
bootstrap_results = compute_bootstrap_for_all(args.benchmark_dir)
output["bootstrap_ci"] = bootstrap_results
ci_file = benchmark_dir / "bootstrap_ci.json"
with open(ci_file, "w") as f:
json.dump(bootstrap_results, f, indent=2)
print(f"\nSaved to {ci_file}")
if not args.skip_ssim:
print("\n" + "=" * 60)
print("Computing SSIM scores...")
print("=" * 60)
ssim_results = {}
for method_dir in sorted(benchmark_dir.iterdir()):
if not method_dir.is_dir():
continue
html_dir = method_dir / "html_predictions"
if not html_dir.exists():
continue
print(f"\n Processing {method_dir.name}...")
result = compute_ssim_for_method(str(method_dir), args.ref_dir, args.render_cache)
if result:
ssim_results[method_dir.name] = {
"n_samples": result["n_samples"],
"avg_ssim": result["avg_ssim"],
"std_ssim": result["std_ssim"],
}
print(f" SSIM={result['avg_ssim']:.4f} ± {result['std_ssim']:.4f} (n={result['n_samples']})")
output[f"ssim_{method_dir.name}"] = result
ssim_file = benchmark_dir / "ssim_scores.json"
with open(ssim_file, "w") as f:
json.dump(ssim_results, f, indent=2)
print(f"\nSaved to {ssim_file}")
if __name__ == "__main__":
main()