|
|
from __future__ import annotations |
|
|
|
|
|
|
|
|
|
|
|
import argparse |
|
|
import csv |
|
|
import re |
|
|
import sys |
|
|
from pathlib import Path |
|
|
from typing import Dict, Tuple |
|
|
|
|
|
try: |
|
|
from scripts.hydrate_defextra import ( |
|
|
_ensure_trailing_punct, |
|
|
_postprocess_text, |
|
|
) |
|
|
except ModuleNotFoundError as exc: |
|
|
if exc.name != "scripts": |
|
|
raise |
|
|
PROJECT_ROOT = Path(__file__).resolve().parent.parent |
|
|
if str(PROJECT_ROOT) not in sys.path: |
|
|
sys.path.insert(0, str(PROJECT_ROOT)) |
|
|
from scripts.hydrate_defextra import ( |
|
|
_ensure_trailing_punct, |
|
|
_postprocess_text, |
|
|
) |
|
|
|
|
|
|
|
|
def _normalize_text(text: str) -> str: |
|
|
if text is None: |
|
|
return "" |
|
|
value = text.replace("\u00ad", "") |
|
|
value = re.sub(r"([A-Za-z])-\s+([A-Za-z])", r"\1\2", value) |
|
|
value = re.sub(r"\s+", " ", value).strip() |
|
|
return value |
|
|
|
|
|
|
|
|
def _normalize_punct(text: str) -> str: |
|
|
if text is None: |
|
|
return "" |
|
|
value = text.replace("\u00ad", "") |
|
|
value = re.sub(r"[^\w\s]", "", value) |
|
|
value = re.sub(r"\\s+", " ", value).strip() |
|
|
return value |
|
|
|
|
|
|
|
|
def _load(path: Path) -> Dict[Tuple[str, str], Dict[str, str]]: |
|
|
with path.open(encoding="utf-8", newline="") as handle: |
|
|
reader = csv.DictReader(handle) |
|
|
return {(row["paper_id"], row["concept"]): row for row in reader} |
|
|
|
|
|
|
|
|
def _row_flag( |
|
|
row: Dict[str, str], |
|
|
key: str, |
|
|
default: bool = False, |
|
|
) -> bool: |
|
|
value = (row.get(key) or "").strip().lower() |
|
|
if not value: |
|
|
return default |
|
|
return value == "true" |
|
|
|
|
|
|
|
|
def main() -> None: |
|
|
parser = argparse.ArgumentParser( |
|
|
description="Compare DefExtra hydrated CSV against reference.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--ref", |
|
|
type=Path, |
|
|
default=Path("results/paper_results/defextra_hf_tablefix.csv"), |
|
|
help="Reference CSV path.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--hyd", |
|
|
type=Path, |
|
|
default=Path( |
|
|
"results/paper_results/defextra_hydrated_tablefix_test.csv", |
|
|
), |
|
|
help="Hydrated CSV path.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--limit", |
|
|
type=int, |
|
|
default=5, |
|
|
help="Number of mismatches to print.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--report", |
|
|
type=Path, |
|
|
default=None, |
|
|
help="Optional path to write a detailed mismatch report.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--report-limit", |
|
|
type=int, |
|
|
default=0, |
|
|
help="Limit mismatches written to report (0 = all).", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--legal", |
|
|
type=Path, |
|
|
default=Path("results/paper_results/defextra_legal_tablefix.csv"), |
|
|
help="Legal CSV with token counts and linebreak flags.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--postprocess", |
|
|
action="store_true", |
|
|
help="Apply hydration postprocessing to hydrated text before compare.", |
|
|
) |
|
|
args = parser.parse_args() |
|
|
|
|
|
ref = _load(args.ref) |
|
|
hyd = _load(args.hyd) |
|
|
legal = _load(args.legal) if args.postprocess else {} |
|
|
|
|
|
missing = [k for k in ref if k not in hyd] |
|
|
extra = [k for k in hyd if k not in ref] |
|
|
|
|
|
def_mismatch = [] |
|
|
ctx_mismatch = [] |
|
|
def_mismatch_norm = [] |
|
|
ctx_mismatch_norm = [] |
|
|
def_mismatch_punct = [] |
|
|
ctx_mismatch_punct = [] |
|
|
|
|
|
for key, ref_row in ref.items(): |
|
|
hyd_row = hyd.get(key) |
|
|
if not hyd_row: |
|
|
continue |
|
|
if args.postprocess: |
|
|
legal_row = legal.get(key, {}) |
|
|
def_expected = int(legal_row.get("definition_token_count") or 0) |
|
|
ctx_expected = int(legal_row.get("context_token_count") or 0) |
|
|
def_preserve = _row_flag( |
|
|
legal_row, |
|
|
"definition_preserve_linebreaks", |
|
|
) |
|
|
ctx_preserve = _row_flag( |
|
|
legal_row, |
|
|
"context_preserve_linebreaks", |
|
|
) |
|
|
def_preserve_hyphen = _row_flag( |
|
|
legal_row, |
|
|
"definition_preserve_hyphenation", |
|
|
) |
|
|
ctx_preserve_hyphen = _row_flag( |
|
|
legal_row, |
|
|
"context_preserve_hyphenation", |
|
|
) |
|
|
def_keep_bracket = _row_flag( |
|
|
legal_row, |
|
|
"definition_has_bracket_citation", |
|
|
True, |
|
|
) |
|
|
def_keep_paren = _row_flag( |
|
|
legal_row, |
|
|
"definition_has_paren_citation", |
|
|
True, |
|
|
) |
|
|
def_split_letter_digit = not _row_flag( |
|
|
legal_row, |
|
|
"definition_has_letter_digit", |
|
|
) |
|
|
ctx_keep_bracket = _row_flag( |
|
|
legal_row, |
|
|
"context_has_bracket_citation", |
|
|
True, |
|
|
) |
|
|
ctx_keep_paren = _row_flag( |
|
|
legal_row, |
|
|
"context_has_paren_citation", |
|
|
True, |
|
|
) |
|
|
ctx_split_letter_digit = not _row_flag( |
|
|
legal_row, |
|
|
"context_has_letter_digit", |
|
|
) |
|
|
hyd_row = dict(hyd_row) |
|
|
hyd_row["definition"] = _ensure_trailing_punct( |
|
|
_postprocess_text( |
|
|
hyd_row.get("definition", ""), |
|
|
def_expected, |
|
|
def_preserve, |
|
|
def_preserve_hyphen, |
|
|
def_keep_bracket, |
|
|
def_keep_paren, |
|
|
def_split_letter_digit, |
|
|
), |
|
|
legal_row.get("definition_end_punct", ""), |
|
|
) |
|
|
hyd_row["context"] = _ensure_trailing_punct( |
|
|
_postprocess_text( |
|
|
hyd_row.get("context", ""), |
|
|
ctx_expected, |
|
|
ctx_preserve, |
|
|
ctx_preserve_hyphen, |
|
|
ctx_keep_bracket, |
|
|
ctx_keep_paren, |
|
|
ctx_split_letter_digit, |
|
|
), |
|
|
legal_row.get("context_end_punct", ""), |
|
|
) |
|
|
if ref_row.get("definition", "") != hyd_row.get("definition", ""): |
|
|
def_mismatch.append(key) |
|
|
if _normalize_text( |
|
|
ref_row.get("definition", ""), |
|
|
) != _normalize_text( |
|
|
hyd_row.get("definition", ""), |
|
|
): |
|
|
def_mismatch_norm.append(key) |
|
|
if _normalize_punct(ref_row.get("definition", "")) == _normalize_punct( |
|
|
hyd_row.get("definition", ""), |
|
|
): |
|
|
def_mismatch_punct.append(key) |
|
|
if ref_row.get("context", "") != hyd_row.get("context", ""): |
|
|
ctx_mismatch.append(key) |
|
|
if _normalize_text(ref_row.get("context", "")) != _normalize_text( |
|
|
hyd_row.get("context", ""), |
|
|
): |
|
|
ctx_mismatch_norm.append(key) |
|
|
if _normalize_punct(ref_row.get("context", "")) == _normalize_punct( |
|
|
hyd_row.get("context", ""), |
|
|
): |
|
|
ctx_mismatch_punct.append(key) |
|
|
|
|
|
total_ref = len(ref) |
|
|
total_hyd = len(hyd) |
|
|
print(f"Reference rows: {total_ref}") |
|
|
print(f"Hydrated rows: {total_hyd}") |
|
|
print(f"Missing keys: {len(missing)}") |
|
|
print(f"Extra keys: {len(extra)}") |
|
|
print(f"Definition mismatches (exact): {len(def_mismatch)}") |
|
|
print(f"Definition mismatches (normalized): {len(def_mismatch_norm)}") |
|
|
print(f"Context mismatches (exact): {len(ctx_mismatch)}") |
|
|
print(f"Context mismatches (normalized): {len(ctx_mismatch_norm)}") |
|
|
if def_mismatch_punct: |
|
|
print( |
|
|
"Definition mismatches (punctuation-only): " |
|
|
f"{len(def_mismatch_punct)}", |
|
|
) |
|
|
if ctx_mismatch_punct: |
|
|
print( |
|
|
"Context mismatches (punctuation-only): " |
|
|
f"{len(ctx_mismatch_punct)}", |
|
|
) |
|
|
|
|
|
if args.limit <= 0: |
|
|
return |
|
|
|
|
|
shown = 0 |
|
|
for key in def_mismatch: |
|
|
if shown >= args.limit: |
|
|
break |
|
|
ref_row = ref[key] |
|
|
hyd_row = hyd[key] |
|
|
print("\nDefinition mismatch:", key) |
|
|
print("ref:", ref_row.get("definition", "")) |
|
|
print("hyd:", hyd_row.get("definition", "")) |
|
|
shown += 1 |
|
|
|
|
|
shown = 0 |
|
|
for key in ctx_mismatch: |
|
|
if shown >= args.limit: |
|
|
break |
|
|
ref_row = ref[key] |
|
|
hyd_row = hyd[key] |
|
|
print("\nContext mismatch:", key) |
|
|
print("ref:", ref_row.get("context", "")) |
|
|
print("hyd:", hyd_row.get("context", "")) |
|
|
shown += 1 |
|
|
|
|
|
if args.report is not None: |
|
|
report_lines = [] |
|
|
report_lines.append(f"Missing keys: {len(missing)}") |
|
|
report_lines.extend([f"- {k}" for k in missing]) |
|
|
report_lines.append("") |
|
|
report_lines.append( |
|
|
f"Definition mismatches (exact): {len(def_mismatch)}" |
|
|
) |
|
|
report_lines.append( |
|
|
f"Definition mismatches (normalized): {len(def_mismatch_norm)}" |
|
|
) |
|
|
report_lines.append( |
|
|
f"Definition mismatches (punctuation-only): {len(def_mismatch_punct)}" |
|
|
) |
|
|
report_lines.append( |
|
|
f"Context mismatches (exact): {len(ctx_mismatch)}" |
|
|
) |
|
|
report_lines.append( |
|
|
f"Context mismatches (normalized): {len(ctx_mismatch_norm)}" |
|
|
) |
|
|
report_lines.append( |
|
|
f"Context mismatches (punctuation-only): {len(ctx_mismatch_punct)}" |
|
|
) |
|
|
report_lines.append("") |
|
|
|
|
|
def_limit = args.report_limit or len(def_mismatch) |
|
|
ctx_limit = args.report_limit or len(ctx_mismatch) |
|
|
|
|
|
report_lines.append("Definition mismatches:") |
|
|
for key in def_mismatch[:def_limit]: |
|
|
ref_row = ref[key] |
|
|
hyd_row = hyd[key] |
|
|
report_lines.append(f"- {key[0]} | {key[1]}") |
|
|
report_lines.append(f" ref: {ref_row.get('definition','')}") |
|
|
report_lines.append(f" hyd: {hyd_row.get('definition','')}") |
|
|
report_lines.append("") |
|
|
|
|
|
report_lines.append("Context mismatches:") |
|
|
for key in ctx_mismatch[:ctx_limit]: |
|
|
ref_row = ref[key] |
|
|
hyd_row = hyd[key] |
|
|
report_lines.append(f"- {key[0]} | {key[1]}") |
|
|
report_lines.append(f" ref: {ref_row.get('context','')}") |
|
|
report_lines.append(f" hyd: {hyd_row.get('context','')}") |
|
|
|
|
|
args.report.parent.mkdir(parents=True, exist_ok=True) |
|
|
args.report.write_text( |
|
|
"\n".join(report_lines) + "\n", |
|
|
encoding="utf-8", |
|
|
) |
|
|
print(f"Wrote report to {args.report}") |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
main() |
|
|
|