|
|
from __future__ import annotations |
|
|
|
|
|
|
|
|
|
|
|
import argparse |
|
|
import csv |
|
|
import re |
|
|
import time |
|
|
from pathlib import Path |
|
|
from typing import Dict, Tuple |
|
|
|
|
|
try: |
|
|
from scripts.defextra_markers import normalize_paper_id |
|
|
except ModuleNotFoundError as exc: |
|
|
if exc.name != "scripts": |
|
|
raise |
|
|
import sys |
|
|
|
|
|
PROJECT_ROOT = Path(__file__).resolve().parent.parent |
|
|
if str(PROJECT_ROOT) not in sys.path: |
|
|
sys.path.insert(0, str(PROJECT_ROOT)) |
|
|
from scripts.defextra_markers import normalize_paper_id |
|
|
|
|
|
|
|
|
def _normalize_title(title: str) -> str: |
|
|
return " ".join((title or "").lower().split()) |
|
|
|
|
|
|
|
|
def _load_csv(path: Path) -> list[dict]: |
|
|
with path.open(encoding="utf-8", newline="") as handle: |
|
|
return list(csv.DictReader(handle)) |
|
|
|
|
|
|
|
|
def _parse_missing_report(path: Path) -> Tuple[list[str], list[str]]: |
|
|
missing_defs: list[str] = [] |
|
|
missing_ctxs: list[str] = [] |
|
|
if not path.exists(): |
|
|
return missing_defs, missing_ctxs |
|
|
section = None |
|
|
for line in path.read_text(encoding="utf-8").splitlines(): |
|
|
line = line.strip() |
|
|
if line.startswith("Missing definitions"): |
|
|
section = "def" |
|
|
continue |
|
|
if line.startswith("Missing contexts"): |
|
|
section = "ctx" |
|
|
continue |
|
|
if not line.startswith("-"): |
|
|
continue |
|
|
item = line[1:].strip() |
|
|
if section == "def": |
|
|
missing_defs.append(item) |
|
|
elif section == "ctx": |
|
|
missing_ctxs.append(item) |
|
|
return missing_defs, missing_ctxs |
|
|
|
|
|
|
|
|
def _index_recent_pdfs( |
|
|
pdf_dir: Path, |
|
|
cutoff_ts: float, |
|
|
) -> Dict[str, Path]: |
|
|
index: Dict[str, Path] = {} |
|
|
if not pdf_dir.exists(): |
|
|
return index |
|
|
version_re = re.compile(r"^(?P<base>.+?)(v\d+)$", re.IGNORECASE) |
|
|
arxiv_re = re.compile(r"^(?P<base>\d{4}\.\d{4,5})v\d+$", re.IGNORECASE) |
|
|
pii_re = re.compile(r"(S\d{8,})", re.IGNORECASE) |
|
|
for suffix in ("*.pdf", "*.PDF"): |
|
|
for path in pdf_dir.rglob(suffix): |
|
|
try: |
|
|
if path.stat().st_mtime < cutoff_ts: |
|
|
continue |
|
|
except OSError: |
|
|
continue |
|
|
stem = path.stem |
|
|
keys = {stem, stem.lower(), normalize_paper_id(stem)} |
|
|
if stem.startswith("paper_"): |
|
|
stripped = stem[len("paper_") :] |
|
|
keys.update( |
|
|
{stripped, stripped.lower(), normalize_paper_id(stripped)}, |
|
|
) |
|
|
if stem.endswith("_fixed") or stem.endswith("-fixed"): |
|
|
base = ( |
|
|
stem[: -len("_fixed")] |
|
|
if stem.endswith("_fixed") |
|
|
else stem[: -len("-fixed")] |
|
|
) |
|
|
if base: |
|
|
keys.update({base, base.lower(), normalize_paper_id(base)}) |
|
|
match = arxiv_re.match(stem) |
|
|
if match: |
|
|
base = match.group("base") |
|
|
keys.update({base, base.lower(), normalize_paper_id(base)}) |
|
|
match = version_re.match(stem) |
|
|
if match: |
|
|
base = match.group("base") |
|
|
keys.update({base, base.lower(), normalize_paper_id(base)}) |
|
|
pii_match = pii_re.search(stem) |
|
|
if pii_match: |
|
|
pii = pii_match.group(1) |
|
|
keys.update({pii, pii.lower(), normalize_paper_id(pii)}) |
|
|
for key in keys: |
|
|
if key: |
|
|
index.setdefault(key, path) |
|
|
return index |
|
|
|
|
|
|
|
|
def main() -> None: |
|
|
parser = argparse.ArgumentParser( |
|
|
description="Report DefExtra hydration coverage and missing spans.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--legal-csv", |
|
|
type=Path, |
|
|
default=Path("results/paper_results/defextra_legal_tablefix.csv"), |
|
|
help="Legal CSV used for hydration.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--legal-report", |
|
|
type=Path, |
|
|
default=Path( |
|
|
"results/paper_results/defextra_legal_tablefix_report.txt", |
|
|
), |
|
|
help="Report generated by prepare_defextra_legal.py.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--hydrated-csv", |
|
|
type=Path, |
|
|
default=Path( |
|
|
"results/paper_results/defextra_hydrated_tablefix_test.csv", |
|
|
), |
|
|
help="Hydrated CSV from hydrate_defextra.py.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--pdf-dir", |
|
|
type=Path, |
|
|
default=Path("ManualPDFsGROBID/manual_pdfs/manual_pdfs"), |
|
|
help="Directory with user PDFs (used to tag recent downloads).", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--recent-days", |
|
|
type=int, |
|
|
default=7, |
|
|
help="How many days count as 'recent' for PDF downloads.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--output", |
|
|
type=Path, |
|
|
default=None, |
|
|
help="Optional report output path.", |
|
|
) |
|
|
args = parser.parse_args() |
|
|
|
|
|
legal_rows = _load_csv(args.legal_csv) |
|
|
hydrated_rows = ( |
|
|
_load_csv(args.hydrated_csv) if args.hydrated_csv.exists() else [] |
|
|
) |
|
|
|
|
|
ref_ids = { |
|
|
row.get("paper_id", "") for row in legal_rows if row.get("paper_id") |
|
|
} |
|
|
hyd_ids = { |
|
|
row.get("paper_id", "") for row in hydrated_rows if row.get("paper_id") |
|
|
} |
|
|
missing_papers = sorted(ref_ids - hyd_ids) |
|
|
|
|
|
missing_defs, missing_ctxs = _parse_missing_report(args.legal_report) |
|
|
|
|
|
idx = { |
|
|
(row.get("paper_id", ""), row.get("concept", "")): row |
|
|
for row in legal_rows |
|
|
} |
|
|
|
|
|
implicit_defs = [] |
|
|
implicit_ctxs = [] |
|
|
for item in missing_defs: |
|
|
try: |
|
|
pid, concept = [p.strip() for p in item.split("|", 1)] |
|
|
except ValueError: |
|
|
continue |
|
|
row = idx.get((pid, concept)) |
|
|
if ( |
|
|
row |
|
|
and (row.get("definition_type") or "").strip().lower() |
|
|
== "implicit" |
|
|
): |
|
|
implicit_defs.append(item) |
|
|
for item in missing_ctxs: |
|
|
try: |
|
|
pid, concept = [p.strip() for p in item.split("|", 1)] |
|
|
except ValueError: |
|
|
continue |
|
|
row = idx.get((pid, concept)) |
|
|
if ( |
|
|
row |
|
|
and (row.get("definition_type") or "").strip().lower() |
|
|
== "implicit" |
|
|
): |
|
|
implicit_ctxs.append(item) |
|
|
|
|
|
cutoff_ts = time.time() - (args.recent_days * 86400) |
|
|
recent_index = _index_recent_pdfs(args.pdf_dir, cutoff_ts) |
|
|
|
|
|
recent_missing_defs = [] |
|
|
recent_missing_ctxs = [] |
|
|
recent_missing_papers = [] |
|
|
|
|
|
for pid in missing_papers: |
|
|
if pid in recent_index or normalize_paper_id(pid) in recent_index: |
|
|
recent_missing_papers.append(pid) |
|
|
|
|
|
for item in missing_defs: |
|
|
try: |
|
|
pid, concept = [p.strip() for p in item.split("|", 1)] |
|
|
except ValueError: |
|
|
continue |
|
|
if pid in recent_index or normalize_paper_id(pid) in recent_index: |
|
|
recent_missing_defs.append(item) |
|
|
|
|
|
for item in missing_ctxs: |
|
|
try: |
|
|
pid, concept = [p.strip() for p in item.split("|", 1)] |
|
|
except ValueError: |
|
|
continue |
|
|
if pid in recent_index or normalize_paper_id(pid) in recent_index: |
|
|
recent_missing_ctxs.append(item) |
|
|
|
|
|
lines = [] |
|
|
lines.append(f"Missing papers (no hydrated rows): {len(missing_papers)}") |
|
|
for pid in missing_papers: |
|
|
lines.append(f"- {pid}") |
|
|
lines.append("") |
|
|
lines.append( |
|
|
f"Missing definition spans marked implicit: {len(implicit_defs)}", |
|
|
) |
|
|
for item in implicit_defs: |
|
|
lines.append(f"- {item}") |
|
|
lines.append("") |
|
|
lines.append( |
|
|
f"Missing context spans marked implicit: {len(implicit_ctxs)}", |
|
|
) |
|
|
for item in implicit_ctxs: |
|
|
lines.append(f"- {item}") |
|
|
lines.append("") |
|
|
lines.append( |
|
|
f"Missing papers with recent PDFs (<= {args.recent_days} days): " |
|
|
f"{len(recent_missing_papers)}", |
|
|
) |
|
|
for pid in recent_missing_papers: |
|
|
lines.append(f"- {pid}") |
|
|
lines.append("") |
|
|
lines.append( |
|
|
f"Missing definition spans with recent PDFs (<= {args.recent_days} days): " |
|
|
f"{len(recent_missing_defs)}", |
|
|
) |
|
|
for item in recent_missing_defs: |
|
|
lines.append(f"- {item}") |
|
|
lines.append("") |
|
|
lines.append( |
|
|
f"Missing context spans with recent PDFs (<= {args.recent_days} days): " |
|
|
f"{len(recent_missing_ctxs)}", |
|
|
) |
|
|
for item in recent_missing_ctxs: |
|
|
lines.append(f"- {item}") |
|
|
lines.append("") |
|
|
|
|
|
output = "\n".join(lines) + "\n" |
|
|
if args.output is not None: |
|
|
args.output.parent.mkdir(parents=True, exist_ok=True) |
|
|
args.output.write_text(output, encoding="utf-8") |
|
|
print(f"Wrote report to {args.output}") |
|
|
else: |
|
|
print(output) |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
main() |
|
|
|