Datasets:

Languages:
English
ArXiv:
License:
DefExtra / scripts /hydrate_defextra.py
bitwise31337's picture
Upload folder using huggingface_hub
fd7e968 verified
from __future__ import annotations
# ruff: noqa: E402
import argparse
import csv
import re
import subprocess
from pathlib import Path
from typing import Dict, Optional
import sys
try:
from scripts.defextra_markers import (
DocIndex,
HASH_VERSION,
TokenIndex,
build_tei_index,
doi_suffix,
extract_ids_from_tei,
extract_text_from_pdf,
hash_token_sequence,
normalize_arxiv,
normalize_doi,
normalize_paper_id,
strip_citations,
tokenize_text,
)
from scripts.defextra_pdf_aliases import candidate_pdf_aliases
except ModuleNotFoundError as exc:
if exc.name != "scripts":
raise
PROJECT_ROOT = Path(__file__).resolve().parent.parent
if str(PROJECT_ROOT) not in sys.path:
sys.path.insert(0, str(PROJECT_ROOT))
from scripts.defextra_markers import (
DocIndex,
HASH_VERSION,
TokenIndex,
build_tei_index,
doi_suffix,
extract_ids_from_tei,
extract_text_from_pdf,
hash_token_sequence,
normalize_arxiv,
normalize_doi,
normalize_paper_id,
strip_citations,
tokenize_text,
)
from scripts.defextra_pdf_aliases import candidate_pdf_aliases
TRAILING_PUNCT = set(".,;:?!)]}\"'")
END_PUNCT = {".", ",", ";", ":", "?", "!"}
TRAILING_QUOTES = {"'", '"', "”", "’", ")", "]"}
CITATION_BRACKET_RE = re.compile(r"\[[0-9][0-9,;\s\-–]*\]")
CITATION_PAREN_RE = re.compile(r"\([^)]*\d{4}[^)]*\)")
def _extend_span_end(doc_text: str, end: int) -> int:
if end < 0:
return end
limit = len(doc_text)
while end < limit and doc_text[end] in TRAILING_PUNCT:
end += 1
j = end
while j < limit and doc_text[j].isspace():
j += 1
if j < limit and doc_text[j] in TRAILING_PUNCT:
end = j + 1
while end < limit and doc_text[end] in TRAILING_PUNCT:
end += 1
return end
def _extract_with_trailing_punct(
doc_text: str,
start: Optional[int],
end: Optional[int],
) -> str:
if start is None or end is None:
return ""
if start < 0 or end > len(doc_text) or start >= end:
return ""
end = _extend_span_end(doc_text, end)
return doc_text[start:end]
def _token_count(text: str) -> int:
tokens, _ = tokenize_text(text or "", return_spans=False)
return len(tokens)
def _row_flag(row: dict, key: str, default: bool = False) -> bool:
value = (row.get(key) or "").strip().lower()
if not value:
return default
return value == "true"
def _trim_pattern(
text: str,
expected: int,
pattern: re.Pattern[str],
) -> str:
if not text or expected <= 0:
return text
while True:
current = _token_count(text)
best = text
best_diff = abs(current - expected)
improved = False
for match in pattern.finditer(text):
candidate = (
text[: match.start()] + " " + text[match.end() :]
).strip()
diff = abs(_token_count(candidate) - expected)
if diff < best_diff:
best = candidate
best_diff = diff
improved = True
if not improved:
break
text = best
if _token_count(text) <= expected:
break
return text
def _trim_citations(text: str, expected: int) -> str:
if not text or expected <= 0:
return text
current = _token_count(text)
if current <= expected:
return text
text = _trim_pattern(text, expected, CITATION_BRACKET_RE)
if _token_count(text) <= expected:
return text
text = _trim_pattern(text, expected, CITATION_PAREN_RE)
return text
def _trim_to_token_count(text: str, expected: int) -> str:
if not text or expected <= 0:
return text
tokens, spans = tokenize_text(text, return_spans=True)
if not spans or len(spans) <= expected:
return text
end_idx = spans[expected - 1][1]
end_idx = _extend_span_end(text, end_idx)
return text[:end_idx].rstrip()
def _cleanup_spacing(text: str) -> str:
if not text:
return text
value = text
value = value.replace("“", '"').replace("”", '"')
value = value.replace("’", "'").replace("‘", "'")
def _dash_repl(match: re.Match[str]) -> str:
run = match.group(0)
return "--" if len(run) >= 2 else "-"
value = re.sub(r"[\u2010-\u2015\u2212\u2043]+", _dash_repl, value)
value = value.replace("\ufb00", "ff")
value = value.replace("\ufb01", "fi")
value = value.replace("\ufb02", "fl")
value = value.replace("\ufb03", "ffi")
value = value.replace("\ufb04", "ffl")
value = value.replace("…", "...")
value = re.sub(r"-{3,}", "--", value)
value = re.sub(r"([a-z0-9])([.!?])(?=[A-Z])", r"\1\2 ", value)
value = re.sub(r"([A-Za-z]),(?=[A-Za-z])", r"\1, ", value)
value = re.sub(r"([A-Za-z0-9])([;:])(?=[A-Za-z])", r"\1\2 ", value)
value = re.sub(r"(\d)\s+(s)\b", r"\1\2", value)
value = re.sub(r"([0-9])([A-Za-z])", r"\1 \2", value)
value = re.sub(r"[ \t]+([,.;:!?])", r"\1", value)
value = re.sub(r"\(\s+", "(", value)
value = re.sub(r"\s+\)", ")", value)
value = re.sub(r"\[\s+", "[", value)
value = re.sub(r"\s+\]", "]", value)
value = re.sub(r"\)(?=[A-Za-z0-9])", ") ", value)
value = re.sub(r"\](?=[A-Za-z0-9])", "] ", value)
def _space_citation_commas(match: re.Match[str]) -> str:
inner = match.group(1)
inner = re.sub(r",(?!\\s)", ", ", inner)
return f"[{inner}]"
value = re.sub(r"\[([0-9,;\s\-–]+)\]", _space_citation_commas, value)
value = re.sub(r"[ \t]{2,}", " ", value)
return value
def _normalize_whitespace(
text: str,
preserve_linebreaks: bool,
) -> str:
if not text:
return text
value = _cleanup_spacing(text)
if preserve_linebreaks:
value = re.sub(r"[ \t]{2,}", " ", value)
return value
value = re.sub(r"\s+", " ", value).strip()
return value
def _normalize_hyphenation(
text: str,
preserve_hyphenation: bool,
) -> str:
if not text:
return text
if preserve_hyphenation:
return text
return re.sub(r"([A-Za-z])-\s+([A-Za-z])", r"\1\2", text)
def _postprocess_text(
text: str,
expected_tokens: int,
preserve_linebreaks: bool,
preserve_hyphenation: bool,
keep_bracket_citations: bool = True,
keep_paren_citations: bool = True,
split_letter_digit: bool = True,
) -> str:
value = text
if not keep_bracket_citations:
value = CITATION_BRACKET_RE.sub(" ", value)
if not keep_paren_citations:
value = CITATION_PAREN_RE.sub(" ", value)
value = _trim_citations(value, expected_tokens)
value = _trim_to_token_count(value, expected_tokens)
value = _cleanup_spacing(value)
if split_letter_digit:
value = re.sub(r"([A-Za-z])([0-9])", r"\1 \2", value)
value = _normalize_hyphenation(value, preserve_hyphenation)
return _normalize_whitespace(value, preserve_linebreaks)
def _ensure_trailing_punct(text: str, end_punct: str) -> str:
if not text or not end_punct:
stripped = text.rstrip()
if not stripped:
return text
i = len(stripped) - 1
suffix = ""
while i >= 0 and stripped[i] in TRAILING_QUOTES:
suffix = stripped[i] + suffix
i -= 1
base = stripped[: i + 1]
if base and base[-1] in END_PUNCT:
base = base[:-1]
if base and suffix:
if ")" in suffix and "(" not in base:
suffix = suffix.replace(")", "")
if "]" in suffix and "[" not in base:
suffix = suffix.replace("]", "")
return f"{base}{suffix}"
stripped = text.rstrip()
if not stripped:
return text
i = len(stripped) - 1
suffix = ""
while i >= 0 and stripped[i] in TRAILING_QUOTES:
suffix = stripped[i] + suffix
i -= 1
base = stripped[: i + 1]
if base and base[-1] in END_PUNCT:
base = base[:-1] + end_punct
else:
base = f"{base}{end_punct}"
return f"{base}{suffix}"
def _find_pdf_hash_span(
row: dict,
pdf_token_index: Optional[TokenIndex],
prefix: str,
) -> Optional[tuple[int, int]]:
if pdf_token_index is None:
return None
spec = _select_hash_specs(row, prefix)
if spec:
span = pdf_token_index.find_span_by_hash(*spec)
if span:
return span
return None
def _bool_flag(value: str) -> bool:
return (value or "").strip().lower() == "true"
def _strip_flags(row: dict, prefix: str) -> tuple[bool, bool]:
keep_bracket = _bool_flag(row.get(f"{prefix}_has_bracket_citation", ""))
keep_paren = _bool_flag(row.get(f"{prefix}_has_paren_citation", ""))
return (not keep_bracket), (not keep_paren)
def _span_matches_hash(row: dict, text: str, prefix: str) -> bool:
if not text:
return False
expected_hash = row.get(f"{prefix}_hash64") or ""
expected_sha = row.get(f"{prefix}_sha256") or ""
if not expected_hash or not expected_sha:
return False
strip_brackets, strip_parens = _strip_flags(row, prefix)
check_text = strip_citations(
text,
strip_brackets=strip_brackets,
strip_parens=strip_parens,
)
tokens, _ = tokenize_text(check_text, return_spans=True)
hash64, sha, _ = hash_token_sequence(tokens)
return str(hash64) == str(expected_hash) and sha == expected_sha
def _candidate_ids(paper_id: str, doi: str, arxiv: str) -> list[str]:
candidates = [
paper_id,
normalize_paper_id(paper_id),
]
if doi:
candidates.append(doi)
candidates.append(doi_suffix(doi))
if arxiv:
candidates.append(arxiv)
candidates.append(normalize_arxiv(arxiv))
seen = set()
ordered = []
for item in candidates:
value = (item or "").strip()
if value and value not in seen:
seen.add(value)
ordered.append(value)
for alias in candidate_pdf_aliases(paper_id, doi, arxiv):
value = (alias or "").strip()
if value and value not in seen:
seen.add(value)
ordered.append(value)
return ordered
def _normalize_title(title: str) -> str:
return " ".join(title.lower().split())
def _build_meta_index(
tei_index: Dict[str, Path],
) -> tuple[Dict[str, Path], Dict[str, Path]]:
doi_index: Dict[str, Path] = {}
arxiv_index: Dict[str, Path] = {}
for path in tei_index.values():
doi, arxiv = extract_ids_from_tei(path)
if doi:
doi_index.setdefault(normalize_doi(doi), path)
doi_index.setdefault(doi_suffix(doi), path)
if arxiv:
arxiv_index.setdefault(normalize_arxiv(arxiv), path)
return doi_index, arxiv_index
def _resolve_tei_path(
paper_id: str,
doi: str,
arxiv: str,
tei_index: Dict[str, Path],
doi_index: Dict[str, Path],
arxiv_index: Dict[str, Path],
) -> Optional[Path]:
for candidate in _candidate_ids(paper_id, doi, arxiv):
if candidate in tei_index:
return tei_index[candidate]
if candidate.startswith("paper_"):
stripped = candidate[len("paper_") :]
if stripped in tei_index:
return tei_index[stripped]
if doi:
doi_key = normalize_doi(doi)
if doi_key in doi_index:
return doi_index[doi_key]
doi_key = doi_suffix(doi)
if doi_key in doi_index:
return doi_index[doi_key]
if arxiv:
arxiv_key = normalize_arxiv(arxiv)
if arxiv_key in arxiv_index:
return arxiv_index[arxiv_key]
return None
def _tei_stem(path: Path) -> str:
name = path.name
if name.endswith(".grobid.tei.xml"):
name = name[: -len(".grobid.tei.xml")]
return name
def _build_pdf_index(pdf_dir: Path) -> Dict[str, Path]:
index: Dict[str, Path] = {}
if not pdf_dir.exists():
return index
version_re = re.compile(r"^(?P<base>.+?)(v\d+)$", re.IGNORECASE)
arxiv_re = re.compile(r"^(?P<base>\d{4}\.\d{4,5})v\d+$", re.IGNORECASE)
pii_re = re.compile(r"(S\d{8,})", re.IGNORECASE)
for suffix in ("*.pdf", "*.PDF"):
for path in pdf_dir.rglob(suffix):
stem = path.stem
index.setdefault(stem, path)
index.setdefault(normalize_paper_id(stem), path)
index.setdefault(f"paper_{stem}", path)
if stem.startswith("paper_"):
stripped = stem[len("paper_") :]
if stripped:
index.setdefault(stripped, path)
index.setdefault(normalize_paper_id(stripped), path)
if stem.endswith("_fixed") or stem.endswith("-fixed"):
base = (
stem[: -len("_fixed")]
if stem.endswith("_fixed")
else stem[: -len("-fixed")]
)
if base:
index[base] = path
index[normalize_paper_id(base)] = path
index[f"paper_{base}"] = path
if base.startswith("paper_"):
stripped_base = base[len("paper_") :]
if stripped_base:
index[stripped_base] = path
index[normalize_paper_id(stripped_base)] = path
match = arxiv_re.match(stem)
if match:
base = match.group("base")
index.setdefault(base, path)
index.setdefault(normalize_paper_id(base), path)
match = version_re.match(stem)
if match:
base = match.group("base")
index.setdefault(base, path)
index.setdefault(normalize_paper_id(base), path)
pii_match = pii_re.search(stem)
if pii_match:
pii = pii_match.group(1)
index.setdefault(pii, path)
index.setdefault(normalize_paper_id(pii), path)
return index
def _select_hash_specs(
row: dict,
prefix: str,
) -> Optional[tuple[int, int, str]]:
hash_version = (row.get("hash_version") or "").strip()
if hash_version and hash_version != HASH_VERSION:
return None
count = row.get(f"{prefix}_token_count") or ""
hash64 = row.get(f"{prefix}_hash64") or ""
sha = row.get(f"{prefix}_sha256") or ""
if not count or not hash64 or not sha:
return None
try:
return int(count), int(hash64), sha
except ValueError:
return None
def _select_anchor_specs(
row: dict,
prefix: str,
position: str,
) -> Optional[tuple[int, int, str]]:
hash_version = (row.get("hash_version") or "").strip()
if hash_version and hash_version != HASH_VERSION:
return None
count = row.get(f"{prefix}_{position}_token_count") or ""
hash64 = row.get(f"{prefix}_{position}_hash64") or ""
sha = row.get(f"{prefix}_{position}_sha256") or ""
if not count or not hash64 or not sha:
return None
try:
return int(count), int(hash64), sha
except ValueError:
return None
def _select_anchor_spec_list(
row: dict,
prefix: str,
position: str,
) -> list[tuple[int, int, str]]:
specs: list[tuple[int, int, str]] = []
primary = _select_anchor_specs(row, prefix, position)
if primary is not None:
specs.append(primary)
alt_count = row.get(f"{prefix}_{position}_alt_token_count") or ""
alt_hash = row.get(f"{prefix}_{position}_alt_hash64") or ""
alt_sha = row.get(f"{prefix}_{position}_alt_sha256") or ""
hash_version = (row.get("hash_version") or "").strip()
if hash_version and hash_version != HASH_VERSION:
return specs
if alt_count and alt_hash and alt_sha:
try:
specs.append((int(alt_count), int(alt_hash), alt_sha))
except ValueError:
pass
return specs
def _match_doc_by_hash(
token_index: TokenIndex,
hash_specs: list[tuple[int, int, str]],
) -> int:
score = 0
for window, hash64, sha in hash_specs:
if token_index.find_span_by_hash(window, hash64, sha):
score += 1
return score
def _build_mid_candidates(
token_index: TokenIndex,
mid_specs: Optional[list[tuple[int, int, str]]],
) -> list[tuple[int, int]]:
if not mid_specs:
return []
candidates: list[tuple[int, int]] = []
for spec in mid_specs:
for position in token_index.find_token_positions_by_hash(*spec):
candidates.append((position, spec[0]))
return candidates
def _span_has_mid(
mid_candidates: list[tuple[int, int]],
start_idx: int,
end_idx: int,
) -> bool:
for mid_start, mid_len in mid_candidates:
mid_end = mid_start + mid_len - 1
if mid_start >= start_idx and mid_end <= end_idx:
return True
return False
def _find_span_by_anchors(
token_index: TokenIndex,
head_spec: Optional[tuple[int, int, str]],
tail_spec: Optional[tuple[int, int, str]],
expected_len: int,
mid_specs: Optional[list[tuple[int, int, str]]] = None,
*,
require_mid: bool = False,
) -> Optional[tuple[int, int]]:
if head_spec is None or tail_spec is None:
return None
head_positions = token_index.find_token_positions_by_hash(*head_spec)
tail_positions = token_index.find_token_positions_by_hash(*tail_spec)
if not head_positions or not tail_positions:
return None
mid_candidates = []
if require_mid:
mid_candidates = _build_mid_candidates(token_index, mid_specs)
if not mid_candidates:
return None
best = None
best_diff = None
tol = max(5, int(expected_len * 0.3)) if expected_len else 10
min_len = max(1, expected_len // 2) if expected_len else 1
max_len = expected_len * 3 if expected_len else None
for head_start in head_positions:
head_end = head_start + head_spec[0] - 1
for tail_start in tail_positions:
tail_end = tail_start + tail_spec[0] - 1
if tail_end < head_end:
continue
length = tail_end - head_start + 1
if mid_candidates and not _span_has_mid(
mid_candidates,
head_start,
tail_end,
):
continue
if expected_len:
if length < min_len or (max_len and length > max_len):
continue
if length < expected_len - tol or length > expected_len + tol:
continue
diff = abs(length - expected_len)
else:
diff = length
if best_diff is None or diff < best_diff:
best_diff = diff
best = (head_start, tail_end)
if best is None and expected_len:
for head_start in head_positions:
head_end = head_start + head_spec[0] - 1
for tail_start in tail_positions:
tail_end = tail_start + tail_spec[0] - 1
if tail_end < head_end:
continue
length = tail_end - head_start + 1
if mid_candidates and not _span_has_mid(
mid_candidates,
head_start,
tail_end,
):
continue
if length < min_len or (max_len and length > max_len):
continue
diff = abs(length - expected_len)
if best_diff is None or diff < best_diff:
best_diff = diff
best = (head_start, tail_end)
if best is None:
return None
start_char = token_index.spans[best[0]][0]
end_char = token_index.spans[best[1]][1]
return start_char, end_char
def _find_span_from_anchor(
token_index: TokenIndex,
anchor_spec: Optional[tuple[int, int, str]],
expected_len: int,
position: str,
mid_specs: Optional[list[tuple[int, int, str]]] = None,
*,
require_mid: bool = False,
require_unique: bool = False,
) -> Optional[tuple[int, int]]:
if anchor_spec is None or expected_len <= 0:
return None
positions = token_index.find_token_positions_by_hash(*anchor_spec)
if not positions:
return None
if require_unique and len(positions) != 1:
return None
mid_candidates = []
if require_mid:
mid_candidates = _build_mid_candidates(token_index, mid_specs)
if not mid_candidates:
return None
if position == "tail":
positions = list(reversed(positions))
for anchor_start in positions:
if position == "head":
start_idx = anchor_start
end_idx = anchor_start + expected_len - 1
else:
anchor_end = anchor_start + anchor_spec[0] - 1
end_idx = anchor_end
start_idx = end_idx - expected_len + 1
if start_idx < 0 or end_idx >= len(token_index.tokens):
continue
if mid_candidates and not _span_has_mid(
mid_candidates,
start_idx,
end_idx,
):
continue
start_char = token_index.spans[start_idx][0]
end_char = token_index.spans[end_idx][1]
return start_char, end_char
return None
def _pick_best_doc(
token_indexes: Dict[Path, TokenIndex],
hash_specs: list[tuple[int, int, str]],
) -> Optional[Path]:
best_path = None
best_score = 0
tie = False
for path, token_index in token_indexes.items():
score = _match_doc_by_hash(token_index, hash_specs)
if score > best_score:
best_score = score
best_path = path
tie = False
elif score == best_score and score > 0:
tie = True
if best_score == 0 or tie:
return None
return best_path
def _pick_best_pdf(
pdf_paths: list[Path],
pdf_token_cache: Dict[Path, TokenIndex],
hash_specs: list[tuple[int, int, str]],
) -> Optional[Path]:
best_path = None
best_score = 0
tie = False
for path in pdf_paths:
if path not in pdf_token_cache:
pdf_text = extract_text_from_pdf(path)
pdf_token_cache[path] = TokenIndex.from_text(pdf_text)
token_index = pdf_token_cache[path]
score = _match_doc_by_hash(token_index, hash_specs)
if score > best_score:
best_score = score
best_path = path
tie = False
elif score == best_score and score > 0:
tie = True
if best_score == 0 or tie:
return None
return best_path
def _run_grobid(input_dir: Path, output_dir: Path, config: Path) -> None:
output_dir.mkdir(parents=True, exist_ok=True)
cmd = [
sys.executable,
"scripts/pdf_to_grobid.py",
"--input_folder",
str(input_dir),
"--output_folder",
str(output_dir),
]
if config and config.exists():
cmd.extend(["--config", str(config)])
subprocess.run(cmd, check=True)
def main() -> None:
parser = argparse.ArgumentParser(
description="Hydrate DefExtra legal CSV using user-provided PDFs.",
)
parser.add_argument(
"--legal-csv",
type=Path,
default=Path("results/paper_results/defextra_legal.csv"),
help="Legal DefExtra CSV with markers.",
)
parser.add_argument(
"--pdf-dir",
type=Path,
required=True,
help="Directory with user-provided PDFs.",
)
parser.add_argument(
"--grobid-out",
type=Path,
default=Path("outputs/defextra_grobid"),
help="Output directory for GROBID TEI files.",
)
parser.add_argument(
"--grobid-config",
type=Path,
default=Path("config.json"),
help="Optional GROBID client config path.",
)
parser.add_argument(
"--skip-grobid",
action="store_true",
help="Skip running GROBID (expects TEI files already present).",
)
parser.add_argument(
"--output-csv",
type=Path,
default=Path("results/paper_results/defextra_hydrated.csv"),
help="Output hydrated CSV with excerpts.",
)
parser.add_argument(
"--report",
type=Path,
default=None,
help="Optional report path for missing matches.",
)
parser.add_argument(
"--require-complete",
action="store_true",
help="Exit with error if any definition/context is missing.",
)
parser.add_argument(
"--filter-to-pdfs",
action="store_true",
help="Only process rows that can be mapped to a provided PDF.",
)
parser.add_argument(
"--allow-pdf-hash-mismatch",
action="store_true",
help=(
"Continue when a PDF filename matches but hash markers do not. "
"By default, such PDFs are skipped and reported."
),
)
args = parser.parse_args()
if not args.legal_csv.exists():
raise SystemExit(f"Legal CSV not found: {args.legal_csv}")
if not args.pdf_dir.exists():
raise SystemExit(f"PDF dir not found: {args.pdf_dir}")
if not args.skip_grobid:
try:
_run_grobid(args.pdf_dir, args.grobid_out, args.grobid_config)
except subprocess.CalledProcessError:
raise SystemExit(
"GROBID processing failed. Ensure the GROBID server is running "
"and reachable (default: http://localhost:8070), or supply "
"--grobid-config with the correct server URL.",
)
tei_index = build_tei_index([args.grobid_out])
doi_index, arxiv_index = _build_meta_index(tei_index)
pdf_index = _build_pdf_index(args.pdf_dir)
doc_cache: Dict[str, Optional[DocIndex]] = {}
token_cache: Dict[str, Optional[TokenIndex]] = {}
tei_path_cache: Dict[str, Optional[Path]] = {}
pdf_token_cache: Dict[Path, TokenIndex] = {}
pdf_token_cache_stripped: Dict[tuple[Path, bool, bool], TokenIndex] = {}
tei_token_cache_stripped: Dict[tuple[Path, bool, bool], TokenIndex] = {}
pdf_failed: set[Path] = set()
def _get_stripped_index(
cache: Dict[tuple[Path, bool, bool], TokenIndex],
source_path: Optional[Path],
source_text: str,
strip_brackets: bool,
strip_parens: bool,
) -> Optional[TokenIndex]:
if source_path is None:
return None
if not strip_brackets and not strip_parens:
return None
key = (source_path, strip_brackets, strip_parens)
if key not in cache:
stripped = strip_citations(
source_text,
strip_brackets=strip_brackets,
strip_parens=strip_parens,
)
cache[key] = TokenIndex.from_text(stripped)
return cache[key]
with args.legal_csv.open("r", encoding="utf-8", newline="") as handle:
reader = csv.DictReader(handle)
legal_rows = list(reader)
paper_hashes: Dict[str, list[tuple[int, int, str]]] = {}
title_to_ids: Dict[str, list[str]] = {}
id_to_row: Dict[str, dict] = {}
for row in legal_rows:
paper_id = (row.get("paper_id") or "").strip()
if not paper_id:
continue
if paper_id not in id_to_row:
id_to_row[paper_id] = row
title_key = _normalize_title(row.get("paper_title") or "")
if title_key:
title_to_ids.setdefault(title_key, []).append(paper_id)
specs: list[tuple[int, int, str]] = []
for prefix in ("definition", "context"):
spec = _select_hash_specs(row, prefix)
if spec is None:
continue
token_count, _, _ = spec
if token_count >= 5:
specs.append(spec)
if specs:
paper_hashes.setdefault(paper_id, []).extend(specs)
for prefix in ("definition", "context"):
for position in ("head", "mid", "tail"):
for spec in _select_anchor_spec_list(row, prefix, position):
if spec and spec[0] >= 5:
paper_hashes.setdefault(paper_id, []).append(spec)
tei_token_indexes: Dict[Path, TokenIndex] = {}
allowed_stems = set(pdf_index.keys()) if pdf_index else set()
for tei_path in tei_index.values():
if allowed_stems:
stem = _tei_stem(tei_path)
stem_norm = normalize_paper_id(stem)
stem_stripped = (
stem[len("paper_") :] if stem.startswith("paper_") else stem
)
if (
stem not in allowed_stems
and stem_norm not in allowed_stems
and stem_stripped not in allowed_stems
):
continue
try:
doc_index = DocIndex.from_tei(tei_path)
except Exception:
continue
tei_token_indexes[tei_path] = TokenIndex.from_text(doc_index.doc_text)
output_rows = []
missing_papers = set()
missing_defs = 0
missing_ctxs = 0
hydrated_from_pdf = 0
hydrated_from_anchor = 0
pdf_hash_mismatches: list[dict[str, str]] = []
pdf_hash_mismatch_seen: set[tuple[str, str]] = set()
missing_def_rows: list[dict] = []
missing_ctx_rows: list[dict] = []
for row in legal_rows:
paper_id = (row.get("paper_id") or "").strip()
doi = (row.get("paper_doi") or "").strip()
arxiv = (row.get("paper_arxiv") or "").strip()
if paper_id not in doc_cache:
tei_path = _resolve_tei_path(
paper_id,
doi,
arxiv,
tei_index,
doi_index,
arxiv_index,
)
if tei_path is None:
hash_specs = paper_hashes.get(paper_id, [])
if hash_specs:
tei_path = _pick_best_doc(
tei_token_indexes,
hash_specs,
)
if tei_path is None:
doc_cache[paper_id] = None
token_cache[paper_id] = None
tei_path_cache[paper_id] = None
else:
doc_index = DocIndex.from_tei(tei_path)
doc_cache[paper_id] = doc_index
token_cache[paper_id] = TokenIndex.from_text(
doc_index.doc_text,
)
tei_path_cache[paper_id] = tei_path
doc_index = doc_cache.get(paper_id)
tei_token_index = token_cache.get(paper_id)
definition = ""
context = ""
pdf_token_index: Optional[TokenIndex] = None
pdf_path = None
tei_path = tei_path_cache.get(paper_id)
pdf_direct_match = False
if tei_path is not None:
stem = _tei_stem(tei_path)
pdf_path = pdf_index.get(stem) or pdf_index.get(
normalize_paper_id(stem),
)
if pdf_path is not None:
pdf_direct_match = True
if pdf_path is None:
for candidate in _candidate_ids(paper_id, doi, arxiv):
pdf_path = pdf_index.get(candidate)
if pdf_path:
pdf_direct_match = True
break
if pdf_path is None:
title_key = _normalize_title(row.get("paper_title") or "")
for other_id in title_to_ids.get(title_key, []):
if other_id == paper_id:
continue
other_row = id_to_row.get(other_id, {})
other_doi = (other_row.get("paper_doi") or "").strip()
other_arxiv = (other_row.get("paper_arxiv") or "").strip()
for candidate in _candidate_ids(
other_id,
other_doi,
other_arxiv,
):
pdf_path = pdf_index.get(candidate)
if pdf_path:
pdf_direct_match = True
break
if pdf_path:
break
hash_specs = paper_hashes.get(paper_id, [])
if pdf_path is not None and hash_specs:
if pdf_path not in pdf_token_cache:
pdf_text = extract_text_from_pdf(pdf_path)
pdf_token_cache[pdf_path] = TokenIndex.from_text(pdf_text)
pdf_token_index = pdf_token_cache[pdf_path]
if _match_doc_by_hash(pdf_token_index, hash_specs) == 0:
mismatch_key = (paper_id, str(pdf_path))
if mismatch_key not in pdf_hash_mismatch_seen:
pdf_hash_mismatch_seen.add(mismatch_key)
pdf_hash_mismatches.append(
{"paper_id": paper_id, "pdf": str(pdf_path)},
)
if not args.allow_pdf_hash_mismatch and pdf_direct_match:
print(
f"Warning: PDF hash markers did not match for {pdf_path.name}; "
"skipping PDF (use --allow-pdf-hash-mismatch to override).",
file=sys.stderr,
)
elif args.allow_pdf_hash_mismatch:
print(
f"Warning: PDF hash markers did not match for {pdf_path.name}; "
"continuing with direct filename match.",
file=sys.stderr,
)
if not args.allow_pdf_hash_mismatch:
pdf_path = None
pdf_token_index = None
pdf_direct_match = False
if pdf_path is None and hash_specs:
pdf_paths = list({p for p in pdf_index.values()})
if pdf_paths:
pdf_path = _pick_best_pdf(
pdf_paths,
pdf_token_cache,
hash_specs,
)
if args.filter_to_pdfs and pdf_path is None:
continue
if pdf_path is not None and pdf_path not in pdf_token_cache:
try:
pdf_text = extract_text_from_pdf(pdf_path)
pdf_token_cache[pdf_path] = TokenIndex.from_text(pdf_text)
except Exception as exc:
pdf_token_cache[pdf_path] = TokenIndex.from_text("")
if pdf_path not in pdf_failed:
pdf_failed.add(pdf_path)
print(
f"Warning: PDF text extraction failed for {pdf_path.name}: {exc}",
file=sys.stderr,
)
pdf_token_index = pdf_token_cache.get(pdf_path) if pdf_path else None
if doc_index is None:
missing_papers.add(paper_id)
else:
def_start = row.get("definition_char_start") or ""
def_end = row.get("definition_char_end") or ""
ctx_start = row.get("context_char_start") or ""
ctx_end = row.get("context_char_end") or ""
def_strip_brackets, def_strip_parens = _strip_flags(
row,
"definition",
)
ctx_strip_brackets, ctx_strip_parens = _strip_flags(row, "context")
if not definition and tei_token_index:
spec = _select_hash_specs(row, "definition")
if spec:
span = tei_token_index.find_span_by_hash(*spec)
if span:
definition = _extract_with_trailing_punct(
doc_index.doc_text,
span[0],
span[1],
)
if not definition:
stripped_index = _get_stripped_index(
tei_token_cache_stripped,
tei_path,
doc_index.doc_text,
def_strip_brackets,
def_strip_parens,
)
if stripped_index is not None:
span = stripped_index.find_span_by_hash(*spec)
if span:
definition = _extract_with_trailing_punct(
stripped_index.doc_text,
span[0],
span[1],
)
if not definition and not (def_start and def_end):
head_specs = _select_anchor_spec_list(
row,
"definition",
"head",
)
mid_specs = _select_anchor_spec_list(
row,
"definition",
"mid",
)
tail_specs = _select_anchor_spec_list(
row,
"definition",
"tail",
)
expected_len = int(row.get("definition_token_count") or 0)
for head_spec in head_specs or [None]:
for tail_spec in tail_specs or [None]:
if head_spec is None or tail_spec is None:
continue
span = _find_span_by_anchors(
tei_token_index,
head_spec,
tail_spec,
expected_len,
mid_specs,
require_mid=True,
)
if span is None:
span = _find_span_by_anchors(
tei_token_index,
head_spec,
tail_spec,
expected_len,
)
if span:
definition = _extract_with_trailing_punct(
doc_index.doc_text,
span[0],
span[1],
)
break
if definition:
break
if not definition:
head_specs = _select_anchor_spec_list(
row,
"definition",
"head",
)
mid_specs = _select_anchor_spec_list(
row,
"definition",
"mid",
)
tail_specs = _select_anchor_spec_list(
row,
"definition",
"tail",
)
expected_len = int(row.get("definition_token_count") or 0)
span = None
for head_spec in head_specs:
if mid_specs:
span = _find_span_from_anchor(
tei_token_index,
head_spec,
expected_len,
"head",
mid_specs,
require_mid=True,
)
if span is None:
span = _find_span_from_anchor(
tei_token_index,
head_spec,
expected_len,
"head",
mid_specs,
require_unique=True,
)
else:
span = _find_span_from_anchor(
tei_token_index,
head_spec,
expected_len,
"head",
mid_specs,
)
if span:
break
if span is None:
for tail_spec in tail_specs:
if mid_specs:
span = _find_span_from_anchor(
tei_token_index,
tail_spec,
expected_len,
"tail",
mid_specs,
require_mid=True,
)
if span is None:
span = _find_span_from_anchor(
tei_token_index,
tail_spec,
expected_len,
"tail",
mid_specs,
require_unique=True,
)
elif span is None:
span = _find_span_from_anchor(
tei_token_index,
tail_spec,
expected_len,
"tail",
mid_specs,
)
if span:
break
if span:
definition = _extract_with_trailing_punct(
doc_index.doc_text,
span[0],
span[1],
)
if not definition and def_start and def_end:
candidate = _extract_with_trailing_punct(
doc_index.doc_text,
int(def_start),
int(def_end),
)
if _span_matches_hash(row, candidate, "definition"):
definition = candidate
if not definition and pdf_token_index:
span = _find_pdf_hash_span(row, pdf_token_index, "definition")
if span:
definition = _extract_with_trailing_punct(
pdf_token_index.doc_text,
span[0],
span[1],
)
hydrated_from_pdf += 1
if not definition:
stripped_index = _get_stripped_index(
pdf_token_cache_stripped,
pdf_path,
pdf_token_index.doc_text,
def_strip_brackets,
def_strip_parens,
)
if stripped_index is not None:
span = _find_pdf_hash_span(
row,
stripped_index,
"definition",
)
if span:
definition = _extract_with_trailing_punct(
stripped_index.doc_text,
span[0],
span[1],
)
hydrated_from_pdf += 1
if not context and tei_token_index:
spec = _select_hash_specs(row, "context")
if spec:
span = tei_token_index.find_span_by_hash(*spec)
if span:
context = _extract_with_trailing_punct(
doc_index.doc_text,
span[0],
span[1],
)
if not context:
stripped_index = _get_stripped_index(
tei_token_cache_stripped,
tei_path,
doc_index.doc_text,
ctx_strip_brackets,
ctx_strip_parens,
)
if stripped_index is not None:
span = stripped_index.find_span_by_hash(*spec)
if span:
context = _extract_with_trailing_punct(
stripped_index.doc_text,
span[0],
span[1],
)
if not context and not (ctx_start and ctx_end):
head_specs = _select_anchor_spec_list(
row,
"context",
"head",
)
mid_specs = _select_anchor_spec_list(
row,
"context",
"mid",
)
tail_specs = _select_anchor_spec_list(
row,
"context",
"tail",
)
expected_len = int(row.get("context_token_count") or 0)
for head_spec in head_specs or [None]:
for tail_spec in tail_specs or [None]:
if head_spec is None or tail_spec is None:
continue
span = _find_span_by_anchors(
tei_token_index,
head_spec,
tail_spec,
expected_len,
mid_specs,
require_mid=True,
)
if span is None:
span = _find_span_by_anchors(
tei_token_index,
head_spec,
tail_spec,
expected_len,
)
if span:
context = _extract_with_trailing_punct(
doc_index.doc_text,
span[0],
span[1],
)
break
if context:
break
if not context:
head_specs = _select_anchor_spec_list(
row,
"context",
"head",
)
mid_specs = _select_anchor_spec_list(
row,
"context",
"mid",
)
tail_specs = _select_anchor_spec_list(
row,
"context",
"tail",
)
expected_len = int(row.get("context_token_count") or 0)
span = None
for head_spec in head_specs:
if mid_specs:
span = _find_span_from_anchor(
tei_token_index,
head_spec,
expected_len,
"head",
mid_specs,
require_mid=True,
)
if span is None:
span = _find_span_from_anchor(
tei_token_index,
head_spec,
expected_len,
"head",
mid_specs,
require_unique=True,
)
else:
span = _find_span_from_anchor(
tei_token_index,
head_spec,
expected_len,
"head",
mid_specs,
)
if span:
break
if span is None:
for tail_spec in tail_specs:
if mid_specs:
span = _find_span_from_anchor(
tei_token_index,
tail_spec,
expected_len,
"tail",
mid_specs,
require_mid=True,
)
if span is None:
span = _find_span_from_anchor(
tei_token_index,
tail_spec,
expected_len,
"tail",
mid_specs,
require_unique=True,
)
elif span is None:
span = _find_span_from_anchor(
tei_token_index,
tail_spec,
expected_len,
"tail",
mid_specs,
)
if span:
break
if span:
context = _extract_with_trailing_punct(
doc_index.doc_text,
span[0],
span[1],
)
if not context and ctx_start and ctx_end:
candidate = _extract_with_trailing_punct(
doc_index.doc_text,
int(ctx_start),
int(ctx_end),
)
if _span_matches_hash(row, candidate, "context"):
context = candidate
if not context and pdf_token_index:
span = _find_pdf_hash_span(row, pdf_token_index, "context")
if span:
context = _extract_with_trailing_punct(
pdf_token_index.doc_text,
span[0],
span[1],
)
hydrated_from_pdf += 1
if not context:
stripped_index = _get_stripped_index(
pdf_token_cache_stripped,
pdf_path,
pdf_token_index.doc_text,
ctx_strip_brackets,
ctx_strip_parens,
)
if stripped_index is not None:
span = _find_pdf_hash_span(
row,
stripped_index,
"context",
)
if span:
context = _extract_with_trailing_punct(
stripped_index.doc_text,
span[0],
span[1],
)
hydrated_from_pdf += 1
if not definition and pdf_path is not None and pdf_token_index:
spec = _select_hash_specs(row, "definition")
if spec:
span = pdf_token_index.find_span_by_hash(*spec)
if span:
definition = _extract_with_trailing_punct(
pdf_token_index.doc_text,
span[0],
span[1],
)
hydrated_from_pdf += 1
if not definition:
head_specs = _select_anchor_spec_list(
row,
"definition",
"head",
)
mid_specs = _select_anchor_spec_list(
row,
"definition",
"mid",
)
tail_specs = _select_anchor_spec_list(
row,
"definition",
"tail",
)
expected_len = int(row.get("definition_token_count") or 0)
for head_spec in head_specs or [None]:
for tail_spec in tail_specs or [None]:
if head_spec is None or tail_spec is None:
continue
span = _find_span_by_anchors(
pdf_token_index,
head_spec,
tail_spec,
expected_len,
mid_specs,
require_mid=True,
)
if span is None:
span = _find_span_by_anchors(
pdf_token_index,
head_spec,
tail_spec,
expected_len,
)
if span:
definition = _extract_with_trailing_punct(
pdf_token_index.doc_text,
span[0],
span[1],
)
hydrated_from_pdf += 1
hydrated_from_anchor += 1
break
if definition:
break
if not definition:
head_specs = _select_anchor_spec_list(
row,
"definition",
"head",
)
mid_specs = _select_anchor_spec_list(
row,
"definition",
"mid",
)
tail_specs = _select_anchor_spec_list(
row,
"definition",
"tail",
)
expected_len = int(row.get("definition_token_count") or 0)
span = None
for head_spec in head_specs:
if mid_specs:
span = _find_span_from_anchor(
pdf_token_index,
head_spec,
expected_len,
"head",
mid_specs,
require_mid=True,
)
if span is None:
span = _find_span_from_anchor(
pdf_token_index,
head_spec,
expected_len,
"head",
mid_specs,
require_unique=True,
)
else:
span = _find_span_from_anchor(
pdf_token_index,
head_spec,
expected_len,
"head",
mid_specs,
)
if span:
break
if span is None:
for tail_spec in tail_specs:
if mid_specs:
span = _find_span_from_anchor(
pdf_token_index,
tail_spec,
expected_len,
"tail",
mid_specs,
require_mid=True,
)
if span is None:
span = _find_span_from_anchor(
pdf_token_index,
tail_spec,
expected_len,
"tail",
mid_specs,
require_unique=True,
)
else:
span = _find_span_from_anchor(
pdf_token_index,
tail_spec,
expected_len,
"tail",
mid_specs,
)
if span:
break
if span:
definition = _extract_with_trailing_punct(
pdf_token_index.doc_text,
span[0],
span[1],
)
hydrated_from_pdf += 1
hydrated_from_anchor += 1
if not definition:
missing_defs += 1
missing_def_rows.append(
{
"paper_id": paper_id,
"concept": row.get("concept", ""),
"reason": "missing_definition",
},
)
if not context and pdf_path is not None and pdf_token_index:
spec = _select_hash_specs(row, "context")
if spec:
span = pdf_token_index.find_span_by_hash(*spec)
if span:
context = _extract_with_trailing_punct(
pdf_token_index.doc_text,
span[0],
span[1],
)
hydrated_from_pdf += 1
if not context:
head_specs = _select_anchor_spec_list(
row,
"context",
"head",
)
mid_specs = _select_anchor_spec_list(
row,
"context",
"mid",
)
tail_specs = _select_anchor_spec_list(
row,
"context",
"tail",
)
expected_len = int(row.get("context_token_count") or 0)
for head_spec in head_specs or [None]:
for tail_spec in tail_specs or [None]:
if head_spec is None or tail_spec is None:
continue
span = _find_span_by_anchors(
pdf_token_index,
head_spec,
tail_spec,
expected_len,
mid_specs,
require_mid=True,
)
if span is None:
span = _find_span_by_anchors(
pdf_token_index,
head_spec,
tail_spec,
expected_len,
)
if span:
context = _extract_with_trailing_punct(
pdf_token_index.doc_text,
span[0],
span[1],
)
hydrated_from_pdf += 1
hydrated_from_anchor += 1
break
if context:
break
if not context:
head_specs = _select_anchor_spec_list(
row,
"context",
"head",
)
mid_specs = _select_anchor_spec_list(
row,
"context",
"mid",
)
tail_specs = _select_anchor_spec_list(
row,
"context",
"tail",
)
expected_len = int(row.get("context_token_count") or 0)
span = None
for head_spec in head_specs:
if mid_specs:
span = _find_span_from_anchor(
pdf_token_index,
head_spec,
expected_len,
"head",
mid_specs,
require_mid=True,
)
if span is None:
span = _find_span_from_anchor(
pdf_token_index,
head_spec,
expected_len,
"head",
mid_specs,
require_unique=True,
)
else:
span = _find_span_from_anchor(
pdf_token_index,
head_spec,
expected_len,
"head",
mid_specs,
)
if span:
break
if span is None:
for tail_spec in tail_specs:
if mid_specs:
span = _find_span_from_anchor(
pdf_token_index,
tail_spec,
expected_len,
"tail",
mid_specs,
require_mid=True,
)
if span is None:
span = _find_span_from_anchor(
pdf_token_index,
tail_spec,
expected_len,
"tail",
mid_specs,
require_unique=True,
)
else:
span = _find_span_from_anchor(
pdf_token_index,
tail_spec,
expected_len,
"tail",
mid_specs,
)
if span:
break
if span:
context = _extract_with_trailing_punct(
pdf_token_index.doc_text,
span[0],
span[1],
)
hydrated_from_pdf += 1
hydrated_from_anchor += 1
if not context:
missing_ctxs += 1
missing_ctx_rows.append(
{
"paper_id": paper_id,
"concept": row.get("concept", ""),
"reason": "missing_context",
},
)
def_preserve_lines = _row_flag(
row,
"definition_preserve_linebreaks",
)
ctx_preserve_lines = _row_flag(
row,
"context_preserve_linebreaks",
)
def_preserve_hyph = _row_flag(
row,
"definition_preserve_hyphenation",
)
ctx_preserve_hyph = _row_flag(
row,
"context_preserve_hyphenation",
)
def_keep_bracket = _row_flag(
row,
"definition_has_bracket_citation",
True,
)
def_keep_paren = _row_flag(
row,
"definition_has_paren_citation",
True,
)
def_split_letter_digit = not _row_flag(
row,
"definition_has_letter_digit",
)
ctx_keep_bracket = _row_flag(
row,
"context_has_bracket_citation",
True,
)
ctx_keep_paren = _row_flag(
row,
"context_has_paren_citation",
True,
)
ctx_split_letter_digit = not _row_flag(
row,
"context_has_letter_digit",
)
output_rows.append(
{
"paper_id": paper_id,
"paper_title": row.get("paper_title", ""),
"paper_doi": doi,
"paper_arxiv": arxiv,
"concept": row.get("concept", ""),
"definition": _ensure_trailing_punct(
_postprocess_text(
definition,
int(row.get("definition_token_count") or 0),
def_preserve_lines,
def_preserve_hyph,
def_keep_bracket,
def_keep_paren,
def_split_letter_digit,
),
row.get("definition_end_punct", ""),
),
"context": _ensure_trailing_punct(
_postprocess_text(
context,
int(row.get("context_token_count") or 0),
ctx_preserve_lines,
ctx_preserve_hyph,
ctx_keep_bracket,
ctx_keep_paren,
ctx_split_letter_digit,
),
row.get("context_end_punct", ""),
),
"definition_type": row.get("definition_type", ""),
"source_file": row.get("source_file", ""),
"is_out_of_domain": row.get("is_out_of_domain", ""),
},
)
args.output_csv.parent.mkdir(parents=True, exist_ok=True)
with args.output_csv.open("w", encoding="utf-8", newline="") as handle:
fieldnames = [
"paper_id",
"paper_title",
"paper_doi",
"paper_arxiv",
"concept",
"definition",
"context",
"definition_type",
"source_file",
"is_out_of_domain",
]
writer = csv.DictWriter(handle, fieldnames=fieldnames)
writer.writeheader()
for row in output_rows:
writer.writerow(row)
print(f"Wrote hydrated CSV to {args.output_csv}")
print(f"Missing TEI for {len(missing_papers)} papers")
print(f"Missing definition spans: {missing_defs}")
print(f"Missing context spans: {missing_ctxs}")
print(
"Hydrated from PDF fallback: "
f"{hydrated_from_pdf} (anchors used: {hydrated_from_anchor})",
)
if args.report is not None:
report_lines = []
if missing_papers:
report_lines.append("Missing papers:")
for paper_id in sorted(missing_papers):
report_lines.append(f"- {paper_id}")
report_lines.append("")
if pdf_hash_mismatches:
report_lines.append(
"PDF hash mismatches (filename matched, hash did not):",
)
for item in pdf_hash_mismatches:
report_lines.append(
f"- {item['paper_id']} | {item['pdf']}",
)
report_lines.append(
"Note: rerun with --allow-pdf-hash-mismatch to continue with these PDFs.",
)
report_lines.append("")
report_lines.append(f"Missing definition spans: {missing_defs}")
report_lines.append(f"Missing context spans: {missing_ctxs}")
if missing_def_rows:
report_lines.append("")
report_lines.append("Missing definitions (paper_id | concept):")
for item in missing_def_rows:
report_lines.append(
f"- {item['paper_id']} | {item['concept']}",
)
if missing_ctx_rows:
report_lines.append("")
report_lines.append("Missing contexts (paper_id | concept):")
for item in missing_ctx_rows:
report_lines.append(
f"- {item['paper_id']} | {item['concept']}",
)
args.report.parent.mkdir(parents=True, exist_ok=True)
args.report.write_text(
"\n".join(report_lines) + "\n",
encoding="utf-8",
)
print(f"Wrote report to {args.report}")
if args.require_complete and (missing_defs or missing_ctxs):
raise SystemExit(
"Hydration incomplete: "
f"{missing_defs} definitions, {missing_ctxs} contexts missing.",
)
if __name__ == "__main__":
main()