|
|
from __future__ import annotations |
|
|
|
|
|
|
|
|
|
|
|
import argparse |
|
|
import csv |
|
|
import re |
|
|
import subprocess |
|
|
from pathlib import Path |
|
|
from typing import Dict, Optional |
|
|
|
|
|
import sys |
|
|
|
|
|
try: |
|
|
from scripts.defextra_markers import ( |
|
|
DocIndex, |
|
|
HASH_VERSION, |
|
|
TokenIndex, |
|
|
build_tei_index, |
|
|
doi_suffix, |
|
|
extract_ids_from_tei, |
|
|
extract_text_from_pdf, |
|
|
hash_token_sequence, |
|
|
normalize_arxiv, |
|
|
normalize_doi, |
|
|
normalize_paper_id, |
|
|
strip_citations, |
|
|
tokenize_text, |
|
|
) |
|
|
from scripts.defextra_pdf_aliases import candidate_pdf_aliases |
|
|
except ModuleNotFoundError as exc: |
|
|
if exc.name != "scripts": |
|
|
raise |
|
|
PROJECT_ROOT = Path(__file__).resolve().parent.parent |
|
|
if str(PROJECT_ROOT) not in sys.path: |
|
|
sys.path.insert(0, str(PROJECT_ROOT)) |
|
|
from scripts.defextra_markers import ( |
|
|
DocIndex, |
|
|
HASH_VERSION, |
|
|
TokenIndex, |
|
|
build_tei_index, |
|
|
doi_suffix, |
|
|
extract_ids_from_tei, |
|
|
extract_text_from_pdf, |
|
|
hash_token_sequence, |
|
|
normalize_arxiv, |
|
|
normalize_doi, |
|
|
normalize_paper_id, |
|
|
strip_citations, |
|
|
tokenize_text, |
|
|
) |
|
|
from scripts.defextra_pdf_aliases import candidate_pdf_aliases |
|
|
|
|
|
TRAILING_PUNCT = set(".,;:?!)]}\"'") |
|
|
END_PUNCT = {".", ",", ";", ":", "?", "!"} |
|
|
TRAILING_QUOTES = {"'", '"', "”", "’", ")", "]"} |
|
|
CITATION_BRACKET_RE = re.compile(r"\[[0-9][0-9,;\s\-–]*\]") |
|
|
CITATION_PAREN_RE = re.compile(r"\([^)]*\d{4}[^)]*\)") |
|
|
|
|
|
|
|
|
def _extend_span_end(doc_text: str, end: int) -> int: |
|
|
if end < 0: |
|
|
return end |
|
|
limit = len(doc_text) |
|
|
while end < limit and doc_text[end] in TRAILING_PUNCT: |
|
|
end += 1 |
|
|
j = end |
|
|
while j < limit and doc_text[j].isspace(): |
|
|
j += 1 |
|
|
if j < limit and doc_text[j] in TRAILING_PUNCT: |
|
|
end = j + 1 |
|
|
while end < limit and doc_text[end] in TRAILING_PUNCT: |
|
|
end += 1 |
|
|
return end |
|
|
|
|
|
|
|
|
def _extract_with_trailing_punct( |
|
|
doc_text: str, |
|
|
start: Optional[int], |
|
|
end: Optional[int], |
|
|
) -> str: |
|
|
if start is None or end is None: |
|
|
return "" |
|
|
if start < 0 or end > len(doc_text) or start >= end: |
|
|
return "" |
|
|
end = _extend_span_end(doc_text, end) |
|
|
return doc_text[start:end] |
|
|
|
|
|
|
|
|
def _token_count(text: str) -> int: |
|
|
tokens, _ = tokenize_text(text or "", return_spans=False) |
|
|
return len(tokens) |
|
|
|
|
|
|
|
|
def _row_flag(row: dict, key: str, default: bool = False) -> bool: |
|
|
value = (row.get(key) or "").strip().lower() |
|
|
if not value: |
|
|
return default |
|
|
return value == "true" |
|
|
|
|
|
|
|
|
def _trim_pattern( |
|
|
text: str, |
|
|
expected: int, |
|
|
pattern: re.Pattern[str], |
|
|
) -> str: |
|
|
if not text or expected <= 0: |
|
|
return text |
|
|
while True: |
|
|
current = _token_count(text) |
|
|
best = text |
|
|
best_diff = abs(current - expected) |
|
|
improved = False |
|
|
for match in pattern.finditer(text): |
|
|
candidate = ( |
|
|
text[: match.start()] + " " + text[match.end() :] |
|
|
).strip() |
|
|
diff = abs(_token_count(candidate) - expected) |
|
|
if diff < best_diff: |
|
|
best = candidate |
|
|
best_diff = diff |
|
|
improved = True |
|
|
if not improved: |
|
|
break |
|
|
text = best |
|
|
if _token_count(text) <= expected: |
|
|
break |
|
|
return text |
|
|
|
|
|
|
|
|
def _trim_citations(text: str, expected: int) -> str: |
|
|
if not text or expected <= 0: |
|
|
return text |
|
|
current = _token_count(text) |
|
|
if current <= expected: |
|
|
return text |
|
|
text = _trim_pattern(text, expected, CITATION_BRACKET_RE) |
|
|
if _token_count(text) <= expected: |
|
|
return text |
|
|
text = _trim_pattern(text, expected, CITATION_PAREN_RE) |
|
|
return text |
|
|
|
|
|
|
|
|
def _trim_to_token_count(text: str, expected: int) -> str: |
|
|
if not text or expected <= 0: |
|
|
return text |
|
|
tokens, spans = tokenize_text(text, return_spans=True) |
|
|
if not spans or len(spans) <= expected: |
|
|
return text |
|
|
end_idx = spans[expected - 1][1] |
|
|
end_idx = _extend_span_end(text, end_idx) |
|
|
return text[:end_idx].rstrip() |
|
|
|
|
|
|
|
|
def _cleanup_spacing(text: str) -> str: |
|
|
if not text: |
|
|
return text |
|
|
value = text |
|
|
value = value.replace("“", '"').replace("”", '"') |
|
|
value = value.replace("’", "'").replace("‘", "'") |
|
|
|
|
|
def _dash_repl(match: re.Match[str]) -> str: |
|
|
run = match.group(0) |
|
|
return "--" if len(run) >= 2 else "-" |
|
|
|
|
|
value = re.sub(r"[\u2010-\u2015\u2212\u2043]+", _dash_repl, value) |
|
|
value = value.replace("\ufb00", "ff") |
|
|
value = value.replace("\ufb01", "fi") |
|
|
value = value.replace("\ufb02", "fl") |
|
|
value = value.replace("\ufb03", "ffi") |
|
|
value = value.replace("\ufb04", "ffl") |
|
|
value = value.replace("…", "...") |
|
|
value = re.sub(r"-{3,}", "--", value) |
|
|
value = re.sub(r"([a-z0-9])([.!?])(?=[A-Z])", r"\1\2 ", value) |
|
|
value = re.sub(r"([A-Za-z]),(?=[A-Za-z])", r"\1, ", value) |
|
|
value = re.sub(r"([A-Za-z0-9])([;:])(?=[A-Za-z])", r"\1\2 ", value) |
|
|
value = re.sub(r"(\d)\s+(s)\b", r"\1\2", value) |
|
|
value = re.sub(r"([0-9])([A-Za-z])", r"\1 \2", value) |
|
|
value = re.sub(r"[ \t]+([,.;:!?])", r"\1", value) |
|
|
value = re.sub(r"\(\s+", "(", value) |
|
|
value = re.sub(r"\s+\)", ")", value) |
|
|
value = re.sub(r"\[\s+", "[", value) |
|
|
value = re.sub(r"\s+\]", "]", value) |
|
|
value = re.sub(r"\)(?=[A-Za-z0-9])", ") ", value) |
|
|
value = re.sub(r"\](?=[A-Za-z0-9])", "] ", value) |
|
|
|
|
|
def _space_citation_commas(match: re.Match[str]) -> str: |
|
|
inner = match.group(1) |
|
|
inner = re.sub(r",(?!\\s)", ", ", inner) |
|
|
return f"[{inner}]" |
|
|
|
|
|
value = re.sub(r"\[([0-9,;\s\-–]+)\]", _space_citation_commas, value) |
|
|
value = re.sub(r"[ \t]{2,}", " ", value) |
|
|
return value |
|
|
|
|
|
|
|
|
def _normalize_whitespace( |
|
|
text: str, |
|
|
preserve_linebreaks: bool, |
|
|
) -> str: |
|
|
if not text: |
|
|
return text |
|
|
value = _cleanup_spacing(text) |
|
|
if preserve_linebreaks: |
|
|
value = re.sub(r"[ \t]{2,}", " ", value) |
|
|
return value |
|
|
value = re.sub(r"\s+", " ", value).strip() |
|
|
return value |
|
|
|
|
|
|
|
|
def _normalize_hyphenation( |
|
|
text: str, |
|
|
preserve_hyphenation: bool, |
|
|
) -> str: |
|
|
if not text: |
|
|
return text |
|
|
if preserve_hyphenation: |
|
|
return text |
|
|
return re.sub(r"([A-Za-z])-\s+([A-Za-z])", r"\1\2", text) |
|
|
|
|
|
|
|
|
def _postprocess_text( |
|
|
text: str, |
|
|
expected_tokens: int, |
|
|
preserve_linebreaks: bool, |
|
|
preserve_hyphenation: bool, |
|
|
keep_bracket_citations: bool = True, |
|
|
keep_paren_citations: bool = True, |
|
|
split_letter_digit: bool = True, |
|
|
) -> str: |
|
|
value = text |
|
|
if not keep_bracket_citations: |
|
|
value = CITATION_BRACKET_RE.sub(" ", value) |
|
|
if not keep_paren_citations: |
|
|
value = CITATION_PAREN_RE.sub(" ", value) |
|
|
value = _trim_citations(value, expected_tokens) |
|
|
value = _trim_to_token_count(value, expected_tokens) |
|
|
value = _cleanup_spacing(value) |
|
|
if split_letter_digit: |
|
|
value = re.sub(r"([A-Za-z])([0-9])", r"\1 \2", value) |
|
|
value = _normalize_hyphenation(value, preserve_hyphenation) |
|
|
return _normalize_whitespace(value, preserve_linebreaks) |
|
|
|
|
|
|
|
|
def _ensure_trailing_punct(text: str, end_punct: str) -> str: |
|
|
if not text or not end_punct: |
|
|
stripped = text.rstrip() |
|
|
if not stripped: |
|
|
return text |
|
|
i = len(stripped) - 1 |
|
|
suffix = "" |
|
|
while i >= 0 and stripped[i] in TRAILING_QUOTES: |
|
|
suffix = stripped[i] + suffix |
|
|
i -= 1 |
|
|
base = stripped[: i + 1] |
|
|
if base and base[-1] in END_PUNCT: |
|
|
base = base[:-1] |
|
|
if base and suffix: |
|
|
if ")" in suffix and "(" not in base: |
|
|
suffix = suffix.replace(")", "") |
|
|
if "]" in suffix and "[" not in base: |
|
|
suffix = suffix.replace("]", "") |
|
|
return f"{base}{suffix}" |
|
|
stripped = text.rstrip() |
|
|
if not stripped: |
|
|
return text |
|
|
i = len(stripped) - 1 |
|
|
suffix = "" |
|
|
while i >= 0 and stripped[i] in TRAILING_QUOTES: |
|
|
suffix = stripped[i] + suffix |
|
|
i -= 1 |
|
|
base = stripped[: i + 1] |
|
|
if base and base[-1] in END_PUNCT: |
|
|
base = base[:-1] + end_punct |
|
|
else: |
|
|
base = f"{base}{end_punct}" |
|
|
return f"{base}{suffix}" |
|
|
|
|
|
|
|
|
def _find_pdf_hash_span( |
|
|
row: dict, |
|
|
pdf_token_index: Optional[TokenIndex], |
|
|
prefix: str, |
|
|
) -> Optional[tuple[int, int]]: |
|
|
if pdf_token_index is None: |
|
|
return None |
|
|
spec = _select_hash_specs(row, prefix) |
|
|
if spec: |
|
|
span = pdf_token_index.find_span_by_hash(*spec) |
|
|
if span: |
|
|
return span |
|
|
return None |
|
|
|
|
|
|
|
|
def _bool_flag(value: str) -> bool: |
|
|
return (value or "").strip().lower() == "true" |
|
|
|
|
|
|
|
|
def _strip_flags(row: dict, prefix: str) -> tuple[bool, bool]: |
|
|
keep_bracket = _bool_flag(row.get(f"{prefix}_has_bracket_citation", "")) |
|
|
keep_paren = _bool_flag(row.get(f"{prefix}_has_paren_citation", "")) |
|
|
return (not keep_bracket), (not keep_paren) |
|
|
|
|
|
|
|
|
def _span_matches_hash(row: dict, text: str, prefix: str) -> bool: |
|
|
if not text: |
|
|
return False |
|
|
expected_hash = row.get(f"{prefix}_hash64") or "" |
|
|
expected_sha = row.get(f"{prefix}_sha256") or "" |
|
|
if not expected_hash or not expected_sha: |
|
|
return False |
|
|
strip_brackets, strip_parens = _strip_flags(row, prefix) |
|
|
check_text = strip_citations( |
|
|
text, |
|
|
strip_brackets=strip_brackets, |
|
|
strip_parens=strip_parens, |
|
|
) |
|
|
tokens, _ = tokenize_text(check_text, return_spans=True) |
|
|
hash64, sha, _ = hash_token_sequence(tokens) |
|
|
return str(hash64) == str(expected_hash) and sha == expected_sha |
|
|
|
|
|
|
|
|
def _candidate_ids(paper_id: str, doi: str, arxiv: str) -> list[str]: |
|
|
candidates = [ |
|
|
paper_id, |
|
|
normalize_paper_id(paper_id), |
|
|
] |
|
|
if doi: |
|
|
candidates.append(doi) |
|
|
candidates.append(doi_suffix(doi)) |
|
|
if arxiv: |
|
|
candidates.append(arxiv) |
|
|
candidates.append(normalize_arxiv(arxiv)) |
|
|
seen = set() |
|
|
ordered = [] |
|
|
for item in candidates: |
|
|
value = (item or "").strip() |
|
|
if value and value not in seen: |
|
|
seen.add(value) |
|
|
ordered.append(value) |
|
|
for alias in candidate_pdf_aliases(paper_id, doi, arxiv): |
|
|
value = (alias or "").strip() |
|
|
if value and value not in seen: |
|
|
seen.add(value) |
|
|
ordered.append(value) |
|
|
return ordered |
|
|
|
|
|
|
|
|
def _normalize_title(title: str) -> str: |
|
|
return " ".join(title.lower().split()) |
|
|
|
|
|
|
|
|
def _build_meta_index( |
|
|
tei_index: Dict[str, Path], |
|
|
) -> tuple[Dict[str, Path], Dict[str, Path]]: |
|
|
doi_index: Dict[str, Path] = {} |
|
|
arxiv_index: Dict[str, Path] = {} |
|
|
for path in tei_index.values(): |
|
|
doi, arxiv = extract_ids_from_tei(path) |
|
|
if doi: |
|
|
doi_index.setdefault(normalize_doi(doi), path) |
|
|
doi_index.setdefault(doi_suffix(doi), path) |
|
|
if arxiv: |
|
|
arxiv_index.setdefault(normalize_arxiv(arxiv), path) |
|
|
return doi_index, arxiv_index |
|
|
|
|
|
|
|
|
def _resolve_tei_path( |
|
|
paper_id: str, |
|
|
doi: str, |
|
|
arxiv: str, |
|
|
tei_index: Dict[str, Path], |
|
|
doi_index: Dict[str, Path], |
|
|
arxiv_index: Dict[str, Path], |
|
|
) -> Optional[Path]: |
|
|
for candidate in _candidate_ids(paper_id, doi, arxiv): |
|
|
if candidate in tei_index: |
|
|
return tei_index[candidate] |
|
|
if candidate.startswith("paper_"): |
|
|
stripped = candidate[len("paper_") :] |
|
|
if stripped in tei_index: |
|
|
return tei_index[stripped] |
|
|
if doi: |
|
|
doi_key = normalize_doi(doi) |
|
|
if doi_key in doi_index: |
|
|
return doi_index[doi_key] |
|
|
doi_key = doi_suffix(doi) |
|
|
if doi_key in doi_index: |
|
|
return doi_index[doi_key] |
|
|
if arxiv: |
|
|
arxiv_key = normalize_arxiv(arxiv) |
|
|
if arxiv_key in arxiv_index: |
|
|
return arxiv_index[arxiv_key] |
|
|
return None |
|
|
|
|
|
|
|
|
def _tei_stem(path: Path) -> str: |
|
|
name = path.name |
|
|
if name.endswith(".grobid.tei.xml"): |
|
|
name = name[: -len(".grobid.tei.xml")] |
|
|
return name |
|
|
|
|
|
|
|
|
def _build_pdf_index(pdf_dir: Path) -> Dict[str, Path]: |
|
|
index: Dict[str, Path] = {} |
|
|
if not pdf_dir.exists(): |
|
|
return index |
|
|
version_re = re.compile(r"^(?P<base>.+?)(v\d+)$", re.IGNORECASE) |
|
|
arxiv_re = re.compile(r"^(?P<base>\d{4}\.\d{4,5})v\d+$", re.IGNORECASE) |
|
|
pii_re = re.compile(r"(S\d{8,})", re.IGNORECASE) |
|
|
for suffix in ("*.pdf", "*.PDF"): |
|
|
for path in pdf_dir.rglob(suffix): |
|
|
stem = path.stem |
|
|
index.setdefault(stem, path) |
|
|
index.setdefault(normalize_paper_id(stem), path) |
|
|
index.setdefault(f"paper_{stem}", path) |
|
|
if stem.startswith("paper_"): |
|
|
stripped = stem[len("paper_") :] |
|
|
if stripped: |
|
|
index.setdefault(stripped, path) |
|
|
index.setdefault(normalize_paper_id(stripped), path) |
|
|
if stem.endswith("_fixed") or stem.endswith("-fixed"): |
|
|
base = ( |
|
|
stem[: -len("_fixed")] |
|
|
if stem.endswith("_fixed") |
|
|
else stem[: -len("-fixed")] |
|
|
) |
|
|
if base: |
|
|
index[base] = path |
|
|
index[normalize_paper_id(base)] = path |
|
|
index[f"paper_{base}"] = path |
|
|
if base.startswith("paper_"): |
|
|
stripped_base = base[len("paper_") :] |
|
|
if stripped_base: |
|
|
index[stripped_base] = path |
|
|
index[normalize_paper_id(stripped_base)] = path |
|
|
match = arxiv_re.match(stem) |
|
|
if match: |
|
|
base = match.group("base") |
|
|
index.setdefault(base, path) |
|
|
index.setdefault(normalize_paper_id(base), path) |
|
|
match = version_re.match(stem) |
|
|
if match: |
|
|
base = match.group("base") |
|
|
index.setdefault(base, path) |
|
|
index.setdefault(normalize_paper_id(base), path) |
|
|
pii_match = pii_re.search(stem) |
|
|
if pii_match: |
|
|
pii = pii_match.group(1) |
|
|
index.setdefault(pii, path) |
|
|
index.setdefault(normalize_paper_id(pii), path) |
|
|
return index |
|
|
|
|
|
|
|
|
def _select_hash_specs( |
|
|
row: dict, |
|
|
prefix: str, |
|
|
) -> Optional[tuple[int, int, str]]: |
|
|
hash_version = (row.get("hash_version") or "").strip() |
|
|
if hash_version and hash_version != HASH_VERSION: |
|
|
return None |
|
|
count = row.get(f"{prefix}_token_count") or "" |
|
|
hash64 = row.get(f"{prefix}_hash64") or "" |
|
|
sha = row.get(f"{prefix}_sha256") or "" |
|
|
if not count or not hash64 or not sha: |
|
|
return None |
|
|
try: |
|
|
return int(count), int(hash64), sha |
|
|
except ValueError: |
|
|
return None |
|
|
|
|
|
|
|
|
def _select_anchor_specs( |
|
|
row: dict, |
|
|
prefix: str, |
|
|
position: str, |
|
|
) -> Optional[tuple[int, int, str]]: |
|
|
hash_version = (row.get("hash_version") or "").strip() |
|
|
if hash_version and hash_version != HASH_VERSION: |
|
|
return None |
|
|
count = row.get(f"{prefix}_{position}_token_count") or "" |
|
|
hash64 = row.get(f"{prefix}_{position}_hash64") or "" |
|
|
sha = row.get(f"{prefix}_{position}_sha256") or "" |
|
|
if not count or not hash64 or not sha: |
|
|
return None |
|
|
try: |
|
|
return int(count), int(hash64), sha |
|
|
except ValueError: |
|
|
return None |
|
|
|
|
|
|
|
|
def _select_anchor_spec_list( |
|
|
row: dict, |
|
|
prefix: str, |
|
|
position: str, |
|
|
) -> list[tuple[int, int, str]]: |
|
|
specs: list[tuple[int, int, str]] = [] |
|
|
primary = _select_anchor_specs(row, prefix, position) |
|
|
if primary is not None: |
|
|
specs.append(primary) |
|
|
alt_count = row.get(f"{prefix}_{position}_alt_token_count") or "" |
|
|
alt_hash = row.get(f"{prefix}_{position}_alt_hash64") or "" |
|
|
alt_sha = row.get(f"{prefix}_{position}_alt_sha256") or "" |
|
|
hash_version = (row.get("hash_version") or "").strip() |
|
|
if hash_version and hash_version != HASH_VERSION: |
|
|
return specs |
|
|
if alt_count and alt_hash and alt_sha: |
|
|
try: |
|
|
specs.append((int(alt_count), int(alt_hash), alt_sha)) |
|
|
except ValueError: |
|
|
pass |
|
|
return specs |
|
|
|
|
|
|
|
|
def _match_doc_by_hash( |
|
|
token_index: TokenIndex, |
|
|
hash_specs: list[tuple[int, int, str]], |
|
|
) -> int: |
|
|
score = 0 |
|
|
for window, hash64, sha in hash_specs: |
|
|
if token_index.find_span_by_hash(window, hash64, sha): |
|
|
score += 1 |
|
|
return score |
|
|
|
|
|
|
|
|
def _build_mid_candidates( |
|
|
token_index: TokenIndex, |
|
|
mid_specs: Optional[list[tuple[int, int, str]]], |
|
|
) -> list[tuple[int, int]]: |
|
|
if not mid_specs: |
|
|
return [] |
|
|
candidates: list[tuple[int, int]] = [] |
|
|
for spec in mid_specs: |
|
|
for position in token_index.find_token_positions_by_hash(*spec): |
|
|
candidates.append((position, spec[0])) |
|
|
return candidates |
|
|
|
|
|
|
|
|
def _span_has_mid( |
|
|
mid_candidates: list[tuple[int, int]], |
|
|
start_idx: int, |
|
|
end_idx: int, |
|
|
) -> bool: |
|
|
for mid_start, mid_len in mid_candidates: |
|
|
mid_end = mid_start + mid_len - 1 |
|
|
if mid_start >= start_idx and mid_end <= end_idx: |
|
|
return True |
|
|
return False |
|
|
|
|
|
|
|
|
def _find_span_by_anchors( |
|
|
token_index: TokenIndex, |
|
|
head_spec: Optional[tuple[int, int, str]], |
|
|
tail_spec: Optional[tuple[int, int, str]], |
|
|
expected_len: int, |
|
|
mid_specs: Optional[list[tuple[int, int, str]]] = None, |
|
|
*, |
|
|
require_mid: bool = False, |
|
|
) -> Optional[tuple[int, int]]: |
|
|
if head_spec is None or tail_spec is None: |
|
|
return None |
|
|
head_positions = token_index.find_token_positions_by_hash(*head_spec) |
|
|
tail_positions = token_index.find_token_positions_by_hash(*tail_spec) |
|
|
if not head_positions or not tail_positions: |
|
|
return None |
|
|
mid_candidates = [] |
|
|
if require_mid: |
|
|
mid_candidates = _build_mid_candidates(token_index, mid_specs) |
|
|
if not mid_candidates: |
|
|
return None |
|
|
best = None |
|
|
best_diff = None |
|
|
tol = max(5, int(expected_len * 0.3)) if expected_len else 10 |
|
|
min_len = max(1, expected_len // 2) if expected_len else 1 |
|
|
max_len = expected_len * 3 if expected_len else None |
|
|
for head_start in head_positions: |
|
|
head_end = head_start + head_spec[0] - 1 |
|
|
for tail_start in tail_positions: |
|
|
tail_end = tail_start + tail_spec[0] - 1 |
|
|
if tail_end < head_end: |
|
|
continue |
|
|
length = tail_end - head_start + 1 |
|
|
if mid_candidates and not _span_has_mid( |
|
|
mid_candidates, |
|
|
head_start, |
|
|
tail_end, |
|
|
): |
|
|
continue |
|
|
if expected_len: |
|
|
if length < min_len or (max_len and length > max_len): |
|
|
continue |
|
|
if length < expected_len - tol or length > expected_len + tol: |
|
|
continue |
|
|
diff = abs(length - expected_len) |
|
|
else: |
|
|
diff = length |
|
|
if best_diff is None or diff < best_diff: |
|
|
best_diff = diff |
|
|
best = (head_start, tail_end) |
|
|
if best is None and expected_len: |
|
|
for head_start in head_positions: |
|
|
head_end = head_start + head_spec[0] - 1 |
|
|
for tail_start in tail_positions: |
|
|
tail_end = tail_start + tail_spec[0] - 1 |
|
|
if tail_end < head_end: |
|
|
continue |
|
|
length = tail_end - head_start + 1 |
|
|
if mid_candidates and not _span_has_mid( |
|
|
mid_candidates, |
|
|
head_start, |
|
|
tail_end, |
|
|
): |
|
|
continue |
|
|
if length < min_len or (max_len and length > max_len): |
|
|
continue |
|
|
diff = abs(length - expected_len) |
|
|
if best_diff is None or diff < best_diff: |
|
|
best_diff = diff |
|
|
best = (head_start, tail_end) |
|
|
if best is None: |
|
|
return None |
|
|
start_char = token_index.spans[best[0]][0] |
|
|
end_char = token_index.spans[best[1]][1] |
|
|
return start_char, end_char |
|
|
|
|
|
|
|
|
def _find_span_from_anchor( |
|
|
token_index: TokenIndex, |
|
|
anchor_spec: Optional[tuple[int, int, str]], |
|
|
expected_len: int, |
|
|
position: str, |
|
|
mid_specs: Optional[list[tuple[int, int, str]]] = None, |
|
|
*, |
|
|
require_mid: bool = False, |
|
|
require_unique: bool = False, |
|
|
) -> Optional[tuple[int, int]]: |
|
|
if anchor_spec is None or expected_len <= 0: |
|
|
return None |
|
|
positions = token_index.find_token_positions_by_hash(*anchor_spec) |
|
|
if not positions: |
|
|
return None |
|
|
if require_unique and len(positions) != 1: |
|
|
return None |
|
|
mid_candidates = [] |
|
|
if require_mid: |
|
|
mid_candidates = _build_mid_candidates(token_index, mid_specs) |
|
|
if not mid_candidates: |
|
|
return None |
|
|
if position == "tail": |
|
|
positions = list(reversed(positions)) |
|
|
for anchor_start in positions: |
|
|
if position == "head": |
|
|
start_idx = anchor_start |
|
|
end_idx = anchor_start + expected_len - 1 |
|
|
else: |
|
|
anchor_end = anchor_start + anchor_spec[0] - 1 |
|
|
end_idx = anchor_end |
|
|
start_idx = end_idx - expected_len + 1 |
|
|
if start_idx < 0 or end_idx >= len(token_index.tokens): |
|
|
continue |
|
|
if mid_candidates and not _span_has_mid( |
|
|
mid_candidates, |
|
|
start_idx, |
|
|
end_idx, |
|
|
): |
|
|
continue |
|
|
start_char = token_index.spans[start_idx][0] |
|
|
end_char = token_index.spans[end_idx][1] |
|
|
return start_char, end_char |
|
|
return None |
|
|
|
|
|
|
|
|
def _pick_best_doc( |
|
|
token_indexes: Dict[Path, TokenIndex], |
|
|
hash_specs: list[tuple[int, int, str]], |
|
|
) -> Optional[Path]: |
|
|
best_path = None |
|
|
best_score = 0 |
|
|
tie = False |
|
|
for path, token_index in token_indexes.items(): |
|
|
score = _match_doc_by_hash(token_index, hash_specs) |
|
|
if score > best_score: |
|
|
best_score = score |
|
|
best_path = path |
|
|
tie = False |
|
|
elif score == best_score and score > 0: |
|
|
tie = True |
|
|
if best_score == 0 or tie: |
|
|
return None |
|
|
return best_path |
|
|
|
|
|
|
|
|
def _pick_best_pdf( |
|
|
pdf_paths: list[Path], |
|
|
pdf_token_cache: Dict[Path, TokenIndex], |
|
|
hash_specs: list[tuple[int, int, str]], |
|
|
) -> Optional[Path]: |
|
|
best_path = None |
|
|
best_score = 0 |
|
|
tie = False |
|
|
for path in pdf_paths: |
|
|
if path not in pdf_token_cache: |
|
|
pdf_text = extract_text_from_pdf(path) |
|
|
pdf_token_cache[path] = TokenIndex.from_text(pdf_text) |
|
|
token_index = pdf_token_cache[path] |
|
|
score = _match_doc_by_hash(token_index, hash_specs) |
|
|
if score > best_score: |
|
|
best_score = score |
|
|
best_path = path |
|
|
tie = False |
|
|
elif score == best_score and score > 0: |
|
|
tie = True |
|
|
if best_score == 0 or tie: |
|
|
return None |
|
|
return best_path |
|
|
|
|
|
|
|
|
def _run_grobid(input_dir: Path, output_dir: Path, config: Path) -> None: |
|
|
output_dir.mkdir(parents=True, exist_ok=True) |
|
|
cmd = [ |
|
|
sys.executable, |
|
|
"scripts/pdf_to_grobid.py", |
|
|
"--input_folder", |
|
|
str(input_dir), |
|
|
"--output_folder", |
|
|
str(output_dir), |
|
|
] |
|
|
if config and config.exists(): |
|
|
cmd.extend(["--config", str(config)]) |
|
|
subprocess.run(cmd, check=True) |
|
|
|
|
|
|
|
|
def main() -> None: |
|
|
parser = argparse.ArgumentParser( |
|
|
description="Hydrate DefExtra legal CSV using user-provided PDFs.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--legal-csv", |
|
|
type=Path, |
|
|
default=Path("results/paper_results/defextra_legal.csv"), |
|
|
help="Legal DefExtra CSV with markers.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--pdf-dir", |
|
|
type=Path, |
|
|
required=True, |
|
|
help="Directory with user-provided PDFs.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--grobid-out", |
|
|
type=Path, |
|
|
default=Path("outputs/defextra_grobid"), |
|
|
help="Output directory for GROBID TEI files.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--grobid-config", |
|
|
type=Path, |
|
|
default=Path("config.json"), |
|
|
help="Optional GROBID client config path.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--skip-grobid", |
|
|
action="store_true", |
|
|
help="Skip running GROBID (expects TEI files already present).", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--output-csv", |
|
|
type=Path, |
|
|
default=Path("results/paper_results/defextra_hydrated.csv"), |
|
|
help="Output hydrated CSV with excerpts.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--report", |
|
|
type=Path, |
|
|
default=None, |
|
|
help="Optional report path for missing matches.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--require-complete", |
|
|
action="store_true", |
|
|
help="Exit with error if any definition/context is missing.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--filter-to-pdfs", |
|
|
action="store_true", |
|
|
help="Only process rows that can be mapped to a provided PDF.", |
|
|
) |
|
|
parser.add_argument( |
|
|
"--allow-pdf-hash-mismatch", |
|
|
action="store_true", |
|
|
help=( |
|
|
"Continue when a PDF filename matches but hash markers do not. " |
|
|
"By default, such PDFs are skipped and reported." |
|
|
), |
|
|
) |
|
|
args = parser.parse_args() |
|
|
|
|
|
if not args.legal_csv.exists(): |
|
|
raise SystemExit(f"Legal CSV not found: {args.legal_csv}") |
|
|
if not args.pdf_dir.exists(): |
|
|
raise SystemExit(f"PDF dir not found: {args.pdf_dir}") |
|
|
|
|
|
if not args.skip_grobid: |
|
|
try: |
|
|
_run_grobid(args.pdf_dir, args.grobid_out, args.grobid_config) |
|
|
except subprocess.CalledProcessError: |
|
|
raise SystemExit( |
|
|
"GROBID processing failed. Ensure the GROBID server is running " |
|
|
"and reachable (default: http://localhost:8070), or supply " |
|
|
"--grobid-config with the correct server URL.", |
|
|
) |
|
|
|
|
|
tei_index = build_tei_index([args.grobid_out]) |
|
|
doi_index, arxiv_index = _build_meta_index(tei_index) |
|
|
pdf_index = _build_pdf_index(args.pdf_dir) |
|
|
doc_cache: Dict[str, Optional[DocIndex]] = {} |
|
|
token_cache: Dict[str, Optional[TokenIndex]] = {} |
|
|
tei_path_cache: Dict[str, Optional[Path]] = {} |
|
|
pdf_token_cache: Dict[Path, TokenIndex] = {} |
|
|
pdf_token_cache_stripped: Dict[tuple[Path, bool, bool], TokenIndex] = {} |
|
|
tei_token_cache_stripped: Dict[tuple[Path, bool, bool], TokenIndex] = {} |
|
|
pdf_failed: set[Path] = set() |
|
|
|
|
|
def _get_stripped_index( |
|
|
cache: Dict[tuple[Path, bool, bool], TokenIndex], |
|
|
source_path: Optional[Path], |
|
|
source_text: str, |
|
|
strip_brackets: bool, |
|
|
strip_parens: bool, |
|
|
) -> Optional[TokenIndex]: |
|
|
if source_path is None: |
|
|
return None |
|
|
if not strip_brackets and not strip_parens: |
|
|
return None |
|
|
key = (source_path, strip_brackets, strip_parens) |
|
|
if key not in cache: |
|
|
stripped = strip_citations( |
|
|
source_text, |
|
|
strip_brackets=strip_brackets, |
|
|
strip_parens=strip_parens, |
|
|
) |
|
|
cache[key] = TokenIndex.from_text(stripped) |
|
|
return cache[key] |
|
|
|
|
|
with args.legal_csv.open("r", encoding="utf-8", newline="") as handle: |
|
|
reader = csv.DictReader(handle) |
|
|
legal_rows = list(reader) |
|
|
|
|
|
paper_hashes: Dict[str, list[tuple[int, int, str]]] = {} |
|
|
title_to_ids: Dict[str, list[str]] = {} |
|
|
id_to_row: Dict[str, dict] = {} |
|
|
for row in legal_rows: |
|
|
paper_id = (row.get("paper_id") or "").strip() |
|
|
if not paper_id: |
|
|
continue |
|
|
if paper_id not in id_to_row: |
|
|
id_to_row[paper_id] = row |
|
|
title_key = _normalize_title(row.get("paper_title") or "") |
|
|
if title_key: |
|
|
title_to_ids.setdefault(title_key, []).append(paper_id) |
|
|
specs: list[tuple[int, int, str]] = [] |
|
|
for prefix in ("definition", "context"): |
|
|
spec = _select_hash_specs(row, prefix) |
|
|
if spec is None: |
|
|
continue |
|
|
token_count, _, _ = spec |
|
|
if token_count >= 5: |
|
|
specs.append(spec) |
|
|
if specs: |
|
|
paper_hashes.setdefault(paper_id, []).extend(specs) |
|
|
for prefix in ("definition", "context"): |
|
|
for position in ("head", "mid", "tail"): |
|
|
for spec in _select_anchor_spec_list(row, prefix, position): |
|
|
if spec and spec[0] >= 5: |
|
|
paper_hashes.setdefault(paper_id, []).append(spec) |
|
|
|
|
|
tei_token_indexes: Dict[Path, TokenIndex] = {} |
|
|
allowed_stems = set(pdf_index.keys()) if pdf_index else set() |
|
|
for tei_path in tei_index.values(): |
|
|
if allowed_stems: |
|
|
stem = _tei_stem(tei_path) |
|
|
stem_norm = normalize_paper_id(stem) |
|
|
stem_stripped = ( |
|
|
stem[len("paper_") :] if stem.startswith("paper_") else stem |
|
|
) |
|
|
if ( |
|
|
stem not in allowed_stems |
|
|
and stem_norm not in allowed_stems |
|
|
and stem_stripped not in allowed_stems |
|
|
): |
|
|
continue |
|
|
try: |
|
|
doc_index = DocIndex.from_tei(tei_path) |
|
|
except Exception: |
|
|
continue |
|
|
tei_token_indexes[tei_path] = TokenIndex.from_text(doc_index.doc_text) |
|
|
|
|
|
output_rows = [] |
|
|
missing_papers = set() |
|
|
missing_defs = 0 |
|
|
missing_ctxs = 0 |
|
|
hydrated_from_pdf = 0 |
|
|
hydrated_from_anchor = 0 |
|
|
pdf_hash_mismatches: list[dict[str, str]] = [] |
|
|
pdf_hash_mismatch_seen: set[tuple[str, str]] = set() |
|
|
missing_def_rows: list[dict] = [] |
|
|
missing_ctx_rows: list[dict] = [] |
|
|
|
|
|
for row in legal_rows: |
|
|
paper_id = (row.get("paper_id") or "").strip() |
|
|
doi = (row.get("paper_doi") or "").strip() |
|
|
arxiv = (row.get("paper_arxiv") or "").strip() |
|
|
|
|
|
if paper_id not in doc_cache: |
|
|
tei_path = _resolve_tei_path( |
|
|
paper_id, |
|
|
doi, |
|
|
arxiv, |
|
|
tei_index, |
|
|
doi_index, |
|
|
arxiv_index, |
|
|
) |
|
|
if tei_path is None: |
|
|
hash_specs = paper_hashes.get(paper_id, []) |
|
|
if hash_specs: |
|
|
tei_path = _pick_best_doc( |
|
|
tei_token_indexes, |
|
|
hash_specs, |
|
|
) |
|
|
if tei_path is None: |
|
|
doc_cache[paper_id] = None |
|
|
token_cache[paper_id] = None |
|
|
tei_path_cache[paper_id] = None |
|
|
else: |
|
|
doc_index = DocIndex.from_tei(tei_path) |
|
|
doc_cache[paper_id] = doc_index |
|
|
token_cache[paper_id] = TokenIndex.from_text( |
|
|
doc_index.doc_text, |
|
|
) |
|
|
tei_path_cache[paper_id] = tei_path |
|
|
|
|
|
doc_index = doc_cache.get(paper_id) |
|
|
tei_token_index = token_cache.get(paper_id) |
|
|
definition = "" |
|
|
context = "" |
|
|
pdf_token_index: Optional[TokenIndex] = None |
|
|
pdf_path = None |
|
|
tei_path = tei_path_cache.get(paper_id) |
|
|
pdf_direct_match = False |
|
|
if tei_path is not None: |
|
|
stem = _tei_stem(tei_path) |
|
|
pdf_path = pdf_index.get(stem) or pdf_index.get( |
|
|
normalize_paper_id(stem), |
|
|
) |
|
|
if pdf_path is not None: |
|
|
pdf_direct_match = True |
|
|
if pdf_path is None: |
|
|
for candidate in _candidate_ids(paper_id, doi, arxiv): |
|
|
pdf_path = pdf_index.get(candidate) |
|
|
if pdf_path: |
|
|
pdf_direct_match = True |
|
|
break |
|
|
if pdf_path is None: |
|
|
title_key = _normalize_title(row.get("paper_title") or "") |
|
|
for other_id in title_to_ids.get(title_key, []): |
|
|
if other_id == paper_id: |
|
|
continue |
|
|
other_row = id_to_row.get(other_id, {}) |
|
|
other_doi = (other_row.get("paper_doi") or "").strip() |
|
|
other_arxiv = (other_row.get("paper_arxiv") or "").strip() |
|
|
for candidate in _candidate_ids( |
|
|
other_id, |
|
|
other_doi, |
|
|
other_arxiv, |
|
|
): |
|
|
pdf_path = pdf_index.get(candidate) |
|
|
if pdf_path: |
|
|
pdf_direct_match = True |
|
|
break |
|
|
if pdf_path: |
|
|
break |
|
|
hash_specs = paper_hashes.get(paper_id, []) |
|
|
if pdf_path is not None and hash_specs: |
|
|
if pdf_path not in pdf_token_cache: |
|
|
pdf_text = extract_text_from_pdf(pdf_path) |
|
|
pdf_token_cache[pdf_path] = TokenIndex.from_text(pdf_text) |
|
|
pdf_token_index = pdf_token_cache[pdf_path] |
|
|
if _match_doc_by_hash(pdf_token_index, hash_specs) == 0: |
|
|
mismatch_key = (paper_id, str(pdf_path)) |
|
|
if mismatch_key not in pdf_hash_mismatch_seen: |
|
|
pdf_hash_mismatch_seen.add(mismatch_key) |
|
|
pdf_hash_mismatches.append( |
|
|
{"paper_id": paper_id, "pdf": str(pdf_path)}, |
|
|
) |
|
|
if not args.allow_pdf_hash_mismatch and pdf_direct_match: |
|
|
print( |
|
|
f"Warning: PDF hash markers did not match for {pdf_path.name}; " |
|
|
"skipping PDF (use --allow-pdf-hash-mismatch to override).", |
|
|
file=sys.stderr, |
|
|
) |
|
|
elif args.allow_pdf_hash_mismatch: |
|
|
print( |
|
|
f"Warning: PDF hash markers did not match for {pdf_path.name}; " |
|
|
"continuing with direct filename match.", |
|
|
file=sys.stderr, |
|
|
) |
|
|
if not args.allow_pdf_hash_mismatch: |
|
|
pdf_path = None |
|
|
pdf_token_index = None |
|
|
pdf_direct_match = False |
|
|
if pdf_path is None and hash_specs: |
|
|
pdf_paths = list({p for p in pdf_index.values()}) |
|
|
if pdf_paths: |
|
|
pdf_path = _pick_best_pdf( |
|
|
pdf_paths, |
|
|
pdf_token_cache, |
|
|
hash_specs, |
|
|
) |
|
|
|
|
|
if args.filter_to_pdfs and pdf_path is None: |
|
|
continue |
|
|
|
|
|
if pdf_path is not None and pdf_path not in pdf_token_cache: |
|
|
try: |
|
|
pdf_text = extract_text_from_pdf(pdf_path) |
|
|
pdf_token_cache[pdf_path] = TokenIndex.from_text(pdf_text) |
|
|
except Exception as exc: |
|
|
pdf_token_cache[pdf_path] = TokenIndex.from_text("") |
|
|
if pdf_path not in pdf_failed: |
|
|
pdf_failed.add(pdf_path) |
|
|
print( |
|
|
f"Warning: PDF text extraction failed for {pdf_path.name}: {exc}", |
|
|
file=sys.stderr, |
|
|
) |
|
|
pdf_token_index = pdf_token_cache.get(pdf_path) if pdf_path else None |
|
|
|
|
|
if doc_index is None: |
|
|
missing_papers.add(paper_id) |
|
|
else: |
|
|
def_start = row.get("definition_char_start") or "" |
|
|
def_end = row.get("definition_char_end") or "" |
|
|
ctx_start = row.get("context_char_start") or "" |
|
|
ctx_end = row.get("context_char_end") or "" |
|
|
def_strip_brackets, def_strip_parens = _strip_flags( |
|
|
row, |
|
|
"definition", |
|
|
) |
|
|
ctx_strip_brackets, ctx_strip_parens = _strip_flags(row, "context") |
|
|
|
|
|
if not definition and tei_token_index: |
|
|
spec = _select_hash_specs(row, "definition") |
|
|
if spec: |
|
|
span = tei_token_index.find_span_by_hash(*spec) |
|
|
if span: |
|
|
definition = _extract_with_trailing_punct( |
|
|
doc_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
if not definition: |
|
|
stripped_index = _get_stripped_index( |
|
|
tei_token_cache_stripped, |
|
|
tei_path, |
|
|
doc_index.doc_text, |
|
|
def_strip_brackets, |
|
|
def_strip_parens, |
|
|
) |
|
|
if stripped_index is not None: |
|
|
span = stripped_index.find_span_by_hash(*spec) |
|
|
if span: |
|
|
definition = _extract_with_trailing_punct( |
|
|
stripped_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
if not definition and not (def_start and def_end): |
|
|
head_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"definition", |
|
|
"head", |
|
|
) |
|
|
mid_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"definition", |
|
|
"mid", |
|
|
) |
|
|
tail_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"definition", |
|
|
"tail", |
|
|
) |
|
|
expected_len = int(row.get("definition_token_count") or 0) |
|
|
for head_spec in head_specs or [None]: |
|
|
for tail_spec in tail_specs or [None]: |
|
|
if head_spec is None or tail_spec is None: |
|
|
continue |
|
|
span = _find_span_by_anchors( |
|
|
tei_token_index, |
|
|
head_spec, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
mid_specs, |
|
|
require_mid=True, |
|
|
) |
|
|
if span is None: |
|
|
span = _find_span_by_anchors( |
|
|
tei_token_index, |
|
|
head_spec, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
) |
|
|
if span: |
|
|
definition = _extract_with_trailing_punct( |
|
|
doc_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
break |
|
|
if definition: |
|
|
break |
|
|
if not definition: |
|
|
head_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"definition", |
|
|
"head", |
|
|
) |
|
|
mid_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"definition", |
|
|
"mid", |
|
|
) |
|
|
tail_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"definition", |
|
|
"tail", |
|
|
) |
|
|
expected_len = int(row.get("definition_token_count") or 0) |
|
|
span = None |
|
|
for head_spec in head_specs: |
|
|
if mid_specs: |
|
|
span = _find_span_from_anchor( |
|
|
tei_token_index, |
|
|
head_spec, |
|
|
expected_len, |
|
|
"head", |
|
|
mid_specs, |
|
|
require_mid=True, |
|
|
) |
|
|
if span is None: |
|
|
span = _find_span_from_anchor( |
|
|
tei_token_index, |
|
|
head_spec, |
|
|
expected_len, |
|
|
"head", |
|
|
mid_specs, |
|
|
require_unique=True, |
|
|
) |
|
|
else: |
|
|
span = _find_span_from_anchor( |
|
|
tei_token_index, |
|
|
head_spec, |
|
|
expected_len, |
|
|
"head", |
|
|
mid_specs, |
|
|
) |
|
|
if span: |
|
|
break |
|
|
if span is None: |
|
|
for tail_spec in tail_specs: |
|
|
if mid_specs: |
|
|
span = _find_span_from_anchor( |
|
|
tei_token_index, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
"tail", |
|
|
mid_specs, |
|
|
require_mid=True, |
|
|
) |
|
|
if span is None: |
|
|
span = _find_span_from_anchor( |
|
|
tei_token_index, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
"tail", |
|
|
mid_specs, |
|
|
require_unique=True, |
|
|
) |
|
|
elif span is None: |
|
|
span = _find_span_from_anchor( |
|
|
tei_token_index, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
"tail", |
|
|
mid_specs, |
|
|
) |
|
|
if span: |
|
|
break |
|
|
if span: |
|
|
definition = _extract_with_trailing_punct( |
|
|
doc_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
if not definition and def_start and def_end: |
|
|
candidate = _extract_with_trailing_punct( |
|
|
doc_index.doc_text, |
|
|
int(def_start), |
|
|
int(def_end), |
|
|
) |
|
|
if _span_matches_hash(row, candidate, "definition"): |
|
|
definition = candidate |
|
|
if not definition and pdf_token_index: |
|
|
span = _find_pdf_hash_span(row, pdf_token_index, "definition") |
|
|
if span: |
|
|
definition = _extract_with_trailing_punct( |
|
|
pdf_token_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
hydrated_from_pdf += 1 |
|
|
if not definition: |
|
|
stripped_index = _get_stripped_index( |
|
|
pdf_token_cache_stripped, |
|
|
pdf_path, |
|
|
pdf_token_index.doc_text, |
|
|
def_strip_brackets, |
|
|
def_strip_parens, |
|
|
) |
|
|
if stripped_index is not None: |
|
|
span = _find_pdf_hash_span( |
|
|
row, |
|
|
stripped_index, |
|
|
"definition", |
|
|
) |
|
|
if span: |
|
|
definition = _extract_with_trailing_punct( |
|
|
stripped_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
hydrated_from_pdf += 1 |
|
|
|
|
|
if not context and tei_token_index: |
|
|
spec = _select_hash_specs(row, "context") |
|
|
if spec: |
|
|
span = tei_token_index.find_span_by_hash(*spec) |
|
|
if span: |
|
|
context = _extract_with_trailing_punct( |
|
|
doc_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
if not context: |
|
|
stripped_index = _get_stripped_index( |
|
|
tei_token_cache_stripped, |
|
|
tei_path, |
|
|
doc_index.doc_text, |
|
|
ctx_strip_brackets, |
|
|
ctx_strip_parens, |
|
|
) |
|
|
if stripped_index is not None: |
|
|
span = stripped_index.find_span_by_hash(*spec) |
|
|
if span: |
|
|
context = _extract_with_trailing_punct( |
|
|
stripped_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
if not context and not (ctx_start and ctx_end): |
|
|
head_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"context", |
|
|
"head", |
|
|
) |
|
|
mid_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"context", |
|
|
"mid", |
|
|
) |
|
|
tail_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"context", |
|
|
"tail", |
|
|
) |
|
|
expected_len = int(row.get("context_token_count") or 0) |
|
|
for head_spec in head_specs or [None]: |
|
|
for tail_spec in tail_specs or [None]: |
|
|
if head_spec is None or tail_spec is None: |
|
|
continue |
|
|
span = _find_span_by_anchors( |
|
|
tei_token_index, |
|
|
head_spec, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
mid_specs, |
|
|
require_mid=True, |
|
|
) |
|
|
if span is None: |
|
|
span = _find_span_by_anchors( |
|
|
tei_token_index, |
|
|
head_spec, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
) |
|
|
if span: |
|
|
context = _extract_with_trailing_punct( |
|
|
doc_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
break |
|
|
if context: |
|
|
break |
|
|
if not context: |
|
|
head_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"context", |
|
|
"head", |
|
|
) |
|
|
mid_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"context", |
|
|
"mid", |
|
|
) |
|
|
tail_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"context", |
|
|
"tail", |
|
|
) |
|
|
expected_len = int(row.get("context_token_count") or 0) |
|
|
span = None |
|
|
for head_spec in head_specs: |
|
|
if mid_specs: |
|
|
span = _find_span_from_anchor( |
|
|
tei_token_index, |
|
|
head_spec, |
|
|
expected_len, |
|
|
"head", |
|
|
mid_specs, |
|
|
require_mid=True, |
|
|
) |
|
|
if span is None: |
|
|
span = _find_span_from_anchor( |
|
|
tei_token_index, |
|
|
head_spec, |
|
|
expected_len, |
|
|
"head", |
|
|
mid_specs, |
|
|
require_unique=True, |
|
|
) |
|
|
else: |
|
|
span = _find_span_from_anchor( |
|
|
tei_token_index, |
|
|
head_spec, |
|
|
expected_len, |
|
|
"head", |
|
|
mid_specs, |
|
|
) |
|
|
if span: |
|
|
break |
|
|
if span is None: |
|
|
for tail_spec in tail_specs: |
|
|
if mid_specs: |
|
|
span = _find_span_from_anchor( |
|
|
tei_token_index, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
"tail", |
|
|
mid_specs, |
|
|
require_mid=True, |
|
|
) |
|
|
if span is None: |
|
|
span = _find_span_from_anchor( |
|
|
tei_token_index, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
"tail", |
|
|
mid_specs, |
|
|
require_unique=True, |
|
|
) |
|
|
elif span is None: |
|
|
span = _find_span_from_anchor( |
|
|
tei_token_index, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
"tail", |
|
|
mid_specs, |
|
|
) |
|
|
if span: |
|
|
break |
|
|
if span: |
|
|
context = _extract_with_trailing_punct( |
|
|
doc_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
if not context and ctx_start and ctx_end: |
|
|
candidate = _extract_with_trailing_punct( |
|
|
doc_index.doc_text, |
|
|
int(ctx_start), |
|
|
int(ctx_end), |
|
|
) |
|
|
if _span_matches_hash(row, candidate, "context"): |
|
|
context = candidate |
|
|
if not context and pdf_token_index: |
|
|
span = _find_pdf_hash_span(row, pdf_token_index, "context") |
|
|
if span: |
|
|
context = _extract_with_trailing_punct( |
|
|
pdf_token_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
hydrated_from_pdf += 1 |
|
|
if not context: |
|
|
stripped_index = _get_stripped_index( |
|
|
pdf_token_cache_stripped, |
|
|
pdf_path, |
|
|
pdf_token_index.doc_text, |
|
|
ctx_strip_brackets, |
|
|
ctx_strip_parens, |
|
|
) |
|
|
if stripped_index is not None: |
|
|
span = _find_pdf_hash_span( |
|
|
row, |
|
|
stripped_index, |
|
|
"context", |
|
|
) |
|
|
if span: |
|
|
context = _extract_with_trailing_punct( |
|
|
stripped_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
hydrated_from_pdf += 1 |
|
|
|
|
|
if not definition and pdf_path is not None and pdf_token_index: |
|
|
spec = _select_hash_specs(row, "definition") |
|
|
if spec: |
|
|
span = pdf_token_index.find_span_by_hash(*spec) |
|
|
if span: |
|
|
definition = _extract_with_trailing_punct( |
|
|
pdf_token_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
hydrated_from_pdf += 1 |
|
|
if not definition: |
|
|
head_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"definition", |
|
|
"head", |
|
|
) |
|
|
mid_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"definition", |
|
|
"mid", |
|
|
) |
|
|
tail_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"definition", |
|
|
"tail", |
|
|
) |
|
|
expected_len = int(row.get("definition_token_count") or 0) |
|
|
for head_spec in head_specs or [None]: |
|
|
for tail_spec in tail_specs or [None]: |
|
|
if head_spec is None or tail_spec is None: |
|
|
continue |
|
|
span = _find_span_by_anchors( |
|
|
pdf_token_index, |
|
|
head_spec, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
mid_specs, |
|
|
require_mid=True, |
|
|
) |
|
|
if span is None: |
|
|
span = _find_span_by_anchors( |
|
|
pdf_token_index, |
|
|
head_spec, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
) |
|
|
if span: |
|
|
definition = _extract_with_trailing_punct( |
|
|
pdf_token_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
hydrated_from_pdf += 1 |
|
|
hydrated_from_anchor += 1 |
|
|
break |
|
|
if definition: |
|
|
break |
|
|
if not definition: |
|
|
head_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"definition", |
|
|
"head", |
|
|
) |
|
|
mid_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"definition", |
|
|
"mid", |
|
|
) |
|
|
tail_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"definition", |
|
|
"tail", |
|
|
) |
|
|
expected_len = int(row.get("definition_token_count") or 0) |
|
|
span = None |
|
|
for head_spec in head_specs: |
|
|
if mid_specs: |
|
|
span = _find_span_from_anchor( |
|
|
pdf_token_index, |
|
|
head_spec, |
|
|
expected_len, |
|
|
"head", |
|
|
mid_specs, |
|
|
require_mid=True, |
|
|
) |
|
|
if span is None: |
|
|
span = _find_span_from_anchor( |
|
|
pdf_token_index, |
|
|
head_spec, |
|
|
expected_len, |
|
|
"head", |
|
|
mid_specs, |
|
|
require_unique=True, |
|
|
) |
|
|
else: |
|
|
span = _find_span_from_anchor( |
|
|
pdf_token_index, |
|
|
head_spec, |
|
|
expected_len, |
|
|
"head", |
|
|
mid_specs, |
|
|
) |
|
|
if span: |
|
|
break |
|
|
if span is None: |
|
|
for tail_spec in tail_specs: |
|
|
if mid_specs: |
|
|
span = _find_span_from_anchor( |
|
|
pdf_token_index, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
"tail", |
|
|
mid_specs, |
|
|
require_mid=True, |
|
|
) |
|
|
if span is None: |
|
|
span = _find_span_from_anchor( |
|
|
pdf_token_index, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
"tail", |
|
|
mid_specs, |
|
|
require_unique=True, |
|
|
) |
|
|
else: |
|
|
span = _find_span_from_anchor( |
|
|
pdf_token_index, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
"tail", |
|
|
mid_specs, |
|
|
) |
|
|
if span: |
|
|
break |
|
|
if span: |
|
|
definition = _extract_with_trailing_punct( |
|
|
pdf_token_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
hydrated_from_pdf += 1 |
|
|
hydrated_from_anchor += 1 |
|
|
|
|
|
if not definition: |
|
|
missing_defs += 1 |
|
|
missing_def_rows.append( |
|
|
{ |
|
|
"paper_id": paper_id, |
|
|
"concept": row.get("concept", ""), |
|
|
"reason": "missing_definition", |
|
|
}, |
|
|
) |
|
|
|
|
|
if not context and pdf_path is not None and pdf_token_index: |
|
|
spec = _select_hash_specs(row, "context") |
|
|
if spec: |
|
|
span = pdf_token_index.find_span_by_hash(*spec) |
|
|
if span: |
|
|
context = _extract_with_trailing_punct( |
|
|
pdf_token_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
hydrated_from_pdf += 1 |
|
|
if not context: |
|
|
head_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"context", |
|
|
"head", |
|
|
) |
|
|
mid_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"context", |
|
|
"mid", |
|
|
) |
|
|
tail_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"context", |
|
|
"tail", |
|
|
) |
|
|
expected_len = int(row.get("context_token_count") or 0) |
|
|
for head_spec in head_specs or [None]: |
|
|
for tail_spec in tail_specs or [None]: |
|
|
if head_spec is None or tail_spec is None: |
|
|
continue |
|
|
span = _find_span_by_anchors( |
|
|
pdf_token_index, |
|
|
head_spec, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
mid_specs, |
|
|
require_mid=True, |
|
|
) |
|
|
if span is None: |
|
|
span = _find_span_by_anchors( |
|
|
pdf_token_index, |
|
|
head_spec, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
) |
|
|
if span: |
|
|
context = _extract_with_trailing_punct( |
|
|
pdf_token_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
hydrated_from_pdf += 1 |
|
|
hydrated_from_anchor += 1 |
|
|
break |
|
|
if context: |
|
|
break |
|
|
if not context: |
|
|
head_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"context", |
|
|
"head", |
|
|
) |
|
|
mid_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"context", |
|
|
"mid", |
|
|
) |
|
|
tail_specs = _select_anchor_spec_list( |
|
|
row, |
|
|
"context", |
|
|
"tail", |
|
|
) |
|
|
expected_len = int(row.get("context_token_count") or 0) |
|
|
span = None |
|
|
for head_spec in head_specs: |
|
|
if mid_specs: |
|
|
span = _find_span_from_anchor( |
|
|
pdf_token_index, |
|
|
head_spec, |
|
|
expected_len, |
|
|
"head", |
|
|
mid_specs, |
|
|
require_mid=True, |
|
|
) |
|
|
if span is None: |
|
|
span = _find_span_from_anchor( |
|
|
pdf_token_index, |
|
|
head_spec, |
|
|
expected_len, |
|
|
"head", |
|
|
mid_specs, |
|
|
require_unique=True, |
|
|
) |
|
|
else: |
|
|
span = _find_span_from_anchor( |
|
|
pdf_token_index, |
|
|
head_spec, |
|
|
expected_len, |
|
|
"head", |
|
|
mid_specs, |
|
|
) |
|
|
if span: |
|
|
break |
|
|
if span is None: |
|
|
for tail_spec in tail_specs: |
|
|
if mid_specs: |
|
|
span = _find_span_from_anchor( |
|
|
pdf_token_index, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
"tail", |
|
|
mid_specs, |
|
|
require_mid=True, |
|
|
) |
|
|
if span is None: |
|
|
span = _find_span_from_anchor( |
|
|
pdf_token_index, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
"tail", |
|
|
mid_specs, |
|
|
require_unique=True, |
|
|
) |
|
|
else: |
|
|
span = _find_span_from_anchor( |
|
|
pdf_token_index, |
|
|
tail_spec, |
|
|
expected_len, |
|
|
"tail", |
|
|
mid_specs, |
|
|
) |
|
|
if span: |
|
|
break |
|
|
if span: |
|
|
context = _extract_with_trailing_punct( |
|
|
pdf_token_index.doc_text, |
|
|
span[0], |
|
|
span[1], |
|
|
) |
|
|
hydrated_from_pdf += 1 |
|
|
hydrated_from_anchor += 1 |
|
|
|
|
|
if not context: |
|
|
missing_ctxs += 1 |
|
|
missing_ctx_rows.append( |
|
|
{ |
|
|
"paper_id": paper_id, |
|
|
"concept": row.get("concept", ""), |
|
|
"reason": "missing_context", |
|
|
}, |
|
|
) |
|
|
|
|
|
def_preserve_lines = _row_flag( |
|
|
row, |
|
|
"definition_preserve_linebreaks", |
|
|
) |
|
|
ctx_preserve_lines = _row_flag( |
|
|
row, |
|
|
"context_preserve_linebreaks", |
|
|
) |
|
|
def_preserve_hyph = _row_flag( |
|
|
row, |
|
|
"definition_preserve_hyphenation", |
|
|
) |
|
|
ctx_preserve_hyph = _row_flag( |
|
|
row, |
|
|
"context_preserve_hyphenation", |
|
|
) |
|
|
def_keep_bracket = _row_flag( |
|
|
row, |
|
|
"definition_has_bracket_citation", |
|
|
True, |
|
|
) |
|
|
def_keep_paren = _row_flag( |
|
|
row, |
|
|
"definition_has_paren_citation", |
|
|
True, |
|
|
) |
|
|
def_split_letter_digit = not _row_flag( |
|
|
row, |
|
|
"definition_has_letter_digit", |
|
|
) |
|
|
ctx_keep_bracket = _row_flag( |
|
|
row, |
|
|
"context_has_bracket_citation", |
|
|
True, |
|
|
) |
|
|
ctx_keep_paren = _row_flag( |
|
|
row, |
|
|
"context_has_paren_citation", |
|
|
True, |
|
|
) |
|
|
ctx_split_letter_digit = not _row_flag( |
|
|
row, |
|
|
"context_has_letter_digit", |
|
|
) |
|
|
|
|
|
output_rows.append( |
|
|
{ |
|
|
"paper_id": paper_id, |
|
|
"paper_title": row.get("paper_title", ""), |
|
|
"paper_doi": doi, |
|
|
"paper_arxiv": arxiv, |
|
|
"concept": row.get("concept", ""), |
|
|
"definition": _ensure_trailing_punct( |
|
|
_postprocess_text( |
|
|
definition, |
|
|
int(row.get("definition_token_count") or 0), |
|
|
def_preserve_lines, |
|
|
def_preserve_hyph, |
|
|
def_keep_bracket, |
|
|
def_keep_paren, |
|
|
def_split_letter_digit, |
|
|
), |
|
|
row.get("definition_end_punct", ""), |
|
|
), |
|
|
"context": _ensure_trailing_punct( |
|
|
_postprocess_text( |
|
|
context, |
|
|
int(row.get("context_token_count") or 0), |
|
|
ctx_preserve_lines, |
|
|
ctx_preserve_hyph, |
|
|
ctx_keep_bracket, |
|
|
ctx_keep_paren, |
|
|
ctx_split_letter_digit, |
|
|
), |
|
|
row.get("context_end_punct", ""), |
|
|
), |
|
|
"definition_type": row.get("definition_type", ""), |
|
|
"source_file": row.get("source_file", ""), |
|
|
"is_out_of_domain": row.get("is_out_of_domain", ""), |
|
|
}, |
|
|
) |
|
|
|
|
|
args.output_csv.parent.mkdir(parents=True, exist_ok=True) |
|
|
with args.output_csv.open("w", encoding="utf-8", newline="") as handle: |
|
|
fieldnames = [ |
|
|
"paper_id", |
|
|
"paper_title", |
|
|
"paper_doi", |
|
|
"paper_arxiv", |
|
|
"concept", |
|
|
"definition", |
|
|
"context", |
|
|
"definition_type", |
|
|
"source_file", |
|
|
"is_out_of_domain", |
|
|
] |
|
|
writer = csv.DictWriter(handle, fieldnames=fieldnames) |
|
|
writer.writeheader() |
|
|
for row in output_rows: |
|
|
writer.writerow(row) |
|
|
|
|
|
print(f"Wrote hydrated CSV to {args.output_csv}") |
|
|
print(f"Missing TEI for {len(missing_papers)} papers") |
|
|
print(f"Missing definition spans: {missing_defs}") |
|
|
print(f"Missing context spans: {missing_ctxs}") |
|
|
print( |
|
|
"Hydrated from PDF fallback: " |
|
|
f"{hydrated_from_pdf} (anchors used: {hydrated_from_anchor})", |
|
|
) |
|
|
if args.report is not None: |
|
|
report_lines = [] |
|
|
if missing_papers: |
|
|
report_lines.append("Missing papers:") |
|
|
for paper_id in sorted(missing_papers): |
|
|
report_lines.append(f"- {paper_id}") |
|
|
report_lines.append("") |
|
|
if pdf_hash_mismatches: |
|
|
report_lines.append( |
|
|
"PDF hash mismatches (filename matched, hash did not):", |
|
|
) |
|
|
for item in pdf_hash_mismatches: |
|
|
report_lines.append( |
|
|
f"- {item['paper_id']} | {item['pdf']}", |
|
|
) |
|
|
report_lines.append( |
|
|
"Note: rerun with --allow-pdf-hash-mismatch to continue with these PDFs.", |
|
|
) |
|
|
report_lines.append("") |
|
|
report_lines.append(f"Missing definition spans: {missing_defs}") |
|
|
report_lines.append(f"Missing context spans: {missing_ctxs}") |
|
|
if missing_def_rows: |
|
|
report_lines.append("") |
|
|
report_lines.append("Missing definitions (paper_id | concept):") |
|
|
for item in missing_def_rows: |
|
|
report_lines.append( |
|
|
f"- {item['paper_id']} | {item['concept']}", |
|
|
) |
|
|
if missing_ctx_rows: |
|
|
report_lines.append("") |
|
|
report_lines.append("Missing contexts (paper_id | concept):") |
|
|
for item in missing_ctx_rows: |
|
|
report_lines.append( |
|
|
f"- {item['paper_id']} | {item['concept']}", |
|
|
) |
|
|
args.report.parent.mkdir(parents=True, exist_ok=True) |
|
|
args.report.write_text( |
|
|
"\n".join(report_lines) + "\n", |
|
|
encoding="utf-8", |
|
|
) |
|
|
print(f"Wrote report to {args.report}") |
|
|
if args.require_complete and (missing_defs or missing_ctxs): |
|
|
raise SystemExit( |
|
|
"Hydration incomplete: " |
|
|
f"{missing_defs} definitions, {missing_ctxs} contexts missing.", |
|
|
) |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
main() |
|
|
|