ancient-scripts-datasets / scripts /extract_wold_borrowings_v2.py
Nacryos's picture
Update cognate pair scripts + borrowing/similarity data (P0-P2 fixes)
05fda2e unverified
#!/usr/bin/env python3
"""Extract WOLD borrowing pairs from the authoritative BorrowingTable.
Reads sources/wold/cldf/borrowings.csv (21K explicit donor-recipient events)
instead of fabricating pairs from forms.csv Borrowed column.
Output: staging/cognate_pairs/wold_borrowing_pairs.tsv (14-column schema)
"""
from __future__ import annotations
import csv
import io
import re
import sys
from pathlib import Path
sys.stdout = io.TextIOWrapper(sys.stdout.buffer, encoding="utf-8")
sys.stderr = io.TextIOWrapper(sys.stderr.buffer, encoding="utf-8")
ROOT = Path(__file__).resolve().parent.parent
sys.path.insert(0, str(ROOT / "cognate_pipeline" / "src"))
sys.path.insert(0, str(ROOT / "scripts"))
from cognate_pipeline.normalise.sound_class import ipa_to_sound_class # noqa: E402
SOURCES_DIR = ROOT / "sources" / "wold" / "cldf"
STAGING_DIR = ROOT / "staging" / "cognate_pairs"
STAGING_DIR.mkdir(parents=True, exist_ok=True)
HEADER = (
"Lang_A\tWord_A\tIPA_A\tLang_B\tWord_B\tIPA_B\tConcept_ID\t"
"Relationship\tScore\tSource\tRelation_Detail\tDonor_Language\t"
"Confidence\tSource_Record_ID\n"
)
def clean_source_word(raw: str) -> str:
"""Clean a source word for pseudo-IPA use.
Strips parenthetical notes, proto-form asterisks, bracketed annotations,
and takes only the first alternative when multiple are separated by
comma, slash, or tilde. Result is still pseudo-IPA (not real IPA) but
free of annotations that would produce garbage.
"""
if not raw:
return ""
s = raw
# 1. Strip parenthetical notes: "(Written Tibetan)" etc.
s = re.sub(r"\([^)]*\)", "", s)
# 2. Strip bracketed annotations: "[loan]" etc.
s = re.sub(r"\[[^\]]*\]", "", s)
# 3. Strip proto-form leading asterisks
s = re.sub(r"^\*+", "", s.strip())
# 4. Take only first alternative (split on , / ~)
s = re.split(r"[,/~]", s)[0]
# 5. Strip whitespace and lowercase
return s.strip().lower()
def segments_to_ipa(segments: str) -> str:
"""Convert CLDF Segments column to IPA string."""
if not segments:
return ""
# Strip boundary markers
tokens = segments.replace("^", "").replace("$", "").replace("+", " ").replace("#", " ").replace("_", "")
# Join phoneme tokens
return re.sub(r"\s+", "", tokens).strip()
def sca_similarity(ipa_a: str, ipa_b: str) -> float:
"""Compute normalised Levenshtein similarity on SCA strings."""
try:
sca_a = ipa_to_sound_class(ipa_a)
sca_b = ipa_to_sound_class(ipa_b)
except Exception:
return 0.0
if not sca_a or not sca_b:
return 0.0
m, n = len(sca_a), len(sca_b)
if m == 0 or n == 0:
return 0.0
dp = list(range(n + 1))
for i in range(1, m + 1):
prev = dp[0]
dp[0] = i
for j in range(1, n + 1):
temp = dp[j]
if sca_a[i - 1] == sca_b[j - 1]:
dp[j] = prev
else:
dp[j] = 1 + min(prev, dp[j], dp[j - 1])
prev = temp
dist = dp[n]
return round(1.0 - dist / max(m, n), 4)
def main():
print("=" * 60)
print("WOLD Borrowing Extraction v2")
print("=" * 60)
# Step 1: Read languages.csv → Language name → ISO code
lang_path = SOURCES_DIR / "languages.csv"
lang_iso = {}
lang_name_to_iso = {}
with open(lang_path, "r", encoding="utf-8") as f:
reader = csv.DictReader(f)
for row in reader:
lid = row["ID"]
iso = row.get("ISO639P3code", "").strip()
name = row.get("Name", "").strip()
if iso:
lang_iso[lid] = iso
lang_name_to_iso[name] = iso
print(f" Languages with ISO codes: {len(lang_iso)}")
# Step 2: Read parameters.csv → Parameter_ID → concept gloss
params_path = SOURCES_DIR / "parameters.csv"
param_concept = {}
if params_path.exists():
with open(params_path, "r", encoding="utf-8") as f:
reader = csv.DictReader(f)
for row in reader:
pid = row["ID"]
concept = row.get("Concepticon_Gloss", row.get("Name", pid)).strip()
param_concept[pid] = concept
# Step 3: Read forms.csv → Form_ID → {language, word, ipa, concept} + Borrowed score
forms_path = SOURCES_DIR / "forms.csv"
forms = {}
form_borrowed: dict[str, str] = {} # Form_ID → Borrowed score string
with open(forms_path, "r", encoding="utf-8") as f:
reader = csv.DictReader(f)
for row in reader:
fid = row["ID"]
lid = row["Language_ID"]
iso = lang_iso.get(lid, "")
if not iso:
continue
form = row.get("Form", row.get("Value", "")).strip()
segments = row.get("Segments", "").strip()
ipa = segments_to_ipa(segments) if segments else form.lower()
param_id = row.get("Parameter_ID", "").strip()
concept = param_concept.get(param_id, param_id)
borrowed = row.get("Borrowed", "").strip()
form_borrowed[fid] = borrowed
forms[fid] = {
"iso": iso,
"word": form,
"ipa": ipa,
"concept": concept,
}
print(f" Forms loaded: {len(forms)}")
# Step 4: Read borrowings.csv → generate pairs
borrowings_path = SOURCES_DIR / "borrowings.csv"
output_path = STAGING_DIR / "wold_borrowing_pairs.tsv"
pair_count = 0
skipped_no_target = 0
skipped_no_source = 0
skipped_no_evidence = 0
with open(output_path, "w", encoding="utf-8") as out:
out.write(HEADER)
with open(borrowings_path, "r", encoding="utf-8") as f:
reader = csv.DictReader(f)
for row in reader:
borrowing_id = row["ID"]
target_fid = row.get("Target_Form_ID", "").strip()
source_fid = row.get("Source_Form_ID", "").strip()
source_word = row.get("Source_word", "").strip()
source_lang = row.get("Source_languoid", "").strip()
source_certain = row.get("Source_certain", "").strip()
source_relation = row.get("Source_relation", "").strip()
# Target form is required
target = forms.get(target_fid)
if target is None:
skipped_no_target += 1
continue
# Skip entries where target form has "no evidence for borrowing"
target_borrowed = form_borrowed.get(target_fid, "")
if target_borrowed.startswith("5"):
skipped_no_evidence += 1
continue
# Source can come from Source_Form_ID or Source_word
pseudo_ipa = False
if source_fid and source_fid in forms:
source = forms[source_fid]
source_iso = source["iso"]
source_word_str = source["word"]
source_ipa = source["ipa"]
elif source_word:
# Source form not in database — use Source_word + Source_languoid
source_iso = lang_name_to_iso.get(source_lang, "-")
source_word_str = clean_source_word(source_word)
source_ipa = source_word_str # cleaned pseudo-IPA (not real IPA)
pseudo_ipa = True
else:
skipped_no_source += 1
continue
# Donor_Language: always use the WOLD language name (Source_languoid)
# for consistency. The field is a human-readable language name,
# NOT an ISO 639-3 code. Use source_iso (Lang_B) for the code.
donor_lang = source_lang if source_lang else "-"
# Confidence
confidence = "certain" if source_certain == "yes" else (
"uncertain" if source_certain == "no" else source_certain if source_certain else "-"
)
# Score: -1 sentinel when source IPA is pseudo-IPA (cleaned
# orthography, not real IPA) — SCA similarity is unreliable
if pseudo_ipa:
score = -1
else:
score = sca_similarity(target["ipa"], source_ipa)
# Filter self-loans (same language borrowing from itself)
if target["iso"] == source_iso:
continue
# Relation_Detail: distinguish immediate vs earlier borrowings
if source_relation == "immediate":
relation_detail = "borrowed_immediate"
elif source_relation == "earlier":
relation_detail = "borrowed_earlier"
else:
relation_detail = "borrowed"
# Lang_A = target (borrower), Lang_B = source (donor)
out.write(
f"{target['iso']}\t{target['word']}\t{target['ipa']}\t"
f"{source_iso}\t{source_word_str}\t{source_ipa}\t"
f"{target['concept']}\tborrowing\t{score}\twold\t"
f"{relation_detail}\t{donor_lang}\t{confidence}\twold_{borrowing_id}\n"
)
pair_count += 1
print(f"\n Total borrowing pairs: {pair_count:,}")
print(f" Skipped (no target form): {skipped_no_target}")
print(f" Skipped (no source info): {skipped_no_source}")
print(f" Skipped (no evidence for borrowing): {skipped_no_evidence}")
print(f" Output: {output_path}")
print("=" * 60)
if __name__ == "__main__":
main()